text
stringlengths
14
1.76M
# Joint Secure Transmit Beamforming Designs for Integrated Sensing and Communication Systems ††thanks: Part of this paper has been presented in the IEEE Wireless Communications and Networking Conference (WCNC), 2022 [1]. ††thanks: J. Chu, R. Liu, M. Li, and Y. Liu are with the School of Information and Communication Engineering, Dalian University of Technology, Dalian 116024, China (e-mail<EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS>[email protected]). ††thanks: Q. Liu is with the School of Computer Science and Technology, Dalian University of Technology, Dalian 116024, China (e-mail: [email protected]). Jinjin Chu, Rang Liu, Ming Li, Yang Liu, and Qian Liu ###### Abstract Integrated sensing and communication (ISAC), which allows individual radar and communication systems to share the same spectrum bands, is an emerging and promising technique for alleviating spectrum congestion problems. In this paper, we investigate how to exploit the inherent interference from strong radar signals to ensure the physical layer security (PLS) for the considered multi-user multi-input single-output (MU-MISO) communication and colocated multi-input multi-output (MIMO) radar coexistence system. In particular, with known eavesdroppers’ channel state information (CSI), we propose to jointly design the transmit beamformers of communication and radar systems to minimize the maximum eavesdropping signal-to-interference-plus-noise ratio (SINR) on multiple legitimate users, while guaranteeing the communication quality-of- service (QoS) of legitimate transmissions, the requirement of radar detection performance, and the transmit power constraints of both radar and communication systems. When eavesdroppers’ CSI is unavailable, we develop a joint artificial noise (AN)-aided transmit beamforming design scheme, which utilizes residual available power to generate AN for disrupting malicious receptions as well as satisfying the requirements of both legitimate transmissions and radar target detection. Extensive simulations verify the advantages of the proposed joint beamforming designs for ISAC systems on secure transmissions and the effectiveness of the developed algorithms. ###### Index Terms: Integrated sensing and communication (ISAC), physical layer security (PLS), multi-user multi-input single-output (MU-MISO), artificial noise (AN), interference exploitation. ## I Introduction With the explosive growth of wireless devices, exponentially increased bandwidth is required to support a variety of high data-rate services. Consequently, spectrum resources have been increasingly scarce, which motivates the development of advanced spectrum sharing technologies [2]. Since large portions of the spectrum are available at radar frequency bands, spectrum sharing between radar and communication systems has led to substantial research interest [3]-[6]. This line of research is referred to as integrated sensing and communication (ISAC), which is also known as joint radar-communication (JRC), joint communication and radar (JCR), joint communication and radar sensing (JCAS), etc. ISAC allows radar and communication systems to share the same spectrum bands, which can significantly improve the spectrum efficiency and thus alleviates the spectrum congestion problem. It is foreseeable that ISAC will become a critical enabling technology for future wireless networks, supporting various vital applications, including vehicular networks [7], Internet of Things (IoT) [8], etc. Research on ISAC can be generally categorized into two main directions: Dual- functional radar-communication (DFRC) and radar and communication coexistence (RCC). The former focuses on using the same signals transmitted from one fully-shared hardware platform to simultaneously perform communication and radar sensing functionalities [9]-[12]. Although it has the benefits of low power consumption and smaller-size platform, the trade-off between the radar and communication functionalities requires sophisticated optimizations on the unified dual-functional waveform. Moreover, the resulting hardware complexity significantly hinders practical applications. On the contrary, RCC enables separately deployed communication and radar platforms to cooperatively perform their respective functions using independent transmitted signals. Therefore, RCC is more suitable for many existing scenarios such as sharing spectrum between air-borne early warning radars and 3.5 GHz time-division duplex long- term evolution (TDD-LTE) [13], [14], between battlefield/ground surveillance and vessel traffic service (VTS) radars and WLAN networks [15], [16], etc. In RCC systems, the interference management between the non-colocated base station (BS) and radar transmitter is vital for achieving good communication and radar sensing performance. Therefore, the cooperative design of these two systems is necessary for practical applications [17]-[28]. Meanwhile, the multi-input multi-output (MIMO) architecture has been widely deployed in both radar and communication systems to provide additional spatial degrees of freedom (DoFs) for pursuing more considerable beamforming gains. Various signal processing techniques have been proposed to design transmit beamformers for multi-antenna BS and MIMO radar to realize efficient interference management. A null space projection (NSP) method was proposed in [29], [30], where the radar transmit beamforming is designed to project the radar signals onto the null space of the effective interference channels to eliminate the interference at the communication receivers. However, the target echo may fall into the row space of the interference channels, resulting in degraded radar performance. As a step further, the authors of [31] expanded the projection space to include the subspace corresponding to the small non- zero singular values under a specified threshold, expecting to control the interference to communication systems and achieve different trade-offs between radar and communication performance. However, these NSP-based beamforming designs significantly reduce the DoFs for optimization and cannot guarantee to meet specific radar sensing requirements. In order to overcome these disadvantages, investigations on the joint design of the transmit beamformers for both radar and communication systems have been proposed in [32], [33], where different performance metrics of these two systems are satisfied with controllable constraints. In addition to the above scenarios with perfectly known channel state information (CSI), a robust beamforming design was proposed in [34] under the assumption of imperfect CSI. As mentioned above, the existing literature on beamforming designs for RCC systems essentially aims at suppressing the interference between radar and communication systems. While radar interference is usually deemed as the most significant harmful component to the communication system, from another perspective, it can be utilized to disrupt illegal receivers to safeguard confidential information against eavesdropping. Utilizing interfering/jamming signals to disrupt potential eavesdroppers has been widely considered in the literature on physical layer security (PLS) [35]-[44]. In [35], constructive interference was leveraged to implement secure beamforming using a symbol- level precoding approach. In [36], a multi-antenna cooperative jammer was employed to assist the secure communications. In order to confuse the eavesdropper, the authors in [37] utilized the idea of destructive interference to push the received symbols at the eavesdropper towards the destructive region where the wrong symbol will be detected. The authors in [38] proposed a cooperative secure transmission scheme, in which the legitimate information and interference signals lie in different subspaces at the destination of the confidential transmission, but are aligned along the same subspace at the eavesdropper. Noting that the above secure beamformer designs require the knowledge of the eavesdropper’s CSI, which is not always available in practical applications. In such cases, artificial noise (AN) technology was introduced to realize PLS [39], which uses a large amount of additional energy to generate interfering signals for disrupting potential malicious receptions. The existing communication literature generally forces the AN to be uniformly distributed onto the null space of the confidential transmission channels to disturb the eavesdropper’s reception but not harm legitimate users [40]. Inspired by this concept, in addition to realizing the respective performance requirements of communication and radar systems, the residual power can be used to generate AN. However, this null space projection design can only exploit limited spatial DoFs to generate AN, especially for a system with many users. In recent works [41], [42], the AN and transmit beamforming were jointly optimized to implement secure DFRC transmissions, making full use of available DoFs for generating AN. In addition, the authors in [43] proposed to enhance secure performance by deploying an RIS in DFRC systems. However, how to ensure the physical layer security in RCC systems remains an open problem. Instead of consuming additional power to generate AN, the authors in [44] proposed to exploit the inherent multi-user interference as a helpful resource by converting it to act like AN or distributed friendly jammers to improve the security performance. Inspired by the concept of interference exploitation, the radar signal, which usually has very strong signal power, is an up-and- coming candidate as the jamming signal to enhance the physical layer security performance for the considered RCC system. Motivated by the above findings, we investigate the PLS problem for multi-user multi-input single-output (MU-MISO) communication and colocated MIMO radar coexistence systems in this paper. In particular, the considered RCC system includes a multi-antenna BS serving multiple single-antenna users in the presence of multiple eavesdroppers and a colocated MIMO radar attempting to detect a point-like target. We aim to exploit the coexisted strong radar signals as inherent jamming signals to disrupt the eavesdroppers’ reception111Since the radar signal that is only used to detect the target does not contain any confidential information, it will not cause security/privacy problem to the considered RCC system. . The transmit beamformers of communication and radar systems and radar receive filter are jointly designed to ensure security performance and satisfy the requirements of legitimate transmissions and radar target detection. The main contributions of this paper can be summarized as follows: * • We consider the physical layer security in an ISAC system and innovatively propose to exploit the coexisted strong radar signals as inherent jamming/interfering signals to weaken the reception of potential eavesdroppers and enhance the transmission security. Joint secure transmit beamforming and radar receive filter designs are investigated to achieve this goal. * • With the knowledge of eavesdroppers’ CSI, we jointly design the transmit beamforming and radar receive filter to minimize the maximum eavesdropping signal-to-interference-plus-noise ratio (SINR) on multiple legitimate users, while satisfying the quality-of-service (QoS) requirements of the legitimate users, the radar output SINR constraint, and the transmit power budgets of radar and communication systems. An efficient algorithm based on the block coordinate descent (BCD), fractional programming (FP), and semi-definite relaxation (SDR) methods is developed to solve the resulting non-convex optimization problem. * • When eavesdroppers’ CSI is unavailable, we propose to jointly design the AN- aided transmit beamforming and radar receive filter to maximize the power of AN under the same constraints. In this case, all available transmit power of the BS and radar should be utilized to generate AN as much as possible to destruct eavesdroppers’ reception. A double-loop BCD and SDR based algorithm is employed to convert the resulting complicated non-convex optimization problem into two more tractable sub-problems that can be alternatively solved. * • Extensive simulation results show that the eavesdropping SINR is generally several orders of magnitude smaller than the SINR of legitimate communication users, which verify the significant advancement of utilizing radar signals as inherent jamming/interference signals to enhance the secure transmissions for ISAC systems and the effectiveness of the proposed joint secure transmit beamforming design algorithms. The rest of this paper is organized as follows. Section ii@ introduces the system model of the ISAC system in the presence of eavesdroppers and develops a joint transmit beamforming and radar receive filter design with known eavesdroppers’ CSI. Section iii@ investigates the joint AN-aided transmit beamforming and radar receive filter design without eavesdroppers’ CSI. Simulation results are demonstrated in Section iv@, and finally, conclusions are provided in Section<EMAIL_ADDRESS> Notations: Boldface lower-case and upper-case letters indicate column vectors and matrices, respectively. $(\cdot)^{H}$ and $(\cdot)^{-1}$ denote the transpose-conjugate and inverse operations, respectively. $\mathbb{C}$ denotes the set of all complex numbers. $|a|$, $\|\mathbf{a}\|$, and $\|\mathbf{A}\|_{F}$ are the magnitude of a scalar $a$, the norm of a vector $\mathbf{a}$, and the Frobenius norm of a matrix $\mathbf{A}$, respectively. $\mathbb{E}\\{\cdot\\}$ represents statistical expectation. $\text{Tr}\\{\mathbf{A}\\}$ and $\text{Rank}\\{\mathbf{A}\\}$ are the trace and rank of matrix $\mathbf{A}$, respectively. Figure 1: An ISAC system at the presence of eavesdroppers. ## II Joint Transmit Beamforming Design with Known Eavesdroppers’ CSI ### II-A System Model and Problem Formulation We consider an ISAC system in which an MU-MISO communication system coexists with a colocated MIMO radar system operating on the same frequency band, as shown in Fig. 1. In particular, a BS equipped with $N$ antennas in a uniform linear array (ULA) serves $K$ single-antenna users in the presence of $I$ eavesdroppers who attempt to intercept the confidential information transmissions between the BS and the legitimate users. Meanwhile, a colocated MIMO radar with $M$ transmit/receive antennas in a ULA attempts to detect a point-like target. For convenience, the BS, the legitimate communication users, and the eavesdroppers are referred to as Alice, Bobs, and Eves, respectively. In this paper, we aim to exploit the coexisted strong radar signals as inherent jamming signals to disrupt eavesdroppers’ reception under different assumptions about the availability of Eves’ CSI. We first assume that the CSI of Eves is perfectly known. This assumption is valid, for example, Eves may be legitimate users who intend to overhear other users’ confidential information. With known Eves’ CSI, we jointly design the transmit beamformers of the BS and the radar to ensure secure transmissions and satisfactory radar target detection performance for the ISAC system. The received signal of the $k$-th Bob can be written as $\displaystyle y_{k}=\mathbf{h}_{k}^{H}\mathbf{W}\mathbf{s}+\mathbf{g}_{k}^{H}\mathbf{F}\mathbf{c}+n_{k},$ (1) where $\mathbf{h}_{k}\in\mathbb{C}^{N}$ is the channel vector between the BS and the $k$-th Bob, $\mathbf{W}\triangleq\left[\mathbf{w}_{1},\mathbf{w}_{2},\ldots,\mathbf{w}_{K}\right]\in\mathbb{C}^{N\times K}$ is the beamforming matrix with $\mathbf{w}_{k}$ representing the beamforming vector of the $k$-th Bob, $\mathbf{s}\in\mathbb{C}^{K}$ is the symbol vector with $\mathbb{E}\\{\mathbf{s}\mathbf{s}^{H}\\}=\mathbf{I}_{K}$, $\mathbf{g}_{k}\in\mathbb{C}^{M}$ is the channel vector between the radar and the $k$-th Bob, $\mathbf{F}\in\mathbb{C}^{M\times M}$ is the radar beamforming matrix, $\mathbf{c}\in\mathbb{C}^{M}$ is the radar transmit waveform with $\mathbb{E}\\{\mathbf{c}\mathbf{c}^{H}\\}=\mathbf{I}_{M}$, and $n_{k}$ is the additive white Gaussian noise (AWGN) with $n_{k}\sim\mathcal{CN}(0,\sigma_{k}^{2})$. We assume that the information symbol vector $\mathbf{s}$ is statistically independent with the radar waveform $\mathbf{c}$. Thus, the SINR of the $k$-th Bob can be calculated as $\displaystyle\mathrm{SINR}_{k}$ $\displaystyle=\frac{|\mathbf{h}_{k}^{H}\mathbf{w}_{k}|^{2}}{\sum\limits_{j\neq k}^{K}|\mathbf{h}_{k}^{H}\mathbf{w}_{j}|^{2}+\|\mathbf{g}_{k}^{H}\mathbf{F}\|^{2}+\sigma_{k}^{2}}$ (2) $\displaystyle=\frac{\mathbf{h}_{k}^{H}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{h}_{k}}{\sum\limits_{j\neq k}^{K}\mathbf{h}_{k}^{H}\mathbf{w}_{j}\mathbf{w}_{j}^{H}\mathbf{h}_{k}+\mathbf{g}_{k}^{H}\mathbf{F}\mathbf{F}^{H}\mathbf{g}_{k}+\sigma_{k}^{2}}.$ Similarly, the received signal of the $i$-th Eve can be expressed as $\displaystyle y_{\mathrm{e},i}=\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{W}\mathbf{s}+\mathbf{g}^{H}_{\mathrm{e},i}\mathbf{F}\mathbf{c}+n_{\mathrm{e},i},$ (3) where $\mathbf{h}_{\mathrm{e},i}\in\mathbb{C}^{N}$ is the channel vector between the BS and the $i$-th Eve, $\mathbf{g}_{\mathrm{e},i}\in\mathbb{C}^{M}$ is the channel vector between the radar and the $i$-th Eve, and $n_{\mathrm{e},i}$ is the AWGN with $n_{\mathrm{e},i}\sim\mathcal{CN}(0,\sigma_{\mathrm{e},i}^{2})$. The eavesdropping SINR of the $i$-th Eve on the $k$-th Bob is thus given by $\displaystyle\mathrm{SINR}_{i,k}^{\mathrm{e}}$ $\displaystyle=\frac{|\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{w}_{k}|^{2}}{\sum\limits_{j\neq k}^{K}|\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{w}_{j}|^{2}+\|\mathbf{g}^{H}_{\mathrm{e},i}\mathbf{F}\|^{2}+\sigma_{\mathrm{e},i}^{2}}$ (4) $\displaystyle=\frac{\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{h}_{\mathrm{e},i}}{\sum\limits_{j\neq k}^{K}\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{w}_{j}\mathbf{w}_{j}^{H}\mathbf{h}_{\mathrm{e},i}+\mathbf{g}^{H}_{\mathrm{e},i}\mathbf{F}\mathbf{F}^{H}\mathbf{g}_{\mathrm{e},i}+\sigma_{\mathrm{e},i}^{2}}.$ To ensure secure communications, we aim to minimize the maximum eavesdropping SINR on the $K$ communication users while guaranteeing the communication QoS of legitimate transmissions. In the radar system, the received signal, which includes the echo signal from the target, interference from the BS, and noise, can be written as $\displaystyle\mathbf{y}_{\mathrm{r}}=\alpha\mathbf{a}_{\mathrm{r}}(\theta_{0})\mathbf{a}_{\mathrm{t}}^{H}(\theta_{0})\mathbf{F}\mathbf{c}+\mathbf{Q}^{H}\mathbf{W}\mathbf{s}+\mathbf{n}_{\mathrm{r}},$ (5) where $\alpha$ is the complex target amplitude with $\mathbb{E}\\{|\alpha|^{2}\\}=\sigma_{0}^{2}$. The vectors $\mathbf{a}_{\mathrm{t}}(\theta_{0})\in\mathbb{C}^{M}$ and $\mathbf{a}_{\mathrm{r}}(\theta_{0})\in\mathbb{C}^{M}$ denote the transmit and receive steering vectors of the radar antenna array, $\displaystyle\mathbf{a}_{\mathrm{t}}(\theta_{0})=\mathbf{a}_{\mathrm{r}}(\theta_{0})\triangleq[1,e^{j\frac{2\pi}{\lambda}\Delta\mathrm{sin}(\theta_{0})},\ldots,e^{j\frac{2\pi}{\lambda}(M-1)\Delta\mathrm{sin}(\theta_{0})}]^{T},$ (6) where $\theta_{0}$ represents the azimuth angle of the target222In radar related literature, the direction of the target is typically known to the transmitter since it can be readily estimated [45]-[47] at previous observations, or given by the center of angular sector-of-interest. , $\Delta$ denotes the antenna spacing and $\lambda$ the wavelength. The matrix $\mathbf{Q}\in\mathbb{C}^{N\times M}$ denotes the interfering channel between the BS and the radar receiver, and $\mathbf{n}_{\mathrm{r}}\in\mathbb{C}^{M}$ is the AWGN with $\mathbf{n}_{\mathrm{r}}\sim\mathcal{CN}(0,\sigma_{\mathrm{r}}^{2}\mathbf{I})$. By defining $\mathbf{A}\triangleq\mathbf{a}_{\mathrm{r}}(\theta_{0})\mathbf{a}_{\mathrm{t}}^{H}(\theta_{0})$, the received signal at the radar can be concisely re-written as $\mathbf{y}_{\mathrm{r}}=\alpha\mathbf{A}\mathbf{F}\mathbf{c}+\mathbf{Q}^{H}\mathbf{W}\mathbf{s}+\mathbf{n}_{\mathrm{r}}.$ (7) To achieve better radar sensing performance, the radar utilizes a receive filter $\mathbf{u}\in\mathbb{C}^{M}$ to suppress the interference from the BS and the noise. The signal after filtering is $\displaystyle y_{\mathrm{r}}=\alpha\mathbf{u}^{H}\mathbf{A}\mathbf{F}\mathbf{c}+\mathbf{u}^{H}\mathbf{Q}^{H}\mathbf{W}\mathbf{s}+\mathbf{u}^{H}\mathbf{n}_{\mathrm{r}},$ (8) and radar output SINR can thus be written as $\displaystyle\mathrm{SINR}_{\mathrm{r}}$ $\displaystyle=\frac{\|\alpha\mathbf{u}^{H}\mathbf{A}\mathbf{F}\|^{2}}{\sum\limits_{k=1}^{K}|\mathbf{u}^{H}\mathbf{Q}^{H}\mathbf{w}_{k}|^{2}+\mathbb{E}\left\\{|\mathbf{u}^{H}\mathbf{n}_{\mathrm{r}}|^{2}\right\\}}$ (9) $\displaystyle=\frac{\sigma_{\mathrm{0}}^{2}\mathbf{u}^{H}\mathbf{A}\mathbf{F}\mathbf{F}^{H}\mathbf{A}^{H}\mathbf{u}}{\mathbf{u}^{H}\Big{(}\sum\limits_{k=1}^{K}\mathbf{Q}^{H}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{Q}+\sigma_{\mathrm{r}}^{2}\mathbf{I}\Big{)}\mathbf{u}}.$ For radar systems, the radar SINR is widely utilized as the metric to evaluate radar sensing performance. Thus, we attempt to design the transmit beamforming and the radar receive filter to guarantee that the radar SINR is no less than a pre-defined threshold for achieving satisfactory radar target detection performance. In the considered MU-MISO communication and MIMO radar coexisted system, the CSI $\mathbf{h}_{k}$, $\mathbf{g}_{k}$, $\forall k$ and $\mathbf{Q}$ are assumed to be perfectly known by appropriate channel estimation approaches333 The CSI $\mathbf{h}_{k}$, $\forall k$ can be obtained by conventional uplink training, i.e., users send orthogonal pilot sequences and the BS estimates CSI by classic channel estimation algorithms. Utilizing the same pilot signals sent by users, the CSI $\mathbf{g}_{k}$, $\forall k$ can also be acquired by the radar without additional signaling overhead. The channel $\mathbf{Q}$ between the BS and the radar requires specific pilot signaling for channel estimation. Fortunately, since the geometric locations of the BS and the radar are fixed, channel $\mathbf{Q}$ is generally quasi-static and requires less estimation, which allows acceptable signaling overhead. . Our objective, in this case, is to jointly design the BS transmit beamformer $\mathbf{W}$, the radar transmit beamformer $\mathbf{F}$ and the radar receive filter $\mathbf{u}$ to minimize the maximum eavesdropping SINR of the $I$ eavesdroppers on the $K$ communication users. Meanwhile, the communication QoS requirements of the legitimate users, the radar target detection performance constraint and the power constraints of both communication and radar systems are satisfied. Therefore, the optimization problem can be formulated as $\displaystyle\underset{\mathbf{W},\mathbf{F},\mathbf{u}}{\min}\quad$ $\displaystyle\underset{i,k}{\max}\quad\mathrm{SINR}_{i,k}^{\mathrm{e}}$ (10a) s.t. $\displaystyle\mathrm{SINR}_{k}\geq\Gamma_{k},~{}\forall k,$ (10b) $\displaystyle\mathrm{SINR}_{\mathrm{r}}\geq\Gamma_{\mathrm{r}},$ (10c) $\displaystyle\left\|\mathbf{W}\right\|_{F}^{2}\leq P_{\mathrm{c}},$ (10d) $\displaystyle\left\|\mathbf{F}\right\|_{F}^{2}\leq P_{\mathrm{r}},$ (10e) where $\Gamma_{k}$ is the SINR requirement of the $k$-th Bob, $\Gamma_{\mathrm{r}}$ is the pre-defined threshold for achieving required target detection performance, $P_{\mathrm{c}}$ and $P_{\mathrm{r}}$ denote the total power budgets of the BS and the radar, respectively. Due to the quadratic fractional objective (10a), the quadratic fractional constraints (10b) and (10c), and the coupled variables, problem (10) is a complicated non- convex problem that cannot be directly solved. In order to tackle these difficulties, in the next subsection, we first utilize the BCD method to convert the original problem into two sub-problems, and then employ efficient algorithms based on FP and SDR methods to iteratively solve them. ### II-B Proposed Joint Transmit Beamforming Design In order to decouple the transmit beamformers $\mathbf{W}$ and $\mathbf{F}$ and the radar receive filter $\mathbf{u}$ in the non-convex constraint (10c), we adopt the BCD method to iteratively solve them, which are described in detail as follows. Update $\mathbf{W}$ and $\mathbf{F}$: With fixed $\mathbf{u}$, the sub-problem for updating $\mathbf{W}$ and $\mathbf{F}$ is re-arranged as $\displaystyle\underset{\mathbf{w}_{k},\forall k,\mathbf{F}}{\min}~{}~{}\underset{i,k}{\max}~{}~{}\frac{\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{h}_{\mathrm{e},i}}{\sum\limits_{j\neq k}^{K}\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{w}_{j}\mathbf{w}_{j}^{H}\mathbf{h}_{\mathrm{e},i}+\mathbf{g}^{H}_{\mathrm{e},i}\mathbf{F}\mathbf{F}^{H}\mathbf{g}_{\mathrm{e},i}+\sigma_{\mathrm{e},i}^{2}}$ (11a) $\displaystyle\text{s.t.}~{}~{}~{}\frac{\mathbf{h}_{k}^{H}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{h}_{k}}{\sum\limits_{j\neq k}^{K}\mathbf{h}_{k}^{H}\mathbf{w}_{j}\mathbf{w}_{j}^{H}\mathbf{h}_{k}+\mathbf{g}_{k}^{H}\mathbf{F}\mathbf{F}^{H}\mathbf{g}_{k}+\sigma_{k}^{2}}\geq\Gamma_{k},\forall k,$ (11b) $\displaystyle~{}~{}~{}~{}~{}~{}\frac{\sigma_{\mathrm{0}}^{2}\mathbf{u}^{H}\mathbf{A}\mathbf{F}\mathbf{F}^{H}\mathbf{A}^{H}\mathbf{u}}{\mathbf{u}^{H}\Big{(}\sum\limits_{k=1}^{K}\mathbf{Q}^{H}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{Q}+\sigma_{\mathrm{r}}^{2}\mathbf{I}\Big{)}\mathbf{u}}\geq\Gamma_{\mathrm{r}},$ (11c) $\displaystyle~{}~{}~{}~{}~{}~{}\sum\limits_{k=1}^{K}\left\|\mathbf{w}_{k}\right\|^{2}\leq P_{\mathrm{c}},$ (11d) $\displaystyle~{}~{}~{}~{}~{}~{}\left\|\mathbf{F}\right\|_{F}^{2}\leq P_{\mathrm{r}},$ (11e) which is a complicated min-max problem. To tackle the min-max problem, we introduce an auxiliary variable $z$ to re-formulate it into a more favorable form as $\displaystyle\underset{\mathbf{w}_{k},\forall k,\mathbf{F},z}{\min}\quad z$ (12a) $\displaystyle\text{s.t.}~{}~{}\frac{\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{h}_{\mathrm{e},i}}{\sum\limits_{j\neq k}^{K}\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{w}_{j}\mathbf{w}_{j}^{H}\mathbf{h}_{\mathrm{e},i}+\mathbf{g}^{H}_{\mathrm{e},i}\mathbf{F}\mathbf{F}^{H}\mathbf{g}_{\mathrm{e},i}+\sigma_{\mathrm{e},i}^{2}}\leq z,~{}\forall i,k,$ (12b) $\displaystyle\qquad(\ref{eq:known update W and F}\text{b})-(\ref{eq:known update W and F}\text{e}),$ (12c) which is a minimization problem but still difficult to solve due to the fractional constraint (12b) and the non-convex constraints (11b) and (11c). Noting that problem (12) has a similar form as the max-min-ratio fractional programming problems [48], Dinkelbach’s transform can be applied to convert it into a more tractable form [49]. Specifically, the fractional constraint (12b) can be converted into a polynomial expression (13) as shown on the top of this page $\displaystyle\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{h}_{\mathrm{e},i}-c_{i,k}\big{(}\sum\limits_{j\neq k}^{K}\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{w}_{j}\mathbf{w}_{j}^{H}\mathbf{h}_{\mathrm{e},i}+\mathbf{g}^{H}_{\mathrm{e},i}\mathbf{F}\mathbf{F}^{H}\mathbf{g}_{\mathrm{e},i}+\sigma_{\mathrm{e},i}^{2}\big{)}\leq z,~{}\forall i,k,$ (13) by introducing an auxiliary variable $c_{i,k}$, which essentially represents the eavesdropping SINR of the $i$-th Eve on the $k$-th Bob and is alternatively updated with the transmit beamformers $\mathbf{w}_{k},\forall k$, and $\mathbf{F}$. With given $\mathbf{w}_{k},\forall k$, and $\mathbf{F}$, the optimal $c_{i,k}^{\star}$ can be easily obtained by $\displaystyle c_{i,k}^{\star}=\frac{\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{h}_{\mathrm{e},i}}{\sum\nolimits_{j\neq k}^{K}\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{w}_{j}\mathbf{w}_{j}^{H}\mathbf{h}_{\mathrm{e},i}+\mathbf{g}^{H}_{\mathrm{e},i}\mathbf{F}\mathbf{F}^{H}\mathbf{g}_{\mathrm{e},i}+\sigma_{\mathrm{e},i}^{2}},~{}\forall i,k.$ (14) Then, the optimization problem for updating $\mathbf{w}_{k}$, $\forall k$, and $\mathbf{F}$ can be formulated as $\displaystyle\underset{\mathbf{w}_{k},\forall k,\mathbf{F},z}{\min}\quad z$ (15a) $\displaystyle\text{s.t.}~{}~{}\eqref{eq:c}$ $\displaystyle~{}~{}~{}~{}~{}~{}\frac{\mathbf{h}_{k}^{H}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{h}_{k}}{\sum\limits_{j\neq k}^{K}\mathbf{h}_{k}^{H}\mathbf{w}_{j}\mathbf{w}_{j}^{H}\mathbf{h}_{k}+\mathbf{g}_{k}^{H}\mathbf{F}\mathbf{F}^{H}\mathbf{g}_{k}+\sigma_{k}^{2}}\geq\Gamma_{k},\forall k,$ (15b) $\displaystyle~{}~{}~{}~{}~{}~{}\frac{\sigma_{\mathrm{0}}^{2}\mathbf{u}^{H}\mathbf{A}\mathbf{F}\mathbf{F}^{H}\mathbf{A}^{H}\mathbf{u}}{\mathbf{u}^{H}\Big{(}\sum\limits_{k=1}^{K}\mathbf{Q}^{H}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{Q}+\sigma_{\mathrm{r}}^{2}\mathbf{I}\Big{)}\mathbf{u}}\geq\Gamma_{\mathrm{r}},$ (15c) $\displaystyle~{}~{}~{}~{}~{}~{}\sum\limits_{k=1}^{K}\left\|\mathbf{w}_{k}\right\|^{2}\leq P_{\mathrm{c}},$ (15d) $\displaystyle~{}~{}~{}~{}~{}~{}\left\|\mathbf{F}\right\|_{F}^{2}\leq P_{\mathrm{r}}.$ (15e) It is easy to see that the constraints (13), (15b), and (15c) are non-convex with respect to variables $\mathbf{w}_{k},\forall k$, and $\mathbf{F}$ and are hard to tackle. Therefore, we apply the SDR method to convert them into primary variables for an easier solution. Specifically, by defining $\displaystyle\mathbf{W}_{k}\triangleq\mathbf{w}_{k}\mathbf{w}_{k}^{H},~{}\forall k,$ (16) $\displaystyle\mathbf{R}_{\mathrm{F}}\triangleq\mathbf{F}\mathbf{F}^{H},$ the quadratic terms $\mathbf{w}_{k}\mathbf{w}_{k}^{H}$ and $\mathbf{F}\mathbf{F}^{H}$ are converted into the primary variables $\mathbf{W}_{k}$ and $\mathbf{R}_{\mathrm{F}}$, respectively. In the meantime, the rank-one Hermitian positive semidefinite matrices $\mathbf{W}_{k}$, $\forall k$, and the Hermitian positive semidefinite matrix $\mathbf{R}_{\mathrm{F}}$ should satisfy $\displaystyle\mathbf{W}_{k}=\mathbf{W}_{k}^{H},~{}\mathbf{W}_{k}\succeq 0,~{}\forall k,$ (17a) $\displaystyle\mathrm{Rank}(\mathbf{W}_{k})=1,$ (17b) $\displaystyle\mathbf{R}_{\mathrm{F}}=\mathbf{R}_{\mathrm{F}}^{H},~{}\mathbf{R}_{\mathrm{F}}\succeq 0.$ (17c) For simplicity, we define the set of all $N\times N$-dimensional Hermitian positive semidefinite matrices as $\mathbb{S}_{N}\triangleq\\{\mathbf{S}|\mathbf{S}=\mathbf{S}^{H},~{}\mathbf{S}\succeq 0\\}$. Afterwards, problem (15) is transformed into $\displaystyle\underset{\mathbf{W}_{k},\forall k,\mathbf{R}_{\mathrm{F}},z}{\min}\quad z$ (18a) $\displaystyle\text{s.t.}~{}~{}\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{W}_{k}\mathbf{h}_{\mathrm{e},i}$ (18b) $\displaystyle\quad~{}~{}~{}-c_{i,k}\big{(}\sum\limits_{j\neq k}^{K}\mathbf{h}^{H}_{\mathrm{e},i}\mathbf{W}_{j}\mathbf{h}_{\mathrm{e},i}+\mathbf{g}^{H}_{\mathrm{e},i}\mathbf{R}_{\mathrm{F}}\mathbf{g}_{\mathrm{e},i}+\sigma_{\mathrm{e},i}^{2}\big{)}\leq z,~{}\forall i,k,$ $\displaystyle\quad~{}~{}~{}\frac{\mathbf{h}_{k}^{H}\mathbf{W}_{k}\mathbf{h}_{k}}{\sum\limits_{j\neq k}^{K}\mathbf{h}_{k}^{H}\mathbf{W}_{j}\mathbf{h}_{k}+\mathbf{g}_{k}^{H}\mathbf{R}_{\mathrm{F}}\mathbf{g}_{k}+\sigma_{k}^{2}}\geq\Gamma_{k},~{}\forall k,$ (18c) $\displaystyle\quad~{}~{}~{}\frac{\sigma_{\mathrm{0}}^{2}\mathbf{u}^{H}\mathbf{A}\mathbf{R}_{\mathrm{F}}\mathbf{A}^{H}\mathbf{u}}{\mathbf{u}^{H}\Big{(}\sum\limits_{k=1}^{K}\mathbf{Q}^{H}\mathbf{W}_{k}\mathbf{Q}+\sigma_{\mathrm{r}}^{2}\mathbf{I}\Big{)}\mathbf{u}}\geq\Gamma_{\mathrm{r}},$ (18d) $\displaystyle\quad~{}~{}~{}\sum_{k=1}^{K}\mathrm{Tr}(\mathbf{W}_{k})\leq P_{\mathrm{c}},$ (18e) $\displaystyle\quad~{}~{}~{}\mathrm{Tr}(\mathbf{R}_{\mathrm{F}})\leq P_{\mathrm{r}},$ (18f) $\displaystyle\quad~{}~{}~{}\mathbf{W}_{k}\in\mathbb{S}_{N},~{}\forall k,~{}\mathbf{R}_{\mathrm{F}}\in\mathbb{S}_{M},$ (18g) $\displaystyle\quad~{}~{}~{}\mathrm{Rank}(\mathbf{W}_{k})=1,~{}\forall k.$ (18h) It is evident that the rank-one constraint (18h) tremendously hinders finding a straightforward solution. Thus, we apply the SDR algorithm by dropping the rank-one constraint (18h) and relaxing the problem (18) as $\displaystyle\underset{\mathbf{W}_{k},\forall k,\mathbf{R}_{\mathrm{F}},z}{\min}~{}~{}$ $\displaystyle z$ (19a) s.t. $\displaystyle(\ref{eq:SDR}\text{b})-(\ref{eq:SDR}\text{g}),$ (19b) which is a semi-definite programming (SDP) problem that can be efficiently solved by various existing algorithms and toolboxes such as CVX. Since the rank-one constraint is temporarily neglected, the optimal objective value of problem (19) only serves as a lower bound. After obtaining $\mathbf{W}_{k},\forall k$, the eigenvalue decomposition (EVD) is usually used to obtain the optimal solution $\mathbf{w}_{k}$ if the resulting $\mathbf{W}_{k}$ satisfies the rank-one constraint. Otherwise, Gaussian randomization is required to convert the high-rank solution to a feasible rank-one solution to the problem (11). In our considered case, the rank-1 solution can be guaranteed, whose proof is given in Appendix A. On the other hand, with resulting optimal $\mathbf{R}_{\text{F}}$, the radar beamforming matrix $\mathbf{F}$ can be obtained by utilizing Cholesky decomposition. Algorithm 1 Joint transmit beamforming design algorithm for solving problem (10) 0: $\mathbf{h}_{\mathrm{e},i}$, $\mathbf{g}_{\mathrm{e},i}$, $\sigma_{\mathrm{e},i}^{2}$, $\forall i$, $\mathbf{h}_{k}$, $\mathbf{g}_{k}$, $\sigma_{k}^{2}$, $\Gamma_{k}$, $\forall k$, $\mathbf{A}$, $\mathbf{Q}$, $\sigma_{\text{0}}^{2}$, $\sigma_{\text{r}}^{2}$, $~{}~{}~{}~{}~{}\Gamma_{\mathrm{r}}$, $P_{\mathrm{c}}$, $P_{\mathrm{r}}$. 0: $\mathbf{w}_{k}^{\star}$, $\forall k$, $\mathbf{F}^{\star}$, and $\mathbf{u}^{\star}$. 1: Initialize $\mathbf{F}$, $\mathbf{u}$, and $c_{i,k}$, $\forall i,k$. 2: while the objective value (10a) does not converge do 3: while no convergence do 4: Calculate $\mathbf{W}_{k},\forall k$, and $\mathbf{R}_{\mathrm{F}}$ by solving (19). 5: Update $\mathbf{w}_{k}$ from $\mathbf{W}_{k},\forall k$, by EVD. 6: Update $\mathbf{F}$ from $\mathbf{R}_{\mathrm{F}}$ by Cholesky decomposition. 7: Update $c_{i,k}$, $\forall i,k$, by (14). 8: end while 9: Update $\mathbf{u}$ by solving (20). 10: end while 11: Return $\mathbf{w}_{k}^{\star},\forall k$, $\mathbf{F}^{\star}$, and $\mathbf{u}^{\star}$. Update $\mathbf{u}$: It can be seen that the variable $\mathbf{u}$ only exists in the constraint (10c) of the problem (10). Therefore, with given $\mathbf{W}$ and $\mathbf{F}$, problem (10) is transformed into a feasibility- check problem. For the sake of leaving enough freedoms for solving $\mathbf{W}$ and $\mathbf{F}$ in the subsequent optimization process and accelerating the convergence, we propose to optimize $\mathbf{u}$ to maximize the radar output SINR. The optimization problem is formulated as $\mathbf{u}^{\mathrm{opt}}=\arg\mathop{\max}\limits_{\mathbf{u}}\frac{\sigma_{\mathrm{0}}^{2}\mathbf{u}^{H}\mathbf{A}\mathbf{F}\mathbf{F}^{H}\mathbf{A}^{H}\mathbf{u}}{\mathbf{u}^{H}\Big{(}\sum\limits_{k=1}^{K}\mathbf{Q}^{H}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{Q}+\sigma_{\mathrm{r}}^{2}\mathbf{I}\Big{)}\mathbf{u}}.$ (20) We observe that problem (20) is a typical generalized Rayleigh quotient, whose optimal solution can be easily obtained as the generalized eigenvector corresponding to the largest eigenvalue of the matrix $\sigma_{\mathrm{0}}^{2}\big{(}\sum\nolimits_{k=1}^{K}\mathbf{Q}^{H}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{Q}+\sigma_{\mathrm{r}}^{2}\mathbf{I}\big{)}^{-1}\mathbf{A}\mathbf{F}\mathbf{F}^{H}\mathbf{A}^{H}$. ### II-C Summary, Initialization, Convergence, and Complexity Analysis #### II-C.1 Summary Based on the above derivations, the proposed joint transmit beamforming and radar receive filter design algorithm is straightforward and summarized in Algorithm 1. With appropriate initialization, problems (11) and (20) are iteratively solved to respectively update $\mathbf{w}_{k}$, $\forall k$, $\mathbf{F}$, and $\mathbf{u}$ until the objective value (10a) converges. For updating $\mathbf{w}_{k}$, $\forall k$, and $\mathbf{F}$, we iteratively update the communication beamforming $\mathbf{w}_{k}$, $\forall k$, and the radar beamforming $\mathbf{F}$ by solving the SDP problem (19) and then recovering feasible solutions using EVD and Cholesky decomposition, and the auxiliary variables $c_{i,k},\forall i,k$, until the convergence of problem (11) is found. #### II-C.2 Initialization In order to solve sub-problem (15) for updating $\mathbf{w}_{k}$, $\forall k$, and $\mathbf{F}$, we need to initialize $\mathbf{u}$ and the auxiliary variables $c_{i,k},~{}\forall i,k$. The initial value of the radar receive filter is selected as $\mathbf{u}=\mathbf{a}_{\mathrm{r}}(\theta_{0})$ via a phase alignment operation for better radar detection performance. Since $c_{i,k}$ represents the eavesdropping SINR, which is generally several orders of magnitude smaller than the communication SINR, initializing $c_{i,k}$ with the pre-defined threshold of communication SINR $\Gamma_{k}$ can guarantee the feasibility. The obtained $\mathbf{w}_{k}$, $\forall k$, and $\mathbf{F}$ by solving (15) can be set as the initial value for solving sub-problem (20). #### II-C.3 Convergence We will briefly prove the convergence of the proposed algorithm as follows. Denote $\eta\big{(}\mathbf{W},\mathbf{F},\mathbf{u}\big{)}$ as the objective value of the original problem (10). First, in the transmit beamformer design, we apply Dinkelbach’s transform to convert it into a more tractable form and transform it into the problem (15). According to [50], it is easy to prove the convergence of the algorithm given the non-increasing property of the auxiliary variable $c_{i,k}$. Since the optimal solution of problem (11) is obtained with given $\mathbf{u}^{t}$, we have $\eta\big{(}\mathbf{W}^{t},\mathbf{F}^{t},\mathbf{u}^{t}\big{)}\geq\eta\big{(}\mathbf{W}^{t+1},\mathbf{F}^{t+1},\mathbf{u}^{t}\big{)},$ (21) where the superscript $t$ denotes the index of iterations. Second, with fixed $\left\\{\mathbf{W},\mathbf{F}\right\\}$, the sub-problem for updating $\mathbf{u}$ is a feasibility-check problem. After solving the problem (20), a better radar output SINR than the original requirement is achieved with the obtained radar receive filter $\mathbf{u}$ in the current iteration, i.e., the feasible domain of the original problem (10) is expanded while the objective value is fixed. In other words, with given $\left\\{\mathbf{W}^{t+1},\mathbf{F}^{t+1}\right\\}$, we have $\eta\big{(}\mathbf{W}^{t+1},\mathbf{F}^{t+1},\mathbf{u}^{t}\big{)}=\eta\big{(}\mathbf{W}^{t+1},\mathbf{F}^{t+1},\mathbf{u}^{t+1}\big{)}.$ (22) Based on the above analysis, we have the relationship of the objective values between iterations as $\eta\big{(}\mathbf{W}^{t},\mathbf{F}^{t},\mathbf{u}^{t}\big{)}\geq\eta\big{(}\mathbf{W}^{t+1},\mathbf{F}^{t+1},\mathbf{u}^{t+1}\big{)},$ (23) which indicates that the objective value of problem (10) is non-increasing during the iterations of Algorithm 1. Since the objective value of problem (10) is greater than zero, the proposed Algorithm 1 can converge to a local optimum point. #### II-C.4 Complexity Analysis In this subsection, the computational complexity of Algorithm 1 is analyzed as follows. We first analyze the computational complexity of solving for $\mathbf{w}_{k}$, $\forall k$, and $\mathbf{F}$. Problem (19) is a convex problem with $K$ $N\times N$-dimensional and an $M\times M$-dimensional variable to be optimized, $(I+1)K+1$ second-order cone (SOC) constraints and $K+1$ linear matrix inequality (LMI) constraints. Using the CVX solver, the computational complexity is of order $\mathcal{O}\\{\text{ln}(1/\xi)2\sqrt{5}IK^{1.5}M^{6}\\}$, where $\xi$ is the convergence threshold. The computational complexity of updating $c_{i,k}$, $\forall i,k$, is of order $\mathcal{O}\\{M^{3}\\}$. Other calculations have much lower complexities. For example, updating $\mathbf{w}_{k}$, $\forall k$, has negligible computational of order $\mathcal{O}\\{N^{3}\\}$. Thus, the total complexity to obtain $\mathbf{w}_{k},\forall k$, and $\mathbf{F}$ is of order $\mathcal{O}\\{N_{\text{FP}}\text{ln}(1/\xi)2\sqrt{5}IK^{1.5}M^{6}\\}$, where $N_{\text{FP}}$ is the number of iterations of the inner loop. The computational complexity of updating $\mathbf{u}$ is of order $\mathcal{O}\\{M^{3}\\}$. Therefore, the total computational complexity of the proposed BCD-FP-SDR algorithm is of order $\mathcal{O}\\{N_{\text{tot}}N_{\text{FP}}\text{ln}(1/\xi)2\sqrt{5}IK^{1.5}M^{6}\\}$, where $N_{\text{tot}}$ is the number of iterations of the outer loop. ## III Joint AN-Aided Transmit Beamforming Design without Eavesdroppers’ CSI ### III-A System Model and Problem Formulation When Eves are pure passive eavesdroppers, Alice is unaware of Eves’ CSI or even their existence. In the sequel, the proposed design in the previous section cannot be adopted to ensure security performance. In such cases, AN is a very effective method to improve the physical layer security by disrupting Eves’ reception. Specifically, in addition to transmitting confidential information or probing signals, the BS and the radar also use the available transmit power to generate AN for disturbing potential eavesdroppers as much as possible. Therefore, the transmit beamforming and AN of both the BS and the radar are jointly designed to guarantee good communication and radar sensing performance while safeguarding the communication system against potential malicious eavesdroppers. Based on the above descriptions, the received signal of the $k$-th Bob can be written as $\displaystyle y_{k}=\mathbf{h}_{k}^{H}(\mathbf{W}\mathbf{s}+\mathbf{z})+\mathbf{g}_{k}^{H}(\mathbf{F}\mathbf{c}+\mathbf{v})+n_{k},$ (24) where $\mathbf{z}\sim\mathcal{CN}(0,\mathbf{R}_{\mathrm{z}})$ and $\mathbf{v}\sim\mathcal{CN}(0,\mathbf{R}_{\mathrm{v}})$ are AN vectors generated by the BS and the radar, respectively. We assume that the information symbol vector $\mathbf{s}$, the AN vector $\mathbf{z}$ generated by the BS, the radar waveform $\mathbf{c}$, and the AN vector $\mathbf{v}$ generated by the radar are statistically independent of each other. Thus, the SINR of the $k$-th Bob can be calculated as $\displaystyle\mathrm{SINR}_{k}$ (25) $\displaystyle=\frac{|\mathbf{h}_{k}^{H}\mathbf{w}_{k}|^{2}}{\sum\limits_{j\neq k}^{K}|\mathbf{h}_{k}^{H}\mathbf{w}_{j}|^{2}+\mathbb{E}\left\\{|\mathbf{h}_{k}^{H}\mathbf{z}|^{2}\right\\}+\|\mathbf{g}_{k}^{H}\mathbf{F}\|^{2}+\mathbb{E}\left\\{|\mathbf{g}_{k}^{H}\mathbf{v}|^{2}\right\\}+\sigma_{k}^{2}}$ $\displaystyle=\frac{\mathbf{h}_{k}^{H}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{h}_{k}}{\mathbf{h}_{k}^{H}(\sum\limits_{j\neq k}^{K}\mathbf{w}_{j}\mathbf{w}_{j}^{H}+\mathbf{R}_{\mathrm{z}})\mathbf{h}_{k}+\mathbf{g}_{k}^{H}(\mathbf{F}\mathbf{F}^{H}+\mathbf{R}_{\mathrm{v}})\mathbf{g}_{k}+\sigma_{k}^{2}}.$ From the communication perspective, the proposed joint AN-aided transmit beamforming design aims to guarantee the communication QoS requirements of legitimate transmission while interfering with Eves as much as possible. On the radar side, the echo wave received by the radar is expressed as $\mathbf{y}_{\mathrm{r}}=\alpha\mathbf{A}(\mathbf{F}\mathbf{c}+\mathbf{v})+\mathbf{Q}^{H}(\mathbf{W}\mathbf{s}+\mathbf{z})+\mathbf{n}_{\mathrm{r}}.$ (26) After passing through the receive filter $\mathbf{u}$, the radar output is $y_{\mathrm{r}}=\alpha\mathbf{u}^{H}\mathbf{A}(\mathbf{F}\mathbf{c}+\mathbf{v})+\mathbf{u}^{H}\mathbf{Q}^{H}(\mathbf{W}\mathbf{s}+\mathbf{z})+\mathbf{u}^{H}\mathbf{n}_{\mathrm{r}}.$ (27) The radar output SINR is thus given by $\displaystyle\mathrm{SINR}_{\mathrm{r}}$ (28) $\displaystyle=\frac{\|\alpha\mathbf{u}^{H}\mathbf{A}\mathbf{F}\|^{2}}{\mathbb{E}\left\\{|\alpha\mathbf{u}^{H}\mathbf{A}\mathbf{v}|^{2}\right\\}+\sum\limits_{k=1}^{K}|\mathbf{u}^{H}\mathbf{Q}^{H}\mathbf{w}_{k}|^{2}+\mathbb{E}\left\\{|\mathbf{u}^{H}\mathbf{Q}^{H}\mathbf{z}|^{2}+|\mathbf{u}^{H}\mathbf{n}_{\mathrm{r}}|^{2}\right\\}}$ $\displaystyle=\frac{\sigma_{\mathrm{0}}^{2}\mathbf{u}^{H}\mathbf{A}\mathbf{F}\mathbf{F}^{H}\mathbf{A}^{H}\mathbf{u}}{\mathbf{u}^{H}\Big{(}\sigma_{\mathrm{0}}^{2}\mathbf{A}\mathbf{R}_{\mathrm{v}}\mathbf{A}^{H}+\mathbf{Q}^{H}(\sum\limits_{k=1}^{K}\mathbf{w}_{k}\mathbf{w}_{k}^{H}+\mathbf{R}_{\mathrm{z}})\mathbf{Q}+\sigma_{\mathrm{r}}^{2}\mathbf{I}\Big{)}\mathbf{u}}.$ From the radar perspective, the radar output SINR is guaranteed to be no less than a pre-defined threshold for achieving satisfactory target detection performance. Meanwhile, as much power as possible is used to generate AN for interfering with Eves. It is intuitive that higher transmission power of the confidential information poses a higher risk of being intercepted, since Eves’ eavesdropping SINR is directly proportional to the transmission power. Thus, the BS should minimize the transmit power to satisfy Bobs’ QoS and use the residual power to generate AN signals. Similarly, for the radar system, the minimum required power is allocated to generate directional signals, whose main lobe points to the direction of the target for achieving satisfactory detection performance. The huge residual power is used to generate omni-directional AN signals, which will bring excellent security performance in the presence of potential eavesdroppers. Therefore, our objective is to jointly design the BS transmit beamformer $\mathbf{W}$, the covariance $\mathbf{R}_{\mathrm{z}}$ of the AN vector $\mathbf{z}$ generated by the BS, the radar transmit beamformer $\mathbf{F}$, the covariance $\mathbf{R}_{\mathrm{v}}$ of the AN vector $\mathbf{v}$ generated by the radar, and the radar receive filter $\mathbf{u}$ to minimize the total transmit power used by the BS and radar beamformers. Meanwhile, the communication QoS requirements of the legitimate users, the radar output SINR constraint, and the power constraints of both communication and radar systems are satisfied. Therefore, the optimization problem is formulated as $\displaystyle\underset{\mathbf{W},\mathbf{R}_{\mathrm{z}},\mathbf{F},\mathbf{R}_{\mathrm{v}},\mathbf{u}}{\min}\quad$ $\displaystyle\left\|\mathbf{W}\right\|_{F}^{2}+\left\|\mathbf{F}\right\|_{F}^{2}$ (29a) s.t. $\displaystyle\mathrm{SINR}_{k}\geq\Gamma_{k},~{}\forall k,$ (29b) $\displaystyle\mathrm{SINR}_{\mathrm{r}}\geq\Gamma_{\mathrm{r}},$ (29c) $\displaystyle\left\|\mathbf{W}\right\|_{F}^{2}+\mathrm{Tr}(\mathbf{R}_{\mathrm{z}})=P_{\mathrm{c}},$ (29d) $\displaystyle\left\|\mathbf{F}\right\|_{F}^{2}+\mathrm{Tr}(\mathbf{R}_{\mathrm{v}})=P_{\mathrm{r}},$ (29e) $\displaystyle\mathbf{R}_{\mathrm{z}}\in\mathbb{S}_{N},~{}\mathbf{R}_{\mathrm{v}}\in\mathbb{S}_{M}.$ (29f) We observe that problem (29) is a non-convex problem that is difficult to solve for the following two reasons. First, the variables are intricately coupled in the constraints (29b) and (29c). Second, these two constraints are quadratic and fractional. In order to efficiently solve this problem, we employ the BCD and SDR algorithms to convert it into two more tractable sub- problems and then alternately solve each of them until convergence is achieved. ### III-B Proposed AN-aided Joint Transmit Beamforming Design In this subsection, we propose an efficient double-loop BCD-SDR algorithm to tackle the non-convex problem (29). It can be seen that problem (29) is very complicated due to the coupling variables in the SINR constraints (29b) and (29c). To this end, we divide this problem into two sub-problems with respect to the radar and communication systems, and utilize a two-block BCD algorithm to iteratively solve them. #### III-B.1 The Sub-problem for Radar System Given the variables $\mathbf{w}_{k}$, $\forall k$, and $\mathbf{R}_{\mathrm{z}}$ of the communication system, the transmit beamformer $\mathbf{F}$, the covariance $\mathbf{R}_{\mathrm{v}}$ of the AN vector $\mathbf{v}$, and the receive filter $\mathbf{u}$ of the radar system are jointly optimized. The problem for updating $\mathbf{F}$, $\mathbf{R}_{\mathrm{v}}$, and $\mathbf{u}$ can be formulated as $\displaystyle\underset{\mathbf{F},\mathbf{R}_{\mathrm{v}},\mathbf{u}}{\min}~{}~{}\left\|\mathbf{F}\right\|_{F}^{2}$ (30a) $\displaystyle\text{s.t.}~{}~{}\frac{\mathbf{h}_{k}^{H}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{h}_{k}}{\mathbf{h}_{k}^{H}(\sum\limits_{j\neq k}^{K}\mathbf{w}_{j}\mathbf{w}_{j}^{H}+\mathbf{R}_{\mathrm{z}})\mathbf{h}_{k}+\mathbf{g}_{k}^{H}(\mathbf{F}\mathbf{F}^{H}+\mathbf{R}_{\mathrm{v}})\mathbf{g}_{k}+\sigma_{k}^{2}}\geq\Gamma_{k},\forall k,$ (30b) $\displaystyle~{}~{}~{}~{}~{}\frac{\sigma_{\mathrm{0}}^{2}\mathbf{u}^{H}\mathbf{A}\mathbf{F}\mathbf{F}^{H}\mathbf{A}^{H}\mathbf{u}}{\mathbf{u}^{H}\Big{(}\sigma_{\mathrm{0}}^{2}\mathbf{A}\mathbf{R}_{\mathrm{v}}\mathbf{A}^{H}+\mathbf{Q}^{H}(\sum\limits_{k=1}^{K}\mathbf{w}_{k}\mathbf{w}_{k}^{H}+\mathbf{R}_{\mathrm{z}})\mathbf{Q}+\sigma_{\mathrm{r}}^{2}\mathbf{I}\Big{)}\mathbf{u}}\geq\Gamma_{\mathrm{r}},$ (30c) $\displaystyle~{}~{}~{}~{}~{}\left\|\mathbf{F}\right\|_{F}^{2}+\mathrm{Tr}(\mathbf{R}_{\mathrm{v}})=P_{\mathrm{r}},$ (30d) $\displaystyle~{}~{}~{}~{}~{}\mathbf{R}_{\mathrm{v}}\in\mathbb{S}_{M}.$ (30e) Since the variables $\mathbf{F}$, $\mathbf{R}_{\mathrm{v}}$, and $\mathbf{u}$ are highly coupled in constraint (30c), which makes problem (30) very difficult to solve, we adopt a two-block BCD scheme to iteratively solve them as follows. Update $\mathbf{F}$ and $\mathbf{R}_{\mathrm{v}}$: With fixed $\mathbf{u}$, the sub-problem for updating $\mathbf{F}$ and $\mathbf{R}_{\mathrm{v}}$ is re- arranged as $\displaystyle\underset{\mathbf{F},\mathbf{R}_{\mathrm{v}}}{\min}~{}~{}$ $\displaystyle\left\|\mathbf{F}\right\|_{F}^{2}$ (31a) s.t. $\displaystyle(\ref{eq:bcd radar}\text{b})-(\ref{eq:bcd radar}\text{e}).$ (31b) Considering that the constraints (30b) and (30c) are still non-convex due to the quadratic terms with respect to $\mathbf{F}$, we transform problem (31) into $\displaystyle\underset{\mathbf{R}_{\mathrm{F}},\mathbf{R}_{\mathrm{v}}}{\min}~{}~{}\mathrm{Tr}(\mathbf{R}_{\mathrm{F}})$ (32a) $\displaystyle\text{s.t.}~{}\frac{\mathbf{h}_{k}^{H}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\mathbf{h}_{k}}{\mathbf{h}_{k}^{H}(\sum\limits_{j\neq k}^{K}\mathbf{w}_{j}\mathbf{w}_{j}^{H}+\mathbf{R}_{\mathrm{z}})\mathbf{h}_{k}+\mathbf{g}_{k}^{H}(\mathbf{R}_{\mathrm{F}}+\mathbf{R}_{\mathrm{v}})\mathbf{g}_{k}+\sigma_{k}^{2}}\geq\Gamma_{k},\forall k,$ (32b) $\displaystyle\frac{\sigma_{\mathrm{0}}^{2}\mathbf{u}^{H}\mathbf{A}\mathbf{R}_{\mathrm{F}}\mathbf{A}^{H}\mathbf{u}}{\mathbf{u}^{H}\Big{(}\sigma_{\mathrm{0}}^{2}\mathbf{A}\mathbf{R}_{\mathrm{v}}\mathbf{A}^{H}+\mathbf{Q}^{H}(\sum\limits_{k=1}^{K}\mathbf{w}_{k}\mathbf{w}_{k}^{H}+\mathbf{R}_{\mathrm{z}})\mathbf{Q}+\sigma_{\mathrm{r}}^{2}\mathbf{I}\Big{)}\mathbf{u}}\geq\Gamma_{\mathrm{r}},$ (32c) $\displaystyle\mathrm{Tr}(\mathbf{R}_{\mathrm{F}})+\mathrm{Tr}(\mathbf{R}_{\mathrm{v}})=P_{\mathrm{r}},$ (32d) $\displaystyle\mathbf{R}_{\mathrm{F}}\in\mathbb{S}_{M},~{}\mathbf{R}_{\mathrm{v}}\in\mathbb{S}_{M}.$ (32e) Obviously, problem (32) is an SDP problem and can be solved by convex tools, e.g., CVX. After solving $\mathbf{R}_{\text{F}}$, the radar beamforming matrix $\mathbf{F}$ can be obtained by utilizing Cholesky decomposition. Update $\mathbf{u}$: Notice that the variable $\mathbf{u}$ only exists in the constraint (30c) of the problem (30). Therefore, with given $\mathbf{F}$ and $\mathbf{R}_{\mathrm{v}}$, problem (30) is transformed into a feasibility- check problem. In order to leave more freedoms for minimizing $\left\|\mathbf{F}\right\|_{F}^{2}$ in the next iteration, we propose to optimize $\mathbf{u}$ to maximize the radar output SINR. The optimization problem is formulated as $\displaystyle\mathbf{u}^{\mathrm{opt}}=$ (33) $\displaystyle\arg\mathop{\max}\limits_{\mathbf{u}}\frac{\sigma_{\mathrm{0}}^{2}\mathbf{u}^{H}\mathbf{A}\mathbf{F}\mathbf{F}^{H}\mathbf{A}^{H}\mathbf{u}}{\mathbf{u}^{H}\Big{(}\sigma_{\mathrm{0}}^{2}\mathbf{A}\mathbf{R}_{\mathrm{v}}\mathbf{A}^{H}+\mathbf{Q}^{H}(\sum\limits_{k=1}^{K}\mathbf{w}_{k}\mathbf{w}_{k}^{H}+\mathbf{R}_{\mathrm{z}})\mathbf{Q}+\sigma_{\mathrm{r}}^{2}\mathbf{I}\Big{)}\mathbf{u}},$ whose optimal solution can be easily obtained as the generalized eigenvector corresponding to the largest eigenvalue of matrix $\sigma_{\mathrm{0}}^{2}\big{[}\sigma_{\mathrm{0}}^{2}\mathbf{A}\mathbf{R}_{\mathrm{v}}\mathbf{A}^{H}\\!+\\!\mathbf{Q}^{H}(\sum\limits_{k=1}^{K}\mathbf{w}_{k}\mathbf{w}_{k}^{H}\\!+\\!\mathbf{R}_{\mathrm{z}})\mathbf{Q}+\sigma_{\mathrm{r}}^{2}\mathbf{I}\big{]}^{-1}\mathbf{A}\mathbf{F}\mathbf{F}^{H}\mathbf{A}^{H}$. Algorithm 2 Joint AN-aided transmit beamforming design algorithm for solving problem (29) 0: $\mathbf{h}_{k}$, $\mathbf{g}_{k}$, $\Gamma_{k}$, $\sigma_{k}^{2}$, $\forall k$, $\mathbf{A}$, $\mathbf{Q}$, $\sigma_{\mathrm{0}}^{2}$, $\sigma_{\text{r}}^{2}$, $\Gamma_{\mathrm{r}}$, $P_{\mathrm{c}}$, $P_{\mathrm{r}}$. 0: $\mathbf{w}_{k}^{\star}$, $\forall k$, $\mathbf{R}_{\mathrm{z}}^{\star}$, $\mathbf{F}^{\star}$, $\mathbf{R}_{\mathrm{v}}^{\star}$, and $\mathbf{u}^{\star}$. 1: Initialize $\mathbf{w}_{k}$, $\forall k$, $\mathbf{R}_{\mathrm{z}}$, and $\mathbf{u}$. 2: while the objective value (29a) does not converge do 3: while the objective value (30a) does not converge do 4: Calculate $\mathbf{R}_{\mathrm{F}}$ and update $\mathbf{R}_{\mathrm{v}}$ by solving (32). 5: Update $\mathbf{F}$ from $\mathbf{R}_{\mathrm{F}}$ by Cholesky decomposition. 6: Update $\mathbf{u}$ by (LABEL:eq:update_u). 7: end while 8: Calculate $\mathbf{W}_{k}$, $\forall k$, and update $\mathbf{R}_{\mathrm{z}}$ by solving (35). 9: Update $\mathbf{w}_{k}$ from $\mathbf{W}_{k}$, $\forall k$, by EVD. 10: end while 11: Return $\mathbf{w}_{k}^{\star}$, $\forall k$, $\mathbf{R}_{\mathrm{z}}^{\star}$, $\mathbf{F}^{\star}$, $\mathbf{R}_{\mathrm{v}}^{\star}$, and $\mathbf{u}^{\star}$. #### III-B.2 The Sub-problem for Communication System With fixed $\mathbf{F}$, $\mathbf{R}_{\mathrm{v}}$, and $\mathbf{u}$, the sub- problem for updating the transmit beamformer $\mathbf{W}$ and the covariance $\mathbf{R}_{\mathrm{z}}$ of the AN vector $\mathbf{z}$ is re-arranged as $\displaystyle\underset{\mathbf{w}_{k},\forall k,\mathbf{R}_{\mathrm{z}}}{\min}~{}~{}$ $\displaystyle\sum_{k=1}^{K}\mathrm{Tr}(\mathbf{w}_{k}\mathbf{w}_{k}^{H})$ (34a) s.t. $\displaystyle(\ref{eq:bcd radar}\text{b}),~{}(\ref{eq:bcd radar}\text{c})$ (34b) $\displaystyle\sum_{k=1}^{K}\mathrm{Tr}(\mathbf{w}_{k}\mathbf{w}_{k}^{H})+\mathrm{Tr}(\mathbf{R}_{\mathrm{z}})=P_{\mathrm{c}},$ (34c) $\displaystyle\mathbf{R}_{\mathrm{z}}\in\mathbb{S}_{N}.$ (34d) Similarly, we utilize the SDR algorithm to convert this problem into an SDP problem by using the definitions in (16) and dropping the rank-one constraint (17b) as $\displaystyle\underset{\mathbf{W}_{k},\forall k,\mathbf{R}_{\mathrm{z}}}{\min}~{}~{}\sum_{k=1}^{K}\mathrm{Tr}(\mathbf{W}_{k})$ (35a) s.t. $\displaystyle\frac{\mathbf{h}_{k}^{H}\mathbf{W}_{k}\mathbf{h}_{k}}{\mathbf{h}_{k}^{H}(\sum\limits_{j\neq k}^{K}\mathbf{W}_{j}+\mathbf{R}_{\mathrm{z}})\mathbf{h}_{k}+\mathbf{g}_{k}^{H}(\mathbf{F}\mathbf{F}^{H}+\mathbf{R}_{\mathrm{v}})\mathbf{g}_{k}+\sigma_{k}^{2}}\geq\Gamma_{k},\forall k,$ (35b) $\displaystyle\frac{\sigma_{\mathrm{0}}^{2}\mathbf{u}^{H}\mathbf{A}\mathbf{F}\mathbf{F}^{H}\mathbf{A}^{H}\mathbf{u}}{\mathbf{u}^{H}\Big{(}\sigma_{\mathrm{0}}^{2}\mathbf{A}\mathbf{R}_{\mathrm{v}}\mathbf{A}^{H}+\mathbf{Q}^{H}(\sum\limits_{k=1}^{K}\mathbf{W}_{k}+\mathbf{R}_{\mathrm{z}})\mathbf{Q}+\sigma_{\mathrm{r}}^{2}\mathbf{I}\Big{)}\mathbf{u}}\geq\Gamma_{\mathrm{r}},$ (35c) $\displaystyle\sum_{k=1}^{K}\mathrm{Tr}(\mathbf{W}_{k})+\mathrm{Tr}(\mathbf{R}_{\mathrm{z}})=P_{\mathrm{c}},$ (35d) $\displaystyle(\ref{eq:Wk constraint}\text{a}),~{}(\ref{eq:bcd communication}\text{d}),$ (35e) and then solve the resulting SDP problem by various existing algorithms or toolboxes, e.g., CVX. The obtained $\mathbf{W}_{k}$, $\forall k$ can also be guaranteed to satisfy the rank-1 constraint in this case. The proof is given in Appendix A. We use the same method as in the previous section to obtain $\mathbf{w}_{k}$, $\forall k$. ### III-C Summary, Initialization, Convergence, and Complexity Analysis #### III-C.1 Summary Based on the above derivations, the proposed joint AN-aided transmit beamforming design is straightforward and summarized in Algorithm 2. In the inner loop, we iteratively solve problems (31) and (LABEL:eq:update_u) to respectively update $\mathbf{F}$, $\mathbf{R}_{\mathrm{v}}$, and $\mathbf{u}$ until the objective value (30a) converges. In the outer loop, we iteratively solve problems (30) and (34) for updating $\mathbf{w}_{k}$, $\forall k$, and $\mathbf{R}_{\text{z}}$ until the objective value (29a) converges. #### III-C.2 Initialization In order to solve sub-problem (31) for updating $\mathbf{F}$ and $\mathbf{R}_{\mathrm{v}}$, we investigate to properly initialize $\mathbf{W}$, $\mathbf{R}_{\mathrm{z}}$, and $\mathbf{u}$. The concept of NSP is utilized to initialize $\mathbf{W}$ and $\mathbf{R}_{\mathrm{z}}$. Specifically, we jointly design the beamforming without AN aiming to guarantee the performance requirements of communication and radar systems, and then project the AN of the communication system onto the null space of the effective interference channels between the BS and Bobs. Similarly, the initial value of the radar receive filter is selected as $\mathbf{u}=\mathbf{a}_{\mathrm{r}}(\theta_{0})$ via a phase alignment operation. The obtained $\mathbf{F}$, $\mathbf{R}_{\mathrm{v}}$, and $\mathbf{u}$ by solving (30) can be set as the initial value for solving sub-problem (34). #### III-C.3 Convergence The convergence of the proposed algorithm will be briefly proven as follows. Denote $\eta\big{(}\mathbf{W},\mathbf{R}_{\mathrm{z}},\mathbf{F},\mathbf{R}_{\mathrm{v}},\mathbf{u}\big{)}$ as the objective value of the original problem (29). First, for the radar system, the sub-problem for updating $\mathbf{F}$ and $\mathbf{R}_{\mathrm{v}}$ is transformed into an SDP problem. It is obvious that the transmit power is non-increasing between iterations, we have $\eta\big{(}\mathbf{W}^{t},\mathbf{R}_{\mathrm{z}}^{t},\mathbf{F}^{t},\mathbf{R}_{\mathrm{v}}^{t},\mathbf{u}^{t}\big{)}\geq\eta\big{(}\mathbf{W}^{t},\mathbf{R}_{\mathrm{z}}^{t},\mathbf{F}^{t+1},\mathbf{R}_{\mathrm{v}}^{t+1},\mathbf{u}^{t}\big{)},$ (36) with fixed $\left\\{\mathbf{W}^{t},\mathbf{R}_{\mathrm{z}}^{t},\mathbf{u}^{t}\right\\}$. Given $\left\\{\mathbf{W}^{t},\mathbf{R}_{\mathrm{z}}^{t},\mathbf{F}^{t+1},\mathbf{R}_{\mathrm{v}}^{t+1}\right\\}$, the sub-problem for updating $\mathbf{u}$ is a feasibility-check problem, which follows that $\eta\big{(}\mathbf{W}^{t},\mathbf{R}_{\mathrm{z}}^{t},\mathbf{F}^{t+1},\mathbf{R}_{\mathrm{v}}^{t+1},\mathbf{u}^{t}\big{)}=\eta\big{(}\mathbf{W}^{t},\mathbf{R}_{\mathrm{z}}^{t},\mathbf{F}^{t+1},\mathbf{R}_{\mathrm{v}}^{t+1},\mathbf{u}^{t+1}\big{)}.$ (37) Second, the sub-problem for the communication system is also transformed into an SDP problem. With given $\left\\{\mathbf{F}^{t+1},\mathbf{R}_{\mathrm{v}}^{t+1},\mathbf{u}^{t+1}\right\\}$, it follows that $\displaystyle\eta\big{(}\mathbf{W}^{t},\mathbf{R}_{\mathrm{z}}^{t},\mathbf{F}^{t+1},\mathbf{R}_{\mathrm{v}}^{t+1},\mathbf{u}^{t+1}\big{)}$ (38) $\displaystyle\geq\eta\big{(}\mathbf{W}^{t+1},\mathbf{R}_{\mathrm{z}}^{t+1},\mathbf{F}^{t+1},\mathbf{R}_{\mathrm{v}}^{t+1},\mathbf{u}^{t+1}\big{)}.$ Based on the above analysis, we have the relationship of the objective values between iterations as $\displaystyle\eta\big{(}\mathbf{W}^{t},\mathbf{R}_{\mathrm{z}}^{t},\mathbf{F}^{t},\mathbf{R}_{\mathrm{v}}^{t},\mathbf{u}^{t}\big{)}$ (39) $\displaystyle\geq\eta\big{(}\mathbf{W}^{t+1},\mathbf{R}_{\mathrm{z}}^{t+1},\mathbf{F}^{t+1},\mathbf{R}_{\mathrm{v}}^{t+1},\mathbf{u}^{t+1}\big{)},$ which indicates that the objective value of problem (29) is non-increasing during iterations of Algorithm 2. Since the objective value of problem (29) is greater than zero, the proposed Algorithm 2 can converge to a local optimum point. #### III-C.4 Complexity Analysis In the inner loop, problem (32) is a convex problem with two $M\times M$-dimensional variables to be optimized, $K+1$ SOC constraints and an LMI constraint. Using the CVX solver, the complexity for updating $\mathbf{F}$ and $\mathbf{R}_{\mathrm{v}}$ is of order $\mathcal{O}\\{\text{ln}(1/\xi)8\sqrt{2}K^{1.5}M^{6}\\}$, and solving problem (LABEL:eq:update_u) for updating $\mathbf{u}$ has the complexity of order $\mathcal{O}\\{M^{3}\\}$. Thus, the total complexity for obtaining $\mathbf{F}$, $\mathbf{R}_{\mathrm{v}}$, and $\mathbf{u}$ is of order $\mathcal{O}\\{N_{\text{inn}}\text{ln}(1/\xi)8\sqrt{2}K^{1.5}M^{6}\\}$, where $N_{\text{inn}}$ is the number of iterations of the inner loop. Similarly, problem (35) is a convex problem with $K+1$ $N\times N$-dimensional variables to be optimized, $K+1$ SOC constraints, and an LMI constraint. The computational complexity for updating $\mathbf{w}_{k}$, $\forall k$, and $\mathbf{R}_{\text{z}}$ is of order $\mathcal{O}\\{\text{ln}(1/\xi)\sqrt{2}K^{4.5}N^{6}\\}$. Other lower complexity calculations are omitted. Thus, the total computational complexity for solving the problem (29) is of order $\mathcal{O}\\{N_{\text{out}}N_{\text{inn}}\text{ln}(1/\xi)8\sqrt{2}K^{1.5}M^{6}\\}$, where $N_{\text{out}}$ is the number of iterations of the outer loop. ## IV Simulation Results In this section, extensive simulation results are provided to demonstrate the performance of our proposed joint secure beamforming designs for the considered ISAC systems under the assumption that Eves’ CSI is available or not. Except for the radar-target link adopting the AoA model, the Rayleigh fading channel model is adopted for all links, i.e., each entry of the channel matrices is assumed to obey the standard complex Gaussian distribution. The number of eavesdroppers is $I$ = 2. The antenna spacing of the radar is $\Delta=\lambda/2$. The noise power at Eves and Bobs are set as $\sigma_{\mathrm{e},i}^{2}=\sigma_{k}^{2}=10\mathrm{dBm},~{}\forall i,k$. The total power budget of the BS is $P_{\mathrm{c}}$ = 10W. For simplicity, we assume the communication QoS requirement of each Bob is the same and denoted by $\Gamma_{\mathrm{c}}$. The target is located at the azimuth angle $\theta_{0}=0^{\circ}$ and the radar cross section (RCS) is $\sigma_{\mathrm{0}}^{2}=1$. The noise power to radar echoes is set as $\sigma_{\mathrm{r}}^{2}=0\mathrm{dB}$444The radar echo passes through the round-trip distance between the radar and the target, while the signal the user receives only passes through the one-way distance from the BS to the user. Thus, the noise power to radar echoes is usually a little larger than that of the users.. Besides, the SINR requirement thresholds should be set appropriately to prevent an infeasible case. Figure 2: Convergence of the proposed algorithm. Figure 3: SINRs of Eve and Bob versus the communication QoS requirement $\Gamma_{\mathrm{c}}$ ($M$ = 16, $\Gamma_{\mathrm{r}}$ = 10dB, $P_{\mathrm{r}}$ = 500W). Figure 4: SINRs of Eve and Bob versus the number of radar transmit/receive antennas $M$ ($\Gamma_{\mathrm{c}}$ = 5dB, $\Gamma_{\mathrm{r}}$ = 10dB, $P_{\mathrm{r}}$ = 500W). Figure 5: SINRs of Eve and Bob versus the radar system transmit power $P_{\mathrm{r}}$ ($M$ = 16, $\Gamma_{\mathrm{c}}$ = 5dB, $\Gamma_{\mathrm{r}}$ = 10dB). ### IV-A Known Eavesdroppers’ CSI In this subsection, we illustrate the performance of our proposed joint secure beamforming design assuming perfect knowledge of Eves’ CSI. We assume that the BS is equipped with $N=4$ antennas to serve $K=4$ Bobs. In all simulations, the SINR of Eve denotes the maximum eavesdropping SINR and the SINR of Bob denotes the minimum legitimate SINR, respectively. We first illustrate the convergence performance of the proposed algorithm in Fig. 2. It can be observed that our proposed algorithm converges very quickly under different settings, which reveals the effectiveness of the proposed algorithm and its great potential in practical applications. The SINRs of Eve and Bob versus the communication QoS requirement $\Gamma_{\mathrm{c}}$ are plotted in Fig. 3 to illustrate the performance of secure communication. The “proposed” scheme denotes our proposed joint transmit beamforming design for secure ISAC systems. For comparisons, we also include the “no radar” scheme, which denotes that there is one single communication system without radar interference, and the “separate” scheme, which represents that the transmit beamformers of the communication system and radar system are separately designed without considering the interference between them. From Fig. 3, we can observe that in the scenario of “no radar”, the required communication QoS of Bobs can be guaranteed, but the security of the communication system is not very satisfactory since Eves’ SINR is not low enough. When the radar joins the system without cooperative joint beamforming design, Eves’ SINR is significantly reduced owing to the strong radar interfering signals, which provides good confidentiality. However, at the same time, legitimate transmissions are also severely damaged by radar interference. Compared with these two scenarios, our proposed joint transmit beamforming design significantly decreases Eves’ SINR to a minimum value, while always maintaining Bobs’ SINR at a required level to satisfy the QoS requirements of the legitimate transmissions. In Fig. 4, we show the SINRs of Eve and Bob versus the number of radar antennas $M$. It is natural that the number of radar antennas does not affect the security and legitimate transmission in the “no radar” and “separate” scenarios. With the proposed joint beamforming design algorithm, Eves’ SINR decreases with increasing $M$ since additional spatial DoFs can be exploited to provide stronger interference to Eve. This phenomenon verifies that a higher DoF of MIMO radar is of great help for secure communication. Next, we present the SINRs of Eve and Bob versus the radar system power budget $P_{\mathrm{r}}$ in Fig. 5, where the same relationship between different scenarios can be observed as that in Figs. 3 and 4. In addition, we observe that the eavesdropping SINRs of both the “proposed” and “separate” schemes decrease with a larger $P_{\mathrm{r}}$ due to stronger jamming signals. Moreover, the proposed scheme achieves the least eavesdropping SINR thanks to the cooperative joint beamforming design, which guarantees the most favorable security performance. ### IV-B Unknown Eavesdroppers’ CSI Figure 6: Convergence of the proposed algorithm. In this subsection, we demonstrate the performance of our proposed joint AN- aided transmit beamforming design for the case that Eves’ CSI is unknown. The BS is equipped with $N=8$ antennas. We first illustrate the convergence performance of the proposed algorithm in Fig. 6. It can be observed that the objective value (29a) monotonically converges within a limited number of iterations under different settings, which verifies the effectiveness of the proposed algorithm. Next, we present the security performance of the proposed algorithm in Figs. 7-9. We also include a scheme that does not consider the physical layer security. In other words, it only uses the least power to satisfy the communication QoS requirements of the legitimate users and the radar output SINR constraint (denoted as “No PLS”), and the scheme that utilizes the tremendous residual power to generate AN in the null space of Bobs’ channels (denoted as “Null space”). The security of the communication system is evaluated in terms of the maximum eavesdropping SINR of the potential eavesdroppers on the $K$ communication users. Figure 7: SINR of Eves versus the communication QoS requirement $\Gamma_{\mathrm{c}}$ ($M$ = 16, $\Gamma_{\mathrm{r}}$ = 10dB, $P_{\mathrm{r}}$ = 1000W). Figure 8: SINR of Eves versus the number of radar antennas $M$ ($\Gamma_{\mathrm{c}}$ = 5dB, $\Gamma_{\mathrm{r}}$ = 10dB, $P_{\mathrm{r}}$ = 500W). Figure 9: SINR of Eves versus the radar system transmit power $P_{\mathrm{r}}$ ($M$ = 16, $\Gamma_{\mathrm{c}}$ = 5dB, $\Gamma_{\mathrm{r}}$ = 10dB). In Fig. 7, we plot the maximum SINR of Eves versus the communication QoS requirement $\Gamma_{\mathrm{c}}$. Not surprisingly, the maximum SINR of Eves increases with increasing communication QoS requirements, because the stricter communication QoS requirements require more power for the beamforming and leave less power to generate AN, which brings higher risks of being eavesdropped. In addition, we observe a considerable performance gap between the “No PLS” scheme and the schemes with PLS design (“Null space” and “Proposed”), since the generated AN greatly confuses the eavesdropper. Moreover, it is seen that the proposed algorithm achieves notably better security performance than the “Null space” scheme since more DoFs are exploited in designing AN. These findings verify the advantages of the proposed AN-aided transmit beamforming design in preventing potential eavesdropping without the knowledge of Eves’ CSI. Fig. 8 shows the maximum SINR of Eves versus the number of radar antennas $M$. Since more antennas can exploit additional spatial DoFs, the transmit power required to achieve the pre-set radar SINR becomes less. Therefore, for the “No PLS” scheme, the jamming/interference from the radar to the eavesdropper becomes weak and the security performance is degraded. On the other hand, for the “Null space” scheme and the proposed scheme, more power is available for generating AN, which can significantly improve the security performance. It is noticed that, compared with Fig. 4, Eves’ SINR will not visibly decrease when the number of radar antennas is greater than 16. This is because the spatial DoFs of a radar system with 16 antennas are sufficient for providing enough signal processing ability. More radar antennas would not significantly reduce the transmit power for achieving the required radar SINR. Consequently, the available power for generating AN will be relatively small with more radar antenna than 16, and the Eves’ SINR will be maintained at almost the same extremely low level, e.g. close to -44dB. These results can still verify that the proposed AN-aided secure beamforming design can provide superior security performance even when the MIMO radar has a moderate number of antennas. Fig. 9 presents the maximum SINR of Eve versus the radar system power budget $P_{\mathrm{r}}$. The same relationship between different design schemes can be observed in Figs. 7 and 8. In addition, since more additional power can be utilized for generating AN, the eavesdropping SINR decreases with the increase of $P_{\mathrm{r}}$ for the “Null space” and the proposed schemes. ## V Conclusions This paper investigated joint secure transmit beamforming designs for ISAC systems. When the eavesdroppers’ channels are available, the maximum eavesdropping SINR was minimized under the communication QoS constraints, the radar detection performance constraint, and the communication and radar power constraints. An efficient BCD-FP-SDR-based algorithm was proposed to solve the non-convex optimization problem. When the eavesdroppers’ channels are unavailable, a joint AN-aided transmit beamforming design was developed to disrupt the eavesdroppers’ reception while guaranteeing the legitimate communication SINR and the radar output SINR requirements by utilizing the available power of radar and communication systems to generate as much AN as possible. We proposed a double-loop BCD-SDR-based algorithm to solve the resulting non-convex optimization problem. Simulation results illustrated the advantages of ISAC systems on secure transmissions and the effectiveness of the proposed algorithms. ## Appendix A Let $\widetilde{\mathbf{W}}_{k}$, $\forall k$, be an arbitrary global optimal solution to problem (18). Then, we construct a new solution $\widehat{\mathbf{W}}_{k}$, $\forall k$, based on $\widetilde{\mathbf{W}}_{k}$, $\forall k$, as $\widehat{\mathbf{W}}_{k}=\widetilde{\mathbf{W}}_{k},~{}\widehat{\mathbf{w}}_{k}=(\mathbf{h}_{k}^{H}\widetilde{\mathbf{W}}_{k}\mathbf{h}_{k})^{-1/2}\widetilde{\mathbf{W}}_{k}\mathbf{h}_{k},~{}\widehat{\mathbf{W}}_{k}=\widehat{\mathbf{w}}_{k}\widehat{\mathbf{w}}_{k}^{H}.$ (40) Clearly, $\widehat{\mathbf{W}}_{k}$, $\forall k$, is rank-one and positive semidefinite which means that constraints (18g) and (18h) hold. Next, we show that $\widehat{\mathbf{W}}_{k}$ is also a global optimal solution to problem (18). Based on (40), we have the following equality $\sum_{k=1}^{K}\widetilde{\mathbf{W}}_{k}=\sum_{k=1}^{K}\widehat{\mathbf{W}}_{k},$ (41) which implies that the constraints (18d) and (18e) hold. Moreover, the constraint (18f) and the objective function hold since they do not contain the variables $\widetilde{\mathbf{W}}_{k}$, $\forall k$. In addition, since $\mathbf{h}_{k}^{H}\widehat{\mathbf{W}}_{k}\mathbf{h}_{k}=\mathbf{h}_{k}^{H}\widehat{\mathbf{w}}_{k}\widehat{\mathbf{w}}_{k}^{H}\mathbf{h}_{k}=\mathbf{h}_{k}^{H}\widetilde{\mathbf{W}}_{k}\mathbf{h}_{k},$ (42) the SINR constraint in (18c) can be re-written as $\displaystyle\frac{(\Gamma_{k}+1)}{\Gamma_{k}}\mathbf{h}_{k}^{H}\widehat{\mathbf{W}}_{k}\mathbf{h}_{k}=\frac{(\Gamma_{k}+1)}{\Gamma_{k}}\mathbf{h}_{k}^{H}\widetilde{\mathbf{W}}_{k}\mathbf{h}_{k}$ (43) $\displaystyle\geq\sum_{j=1}^{K}\mathbf{h}_{k}^{H}\widetilde{\mathbf{W}}_{j}\mathbf{h}_{k}+\mathbf{g}_{k}^{H}\mathbf{R}_{\mathrm{F}}\mathbf{g}_{k}+\sigma_{k}^{2}$ $\displaystyle=\sum_{j=1}^{K}\mathbf{h}_{k}^{H}\widehat{\mathbf{W}}_{j}\mathbf{h}_{k}+\mathbf{g}_{k}^{H}\mathbf{R}_{\mathrm{F}}\mathbf{g}_{k}+\sigma_{k}^{2},$ namely (18c) holds for $\widehat{\mathbf{W}}_{k}$, $\forall k$. Similarly, the SINR constraint (18b) also holds for $\widehat{\mathbf{W}}_{k}$, $\forall k$. Therefore, it is proven that the constructed $\widehat{\mathbf{W}}_{k}$, $\forall k$ in (40) is a feasible and global optimal rank-1 solution to the problem (18). For problem (35), we also let $\widetilde{\mathbf{W}}_{k}$, $\forall k$, be an arbitrary global optimal solution to problem (35). In the same way as (40), we construct a new solution $\widehat{\mathbf{W}}_{k}$, $\forall k$, which implies that (35e) holds. Similar to the previous proof, constraints (35a)-(35d) hold. Thus, the constructed $\widehat{\mathbf{W}}_{k}$, $\forall k$, in (40) being a feasible and global optimal solution to problem (35) is verified. ## References * [1] J. Chu, R. Liu, Y. Liu, M. Li, and Q. Liu, “Joint transmit beamforming design for secure communication and radar coexistence systems,” in Proc. IEEE Wireless Commun. Network Conf. (WCNC), Austin, USA, Apr. 2022. * [2] A. R. Chiriyath, B. Paul, and D. W. Bliss, “Radar-communication convergence: Coexistence, cooperation, and co-design,” IEEE Trans. Cogn. Commun. Netw., vol. 3, no. 1, pp. 1-12, Mar. 2017. * [3] F. Liu, C. Masouros, A. P. Petropulu, H. Griffiths, and L. Hanzo, “Joint radar and communication design: Applications, state-of-the-art, and the road ahead,” IEEE Trans. Commun., vol. 68, no. 6, pp. 3834-3862, Jun. 2020. * [4] N. C. Luong, X. Lu, D. T. Hoang, D. Niyato, and D. I. Kim, “Radio resource management in joint radar and communication: A comprehensive survey,” IEEE Commun. Surv. Tut., vol. 23, no. 2, pp. 780-814, 2nd Quart. 2021. * [5] J. A. Zhang, M. L. Rahman, K. Wu, X. Huang, Y. J. Guo, S. Chen, and J. Yuan, “Enabling joint communication and radar sensing in mobile networks - A survey,” IEEE Commun. Surv. Tut., vol. 24, no. 1, pp. 306-345, 1st Quart. 2022. * [6] A. Liu et al., “A survey on fundamental limits of integrated sensing and communication,” IEEE Commun. Surv. Tut., vol. 24, no. 2, pp. 994-1034, 2nd Quart. 2022. * [7] F. Liu and C. Masouros, “A tutorial on joint radar and communication transmission for vehicular networks—Part I: Background and fundamentals,” IEEE Commun. Lett., vol. 25, no. 2, pp. 322-326, Feb. 2021. * [8] Y. Cui, F. Liu, X. Jing, and J. Mu, “Integrating sensing and communications for ubiquitous IoT: Applications, trends, and challenges,” IEEE Netw., vol. 35, no. 5, pp. 158-167, Sep./Oct. 2021. * [9] F. Liu, L. Zhou, C. Masouros, A. Li, W. Luo, and A. Petropulu, “Toward dual-functional radar-communication systems: Optimal waveform design,” IEEE Trans. Signal Process., vol. 66, no. 16, pp. 4264-4279, Aug. 2018. * [10] F. Liu, C. Masouros, A. Li, H. Sun, and L. Hanzo, “MU-MIMO communications with MIMO radar: From co-existence to joint transmission,” IEEE Trans. Wireless Commun., vol. 17, no. 4, pp. 2755-2770, Apr. 2018. * [11] R. Liu, M. Li, Q. Liu, and A. L. Swindlehurst, “Joint waveform and filter designs for STAP-SLP-based MIMO-DFRC systems,” IEEE J. Sel. Areas Commun., vol. 40, no. 6, pp. 1918-1931, Jun. 2022. * [12] R. Liu, M. Li, Y. Liu, Q. Wu, and Q. Liu, “Joint transmit waveform and passive beamforming design for RIS-aided DFRC Systems,” IEEE J. Sel. Topics Signal Process., vol. 16, no. 5, pp. 995-1010, Aug. 2022. * [13] J. H. Reed et al., “On the co-existence of TD-LTE and radar over 3.5 GHz band: An experimental study,” IEEE Wireless Commun. Lett., vol. 5, no. 4, pp. 368-371, Aug. 2016. * [14] F. Hessar and S. Roy, “Spectrum sharing between a surveillance radar and secondary Wi-Fi networks,” IEEE Trans. Aerosp. Electron. Syst., vol. 52, no. 3, pp. 1434-1448, Jun. 2016. * [15] H. Griffiths et al., “Radar spectrum engineering and management: Technical and regulatory issues,” IEEE Proc., vol. 103, no. 1, pp. 85-102, Jan. 2015. * [16] E. Grossi, M. Lops, and L. Venturino, “Joint design of surveillance radar and MIMO communication in cluttered environments,” IEEE Trans. Signal Process., vol. 68, no. 1, pp. 1544-1557, Dec. 2020. * [17] L. Zheng, M. Lops, X. Wang, and E. Grossi, “Joint design of overlaid communication systems and pulsed radars,” IEEE Trans. Signal Process., vol. 66, no. 1, pp. 139-154, Jan. 2018. * [18] H. Deng and B. Himed, “Interference mitigation processing for spectrum-sharing between radar and wireless communications systems,” IEEE Trans. Aerosp. Electron. Syst., vol. 49, no. 3, pp. 1911-1919, Jul. 2013. * [19] Z. Geng, H. Deng, and B. Himed, “Adaptive radar beamforming for interference mitigation in radar-wireless spectrum sharing,” IEEE Signal Process. Lett., vol. 22, no. 4, pp. 484-488, Apr. 2015. * [20] B. Li and A. Petropulu, “MIMO radar and communication spectrum sharing with clutter mitigation,” in Proc. IEEE Radar Conf. (RadarConf), Philadelphia, USA, May 2016. * [21] F. Liu, C. Masouros, A. Li, T. Ratnarajah, and J. Zhou, “MIMO radar and cellular coexistence: A power-efficient approach enabled by interference exploitation,” IEEE Trans. Signal Process., vol. 66, no. 14, pp. 3681-3695, Jul. 2018. * [22] B. Li, A. P. Petropulu, and W. Trappe, “Optimum co-design for spectrum sharing between matrix completion based MIMO radars and a MIMO communication system,” IEEE Trans. Signal Process., vol. 64, no. 17, pp. 4562-4575, Sep. 2016. * [23] N. Nartasilpa, A. Salim, D. Tuninetti, and N. Devroye, “Communications system performance and design in the presence of radar interference,” IEEE Trans. Commun., vol. 66, no. 9, pp. 4170-4185, Sep. 2018. * [24] L. Zheng, M. Lops, Y. C. Eldar, and X. Wang, “Radar and communication coexistence: An overview: A review of recent methods,” IEEE Signal Process. Mag., vol. 36, no. 5, pp. 85-99, Sep. 2019. * [25] Z. Cheng, B. Liao, S. Shi, Z. He, and J. Li, “Co-design for overlaid MIMO radar and downlink MISO communication systems via Cramér-Rao bound minimization,” IEEE Trans. Signal Process., vol. 67, no. 24, pp. 6227-6240, Dec. 2019. * [26] M. Bic and V. Koivunen, “Radar waveform optimization for target parameter estimation in cooperative radar-communications systems,” IEEE Trans. Aerosp. Electron. Syst., vol. 55, no. 5, pp. 2314-2326, Oct. 2019. * [27] J. Qian, M. Lops, L. Zheng, X. Wang, and Z. He, “Joint system design for coexistence of MIMO radar and MIMO communication,” IEEE Trans. Signal Process., vol. 66, no. 13, pp. 3504-3519, Jul. 2018. * [28] J. Qian, L. Venturino, M. Lops, and X. Wang, “Radar and communication spectral coexistence in range-dependent interference,” IEEE Trans. Signal Process., vol. 69, pp. 5891-5906, Oct. 2021. * [29] S. Sodagari, A. Khawar, T. C. Clancy, and R. McGwier, “A projection based approach for radar and telecommunication systems coexistence,” in Proc. IEEE Global Commun. Conf. (GLOBECOM), Anaheim, USA, Dec. 2012, pp. 5010-5014. * [30] A. Babaei, W. H. Tranter, and T. Bose, “A nullspace-based precoder with subspace expansion for radar/communications coexistence,” in Proc. IEEE Global Commun. Conf. (GLOBECOM), Atlanta, USA, Dec. 2013. * [31] J. A. Mahal, A. Khawar, A. Abdelhadi, and T. C. Clancy, “Spectral coexistence of MIMO radar and MIMO cellular system,” IEEE Trans. Aerosp. Electron. Syst., vol. 53, no. 2, pp. 655-668, Apr. 2017. * [32] M. Rihan and L. Huang, “Optimum co-design of spectrum sharing between MIMO radar and MIMO communication systems: An interference alignment approach,” IEEE Trans. Veh. Technol., vol. 67, no. 12, pp. 11667-11680, Dec. 2018. * [33] X. Liu, T. Huang, N. Shlezinger, Y. Liu, J. Zhou, and Y. C. Eldar, “Joint transmit beamforming for multiuser MIMO communications and MIMO radar,” IEEE Trans. Signal Process., vol. 68, pp. 3929-3944, Jun. 2020. * [34] F. Liu, C. Masouros, A. Li, and T. Ratnarajah, “Robust MIMO beamforming for cellular and radar coexistence,” IEEE Wireless Commun. Lett., vol. 6, no. 3, pp. 374-377, Jun. 2017. * [35] M. Khandaker, C. Masouros, and K.-K. Wong, “Constructive interference based secure precoding: A new dimension in physical layer security,” IEEE Trans. Inf. Forensics Security, vol. 13, no. 9, pp. 2256-2268, Sep. 2018. * [36] Z. Chu, K. Cumanan, Z. Ding, M. Johnston, and S. Y . Le Goff, “Secrecy rate optimizations for a MIMO secrecy channel with a cooperative jammer,” IEEE Trans. Veh. Technol., vol. 64, no. 5, pp. 1833-1847, May 2015. * [37] N. Su, F. Liu, Z. Wei, Y.-F. Liu, and C. Masouros, “Secure dual-functional radar-communication transmission: Exploiting interference for resilience against target eavesdropping,” IEEE Trans. Wireless Commun., vol. 21, no. 9, pp. 7238-7252, Sep. 2022. * [38] L. Li, A. P. Petropulu, Z. Chen, and J. Fang, “Improving wireless physical layer security via exploiting co-channel interference,” IEEE J. Sel. Topics Signal Process., vol. 10, no. 8, pp. 1433-1448, Dec. 2016. * [39] W.-C. Liao, T.-H. Chang, W.-K. Ma, and C.-Y. Chi, “QoS-based transmit beamforming in the presence of eavesdroppers: An optimized artificial-noise-aided approach,” IEEE Trans. Signal Process., vol. 59, no. 3, pp. 1202-1216, Mar. 2011. * [40] S. Goel and R. Negi, “Guaranteeing secrecy using artificial noise,” IEEE Trans. Wireless Commun., vol. 7, no. 6, pp. 2180-2189, Jun. 2008. * [41] N. Su, F. Liu, and C. Masouros, “Secure radar-communication systems with malicious targets: Integrating radar, communications and jamming functionalities,” IEEE Trans. Wireless Commun., vol. 20, no. 1, pp. 83-95, Jan. 2021. * [42] Y. Zhou, H. Zhou, F. Zhou, S. Ma, and R. Q. Hu, “Artificial-noise aided precoding design for MIMO integrated radar and communication systems,” in Proc. Int. Conf. Wireless Commun. Signal Process. (WCSP), Xi’an, China, Dec. 2019. * [43] K. V. Mishra, A. Chattopadhyay, S. S. Acharjee, and A. P. Petropulu, “OptM3Sec: Optimizing multicast IRS-aided multiantenna DFRC secrecy channel with multiple eavesdroppers,” in Proc. IEEE Int. Conf. Acoustics Speech Signal Process. (ICASSP), Singapore, Singapore, May 2022. * [44] M. Li, G. Ti, and Q. Liu, “Secure beamformer designs in MU-MIMO systems with multiuser interference exploitation,” IEEE Trans. Veh. Technol., vol. 67, no. 9, pp. 8288-8301, Sep. 2018. * [45] I. Bekkerman and J. Tabrikian, “Target detection and localization using MIMO radars and sonars,” IEEE Trans. Signal Process., vol. 54, no. 10, pp. 3873-3883, Oct. 2006. * [46] E. Aboutanios and A. Hassanien, “Low-cost beamforming-based DOA estimation with model order determination,” in Proc. IEEE Sensor Array Multichannel Signal Process. Workshop (SAM), Hangzhou, China, Jun. 2020. * [47] P. Stoica and K. C. Sharman, “Maximum likelihood methods for direction-of-arrival estimation,” IEEE Trans. Acousti. Speech Signal Process., vol. 38, no. 7, pp. 1132-1143, Jul. 1990. * [48] K. Shen and W. Yu, “Fractional programming for communication systems—Part I: Power control and beamforming,” IEEE Trans. Signal Process., vol. 66, no. 10, pp. 2616-2630, May 2018. * [49] W. Dinkelbach, “On nonlinear fractional programming,” Manage. Sci., vol. 133, no. 7, pp. 492-498, Mar. 1967. * [50] J.-P. Crouzeix, “Algorithms for generalized fractional programming,” Math. Program., vol. 52, no. 1, pp. 191-207, May 1991.
# Controlled-source electromagnetic modelling using high order finite- difference time-domain method on a nonuniform grid Pengliang Yang1 and Rune Mittet2 1School of Mathematics, Harbin Institute of Technology, Harbin, China, 150001 E-mail<EMAIL_ADDRESS> 2Norwegian University of Science and Technology (NTNU), Norway E-mail<EMAIL_ADDRESS> ###### Abstract Simulation of 3D low-frequency electromagnetic fields propagating in the Earth is computationally expensive. We present a fictitious wave domain high-order finite-difference time-domain (FDTD) modelling method on nonuniform grids to compute frequency-domain 3D controlled-source electromagnetic (CSEM) data. The method overcomes the inconsistency issue widely present in the conventional 2nd order staggered grid finite difference scheme over nonuniform grid, achieving high accuracy with arbitrarily high order scheme. The finite- difference coefficients adaptive to the node spacings, can be accurately computed by inverting a Vandermonde matrix system using efficient algorithm. A generic stability condition applicable to nonuniform grids is established, revealing the dependence of the time step and these finite-difference coefficients. A recursion scheme using fixed point iterations is designed to determine the stretching factor to generate the optimal nonuniform grid. The grid stretching in our method reduces the number of grid points required in the discretization, making it more efficient than the standard high-order FDTD with a densely sampled uniform grid. Instead of stretching in both vertical and horizontal directions, better accuracy of our method is observed when the grid is stretched along the depth without horizontal stretching. The efficiency and accuracy of our method are demonstrated by numerical examples. ## 1 Introduction Marine controlled source electromagnetics (CSEM) provides valuable information about subsurface resistivities and therefore potentially about pore fluids or rocks. It is very useful to decipher subsurface properties to assist energy exploration, in particular when combined with seismic data. The CSEM technology relies on low-frequency electromagnetic field propagation to probe the subsurface. The low-frequency electromagnetic (EM) field propagation does not lend itself to an intuitive understanding in the same manner as seismic field propagation does due to the diffusive nature of the EM field in conductive media. Thus, three-dimensional modeling becomes an important tool for the interpretation of CSEM data. Imaging of marine CSEM data is today mainly done by inversion of the observed electric and/or magnetic fields. The kernel of CSEM inversion is the numerical simulation of 3D electromagnetic field propagation, which is computationally expensive. Reducing the simulation time without compromising the accuracy is important. It can shorten the turnaround time for an imaging project while reducing the investments in computer hardware. The implementation of nonuniform grid schemes is a well known strategy to reduce simulation time. To retain good accuracy we propose a high-order finite-difference approach. There are many studies on diffusive electromagnetic modelling using different methods. Examples are the frequency-domain finite-difference method (Newman and Alumbaugh,, 1995; Smith, 1996a, ; Mulder,, 2006; Streich,, 2009), the frequency-domain finite-element method (Li and Key,, 2007; da Silva et al.,, 2012; Key,, 2016; Rochlitz et al.,, 2019), and the time-domain finite- difference method (Oristaglio and Hohmann,, 1984; Wang and Hohmann,, 1993; Taflove and Hagness,, 2005). A key fact in all numerical modelling methods is that the computational cost and the memory requirement are connected and cannot be splitted. A method can be very efficient if more computer memory is available. The efficiency and accuracy of the modelling can be dramatically hampered when the available computer resources are restricted. Due to the diffusive nature of low-frequency CSEM fields, most of the 3D CSEM modelling schemes resort to the frequency-domain solution of the Maxwell equation to avoid the high computational cost dictated by the restrictive stability condition for the direct solution in the time domain. Time-domain methods are attractive options because they require less amount of memory than frequency-domain modelling within a model of the same size. Another advantage with time-domain solutions is that multiple frequencies can be extracted from the same simulation. Both the frequency-domain finite-difference method and the frequency-domain finite-element method formulate Maxwell equation as a linear equation system, which may be solved using direct (Streich,, 2009) or iterative (Smith, 1996b, ; Mulder,, 2006; Puzyrev et al.,, 2013) solvers. A nice feature with a direct solver is that multiple right-hand sides are fast to calculate after the system matrix has been factorized or inverted. However, there are significant implementation challenges with this approach when realistic size marine CSEM surveys are simulated. The memory requirements are large even if the equation system is sparse. The finite-difference time-domain (FDTD) modelling based on the staggered grid proposed by Yee, (1966) has for several decades been a main workhorse for many EM applications. The implementation of the numerical core is straight forward and the computational efficiency is good for wave phenomena. The computational efficiency for diffusive phenomena is rather poor in the time domain. The system of partial differential equations can be considered stiff in this case (Mittet,, 2010) and a very small time step is required to retain stability. The computational efficiency can be improved significantly due to a correspondence principle for wave and diffusion fields (Lee et al.,, 1989; de Hoop,, 1996; Mittet,, 2010). Maaø, (2007) proposed a mixed wave and diffusion- domain FDTD method to perform numerically efficient CSEM modelling. This method allowed for large time steps compared to a purely diffusion-domain solution. Mittet, (2010) proposed a high-order FDTD scheme by utilizing the fictitious wave to diffusion-domain transformation. The simulation is performed in the wave domain where the propagation velocity is proportional to the square root of resistivity. The Yee grid (staggered grid) FDTD scheme is often the method of choice due to a good agreement with physics. It gives divergence free magnetic fields and electric currents (Smith, 1996a, ). This standard scheme proposed by Yee, (1966) is based on the second-order approximation of the first derivatives assuming an equispaced mesh. On the uniform grid, moving from second-order FDTD to high-order FDTD is straight forward (Mittet,, 2010), and gives improved modelling accuracy due to the reduction of spatial dispersion errors. To improve the modelling efficiency, the use of nonuniform grid is widespread (Newman and Alumbaugh,, 1995; Mulder,, 2006), however it results in inconsistencies for the grid staggering. As illustrated in Figure 1, this inconsistency leads to only first-order local truncation error, even though the global accuracy may be up to second order (Monk and Süli,, 1994). The problem is persistent and has remained unresolved for high-order schemes. Figure 1: The 2nd order staggered-grid finite-difference scheme on (a) uniform grid and (b) nonuniform grid. The grid points on staggered grid are obtained by taking the midpoints from nonuniform grid to ensure the 2nd order accuracy. This accuracy is not guaranteed on nonuniform grid since the midpoints is inconsistent. For example, $x_{3/2}=(x_{1}+x_{2})/2$, $x_{5/2}=(x_{2}+x_{3})/2$, but $\tilde{x}_{2}=(x_{3/2}+x_{5/2})/2\neq x_{2}$ due to uniform grid spacing. This inconsistency leads to first-order local truncation error using staggered-grid finite-difference scheme on the nonuniform grid. We propose an efficient 3D CSEM simulation method with high accuracy using _high-order FDTD on a staggered, nonuniform grid_ , following the fictitious wave domain approach (Mittet,, 2010). To resolve the inconsistency issue in conventional 2nd order staggered grid approach, our key recognition is that the order of local truncation error can be arbitrarily high also on a nonuniform grid if the finite-difference operator coefficients are adapted properly to the variable grid spacing. The derivative operator coefficients are calculated by inverting a Vandermonde matrix system. To gain good modeling efficiency, we transforms the diffusive Maxwell equation into the fictitious wave domain. The efficiency of the method is restricted by stability condition: the stepsize in time is proportional to the inverse of the propagation velocity of the field, affected by the node spacing. The gridding of the same physical domain leads to different number of gridpoints, affecting the size of the linear system to be solved. The use of nonuniform grid helps to reduce the number of gridpoint, thus reducing the computational cost. Unfortunately, the stability condition for high order FDTD over non- uniform grid is non-trivial. An important contribution of this paper is to establish a new stability condition valid for arbitrarily high order FDTD scheme on nonuniform grid. The stability condition shows the strong dependence between time step and the finite difference coefficients computed by inverting the Vandermonde matrix. This is not known in EM geophysics community, as far as we know. To generate the optimal grid using a power law, we design a recursion scheme using fixed point iterations to find the optimal stretching factor. We prove the recursion scheme is guaranteed to converge. The optimal factor found by the recursive scheme allows accurate matching of the computational domain using given number of mesh points. The high accuracy and efficiency of this high-order FDTD method on a nonuniform grid will be exemplified by a number of numerical tests using reference solutions. ## 2 Theory We utilize the correspondence principle for electromagnetic wave and diffusion fields (Lee et al.,, 1989; de Hoop,, 1996; Mittet,, 2010) to calculate the CSEM response efficiently. The key to a high-order local truncation error on a nonuniform grid relies on the solutions to a Vandermonde system giving derivative operator coefficients that adapts to local grid properties. The stability condition is then established for FDTD modelling on the nonuniform grid. The Maxwell equations in a quasi-static regime (i.e., with negligible effect of displacement currents) are written in the time domain as $\displaystyle\nabla\times\mathbf{E}+\mu\partial_{t}\mathbf{H}$ $\displaystyle=$ $\displaystyle-\mathbf{M},$ (1) $\displaystyle-\nabla\times\mathbf{H}+\mathbf{\sigma}\mathbf{E}$ $\displaystyle=$ $\displaystyle-\mathbf{J},$ or in the frequency domain as $\displaystyle\nabla\times\mathbf{E}-\mathrm{i}\omega\mu\mathbf{H}$ $\displaystyle=$ $\displaystyle-\mathbf{M},$ (2) $\displaystyle-\nabla\times\mathbf{H}+\mathbf{\sigma}\mathbf{E}$ $\displaystyle=$ $\displaystyle-\mathbf{J},$ where $\mathbf{E}=(E_{x},E_{y},E_{z})^{T}$ and $\mathbf{H}=(H_{x},H_{y},H_{z})^{T}$ are electric and magnetic fields. The magnetic permeability is $\mu$. The conductivity is a symmetric $3\times 3$ tensor: $\sigma_{ij}=\sigma_{ji},\;i,j\in\\{x,y,z\\}$. An isotropic medium means that only the diagonal elements of the conductivity tensor are non-zeros and the same in all directions: $\sigma_{xx}=\sigma_{yy}=\sigma_{zz}$; $\sigma_{ij}=0,i\neq j$. The vertical transverse isotropic (VTI) medium implemented here still has only diagonal elements, but the vertical and the horizontal conductivities may differ, i.e., $\sigma_{h}:=\sigma_{xx}=\sigma_{yy}$, $\sigma_{v}=\sigma_{zz}$. We use the following Fourier transform convention, $\partial_{t}\leftrightarrow-\mathrm{i}\omega$. To speed up the FDTD modelling, we transform the above system from the diffusion to the wave domain, following Mittet, (2010). The idea is to define a fictitious dielectric permittivity in equation 2 as $\sigma=2\omega_{0}\varepsilon$, yielding $\displaystyle\nabla\times\mathbf{E}-\mathrm{i}\omega\mu\mathbf{H}$ $\displaystyle=$ $\displaystyle-\mathbf{M},$ (3) $\displaystyle-\nabla\times\mathbf{H}+2\omega_{0}\varepsilon\mathbf{E}$ $\displaystyle=$ $\displaystyle-\mathbf{J},$ which gives the following relation after multiplying the second equation with $\sqrt{-\mathrm{i}\omega/2\omega_{0}}$ $\displaystyle\nabla\times\mathbf{E}+\underbrace{\sqrt{-\mathrm{i}2\omega\omega_{0}}}_{-\mathrm{i}\omega^{\prime}}\mu\underbrace{\sqrt{\frac{-\mathrm{i}\omega}{2\omega_{0}}}\mathbf{H}}_{\mathbf{H}^{\prime}}$ $\displaystyle=$ $\displaystyle-\mathbf{M},$ (4) $\displaystyle-\nabla\times\underbrace{\sqrt{\frac{-\mathrm{i}\omega}{2\omega_{0}}}\mathbf{H}}_{\mathbf{H}^{\prime}}+\underbrace{\sqrt{-\mathrm{i}2\omega\omega_{0}}}_{-\mathrm{i}\omega^{\prime}}\varepsilon\mathbf{E}$ $\displaystyle=$ $\displaystyle-\underbrace{\sqrt{\frac{-\mathrm{i}\omega}{2\omega_{0}}}\mathbf{J},}_{\mathbf{J}^{\prime}}$ which translates into the wave and simulation domain as the time dependent system $\displaystyle\nabla\times\mathbf{E}^{\prime}+\mu\partial_{t}\mathbf{H}^{\prime}$ $\displaystyle=$ $\displaystyle-\mathbf{M}^{\prime},$ (5) $\displaystyle-\nabla\times\mathbf{H}^{\prime}+\varepsilon\partial_{t}\mathbf{E}^{\prime}$ $\displaystyle=$ $\displaystyle-\mathbf{J}^{\prime}.$ We have introduced a prime to identify the fields in the wave domain. From the electromagnetic fields in the wave domain, the frequency-domain fields can be computed on the fly during modelling using the fictitious wave transformation, exemplified by the electric field here, $\displaystyle\mathbf{E}^{\prime}(\mathbf{x},\omega^{\prime})=\int_{0}^{T_{\max}}\mathbf{E}^{\prime}(\mathbf{x},t)e^{\mathrm{i}\omega^{\prime}t}\mathrm{d}t,$ (6) where $T_{\max}$ is the final time until the field $E^{\prime}(\mathbf{x},\omega^{\prime})$ reaches its steady state and where $\displaystyle\omega^{\prime}=(1+\mathrm{i})\sqrt{\omega\omega_{0}}.$ (7) In order to have results valid for the frequency domain we need to calculate the Green’s functions. We need the following relation, $\displaystyle\mathbf{E}^{\prime}$ $\displaystyle=$ $\displaystyle\mathbf{E},$ (8) $\displaystyle\mathbf{H}^{\prime}$ $\displaystyle=$ $\displaystyle\sqrt{\frac{-\mathrm{i}\omega}{2\omega_{0}}}\mathbf{H},$ $\displaystyle\mathbf{M}^{\prime}$ $\displaystyle=$ $\displaystyle\mathbf{M},$ $\displaystyle\mathbf{J}^{\prime}$ $\displaystyle=$ $\displaystyle\sqrt{\frac{-\mathrm{i}\omega}{2\omega_{0}}}\mathbf{J}.$ The Green’s functions are then obtained by normalizing the transformed electric and magnetic fields with the source current, $\displaystyle G_{kj}^{E|J}(\mathbf{x},\omega|\mathbf{x}_{s})=\frac{E_{k}(\mathbf{x},\omega|\mathbf{x}_{s})}{J_{j}(\omega)}=\sqrt{\frac{-\mathrm{i}\omega}{2\omega_{0}}}\frac{E_{k}^{\prime}(\mathbf{x},\omega|\mathbf{x}_{s})}{J_{j}^{\prime}(\omega)},$ (9) $\displaystyle G_{kj}^{H|J}(\mathbf{x},\omega|\mathbf{x}_{s})=\frac{H_{k}(\mathbf{x},\omega|\mathbf{x}_{s})}{J_{j}(\omega)}=\frac{H_{k}^{\prime}(\mathbf{x},\omega|\mathbf{x}_{s})}{J_{j}^{\prime}(\omega)},$ where $G_{kj}^{E|J}(\mathbf{x},\omega|\mathbf{x}_{s})$ and $G_{kj}^{H|J}(\mathbf{x},\omega|\mathbf{x}_{s})$ stand for the electrical and magnetic Green’s function for angular frequency $\omega$ at spatial location $\mathbf{x}$ with the source located at $\mathbf{x}_{s}$. Equation 4 is a pure wave-domain equation and the time integration can easily be discretized using the leap-frog method. We let the time be $t_{n}=n\Delta t$ with $n$ the integer time variable and $\Delta t$ the time step. We also introduce $N=n+\frac{1}{2}$ such that, $\displaystyle{\mathbf{H}^{\prime}}^{N}$ $\displaystyle=$ $\displaystyle{\mathbf{H}^{\prime}}^{N-1}+\Delta t\mu^{-1}(-\nabla\times{\mathbf{E}^{\prime}}^{n}-{\mathbf{M}^{\prime}}^{n}),$ (10) $\displaystyle{\mathbf{E}^{\prime}}^{n+1}$ $\displaystyle=$ $\displaystyle{\mathbf{E}^{\prime}}^{n}+\Delta t\epsilon^{-1}(\nabla\times{\mathbf{H}^{\prime}}^{N}-{\mathbf{J}^{\prime}}^{N}).$ The time integration of these equations is second-order accurate. It is shown in Mittet, (2010) that the calculation of the desired fields in the “real world” diffusive domain is independent of the frequency content of the source term used for calculating the fictitious fields. We exploit this fact and achieve good accuracy for the time integration by transmitting a low-frequency signal in the fictitious wave domain. Here we are concerned with the spatial part of the simulation scheme so we turn to this topic next. We use a similar notation for the space variables as for the time variables where we write $x_{i}=x_{i-1}+\Delta x_{i}$ where $\Delta x_{i}$ is the node separation between node $x_{i-1}$ and node $x_{i}$. Likewise, we assume a forward staggered grid such that $x_{I}=x_{I-1}+\Delta x_{I}$ where $\Delta x_{I}$ is the node separation between node $x_{I-1}$ and node $x_{I}$. For a uniform staggered grid we have a constant node separation such that $\Delta x_{I}=\Delta x_{i}=\Delta x$ and $I=i+\frac{1}{2}$. The $y$ and $z$ directions can be described in the same way with lower case and upper case integer arguments. Calculation of the partial derivative of the field $f(x)$ can, in the continuous case, be formulated as an integral operator by $\displaystyle\partial_{x}f(x)=\int_{-\infty}^{\infty}\mathrm{d}x^{\prime}f(x+x^{\prime})\\{-\partial_{x^{\prime}}\delta(x^{\prime})\\}=\int_{-\infty}^{\infty}\mathrm{d}x^{\prime}f(x+x^{\prime})\alpha(x^{\prime}).$ (11) The discrete formulation, with $f(x_{i}):=f(i)$, is, $\displaystyle\partial_{x}f(x_{i})\approx D_{x}f(i)=\sum_{l=-L}^{L}f(i+l)\alpha_{l}(i),$ (12) where $\alpha_{l}(i)$ is a band-limited approximation to the operator $\alpha(x^{\prime})$ in equation 11. The half length of the operator is $L$. The argument $i$ is used to explicitly show that this operator will vary with location for a nonuniform grid. For the staggered grid, we can then define discretized forward, $D_{x}^{+}$, and backward, $D_{x}^{-}$, derivative operators as, $\displaystyle D_{x}^{+}f(i)=\partial_{x}f(I)=\sum_{l=1}^{L}f(i+l)\alpha_{l}(i)-f(i-l+1)\alpha_{-l}(i),$ (13) $\displaystyle D_{x}^{-}f(I)=\partial_{x}f(i)=\sum_{l=1}^{L}f(I+l-1)\alpha_{l}(I)-f(I-l)\alpha_{-l}(I),$ which is the form we need for nonuniform grids and which is investigated here. The $f^{{}^{\prime}}$ implies spatial derivative in equation 13. The operator simplifies for uniform grids where the operator becomes independent of spatial location such that $\alpha_{l}(i)=\alpha_{-l}(i)=\alpha_{l}$ $\displaystyle D_{x}^{+}f(i)=\partial_{x}f(I)=\sum_{l=1}^{L}(f(i+l)-f(i-l+1))\alpha_{l},$ (14) $\displaystyle D_{x}^{-}f(I)=\partial_{x}f(i)=\sum_{l=1}^{L}(f(I+l-1)-f(I-l))\alpha_{l}.$ If we use $L=1$ we have that $\alpha_{1}=1/\Delta x$ and equation 14 formulates the well known second-order accurate partial derivative operations, $\displaystyle D_{x}^{+}f(i)=\partial_{x}f(I)=(f(i+1)-f(i))/\Delta x,$ (15) $\displaystyle D_{x}^{-}f(I)=\partial_{x}f(i)=(f(I)-f(I-1))/\Delta x.$ Let us assume a non-magnetic subsurface so that $\mu$ has the same value as in the vacuum, $\mathbf{M}=0$, while $\epsilon$ is a diagonal tensor $\epsilon=\mbox{diag}(\epsilon_{ii}),i=x,y,z$. The staggering is as in Mittet, (2010), $\displaystyle{{H}^{\prime}}_{x}^{N}(i,J,K),\quad{{H}^{\prime}}_{y}^{N}(I,j,K),\quad{{H}^{\prime}}_{z}^{N}(I,J,k),$ (16) $\displaystyle{{E}^{\prime}}_{x}^{n}(I,j,k),\quad{{E}^{\prime}}_{y}^{n}(i,J,k),\quad{{E}^{\prime}}_{z}^{n}(i,j,K),$ $\displaystyle{{J}^{\prime}}_{x}^{n}(I,j,k),\quad{{J}^{\prime}}_{y}^{n}(i,J,k),\quad{{J}^{\prime}}_{z}^{n}(i,j,K),$ $\displaystyle{\varepsilon}_{xx}(I,j,k),\quad{\varepsilon}_{yy}(i,J,k),\quad{\varepsilon}_{zz}(i,j,K),$ and the scheme implemented is, $\displaystyle{{H}^{\prime}}_{x}^{N}$ $\displaystyle=$ $\displaystyle{{H}^{\prime}}_{x}^{N-1}-\frac{\Delta t}{\mu}(D_{y}^{+}{{E}^{\prime}}_{z}^{n}-D_{z}^{+}{{E}^{\prime}}_{y}^{n}),$ (17) $\displaystyle{{H}^{\prime}}_{y}^{N}$ $\displaystyle=$ $\displaystyle{{H}^{\prime}}_{y}^{N-1}-\frac{\Delta t}{\mu}(D_{z}^{+}{{E}^{\prime}}_{x}^{n}-D_{x}^{+}{{E}^{\prime}}_{z}^{n}),$ $\displaystyle{{H}^{\prime}}_{z}^{N}$ $\displaystyle=$ $\displaystyle{{H}^{\prime}}_{z}^{N-1}-\frac{\Delta t}{\mu}(D_{x}^{+}{{E}^{\prime}}_{y}^{n}-D_{y}^{+}{{E}^{\prime}}_{x}^{n}),$ $\displaystyle{{E}^{\prime}}_{x}^{n+1}$ $\displaystyle=$ $\displaystyle{{E}^{\prime}}_{x}^{n}+\frac{\Delta t}{\epsilon_{xx}}(D_{y}^{-}{{H}^{\prime}}_{z}^{N}-D_{z}^{-}{{H}^{\prime}}_{y}^{N}-{{J}^{\prime}}_{x}^{N}),$ $\displaystyle{{E}^{\prime}}_{y}^{n+1}$ $\displaystyle=$ $\displaystyle{{E}^{\prime}}_{y}^{n}+\frac{\Delta t}{\epsilon_{yy}}(D_{z}^{-}{{H}^{\prime}}_{x}^{N}-D_{x}^{-}{{H}^{\prime}}_{z}^{N}-{{J}^{\prime}}_{y}^{N}),$ $\displaystyle{{E}^{\prime}}_{z}^{n+1}$ $\displaystyle=$ $\displaystyle{{E}^{\prime}}_{z}^{n}+\frac{\Delta t}{\epsilon_{zz}}(D_{x}^{-}{{H}^{\prime}}_{y}^{N}-D_{y}^{-}{{H}^{\prime}}_{x}^{N}-{{J}^{\prime}}_{z}^{N}).$ The conventional second-order staggered grid FDTD scheme discretizes the spatial derivatives as follows (Newman and Alumbaugh,, 1995; Mulder,, 2006): $\displaystyle D_{y}^{-}H^{\prime}_{z}=\frac{H^{\prime}_{z}(I,J,k)-H^{\prime}_{z}(I,J-1,k)}{\Delta y_{J}},\quad D_{z}^{-}H^{\prime}_{y}=\frac{H^{\prime}_{y}(I,j,K)-H^{\prime}_{y}(I,j,K-1)}{\Delta z_{K}},$ $\displaystyle D_{z}^{-}H^{\prime}_{x}=\frac{H^{\prime}_{x}(i,J,K)-H^{\prime}_{x}(i,J,K-1)}{\Delta z_{K}},\quad D_{x}^{-}H^{\prime}_{z}=\frac{H^{\prime}_{z}(I,J,k)-H^{\prime}_{z}(I-1,J,k)}{\Delta x_{I}},$ $\displaystyle D_{x}^{-}H^{\prime}_{y}=\frac{H^{\prime}_{y}(I,j,K)-H^{\prime}_{y}(I-1,j,K)}{\Delta x_{I}},\quad D_{y}^{-}H^{\prime}_{x}=\frac{H^{\prime}_{x}(i,J,K)-H^{\prime}_{x}(i,J-1,K)}{\Delta y_{J}},$ $\displaystyle D_{y}^{+}E^{\prime}_{z}=\frac{E^{\prime}_{z}(i,j+1,K)-E^{\prime}_{z}(i,j,K)}{\Delta y_{j}},\quad D_{z}^{+}E^{\prime}_{y}=\frac{E^{\prime}_{y}(i,J,k+1)-E^{\prime}_{y}(i,J,k)}{\Delta z_{k}},$ $\displaystyle D_{z}^{+}E^{\prime}_{x}=\frac{E^{\prime}_{x}(I,j,k+1)-E^{\prime}_{x}(I,j,k)}{\Delta z_{k}},\quad D_{x}^{+}E^{\prime}_{z}=\frac{E^{\prime}_{z}(i+1,j,K)-E^{\prime}_{z}(i,j,K)}{\Delta x_{i}},$ $\displaystyle D_{x}^{+}E^{\prime}_{y}=\frac{E^{\prime}_{y}(i+1,J,k)-E^{\prime}_{y}(i,J,k)}{\Delta x_{i}},\quad D_{y}^{+}E^{\prime}_{x}=\frac{E^{\prime}_{x}(I,j+1,k)-E^{\prime}_{x}(I,j,k)}{\Delta y_{j}},$ (18) where $\Delta x_{i}$, $\Delta y_{j}$ and $\Delta z_{k}$ are distances between nodes on the reference grid, while $\Delta x_{I}$, $\Delta y_{J}$ and $\Delta z_{K}$ are the distance between the grid points $(I,j,k)$ and $(I-1,j,k)$, the distance between the grid points $(i,J,k)$ and $(i,J-1,k)$, and the distance between the grid points $(i,j,K)$ and $(i,j,K-1)$. This scheme is second-order accurate on a uniform grid with $\Delta x_{i}=\Delta x_{I}$ and likewise for the other spatial directions. The scheme is consistent with using equation 14 with $L=1$ to approximate the derivatives. A standard discretization method for nonuniform grids is to use the same formulation as above, but where the node distance may vary along the same spatial direction. It is well known that there are accuracy issues with this implementation. To achieve second-order accuracy using equation 14 is not possible. The midpoints between nodes do not align after going from the reference grid to the staggered grid and back again. Since $\Delta x_{i}\neq\Delta x_{I}$, we have a situation where the cell center does not match on a nonuniform grid, as is illustrated in Figure 1b. Consequently, the local truncation error of the resulting scheme can only reach first order. Reduced accuracy will also be a problem if we implement a high-order FDTD scheme ($L>1$) on a nonuniform grid, using derivative-operator coefficients designed for a regular grid. This is unfortunate since the nonuniform grid is potentially attractive for efficient modelling due to significant reduction of the number of grid point. However, good accuracy can be restored if equation 13 is used instead of equation 14. The problem that remains is to calculate the operator coefficients for equation 13. ### 2.1 Vandermonde matrix The major difference between FDTD implementations on a uniform grid and on a nonuniform grid lies in the design of the spatial-derivative operator coefficients. The position of each field component on the nonuniform staggered grid has been illustrated in Figure 2, which is similar to the staggered FDTD on a uniform grid. Figure 2: Electrical and magnetic fields on staggered grid (Yee,, 1966). In order to compute the electromagnetic field as well as its derivatives with arbitrary grid spacing, we have to do a polynomial interpolation using a number of knots $x_{0},x_{1},\cdots,x_{n}$. According to the Taylor expansion, we have $\displaystyle f(x_{i})$ $\displaystyle=$ $\displaystyle f(x)+f^{\prime}(x)(x_{i}-x)+\frac{1}{2}f^{\prime\prime}(x)(x_{i}-x)^{2}+\cdots+\frac{1}{n!}f^{(n)}(x)(x_{i}-x)^{n}+\cdots$ (19) $\displaystyle i$ $\displaystyle=$ $\displaystyle 0,1,\cdots,n.$ By defining $a_{i}(x):=f^{(i)}(x)/i!$, we end up with a polynomial of the Newton form, $\displaystyle f(x_{i})$ $\displaystyle=$ $\displaystyle a_{0}(x)+a_{1}(x)(x_{i}-x)+a_{2}(x)(x_{i}-x)^{2}+\cdots+a_{n}(x)(x_{i}-x)^{n}+\cdots$ (20) $\displaystyle i$ $\displaystyle=$ $\displaystyle 0,1,\cdots,n.$ Let us consider $n+1$ distinct nodes $x_{0},x_{1},\cdots,x_{n}$ and drop the terms $O((x_{i}-x)^{n+1})$. This builds a matrix system $\displaystyle\underbrace{\begin{pmatrix}f(x_{0})\\\ f(x_{1})\\\ \vdots\\\ f(x_{n})\end{pmatrix}}_{\mathbf{f}}=\underbrace{\begin{pmatrix}1&x_{0}-x&(x_{0}-x)^{2}&\cdots&(x_{0}-x)^{n}\\\ 1&x_{1}-x&(x_{1}-x)^{2}&\cdots&(x_{1}-x)^{n}\\\ \cdots\\\ 1&x_{n}-x&(x_{n}-x)^{2}&\cdots&(x_{n}-x)^{n}\\\ \end{pmatrix}}_{\mathbf{V}^{T}(x_{0}-x,\cdots,x_{n}-x)}\underbrace{\begin{pmatrix}a_{0}(x)\\\ a_{1}(x)\\\ \vdots\\\ a_{n}(x)\end{pmatrix}}_{\mathbf{a}},$ (21) where $\mathbf{V}^{T}(x_{0}-x,\cdots,x_{n}-x)$ is the transpose of a Vandermonde matrix determined by $x_{0}-x,\cdots,x_{n}-x$. The above expression implies that the function $f(x)$ and its derivatives up to the $n$-th order at arbitrary location $x$ can be found by inverting the Vandermonde matrix: $(f(x),f^{\prime}(x),\cdots,f^{(n)}(x)/n!)^{T}=\mathbf{a}=[\mathbf{V}^{T}]^{-1}\mathbf{f}$. It is well known that the Vandermonde matrix is highly ill-conditioned and direct matrix inversion by Gaussian elimination should be avoided due to numerical instabilities when the matrix size becomes large. Fortunately, there exists an efficient algorithm based on the method of Björck and Pereyra, (1970) to invert the Vandermonde matrix. In fact, the algorithm circumvents the curse of severe ill-conditioning of the Vandermonde matrix to arrive at arbitrarily high accuracy for the inversion (Demmel and Koev,, 2005). Compared with Gauss-elimination of complexity $O(n^{3})$ , the Vandermonde matrix inversion algorithm reduces the computational complexity to $O(n^{2})$. Because the elements of the Vandermonde matrix are fully determined by the interpolation nodes, there is no need to explicitly construct the matrix and store it before inversion. The detailed implementation of this fast algorithm is available in Golub, (1996, Algorithm 4.6.1). Let the $i$-th row, $j$-th column of the inverse matrix $[\mathbf{V}^{T}]^{-1}$ be $w_{ij}$, i.e., $([\mathbf{V}^{T}]^{-1})_{ij}=w_{ij},i,j=0,\cdots,n$. It also follows that $\displaystyle\underbrace{\begin{pmatrix}a_{0}(x)\\\ a_{1}(x)\\\ \vdots\\\ a_{n}(x)\end{pmatrix}}_{\mathbf{a}}:=\underbrace{\begin{pmatrix}w_{00}&w_{01}&\cdots w_{0n}\\\ w_{10}&w_{11}&\cdots w_{1n}\\\ \vdots\\\ w_{n0}&w_{n1}&\cdots w_{nn}\\\ \end{pmatrix}}_{[\mathbf{V}^{T}]^{-1}}\underbrace{\begin{pmatrix}f(x_{0})\\\ f(x_{1})\\\ \vdots\\\ f(x_{n})\end{pmatrix}}_{\mathbf{f}}.$ (22) The $i$-th row gives the explicit expression to find the $i$-th derivative $\displaystyle\frac{1}{i!}f^{(i)}(x)=a_{i}(x)=\sum_{j=0}^{n}w_{ij}f(x_{j})=\langle w_{i\cdot}|\mathbf{f}\rangle,\quad i=0,\cdots,n.$ (23) The first row of the matrix $[\mathbf{V}^{T}]^{-1}$ (i.e., $w_{0j}$, $j=0,\cdots,n$) is $f(x)$ which in analogy with equation 11 can be written in discrete form as $\displaystyle f(x)=\int_{-\infty}^{\infty}\mathrm{d}x^{\prime}f(x+x^{\prime})\\{\delta(x^{\prime})\\}=\int_{-\infty}^{\infty}\mathrm{d}x^{\prime}f(x+x^{\prime})\beta(x^{\prime}).$ (24) The discrete formulation is, $\displaystyle f(x_{i})=\sum_{l=-L}^{L}f(i+l)\beta_{l}(i),$ (25) where $\beta_{l}(i)$ are the coefficients for an interpolation operator adapted to a nonuniform grid and identical to the $w_{0j}$ coefficients. The second row of the matrix $[\mathbf{V}^{T}]^{-1}$ (i.e., $w_{1j}$, $j=0,\cdots,n$) is $\partial_{x}f(x)$ and the continuous and discrete representations are given in equations 11 and 12. The $\alpha_{l}(i)$ coefficients are for a derivative operator adapted to a nonuniform grid and identical to the $w_{1j}$ coefficients. To be explicit, consider the staggered finite-difference approximation of the first derivatives in $x$ direction using $2L$ non-equidistant nodes. The finite-difference coefficients $\alpha_{l}(x_{i})$ and $\alpha_{l}(x_{I})$, $l=-L+1,\cdots,L$ are the 2nd row of the inverse of the matrices $\mathbf{V}^{T}(x_{i+L}-x_{I},\cdots,x_{i-L+1}-x_{I})$ and $\mathbf{V}^{T}(x_{I+L}-x_{i},\cdots,x_{I-L+1}-x_{i})$. Using the $2L$ nodes, we achieve accuracy up to $2L$-th order in space. In general we find that the operator coefficients (interpolation weights) for $f^{(i)}(x)$ are $i!w_{ij}$. Given the points $x_{0},\cdots,x_{n}$ and $x$, the Vandermonde matrix is determined and the operator coefficients can be calculated. For the simulation we only need the derivative-operator coefficients. It is noteworthy to mention that the Vandermonde matrix must be non-singular to be inverted. For the derivative-operator coefficients we have that the Vandermonde matrix is non-singular by construction of the staggered grids. A finite-difference scheme on a staggered grid implies that the node $x_{I}$, whose derivative is computed, will stay in the middle between the selected nodes $x_{i}$, which ensures that the resulting Vandermonde matrix is invertible. The interpolation operators are useful for recording fields at arbitrary locations and for the distribution of source contributions (Mittet,, 2017). Just as for the derivative-operator coefficients, the interpolation coefficients can be pre-calculated and then reused every time step. For the interpolation operator we may find that some of the coordinates $x_{i}$ matches the interpolation points $x$, the inversion of Vandermonde matrix is then not necessary. This might happen when the source or receiver positions coincide with a finite-difference node. In this case, the interpolation weights $w_{0i}$ to evaluate $f(x)$ should be exactly 1 at $x_{i}$ and $0$ elsewhere. In order to do 2D/3D simulation on nonuniform grid, multidimensional interpolant is simply constructed by tensor products of many 1D interpolants. The above procedure is significant as it allows us to use arbitrarily high- order finite-difference scheme to accurately compute the electromagnetic fields and their derivatives, typically with arbitrary grid spacing in the rectilinear grid. This opens the door for CSEM modelling using high-order FDTD on a nonuniform grid in a consistent framework. The computed finite-difference coefficients may also be used to do high-order frequency-domain modelling on a nonuniform grid, while the resulting sparse banded matrix has to be solved accurately if sufficient computational resources are available. ### 2.2 Stability condition Let us write down the FDTD scheme in equation 10 without source terms as follows: $\displaystyle\begin{cases}\mathbf{E}^{{}^{\prime}n+1}=\mathbf{E}^{{}^{\prime}n}+\Delta t\epsilon^{-1}\nabla\times\mathbf{H}^{{}^{\prime}N}\\\ \mathbf{H}^{{}^{\prime}N+1}=\mathbf{H}^{{}^{\prime}N}-\Delta t\mu^{-1}\nabla\times\mathbf{E}^{{}^{\prime}n+1}\end{cases},$ (26) leading to $\displaystyle\begin{split}\begin{bmatrix}\mathbf{E}^{{}^{\prime}n+1}\\\ \mathbf{H}^{{}^{\prime}N+1}\end{bmatrix}=\underbrace{\begin{bmatrix}\mathbf{I}&\Delta t\epsilon^{-1}\nabla\times\\\ -\Delta t\mu^{-1}\nabla\times&\mathbf{I}-\Delta t^{2}\epsilon^{-1}\mu^{-1}\nabla\times\nabla\times\end{bmatrix}}_{\mathbf{A}}\begin{bmatrix}\mathbf{E}^{{}^{\prime}n}\\\ \mathbf{H}^{{}^{\prime}N}\end{bmatrix}.\end{split}$ (27) The numerical stability requires the eigenvalues of the amplification matrix $\mathbf{A}$ to be less than or equal to 1. Assume the eigenvalue decomposition for the amplification matrix is $\mathbf{A}=\mathbf{\bar{V}}\mathbf{\Lambda}\mathbf{\bar{V}}^{T}$, where $\mathbf{\bar{V}}=(\mathbf{V}_{E},\mathbf{V}_{H})^{T}$ is an unitary matrix such that $\mathbf{\bar{V}}^{T}\mathbf{\bar{V}}=\mathbf{I}$. Then we have $\mathbf{A}\mathbf{V}=\mathbf{V}\mathbf{\Lambda}$, yielding $\displaystyle\begin{bmatrix}\mathbf{I}&\Delta t\epsilon^{-1}\nabla\times\\\ -\Delta t\mu^{-1}\nabla\times&\mathbf{I}-\Delta t^{2}\epsilon^{-1}\mu^{-1}\nabla\times\nabla\times\end{bmatrix}\begin{bmatrix}\mathbf{V}_{E}\\\ \mathbf{V}_{H}\end{bmatrix}=\begin{bmatrix}\mathbf{V}_{E}\\\ \mathbf{V}_{H}\end{bmatrix}\Lambda.$ (28) That is, $\displaystyle\begin{cases}\Delta t\epsilon^{-1}\mu^{-1}\nabla\times\mathbf{V}_{H}=\mathbf{V}_{E}(\mathbf{\Lambda}-\mathbf{I})\\\ -\Delta t\mu^{-1}\nabla\times\mathbf{V}_{E}-\Delta t^{2}\epsilon^{-1}\mu^{-1}\nabla\times\nabla\times\mathbf{V}_{H}=\mathbf{V}_{H}(\mathbf{\Lambda}-\mathbf{I}).\end{cases}$ (29) Multiplying the second sub equation $\mathbf{\Lambda}-\mathbf{I}$ from the right and inserting the first sub equation gives $-\Delta t^{2}\mu^{-1}\epsilon^{-1}\nabla\times\nabla\times\mathbf{V}_{H}\mathbf{\Lambda}=\mathbf{V}_{H}(\mathbf{\Lambda}-\mathbf{I})^{2}.$ (30) Denote $\mathbf{V}_{H,j}$ the $j$th column of $\mathbf{V}_{H}$ and $\lambda_{j}$ the $j$th eigenvalue in $\mathbf{\Lambda}$. The above equation reads $\mu^{-1}\epsilon^{-1}\nabla\times\nabla\times\mathbf{V}_{H,j}=-\frac{(\lambda_{j}-1)^{2}}{\Delta t^{2}\lambda_{j}}\mathbf{V}_{H,j},$ (31) which shows that $-\frac{(\lambda_{i}-1)^{2}}{\Delta t^{2}\lambda_{i}}$ is the eigenvalue of the matrix $(\mu\epsilon)^{-1}\nabla\times\nabla\times$ associated with the eigenvector $V_{H,j}$. This leads to $(\lambda_{j}^{2}+(-2+\Delta t^{2}c^{2}\nabla\times\nabla\times)\lambda_{j}+1)\mathbf{V}_{H,j}=(\lambda_{j}^{2}-(2+\Delta t^{2}c^{2}\Delta)\lambda_{j}+1)\mathbf{V}_{H,j}=0,$ (32) where we denote $c:=1/\sqrt{\mu\epsilon}$ and have applied $\nabla\times\nabla\times\mathbf{F}=\nabla\nabla\cdot\mathbf{F}-\nabla\cdot\nabla\mathbf{F}=-\Delta\mathbf{F}$ due to Gauss theorem $\nabla\cdot\mathbf{F}=0$, $\mathbf{F}=\mathbf{E},\mathbf{H}$ in the homogeneous, source free medium. The roots of the above equation are $\lambda_{j;1,2}=1+\frac{\Delta t^{2}c^{2}\Delta}{2}\pm\frac{\mathrm{i}}{2}\sqrt{-\Delta t^{2}c^{2}\Delta(4+\Delta t^{2}c^{2}\Delta)},$ (33) which requires the following condition to be satisfied $0\leq-\Delta t^{2}c^{2}\Delta\leq 4,$ (34) in order to ensure $|\lambda_{j;1,2}|\leq 1$. Finally, we arrive at the same stability condition as equation 41 of Mittet, (2010), $\Delta tc_{\max}\sqrt{(D_{x}^{\max})^{2}+(D_{y}^{\max})^{2}+(D_{z}^{\max})^{2}}\leq 2,$ (35) where $D_{x}^{\max}$, $D_{y}^{\max}$ and $D_{z}^{\max}$ are the maximum value of the the discretized first derivatives along $x$, $y$ and $z$ directions. Let us emphasize this condition applies to both uniform and nonuniform grid. The difference lies in the spatial derivative operator. To proceed with the stability analysis, we represent the fields on the grid via time harmonic plane waves $\mathbf{E}^{\prime},\mathbf{H}^{\prime}\propto e^{-\mathrm{i}(\omega t-k_{x}x-k_{y}y-k_{z}z)},$ (36) where the amplitude has been omitted. Equation 13 becomes $\begin{cases}D_{x}^{+}u(x_{I})=(\alpha_{L}(x_{I})e^{\mathrm{i}k_{x}(x_{i+L}-x_{I})}+\alpha_{L-1}(x_{I})e^{\mathrm{i}k_{x}(x_{i+L-1}-x_{I})}+\cdots+\alpha_{-L+1}(x_{I})e^{\mathrm{i}k_{x}(x_{i-L+1}-x_{I})})u(x_{I})\\\ D_{x}^{-}u(x_{i})=(\alpha_{L}(x_{i})e^{\mathrm{i}k_{x}(x_{I+L-1}-x_{i})}+\alpha_{L-1}(x_{i})e^{\mathrm{i}k_{x}(x_{I+L-1}-x_{i})}+\cdots+\alpha_{-L+1}(x_{i})e^{\mathrm{i}k_{x}(x_{I-L}-x_{i})})u(x_{i})\end{cases}.$ (37) Hence, we end up with the maximum possible values for discrete first derivative operators $D_{x}^{\max}=\max\left(\sum_{l=-L+1}^{L}|\alpha_{l}(x_{I})|,\sum_{l=-L+1}^{L}|\alpha_{i}(x_{l})|\right)$ (38) and similar estimations for $D_{y}^{\max}$ and $D_{z}^{\max}$. In case of a uniform grid, the above expressions becomes much simpler $\begin{cases}D_{x}^{+}u(x_{I})=(\alpha_{L}e^{\mathrm{i}k_{x}(L-1/2)\Delta x}+\alpha_{L-1}e^{\mathrm{i}k_{x}(L-3/2)\Delta x}+\cdots+\alpha_{-L+1}e^{\mathrm{i}k_{x}(-L+1/2)\Delta x}u(x_{I})\\\ D_{x}^{-}u(x_{i})=(\alpha_{L}e^{\mathrm{i}k_{x}(L-1/2)\Delta x}+\alpha_{L-1}e^{\mathrm{i}k_{x}(L-3/2)\Delta x}+\cdots+\alpha_{-L+1}e^{\mathrm{i}k_{x}(-L+1/2)\Delta x}u(x_{i})\\\ \end{cases},$ where $\Delta x$ stands for the uniform grid spacing in $x$ direction, while the coefficients $\alpha_{l}$ (which can be computed by inverting a Vandermonde matrix system according to Appendix AComputing uniform staggered- grid finite difference coefficients via Vandermonde matrix inversion) are independent of the location $x_{i}$. ### 2.3 Grid stretching Our finite-difference modelling is carried out on a rectilinear mesh, which can be generated from the tensor (outer) product of 1D non-equispaced meshes. We use the geometrical progression to generate the 1D nonuniform grid, following the work of Mulder, (2006, Appendix C). This is also often referred to as power law grid stretching since the cell sizes stretch exponentially to guarantee a smooth extension of the grid. Assume we have the total grid length $L_{x}$ divided into $n$ intervals ($n+1$ nodes) with a common ratio $r>1$. Denote the smallest interval $\Delta x=x_{1}-x_{0}$. Thus, the relation between $L_{x}$ and $\Delta x$ is $L_{x}=(x_{1}-x_{0})+(x_{2}-x_{1})+\cdots+(x_{n}-x_{n-1})=\Delta x(1+r+\cdots+\cdot r^{n-1})=\Delta x\frac{r^{n}-1}{r-1}.$ (39) Given the total distance $L_{x}$, the smallest grid spacing $\Delta x$ and the stretching factor $r$, we can compute an approximate value for the number of nodes $n=\left\lceil\frac{\ln(1+\frac{L_{x}}{\Delta x}(r-1))}{\ln(r)}\right\rceil$ following Mulder, (2006), where $\left\lceil\cdot\right\rceil$ takes the ceiling integer value. This strategy yields approximate solution as the value of $L_{x}$ is not exactly preserved. Due to the stability requirement and the resulting computational cost in the modelling, we are restricted to the smallest interval $\Delta x$ and a given number of intervals $n$ to discretize over a certain distance $L_{x}$. The question boils down to finding an optimal growth factor $r$. This problem is more complicated since equation 39 does not yield an explicit expression for the stretching factor $r$. The relation in equation 39 is equivalent to $r=\underbrace{\left(\frac{L_{x}}{\Delta x}(r-1)+1\right)^{\frac{1}{n}}}_{g(r)},$ (40) which inspires us to carry out a number of fixed point iterations until convergence: $r^{k+1}=g(r^{k}),\quad k=0,1,\cdots.$ (41) Assume $r^{*}$ is the analytic solution such that $r^{*}=g(r^{*})$. Thanks to Lagrange mean value theorem, the error estimation at ($k+1$)-th iteration is linked with the error at $k$-th iteration via $|e^{k+1}|=|r^{k+1}-r^{*}|=|g(r^{k})-g(r^{*})|=|g^{\prime}(\xi)(r^{k}-r^{*})|=|g^{\prime}(\xi)||e^{k}|,\quad\xi\mbox{ between }r^{k}\mbox{ and }r^{*}.$ (42) It becomes evident that $e^{k}\rightarrow 0\;(k\rightarrow\infty)$ provided that $|g^{\prime}(r)|<1$. Starting from any initial guess $r^{0}>1$, the fixed point iteration scheme in equation 41 is guaranteed to converge since $|g^{\prime}(r)|=\frac{L_{x}}{n\Delta x}\left(\frac{L_{x}}{\Delta x}(r-1)+1\right)^{1/n-1}=\frac{1}{n}(r^{-(n-1)}+r^{-(n-2)}+\cdots+1)<1,$ thanks to the relations in equations 39 and 40. We note that within areas of constant $r$, the derivative-operator coefficients in equation 13 can be calculated from the $2L$ coefficients $a_{-l}$ and $a_{l}$, such that $\displaystyle\alpha_{-l}(i)$ $\displaystyle=$ $\displaystyle\frac{a_{-l}}{\Delta x\>r^{i}},$ (43) $\displaystyle\alpha_{l}(i)$ $\displaystyle=$ $\displaystyle\frac{a_{l}}{\Delta x\>r^{i}}.$ Two solutions of the Vandermonde system are required in this case, one for operators valid on the reference grid and one for operators valid on the staggered grid. ### 2.4 Implementation A number of techniques have been applied to achieve efficient and accurate 3D CSEM simulation. According to equation 6, the frequency-domain CSEM response is obtained by a time to frequency transform. The transform is using the complex frequency given in equation 7. This gives exponential damping of late arrivals as is discussed in more detail in Mittet, (2015). The lowest frequency experiences the least damping in this transform and by that requires the longest simulation time. This allows us to bound the number of times steps to terminate the simulation when the lowest frequency component has converged. The convergence means that the frequency-domain field obtained by time integration has reached its steady state and later arrivals are damped to the degree that they do not contribute to the time integral. The source and receiver locations may be arbitrarily distributed over the whole computational domain, not necessarily located at the nodes of the finite-difference grid. In case they do, interpolation is not required and we directly take the field from the grid; otherwise, we need interpolation operators extracted from the first row of the relevant $[\mathbf{V}^{T}]^{-1}$ matrices. The CFL condition to achieve stable FDTD modelling dictates the timestep for a given spatial sampling. The air-wave travels at extremely high speed which does not allow us to use the local finite-difference stencil to simulate it. The air-water boundary condition is implemented in the Fourier-wavenumber domain following the method proposed by Oristaglio and Hohmann, (1984) and Wang and Hohmann, (1993) and using the extension to high-order schemes given in Mittet, (2010). To mimic wave propagation in unbounded domain, we use a truncated computation mesh surrounded by an artificial absorbing boundary using convolutional perfectly matched layer (PML) technique (Roden and Gedney,, 2000), except the top air-water interface. To ease the implementation, we extend the domain with equal spacing based on the last finite-difference cell size in the interior domain. Our CPML implementation is very standard using all the parameter setups given in Komatitsch and Martin, (2007). ## 3 Numerical examples We now present three examples to demonstrate the merits of high-order finite differences using nonuniform staggered grids. The first two examples are using 1D resistivity models under shallow water and deep water scenarios. In the 1D case, the semi-analytic solution can be computed in the frequency-wavenumber domain as a reference to benchmark our results. The third example takes into account varying seafloor topography, in which a reference solution can be computed using `emg3d` software (Werthmüller et al.,, 2019). In all of our modelling, we use x-directed electrical dipole source. The $E_{x}$ component at three commonly used frequencies - 0.25 Hz, 0.75 Hz and 1.25 Hz, are modelled. In total 12 layers of CPML are sufficient to achieve nearly perfect absorbing effect. The computed electromagnetic fields are normalized by the source current. The convergence check has been conducted every 100 time steps to avoid modelling after the frequency-domain EM fields stop evolving. We examine the amplitude error by inspecting the ratio of the modelled field to the reference solution, $|E_{x}^{FD}|/|E_{x}^{ref}|$, which should be close to unity if the modelling is precise. The phase difference is computed by $\angle E_{x}^{FD}-\angle E_{x}^{ref}$ in degrees. ### 3.1 Moving to high-order schemes The model shown in Figure 3 includes 5 layers: the top is the air, then a 325 m layer of sea water with a resistivity of 0.3 $\Omega$m, followed by three layers of formation with increasing resistivity in the depth direction. The whole model extends down to 5 km depth below the sea surface. A horizontal electrical dipole source is deployed at 275 m water depth, its lateral position being in the middle of the model. The EM fields are recorded using 201 receiver positions at the seafloor. The offsets range from -10 km to +10 km (receiver separation equals 100 m). Figure 3: The resistivity model with air ($\rho=10^{12}$ $\Omega$m), shallow column (325 m) of 0.3 $\Omega$m and 3 sediment layers ($\rho=1$ $\Omega$m in [325, 1025] m; $\rho=2$ $\Omega$m in [1025, 1525] m; $\rho=4$ $\Omega$m downwards). We first compare the 3D CSEM response simulated by FDTD with the reference solution calculated using `empymod` program (Werthmüller,, 2017). To limit the factors affecting the modelling accuracy, this experiment has been done using isotropic modeling and uniform grid spacing ($\Delta x=\Delta y=150$ m, $\Delta z=50$ m). Panels (a) and (b) in Figure 4 shows the amplitude and the phase of the modelled EM fields by FDTD of 2nd order, which is highly consistent to the reference solution. At the offset beyond 1 km, the agreement between finite-difference and the analytical solution is good. The kink in the transition of the near-to-far offset in the amplitude response for all frequencies is a manifestation of the strong air-wave effect. In the very near field, where the receivers are located just below the transmitter, the phase exhibits a $180^{o}$ jump due to the change in the direction of the electric field immediately below the electric dipole source. Due to the extension of finite-difference stencil, this phase rollover becomes smeared compared to the reference solution. For the same reason, finite difference method introduces significant errors in the near field in amplitude (which becomes more evident for high-order schemes). Fortunately it is not an issue for practical 3D CSEM applications since the CSEM inversion to deduce the subsurface resistivity is mainly driven by far offset refractions. We thus focus on the analysis of the amplitude and phase error beyond 1 km offset in the following. Figure 4: Comparison between 2nd order FDTD (solid line) and reference solution (dash line) for 3D CSEM simulation in the shallow water scenario. The horizontal coordinates are offsets, while the vertical coordinates are (a) Amplitude; (b) Phase. The modelling results by FDTD of higher orders are not displayed in Figure 4, since they are visually very similar to the reference solution. Instead, the amplitude and phase error are computed using 1D reference solution to examine the accuracy of our methods. In Figure 5a, c and e, we clearly see that the 2nd order FDTD gives the largest amplitude error for all frequencies; moving from 2nd order to 4th order significantly reduces the amplitude error, while moving to 6th order behaves even better, although the accuracy improvement becomes less. The phase errors exhibit a similar behavior. Figure 5: The amplitude and phase error of FDTD compared with 1D reference solution for 0.25 Hz (a,b), 0.75 Hz (c,d) and 1.25 Hz (e,f). In principle, the computational cost of the 4th order and the 6th order finite differences will be double and triple of that for the 2nd order scheme. Table 1 lists the CPU time for these modelling exercises, running on a laptop possessing Intel(R) Core(TM) i7-4710HQ CPU @ 2.50GHz. It shows the increase of the computing cost by increasing the order of FDTD scheme. The computing time of the 4th order scheme is less than two times the computing time of the 2nd order scheme. Further increase of the FDTD order results in significant increase of CPU time. Since moving to the 6th order scheme demands more computation while the accuracy improvement is marginal, we stay with the 4th order scheme from now on, as it gives the best compromise between increased accuracy and computational load. Table 1: Comparison of computing time using FDTD of orders 2, 4 and 6. FDTD | Order-2 | Order-4 | Order-6 ---|---|---|--- Time (s) | 422.3 | 715.5 | 1090.9 ### 3.2 The impact of nonuniform grid The above example demonstrates the importance of higher order scheme to achieve accurate CSEM modeling in the presence of strong air-wave. We now examine the impact of grid non uniformity in achieving computational efficiency for high-order staggered grid FDTD. To get rid of the impact of air-wave, we consider a resistivity model in deep water scenario. As shown in Figure 6a, the model has 1020 m water column of 0.3 $\Omega$m, followed by formation of $1\Omega\mbox{m}$ down to 1900 m, and 120 m thickness of resistor of 50 $\Omega$m. The background resistivity below the resistor is 2.5 $\Omega$m. To mimic a vertical transverse isotropic (VTI) Earth, all layers below the seabed are assigned with an anisotropy ratio (defined as $\lambda=\rho_{v}/\rho_{h}$ in this paper) $\lambda=1.5$. The source is placed in the middle of the model, 40 m above the receivers sitting on the seabed. The resistivity in the water column and the formation above 1200 m was discretized with constant grid spacing: $\Delta x=\Delta y=150$ m, $\Delta z=40$ m. From 1200 m down to the bottom of the resistivity model, the grid has been stretched with different growing factors. Figure 6: (a) The resistivity model in deep water: the 1020 m water column of 0.3 $\Omega$m followed by formation of $1\Omega\mbox{m}$ down to 1900 m, and 120 m thickness of resistor of 50 $\Omega$m, while the background resistivity below the resistor is 2.5 $\Omega$m; (b) VTI anisotropy below seabed is 1.5. Figure 7a displays the stretching factor $r$ of the nonuniform grid for different number of grid points $n_{z}$. Note that the cell size grows very fast with the factor $r$. For example, with $r=1.05$ after 40 cells we obtain $r^{40}>7$ times the size of the first cell. Let us now increase $r$ (hence decrease $n_{z}$ correspondingly) and analyze how it changes the total modelling time for the resistivity model of the same physical length in $z$ direction. Figure 7b displays a significant decrease of the modelling time associated with the reduction of the number of grid points $n_{z}$. We indeed see that modelling using $n_{z}=66$ takes almost half of the simulation time compared to modelling using $n_{z}=126$. Figure 7: (a) The exponentially growing factor with the decreasing number of grid points $n_{z}$; (b) With the decreasing number of grid points $n_{z}$, the modelling time decreases dramatically. Figure 8 overlays the FDTD modelled EM fields with the reference solution. We see a very good agreement between the two in terms of both the amplitude and phase. To gain an idea of the modelling accuracy reduction when using less computing effort with nonuniform grid, we plotted the amplitude and phase errors in Figure 9 corresponding to three different frequencies. These figures clearly show that both the amplitude error and the phase error increase with the increase of the frequencies. It is also interesting to note that stretching the grid does not necessarily increase the amplitude error, but does increase the phase discrepancy. This highlights the importance of combination in examing both amplitude and phase. Figure 8: Comparison of (a) amplitude and (b) phase between 4th order FDTD (solid line) and reference solution (dash line) for 3D CSEM simulation in the deep water scenario. Figure 9: The amplitude and phase error of FDTD compared with 1D reference solution for 0.25 Hz (a,b), 0.75 Hz (c,d) and 1.25 Hz (e,f) in deep water. ### 3.3 3D modelling with seafloor bathymetry The above 1D example highlights the importance of nonuniform grid in combination with high-order schemes to achieve efficient numerical modelling with sufficient accuracy. Here we consider a more realistic 3D resistivity model with bathymetry variations in horizontal directions, as shown in Figure 10a. The model has seawater of 0.3 $\Omega$m, followed by 2 formation layers of resistivity - 1 $\Omega$m and 2 $\Omega$m. A resistor of 50 $\Omega$m was buried in the last formation layer to mimic a hydrocarbon bearing formation located between 1800 m and 2000 m in depth, with the offset expanding from -3000 m to 3000 m in both x and y directions. Figure 10: The 3D resistivity model with seafloor bathymetry, followed by 2 formation layers of resistivity values 1 $\Omega$m and 2 $\Omega$m. The last sediment layer includes a strong resistor of 100 $\Omega$m mimicking a hydrocarbon bearing formation located within the depth range [1800, 2000] m and the offset range [-3000, 3000] m in both x and y directions. The source was placed at the center of the model while 101 receivers with 200 m spacing are deployed along x direction. In order to validate the modelling accuracy of the proposed method, we simulated a reference solution by finite integration method in frequency domain using `emg3d` (Werthmüller et al.,, 2019) The finite-integration modelling extends the model tens of kilometers in each direction, to mimic that the EM fields propagate to very far distance while avoiding possible edge/boundary effects. In our finite-difference modelling, the PML boundary condition attenuates the artificial reflections in the computational domain within ten grid nodes to achieve the same behavior. A dipole source was placed at 650 m depth in the middle of the model, while 101 receivers are sitting on the curved seabed. Figure 11 displays the comparison of the 3D CSEM modelling between our method and the result from `emg3d`. From the Figures 11a and 11b, both the amplitude and the phase from our method match very well with the reference solution. The maximum amplitude discrepancy for all frequencies in Figure 11c is bounded within 5% at most of the relevant offset. The maximum mismatch in phase is less than 1 degree for 0.25 Hz and 2 degrees for 0.75 Hz and 1.25 Hz for the EM field above the noise level (1e-15 V/m2), as can be seen in Figure 11d. These demonstrate the good accuracy achieved by our method. Figure 11: Comparison of the modelling results between the proposed method and the reference solution. ## 4 Discussion A natural idea to achieve higher modelling accuracy is to use denser sampling. However, increasing the number of grid points in each dimension will lead to exponentially growing computational overhead, i.e., double the sampling in x, y and z coordinates results in eight times more nodes in the simulation. Over the same domain the node separations are reduced with a factor of two. By that, the stability criterion dictates a reduction in the time step by a factor of two. The net result is an increase of computational cost with a factor of sixteen. This is significantly more costly than considering high- order FDTD. For a fixed error requirement, high-order FDTD has been demonstrated to be much more efficient than a lower-order scheme with dense sampling (Yefet and Petropoulos,, 2001). Assuming a linear scaling dependency between computing time and the number of grid points, doubling the length of the finite-difference operator in x, y and z directions will simply double the computational cost. Stretching along $z$-direction seems always to be beneficial in terms of computational efficiency. It reduces the computational cost in two ways: first, it leads to increased grid spacing, hence less grid points and larger node spacing to discretize the resistivity model for simulation on the same physical size. Meanwhile, larger grid spacing permits to use a larger temporal step in terms of stability condition. This decreases the number of time steps needed to reach the steady state of the frequency-domain EM fields. It is natural to stretch the nonuniform grid in all three spatial directions, with the motivation to decrease the computational cost further. In case there is no a priori knowledge about the subsurface, a possible practice is to start stretching from a given offset from the source location in the horizontal directions. This approach is not followed up here due to the fact that grid stretching complicates the calculation of the air-wave for time-domain codes. Grid stretching is applicable also for frequency-domain finite-difference codes. The airwave implementation is very different for frequency-domain codes, where the air layer is part of the simulation domain. Frequency-domain codes lend themselves easily to both vertical and horizontal grid stretching. More research must be invested with respect to increasing the simulation efficiency of time-domain finite-difference schemes focusing on airwave implementation combined with horizontally nonuniform grids. The root of the problem is that the most common airwave implementations require fast Fourier transform (FFT) on regular grids. The fields are transformed to the wavenumber domain, propagated into the air layer in this domain and then transformed back to the space domain. A straight forward approach is to use interpolation between the uniform grid and the nonuniform one, as illustrated in Figure 12. The grid stretching in $x$\- and $y$\- directions certainly complicates the implementation while introducing additional computational cost, which is opposite to what we want to achieve. We have tested horizontal stretching for the previous 1D model using the stretching factor 1.05 in both $x$\- and $y$\- directions. The resulting amplitude and phase error in Figure 13 shows that the numerical accuracy is highly degraded (maximum amplitude error for 0.25 Hz is around 4%) compared with the result for a horizontally uniform grid (maximum amplitude error is less than 1.5% for all calculated frequencies). The running time is in fact longer than for the horizontally uniform grid which has a higher number of nodes. Due to the nonuniform grid staggering, several nodes on the uniform grid may reside in the same interval between two neighboring nodes on the nonuniform grid. The error panel displays an unsymmetrical pattern in Figure 13. Interpolating the fields from the uniform grid back to nonuniform grid can thus produce less accurate solutions. No stretching is therefore recommended along $x$\- and $y$\- for efficiency and accuracy considerations until a more accurate solution to this problem is developed. Figure 12: Horizontal grid stretching requires interpolating between the coarse nonuniform grid and a dense uniform grid, due to the equidistance requirement of FFT in airwave manipulation. Figure 13: The amplitude and phase error after horizontal stretching of the 1D model with a stretching factor 1.05 along both $x$\- and $y$\- directions. It is noteworthy that fictitious wave domain is simply a mathematical tool to compute correct frequency domain EM response efficiently by time stepping. The fictitious time is different from real time. To compute the true EM time series correctly, one needs to simulate a large number frequencies of the CSEM fields and then perform inverse Fourier transform, as has been done in Mittet, (2010) and Rochlitz et al., (2021). ## 5 Conclusion We have presented a 3D CSEM modelling method using high-order FDTD on a nonuniform grid. The key problem addressed in this work is the low accuracy and inconsistency issue in standard 2nd order staggered FDTD scheme on nonuniform grid. The strategy we propose is to adapt the interpolation weights depending on the nodal distance. These finite difference coefficients can be computed by inverting a Vandermonde matrix in an accurate and efficient manner. This makes our approach different from the commonly used EM modelling approaches. A new yet more generic stability condition has been established in order to achieve stable FDTD modelling. In designing the nonuniform grid based on geometrical progression, we develop a fixed point iteration to compute the optimal growing factor which allows good match of the modelling domain in case of grid stretching. The numerical examples demonstrate that there is a significant improvement in accuracy by using a high-order FDTD scheme, while combining it with a nonuniform grid reduces the computational cost without a significant sacrifice of accuracy. We conclude that high-order finite differences on nonuniform grid is a viable tool for full scale 3D CSEM modelling applications. Since the key idea is to use high order finite- difference coefficients adaptive to the node spacing, the method is expected to be applicable also for finite-difference frequency-domain schemes. ## Acknowledgments Pengliang Yang was supported by Chinese Fundamental Research Funds for the Central Universities (AUGA5710010121) and National Natural Science Fundation of China (42274156). Pengliang Yang thanks Dieter Werthermullter for the assistance to produce the reference solution using `empymod` and `emg3d` to validate the accuracy of the proposed method. The source code of this work can be found in the github repository: https://github.com/yangpl/libEMM. ## Computing uniform staggered-grid finite difference coefficients via Vandermonde matrix inversion The method to invert Vandermonde matrices gives us a generic approach to compute finite-difference coefficients with arbitrary grid spacing. A special case is the regular grid spacing. In what follows, we show how the standard staggered grid finite difference coefficients can be accurately computed also within the same framework. The Taylor series expansion of a function $f(x)$ can be written as $\begin{cases}f(x+h)=f(x)+\frac{\partial f(x)}{\partial x}h+\frac{1}{2!}\frac{\partial^{2}f(x)}{\partial x^{2}}h^{2}+\frac{1}{3!}\frac{\partial^{3}f(x)}{\partial x^{3}}h^{3}+\ldots\\\ f(x-h)=f(x)-\frac{\partial f(x)}{\partial x}h+\frac{1}{2!}\frac{\partial^{2}f(x)}{\partial x^{2}}h^{2}-\frac{1}{3!}\frac{\partial^{3}f(x)}{\partial x^{3}}h^{3}+\ldots\end{cases}.$ (44) It leads to $\begin{cases}\frac{f(x+h)+f(x-h)}{2}&=f(x)+\frac{1}{2!}\frac{\partial^{2}f(x)}{\partial x^{2}}h^{2}+\frac{1}{4!}\frac{\partial^{4}f(x)}{\partial x^{4}}h^{4}+\ldots\\\ \frac{f(x+h)-f(x-h)}{2}&=\frac{\partial f(x)}{\partial x}h+\frac{1}{3!}\frac{\partial^{3}f(x)}{\partial x^{3}}h^{3}+\frac{1}{5!}\frac{\partial^{5}f(x)}{\partial x^{5}}h^{5}+\ldots\end{cases}.$ (45) Let $h=\Delta x/2$. This implies the 2nd order accuracy of centered finite difference scheme using only two staggered nodes $\begin{cases}\frac{\partial f(x)}{\partial x}=\frac{f(x+\Delta x/2)-f(x-\Delta x/2)}{\Delta x}+O(\Delta x^{2})\\\ f(x)=\frac{f(x+\Delta x/2)+f(x-\Delta x/2)}{2}+O(\Delta x^{2})\end{cases}.$ (46) To approximate the 1st order derivatives as accurate as possible, we express it using more consecutive nodes. Due to regular grid staggering, the coefficients lying on symmetric positions should have the same coefficients. This means the first-order derivative reads in the following form $\begin{split}\frac{\partial f}{\partial x}=&b_{1}\frac{f(x+\Delta x/2)-f(x-\Delta x/2)}{\Delta x}+\\\ &b_{2}\frac{f(x+3\Delta x/2)-f(x-3\Delta x/2)}{3\Delta x}+\\\ &b_{3}\frac{f(x+5\Delta x/2)-f(x-5\Delta x/2)}{5\Delta x}+\cdots.\end{split}$ (47) Substituting the $f(x+h)$ and $f(x-h)$ with equation 44 for $h=\Delta x/2,3\Delta x/2,\ldots$ results in $\begin{split}\frac{\partial f}{\partial x}=&b_{1}\cdot 2\left(\frac{\Delta x}{2}\frac{\partial f}{\partial x}+\frac{1}{3!}(\frac{\Delta x}{2})^{3}\frac{\partial^{3}f}{\partial x^{3}}+\cdots\right)/{\Delta x}\\\ &+b_{2}\cdot 2\left(\frac{3\Delta x}{2}\frac{\partial f}{\partial x}+\frac{1}{3!}(\frac{3\Delta x}{2})^{3}\frac{\partial^{3}f}{\partial x^{3}}+\cdots\right)/{3\Delta x}\\\ &+b_{3}\cdot 2\left(\frac{5\Delta x}{2}\frac{\partial f}{\partial x}+\frac{1}{3!}(\frac{5\Delta x}{2})^{3}\frac{\partial^{3}f}{\partial x^{3}}+\cdots\right)/{5\Delta x}+\ldots\\\ =&(b_{1}+b_{2}+b_{3}+b_{4}+\cdots)\frac{\partial f}{\partial x}\\\ &+\frac{\Delta x^{2}}{3!\cdot 2^{2}}(b_{1}+3^{2}b_{2}+5^{2}b_{3}+7^{2}b_{4}+\cdots)\frac{\partial^{3}f}{\partial x^{3}}\\\ &+\frac{\Delta x^{4}}{5!\cdot 2^{4}}(b_{1}+3^{4}b_{2}+5^{4}b_{3}+7^{4}b_{4}+\cdots)\frac{\partial^{5}f}{\partial x^{5}}+\cdots.\end{split}$ (48) Thus, taking first $L$ terms (corresponding to using $2L$ nodes) requires $\begin{cases}b_{1}+b_{2}+b_{3}+\cdots+b_{L}&=1\\\ b_{1}+3^{2}b_{2}+5^{2}b_{3}+\cdots+(2L-1)^{2}b_{L}&=0\\\ b_{1}+3^{4}b_{2}+5^{4}b_{3}+\cdots+(2L-1)^{4}b_{L}&=0\\\ \cdots&\\\ b_{1}+3^{2L-2}b_{2}+5^{2L-2}b_{3}+\cdots+(2L-1)^{2L-2}b_{L}&=0\\\ \end{cases},$ (49) which again builds up a Vandermonde-like system $\displaystyle\underbrace{\begin{bmatrix}1&1&\ldots&1\\\ x_{1}&x_{2}&\ldots&x_{L}\\\ \vdots&&\ddots&\vdots\\\ x_{1}^{L-1}&x_{2}^{L-1}&\ldots&x_{L}^{L-1}\\\ \end{bmatrix}}_{\textbf{V}}\underbrace{\begin{bmatrix}b_{1}\\\ b_{2}\\\ \vdots\\\ b_{L}\\\ \end{bmatrix}}_{\textbf{b}}=\underbrace{\begin{bmatrix}1\\\ 0\\\ \vdots\\\ 0\\\ \end{bmatrix}}_{\textbf{z}},$ (50) in which $x_{i}=(2i-1)^{2}$, $i=1,\cdots,L$. The finite difference weights can then be easily computed using Golub, (1996, Algorithm 4.6.2), see an Octave/Matlab script for computing them in Yang, (2014, section 2.5). These numerically computed weights may be cross-validated with the generic method by Fornberg, (1988, Table 2). With regular grid spacing $\Delta x$, the weights $b_{i}$ are dimensionless and can be connected to the coefficients $\alpha_{i}$ in equation 13 involving a scaling factor $\Delta x$. ## References * Björck and Pereyra, (1970) Björck, A., and V. Pereyra, 1970, Solution of Vandermonde systems of equations: Mathematics of computation, 24, 893–903. * da Silva et al., (2012) da Silva, N. V., J. V. Morgan, L. MacGregor, and M. Warner, 2012, A finite element multifrontal method for 3D CSEM modeling in the frequency domain: Geophysics, 77, E101–E115. * de Hoop, (1996) de Hoop, A. T., 1996, A general correspondence principle for time-domain electromagnetic wave and diffusion fields: Geophysical Journal International, 127, 757–761. * Demmel and Koev, (2005) Demmel, J., and P. Koev, 2005, The accurate and efficient solution of a totally positive generalized vandermonde linear system: SIAM Journal on Matrix Analysis and Applications, 27, 142–152. * Fornberg, (1988) Fornberg, B., 1988, Generation of finite difference formulas on arbitrarily spaced grids: Mathematics of Computation, 51, 699–706. * Golub, (1996) Golub, G. H., 1996, Matrix computation, third edition: Johns Hopkins Studies in Mathematical Sciences. * Key, (2016) Key, K., 2016, MARE2DEM: a 2-D inversion code for controlled-source electromagnetic and magnetotelluric data: Geophysical Journal International, 207, 571–588. * Komatitsch and Martin, (2007) Komatitsch, D., and R. Martin, 2007, An unsplit convolutional perfectly matched layer improved at grazing incidence for the seismic wave equation: Geophysics, 72, SM155–SM167. * Lee et al., (1989) Lee, K. H., G. Liu, and H. Morrison, 1989, A new approach to modeling the electromagnetic response of conductive media: Geophysics, 54, 1180–1192. * Li and Key, (2007) Li, Y., and K. Key, 2007, 2D marine controlled-source electromagnetic modeling: Part 1—an adaptive finite-element algorithm: Geophysics, 72, WA51–WA62. * Maaø, (2007) Maaø, F., 2007, Fast finite-difference time-domain modeling for marine subsurface electromagnetic problems: Geophysics, 72, A19–A23. * Mittet, (2010) Mittet, R., 2010, High-order finite-difference simulations of marine CSEM surveys using a correspondence principle for wave and diffusion fields: Geophysics, 75, F33–F50. * Mittet, (2015) ——–, 2015, Seismic wave propagation concepts applied to the interpretation of marine controlled-source electromagnetics: Geophysics, 80, E63–E81. * Mittet, (2017) ——–, 2017, On the internal interfaces in finite-difference schemes: Geophysics, 82, T159–T182. * Monk and Süli, (1994) Monk, P., and E. Süli, 1994, A convergence analysis of Yee’s scheme on nonuniform grids: SIAM Journal on Numerical Analysis, 31, 393–412. * Mulder, (2006) Mulder, W., 2006, A multigrid solver for 3D electromagnetic diffusion: Geophysical prospecting, 54, 633–649. * Newman and Alumbaugh, (1995) Newman, G. A., and D. L. Alumbaugh, 1995, Frequency-domain modelling of airborne electromagnetic responses using staggered finite differences: Geophysical Prospecting, 43, 1021–1042. * Oristaglio and Hohmann, (1984) Oristaglio, M. L., and G. W. Hohmann, 1984, Diffusion of electromagnetic fields into a two-dimensional earth: A finite-difference approach: Geophysics, 49, 870–894. * Puzyrev et al., (2013) Puzyrev, V., J. Koldan, J. de la Puente, G. Houzeaux, M. Vázquez, and J. M. Cela, 2013, A parallel finite-element method for three-dimensional controlled-source electromagnetic forward modelling: Geophysical Journal International, 193, 678–693. * Rochlitz et al., (2021) Rochlitz, R., M. Seidel, and R.-U. Börner, 2021, Evaluation of three approaches for simulating 3-D time-domain electromagnetic data: Geophysical Journal International, 227, 1980–1995. * Rochlitz et al., (2019) Rochlitz, R., N. Skibbe, and T. Günther, 2019, custem: Customizable finite-element simulation of complex controlled-source electromagnetic data: Geophysics, 84, F17–F33. * Roden and Gedney, (2000) Roden, J. A., and S. D. Gedney, 2000, Convolution PML (CPML): An efficient FDTD implementation of the CFS–PML for arbitrary media: Microwave and optical technology letters, 27, 334–339. * (23) Smith, J. T., 1996a, Conservative modeling of 3-D electromagnetic fields, Part i: Properties and error analysis: Geophysics, 61, 1308–1318. * (24) ——–, 1996b, Conservative modeling of 3-D electromagnetic fields, Part II: Biconjugate gradient solution and an accelerator: Geophysics, 61, 1319–1324. * Streich, (2009) Streich, R., 2009, 3D finite-difference frequency-domain modeling of controlled-source electromagnetic data: Direct solution and optimization for high accuracy: Geophysics, 74, F95–F105. * Taflove and Hagness, (2005) Taflove, A., and S. C. Hagness, 2005, Computational electrodynamics: The finite-difference time-domain method, 3rd ed.: Artech House. * Wang and Hohmann, (1993) Wang, T., and G. W. Hohmann, 1993, A finite-difference, time-domain solution for three-dimensional electromagnetic modeling: Geophysics, 58, 797–809. * Werthmüller, (2017) Werthmüller, D., 2017, An open-source full 3D electromagnetic modeler for 1D VTI media in Python: empymod: Geophysics, 82, WB9–WB19. * Werthmüller et al., (2019) Werthmüller, D., W. Mulder, and E. Slob, 2019, emg3d: A multigrid solver for 3d electromagnetic diffusion: Journal of Open Source Software, 4, 1463\. * Yang, (2014) Yang, P., 2014, A numerical tour of wave propagation: Technical report, Xi’an Jiaotong University. * Yee, (1966) Yee, K. S., 1966, Numerical solution of initial boundary value problems involving Maxwell’s equations in isotropic media: IEEE Transactions on Antennas and Propagation, 14, 302–307. * Yefet and Petropoulos, (2001) Yefet, A., and P. G. Petropoulos, 2001, A staggered fourth-order accurate explicit finite difference scheme for the time-domain maxwell’s equations: Journal of Computational Physics, 168, 286–315.
# Tight Conditional Lower Bounds for Vertex Connectivity Problems Zhiyi Huang Tsinghua University Yaowei Long University of Michigan Thatchaphol Saranurak Supported by NSF CAREER grant 2238138. University of Michigan Benyu Wang Tsinghua University (April, 2023) ###### Abstract We study the fine-grained complexity of graph connectivity problems in unweighted undirected graphs. Recent development shows that all variants of _edge connectivity_ problems, including single-source-single-sink, global, Steiner, single-source, and all-pairs connectivity, are solvable in $m^{1+o(1)}$ time, collapsing the complexity of these problems into the almost-linear-time regime. While, historically, _vertex connectivity_ has been much harder, the recent results showed that both single-source-single-sink and global vertex connectivity can be solved in $m^{1+o(1)}$ time, raising the hope of putting all variants of vertex connectivity problems into the almost- linear-time regime too. We show that this hope is impossible, assuming conjectures on finding 4-cliques. Moreover, we essentially settle the complexity landscape by giving tight bounds for _combinatorial_ algorithms in dense graphs. There are three separate regimes: 1. 1. all-pairs and Steiner vertex connectivity have complexity $\hat{\Theta}(n^{4})$, 2. 2. single-source vertex connectivity has complexity $\hat{\Theta}(n^{3})$, and 3. 3. single-source-single-sink and global vertex connectivity have complexity $\hat{\Theta}(n^{2})$. For graphs with general density, we obtain tight bounds of $\hat{\Theta}(m^{2})$, $\hat{\Theta}(m^{1.5})$, $\hat{\Theta}(m)$, respectively, assuming Gomory-Hu trees for element connectivity can be computed in almost-linear time. ## 1 Introduction Vertex connectivity and edge connectivity are central concepts in graph theory. In an unweighted undirected graph $G$ with $n$ vertices and $m$ edges, the _vertex connectivity_(_edge connectivity_) between two vertices $u,v$ is the maximum number of internally vertex-disjoint (edge-disjoint) paths from $u$ to $v$. Efficient algorithms for variants of both problems have been extensively studied for at least half a century [FF56, Kle69, GH61, GR98, CKM+11, She13, KLOS14, Mad16, vdBLN+20, CKL+22]. Until very recently, the graph algorithm community has reached a very satisfying conclusion on the complexity of edge connectivity problems: _all variants of edge connectivity problems can be solved in almost-linear time_. There are five variants of the connectivity problems studied in the literature, including (1) global, (2) single-source single-sink, (3) Steiner, (4) single-source, and (5) all-pairs connectivity. See the detailed definitions in Section 2. Among these problems, all-pairs connectivity is the hardest via straightforward reductions. For the global edge connectivity problem, Karger showed in 2000 that the problem admits a near-linear time algorithm [Kar00]. In the recent line of work [LP20, AKT21b, LP21, AKT21a, LPS21, AKL+22], Abboud et al. [AKL+22] finally showed that even the all-pairs edge connectivity case could be reduced to the single-source single-sink case, which solvable in almost-linear time via the recent max flow result by Chen et al. [CKL+22]. This finally puts all five problems on edge connectivity into the almost-linear time regime and raises the following hope: _Can we solve all variants of vertex connectivity problems in almost-linear time too?_ Historically, the vertex connectivity problems have been much more difficult than their edge connectivity counterpart. Furthermore, Abboud et al. [AKT20] showed that the all-pairs vertex connectivity in _weighted_ graphs with $\tilde{O}(n)$ edges requires $\hat{\Omega}(n^{3})$ time assuming SETH.111We use $\tilde{O}(\cdot)$ to hide a $\mathrm{polylog}(n)$ factor and $\hat{O}(\cdot)$ to hide a $n^{o(1)}$ factor. In _directed_ unweighted graphs, Abboud et al. [AGI+19] also showed that the problem even requires $\hat{\Omega}(n^{4})$ for combinatorial algorithms and $\hat{\Omega}(n^{\omega+1})$ time for general algorithms. Thus, at least in weighted or directed graphs, the problem does not admit almost-linear algorithms. But, again, the recent development on vertex connectivity in the standard _unweighted undirected_ graphs raises the hope for almost-linear time algorithms. Li et al. [LNP+21] showed how to compute global vertex connectivity using polylogarithmic calls to max flows, which implied an $\hat{O}(m)$-time algorithm via the max flow algorithm of [CKL+22] and improved upon the known $\tilde{O}(mn)$ bound by [HRG00]. Indeed, the recent max flow algorithm [CKL+22] also implies a $\hat{O}(m)$-time algorithm for the $(s,t)$-vertex connectivity problem. Until now, there are still no nontrivial lower bounds for Steiner, single-source, and all-pairs vertex connectivity in unweighted graphs, and the technique of [AKT20] is quite specific for weighted graphs. The complexity landscape for vertex connectivity problems is still very unclear. ### 1.1 Our Results We give a firm negative answer to the above open problem. Based on well-known conjectures, we settle the complexity landscape for all five vertex connectivity problems in dense graphs by giving tight bounds for _combinatorial_ algorithms, which generally refer to algorithms that do not use fast matrix multiplication. We can obtain tight bounds even in general graphs by assuming a possible hypothesis about computing Gomory-Hu trees for element connectivity. Below, we discuss our results in more detail. #### All-pairs vertex connectivity (Section 3). Our first result is a conditional lower bound of the all-pair vertex connectivity (APVC) problem based on the _4-clique conjecture_ , which postulates that the running time for deciding the existence of a 4-clique in a graph must be at least $\hat{\Omega}(n^{4})$ time for combinatorial algorithms [BGL17] and $\hat{\Omega}(n^{\omega(1,2,1)})=\Omega(n^{3.25})$ time for general algorithms [DW22].222$\omega(1,2,1)$ is the exponent of multiplying an $n\times n^{2}$ matrix by an $n^{2}\times n$ matrix. ###### Theorem 1.1. Assuming the 4-clique conjecture, the all-pairs vertex connectivity problem on an undirected unweighted graph with $n$ vertices requires $\hat{\Omega}(n^{4})$ time for combinatorial algorithms, and $\Omega(n^{3.25})$ time for general algorithms. Theorem 1.1 gives the first super cubic lower bounds for all-pairs vertex connectivity problems in the standard undirected case. Moreover, the bound is tight for _combinatorial_ algorithms. Indeed, a naive algorithm is to call max flow $O(n^{2})$ times, one for each pair of vertices, which takes $\hat{O}(mn^{2})=\hat{O}(n^{4})$ time. It is interesting to compare APVC with the _all-pairs shortest paths_ (APSP) problem, a central problem in fine-grained complexity. It is conjectured that the right complexity of APSP in weighted graphs is $\hat{\Theta}(n^{3})$ (see e.g. [WW18]). Theorem 1.1 shows that, for general algorithms, APVC in unweighted graphs is strictly harder than APSP in weighted graphs, assuming $\omega>2$, #### Steiner vertex connectivity (Section 4). Next, we study the _Steiner vertex connectivity_ problem. In this problem, given a set of vertices $T$, we need to compute the minimum vertex connectivity among all pairs of vertices in $T$. Even though Steiner vertex connectivity looks much easier than APVC, we can extend the same lower bound of Theorem 1.1 for APVC to work for the Steiner case too. Towards this hardness, we propose a variant of the 4-clique conjecture, the edge-universal 4-clique conjecture, which postulates that, given any subset of demand edges $E_{\mathrm{dem}}\subseteq E(G)$, checking if every edge in $E_{\mathrm{dem}}$ is contained in a 4-clique requires $\hat{\Omega}(n^{4})$ time for combinatorial algorithms (see 2.3 for details).333Note that this problem is at least as hard as the problem when $E_{\mathrm{dem}}=E(G)$, i.e. when we need to check if every edge is contained in a 4-clique. ###### Theorem 1.2. Assuming the edge-universal 4-clique conjecture, the Steiner vertex connectivity problem on an undirected unweighted graph with $n$ vertices requires $\hat{\Omega}(n^{4})$ time for combinatorial algorithms. We note that our reduction would imply hardness for general algorithms too if we conjectured the hardness of the edge-universal 4-clique problem for general algorithms. #### Upper bounds for general density (Section 5). On sparse graphs, we observe that there is still some discrepancy between our lower bounds and the naive algorithm. More precisely, we observe that our lower bounds for combinatorial algorithms for all-pairs and Steiner vertex connectivity can be easily extended to $\hat{\Omega}(m^{2})$ in a graph with $m$ edges. However, the naive algorithm requires $\hat{O}(mn^{2})$ time. For example, in sparse graphs, the naive algorithm takes $\hat{O}(n^{3})$ time, while our lower bound is $\hat{\Omega}(n^{2})$. We fix the above discrepancy by showing improved algorithms in sparse graphs via the following result. ###### Theorem 1.3. Given an $n$-vertex $m$-edge unweighted graph, there is an algorithm to solve the APVC problem in $\hat{O}(m^{11/5})$ time with high probability. Assuming that the _element connectivity Gomory-Hu tree_ can be constructed in $\hat{O}(m)$ time, the running time can be improved to $\hat{O}(m^{2})$. Theorem 1.3 gives the first subcubic algorithm for computing all-pairs vertex connectivity in sparse graphs. Moreover, the bound $\hat{\Theta}(m^{2})$ is tight with our lower bounds for all density regimes, assuming the almost- linear-time construction of the element connectivity Gomory-Hu tree. We note that the _element connectivity Gomory-Hu tree_ is a generalization of the well-known _edge connectivity Gomory-Hu tree_ [GH61], whose almost-linear- time construction was very recently shown by Abboud et al. [AKL+22]. It is quite believable that an element connectivity Gomory-Hu tree can also be constructed in almost-linear time. #### The complexity landscape of vertex connectivity. Based on these main results, we obtain some other results and corollaries on all variants of vertex connectivity problems, as summarized in Table 1. In contrast to the edge connectivity problems which can all be solved in almost- linear time, there are three separate regimes for vertex connectivity. 1. 1. all-pairs and Steiner vertex connectivity have complexity $\hat{\Theta}(n^{4})$, 2. 2. single-source vertex connectivity has complexity $\hat{\Theta}(n^{3})$, and 3. 3. single-source-single-sink and global vertex connectivity have complexity $\hat{\Theta}(n^{2})$. For graphs with general density, we obtain tight bounds of $\hat{\Theta}(m^{2})$, $\hat{\Theta}(m^{1.5})$, $\hat{\Theta}(m)$, respectively, assuming Gomory-Hu trees for element connectivity can be computed in almost-linear time. Table 1: Upper bounds and lower bounds for connectivity problems | Global | Single-Source | all-pairs | Steiner ---|---|---|---|--- edge connectivity, unweighted graphs | $\tilde{O}(m)$ [Kar00] | $\hat{O}(m)$ [AKL+22] | $\hat{O}(m)$ [AKL+22] | $\hat{O}(m)$ [LP20] vertex connectivity, unweighted graphs with general density | | $\hat{\Omega}(m^{1.5})$ for comb. algo., Corollary 3.11 | $\hat{\Omega}(m^{2})$ for comb. algo., Corollary 3.10 | $\hat{\Omega}(m^{2})$ for comb. algo., Corollary of Theorem 4.1 $\hat{O}(m)$ [LNP+21] | $\hat{O}(m^{1.5})$, assuming 2.5, Theorem 5.2 | $\hat{O}(m^{2})$, assuming 2.5, Theorem 5.1 | $\hat{O}(m^{2})$, assuming 2.5, Theorem 5.1 | | $\hat{O}(m^{5/3})$, Theorem 5.2 | $\hat{O}(m^{11/5})$, Theorem 5.1 | $\hat{O}(m^{11/5})$, Theorem 5.1 vertex connectivity, dense unweighted graphs $m=\Theta(n^{2})$ | | $\hat{\Omega}(n^{3})$ for comb. algo., Corollary 3.9 | $\hat{\Omega}(n^{4})$ for comb. algo., Theorem 3.1 | $\hat{\Omega}(n^{4})$ for comb. algo. Theorem 4.1 | | $\hat{\Omega}(n^{\omega(1,2,1)})$ for all algo., Remark 3.8 | $\hat{O}(n^{2})$ [LNP+21] | $\hat{O}(n^{3})$ trivially | $\hat{O}(n^{4})$ trivially | $\hat{O}(n^{4})$ trivially vertex connectivity, sparse weighted graphs, $m=\tilde{O}(n)$ | | | $\hat{\Omega}(n^{3})$ [AKT20] | #### Discussions and Open Problems. Our lower bounds for combinatorial algorithms are particularly relevant to the context of vertex connectivity since basically all algorithms for the problems are indeed combinatorial. There are a few exceptions [LLW88, AGI+19], but these algorithms are still far from optimal (even cannot break our combinatorial lower bounds). It is a very interesting open problem whether one can bypass our combinatorial lower bounds using fast matrix multiplications, or show conditional lower bounds for general algorithms that match our combinatorial lower bounds. ### 1.2 Technical Overview To prove Theorem 1.1, we are will reduce the 4-clique problem to the APVC problem. Previously, Abboud et al. [AGI+19] showed the hardness of all-pairs vertex connectivity in _directed_ graphs using the 4-clique problem, which inspired our reduction. However, the techniques are not strong enough to work on undirected graphs. In more details, the hard instance of [AGI+19] is a directed acyclic graph with four layers, and so they only need to consider directed paths of length at most 3. In contrast, when we consider undirected graphs, paths connecting sources and sinks can be much more complex, which requires more advanced techniques and more careful arguments. Let us sketch our construction below. Starting from a 4-clique instance $G$, it is helpful to consider its 4-partite version $G_{4p}$, which is simply constructed by duplicating $V(G)$ into 4 groups $A,B,C,D$ and copying $E(G)$ for each pair of different groups (see Definition 2.2 for a formal definition). A natural way to answer the 4-clique problem on $G$ is then checking for each pair of adjacent $a\in A$ and $d\in D$, whether there exists an adjacent pair of vertices $b\in B$ and $c\in C$ that is adjacent to both $a$ and $d$ (call such $(b,c)$ a _4-clique witness_ of $(a,d)$). To correspond this to a vertex connectivity problem, for each pair of $a$ and $d$, consider a 4-layer graph $\hat{H}_{ad}$ defined as follows. Let $B_{a}$ denote the set of vertices in $B$ adjacent to $a$ (also define $B_{d}$, $C_{a}$ and $C_{d}$ similarly). Then $B_{a}\cap B_{d}$ (resp. $C_{a}\cap C_{d}$) are vertices in $B$ (resp. $C$) adjacent to both $a$ and $d$. The vertices of $\hat{H}_{ad}$ are $V(\hat{H}_{ad})=\\{a\\}\cup(B_{a}\cap B_{d})\cup(C_{a}\cap C_{d})\cup\\{d\\}$, where the first layer (resp. the last layer) has only a single vertex $a$ (resp. $d$), and the second layer (resp. the third layer) has vertices $B_{a}\cap B_{d}$ (resp. $C_{a}\cap C_{d}$). The edge set of $\hat{H}_{ad}$ is $E(\hat{H}_{ad})=\\{(a,b)\mid b\in B_{a}\cap B_{d}\\}\cup E_{G_{4p}}(B_{a}\cap B_{d},C_{a}\cap C_{d})\cup\\{(c,d)\mid c\in C_{a}\cap C_{d}\\}$444Here $E_{G_{4p}}(B_{a}\cap B_{d},C_{a}\cap C_{d})\subseteq E(G_{4p})$ denote the set of edges connecting $B_{a}\cap B_{d}$ and $C_{a}\cap C_{d}$ in $G_{4p}$., which connects vertex $a$ (resp. $d$) to each second-layer (resp. third layer) vertex, and connects the second layer $B_{a}\cap B_{d}$ and the third layer $C_{a}\cap C_{d}$ using the same edges in $G_{4p}$. One can simply observe that a 4-clique witness $(b,c)$ of $(a,d)$ exists if and only if $\kappa_{\hat{H}_{ad}}(a,d)\geq 1$555In the overview, we use $\kappa_{H}(a,d)$ to denote the vertex connectivity between $a$ and $d$ in a graph $H$.. To check the existence of 4-clique witnesses for all pairs $(a,d)$ simultaneously, our final APVC instance $H$ will be a combination of all $\hat{H}_{ad}$, and the main challenge is to make the combination compact. We overcome this challenge by introducing two modules called the _source-sink isolating gadgets_ and the _set-intersection filter_. Interestingly, our technique for proving time lower bounds is inspired by the space-lower bound techniques. More specifically, the source-sink isolating gadget is inspired by the construction in [PSY22]. The intuition behind these two modules is as follows. When considering $\kappa_{\hat{H}_{ad}}(a,d)$ of a specific pair of $a\in A$ and $d\in D$, we will somehow “remove” redundant vertices in $H\setminus\hat{H}_{ad}$ so that $\kappa_{\hat{H}_{ad}}(a,d)$ can be derived from $\kappa_{H}(a,d)$. The high- level idea to remove a redundant vertex $v$ in $H$ is to create a “flow” path from $a$ to $d$ through $v$ (e.g. a simple path made up of two edges $(a,v)$ and $(v,d)$). By a simple flow-cut argument, this path will enforce that $v$ appears in any $(a,d)$-vertex cut, while bringing some additive deviations to estimate $\kappa_{\hat{H}_{ad}}(a,d)$ as $\kappa_{H}(a,d)$. These modules apply this simple rule in a more general way. The source-sink isolating gadget will remove all vertices in $A$ and $D$ except $a$ and $d$, and the intersection patterns will generate $B_{a}\cap B_{d}$ and $C_{a}\cap C_{d}$ by removing other vertices in $B$ and $C$. The remaining graph will then be exactly $\hat{H}_{ad}$, and the additive deviations between $\kappa_{\hat{H}_{ad}}(a,d)$ and $\kappa_{H}(a,d)$ can be computed and subtracted easily. The proof of Theorem 1.2 extends the above idea to reduce a Steiner vertex connectivity problem to an edge-universal 4-clique problem. Consider an edge- universal 4-clique instance $G$. Based on the above construction of $H$, by creating more “flow” paths, we can guarantee that the additive deviations between $\kappa_{\hat{H}_{ad}}(a,d)$ and $\kappa_{H}(a,d)$ for all pair of $(a,d)$ will be the same, say a value $K$. Therefore, for each pair $(a,d)$, $\kappa_{H}(a,d)$ is either at least $K+1$ or equal to $K$, and there is no 4-clique containing $a$ and $d$ in the original graph $G$ if and only if the latter case happens. Finally, checking the Steiner vertex connectivity of terminal set $A\cup D$ suffices to answer the edge-universal 4-clique problem on $G$. Towards the upper bound of the APVC problem on sparse graphs in Theorem 1.3, our algorithm uses the following scheme. Let the input graph be $G$ and let $k$ be a degree threshold to separate vertices into two parts, the high-degree part $V_{h}=\\{v\in V(G)\mid\deg_{G}(v)>k\\}$ and the low-degree part $V_{l}=\\{v\in V(G)\mid\deg_{G}(v)\leq k\\}$. For pairs $(u,v)$ such that $u,v\in V_{h}$, we can compute $\kappa_{G}(u,v)$ by simply calling max flows, which takes totally $O(m^{2}/k^{2})$ calls since $|V_{h}|=O(m/k)$. For other pairs $(u,v)$ with $u\in V_{l}$ or $v\in V_{l}$, there will be $\kappa_{G}(u,v)\leq k$, because the vertex connectivity of $u$ and $v$ is upper bounded by their degrees. The vertex connectivity oracle in [PSY22] can exactly capture all-pairs vertex connectivity bounded by $k$, which takes $\tilde{O}(k^{2})$ black-box calls to Gomory-Hu trees for element connectivity whose construction time is currently $\hat{O}(mk)$. The whole algorithm takes $\hat{O}(m^{11/5})$ time by choosing a proper $k$, and the running time will be immediately improved to $\hat{O}(m^{2})$ if Gomory-Hu trees for element connectivity can be constructed in almost linear time. ### 1.3 Organization We will start with some basic notations and introduce conjectures in Section 2. In Section 3, we will show the conditional lower bound of the APVC problem. In Section 4, we will extend the approach in Section 3 to show the conditional lower bound of the Steiner vertex connectivity problem. In Section 5, we will show a simple APVC algorithm for graphs with general density. ## 2 Preliminaries In this paper, we use standard graph theoretic notations. For a graph $G$, we use $V(G)$ and $E(G)$ to denote its vertex set and edge set. For any vertex set $V^{\prime}\subseteq V(G)$, we let $G[V^{\prime}]$ denote the subgraph induced by $V^{\prime}$ and let $G\setminus V^{\prime}$ be a short form of $G[V(G)\setminus V^{\prime}]$. For two graphs $G_{1}$ and $G_{2}$ which vertex sets $V(G_{1})$ and $V(G_{2})$ may intersect, we let $G_{1}\cup G_{2}$ denote the graph with vertex set $V(G_{1})\cup V(G_{2})$ and edge set $E(G_{1})\cup E(G_{2})$. For two subset of vertices $V_{1},V_{2}\subseteq V(G)$, we let $E_{G}(V_{1},V_{2})$ denote the set of edges directly connecting $V_{1}$ and $V_{2}$. For a vertex $v\in G$, we let $N_{G}(v)=\\{u\mid(u,v)\in E(G)\\}$ denote its neighbor set, and let $\bar{N}_{G}(v)=V(G)\setminus N_{G}(v)$ denote its non-neighbors. #### Vertex connectivity. In a graph $G$, the vertex connectivity for two vertices $u,v\in V(G)$, denoted by $\kappa_{G}(u,v)$, is the maximum number of internally vertex- disjoint paths from $u$ to $v$. By Menger’s theorem, $\kappa_{G}(u,v)$ is equal to the size of minimized subsets $C\subseteq(V(G)\setminus\\{u,v\\})\cup E(G)$ deleting which from $G$ will disconnect $u$ and $v$. #### Vertex Connectivity problems. In this paper, we will consider four vertex connectivity problems, i.e. the global, single-source, all-pairs, and Steiner vertex connectivity problems. The edge connectivity versions of these problems are analogous. * • The global vertex connectivity problem. Given an undirected unweighted graph $G$, the global vertex connectivity problem (the global-VC problem) asks the global vertex connectivity, denoted by $\kappa_{G}$, where $\kappa_{G}=\min_{u,v\in G}\kappa_{G}(u,v)$. * • The single-source vertex connectivity problem. Given an undirected unweighted graph $G$ with a source vertex $s$, a single source vertex connectivity problem (the SSVC problem) asks $\kappa_{G}(s,v)$ for all other vertices $v\in G$. * • The all-pairs vertex connectivity problem. Given an undirected unweighted graph $G$, the all-pairs vertex connectivity problem (the APVC problem) asks $\kappa_{G}(u,v)$ for all pairs of $u,v\in G$. * • The Steiner vertex connectivity problem. Given an undirected unweighted graph $G$ with a terminal set $T\subseteq V(G)$, the Steiner vertex connectivity problem (the Steiner-VC problem) asks the Steiner vertex connectivity of $T$, denoted by $\kappa_{G}(T)$, where $\kappa_{G}(T)=\min_{u,v\in T}\kappa_{G}(u,v)$. #### The 4-clique conjecture. Given an $n$-vertex undirected graph $G$, the $k$-clique problem is to decide whether there is a clique with $k$ vertices in $G$. The $k$-clique problem can be solved in $O(n^{k})$ time trivially, and a more efficient combinatorial algorithm takes running time $O(n^{k}/\log^{k}n)$ [Vas09]. The popular $k$-clique conjecture (see e.g. [BGL17]) suggests that there is no combinatorial algorithm for the $k$-clique problem with $O(n^{k-\epsilon})$ running time for any constant $\epsilon>0$. In Section 3, we will use this conjecture in the case $k=4$. ###### Conjecture 2.1 (4-clique conjecture). There is no combinatorial algorithm that solves the 4-clique problem for $n$-vertex graphs in $O(n^{4-\epsilon})$ time for any constant $\epsilon>0$. For each 4-clique instance $G$, it is equivalent to consider its 4-partite form $G_{4p}$ as defined below. Note that a 4-clique in a 4-partite graph should contain exactly one vertex from each group, and the original graph has a 4-clique if and only if the 4-partite graph $G_{4p}$ has a 4-clique. ###### Definition 2.2 (4-partite graph $G_{4p}$). Given an undirected graph $G$, the 4-partite graph $G_{4p}$ of $G$ has vertex set $V(G_{4p})=\\{v_{A},v_{B},v_{C},v_{D}\mid v\in V(G)\\}$, and we let $A=\\{v_{A}\mid v\in V(G)\\},B=\\{v_{B}\mid v\in V(G)\\},C=\\{v_{C}\mid v\in V(G)\\},D=\\{v_{D}\mid v\in V(G)\\}$ be four groups partitioning $V(G_{4p})$. The edge set $E(G_{4p})=\\{(u_{X},v_{Y})\mid(u,v)\in E(G),X,Y\in\\{A,B,C,D\\},X\neq Y\\}$. #### The edge-universal 4-clique conjecture. We consider a variant of the 4-clique problem, called the _edge-universal 4-clique problem_. Given an undirected graph $G$ and a subset of _demand_ edges $E_{\mathrm{dem}}\subseteq E(G)$, this problem asks if every edge in $E_{\mathrm{dem}}$ is contained by a 4-clique. We suggest the following conjecture on this problem. ###### Conjecture 2.3 (Edge-universal 4-clique conjecture). There is no combinatorial algorithm that answers the all edge 4-clique problem for $n$-vertex graphs in $O(n^{4-\epsilon})$ time for any constant $\epsilon>0$. We note our formulation of the edge-universal 4-clique problem is at least as hard as the problem of checking if every edge is contained in some 4-clique by fixing $E_{\mathrm{dem}}=E(G)$. We choose to present this formulation that allows any $E_{\mathrm{dem}}\subseteq E(G)$ because it only strengthens our hardness result and shows the flexibility of our reduction. The difference between the edge-universal 4-clique conjecture vs. the 4-clique conjecture is that we switch the quantifier in the definition of the problems. This allows us to obtain new hardness results. This strategy for proving conditional lower bounds has been studied and done several times in the literature of fine-grained complexity [GIKW19, AWW16, ABHS22]. For example, the relationship between the edge-universal 4-clique problem vs. the 4-clique problem is the same as the relationship between the well-known _orthogonal vector_ problem vs. the _hitting set_ problem introduced in [AWW16], and _SETH_ vs. _quantified SETH_ introduced in [ABHS22]. #### Gomory-Hu trees for element connectivity Gomory-Hu trees are cut-equivalent trees introduced by Gomory and Hu [GH61] to _capture_ all-pairs edge connectivity in weighted graphs. More precisely, given a Gomory-Hu tree, the edge connectivity of any given pair of vertices can be queried in nearly constant time. Very recently, a breakthrough by [AKL+22] showed that a Gomory-Hu tree can be constructed in $\tilde{O}(n^{2})$ time for a weighted graph and $\hat{O}(m)$ time for an unweighted graph. For vertex connectivity, it has been shown by [Ben95] that there are no such cut- equivalent trees. See also [PSY22] for a more general space lower bound forbidding the existence of such trees for vertex connectivity. Element connectivity is the notion of connectivity that is related to vertex connectivity, and yet Gomory-Hu trees have been shown to exist for element connectivity (see e.g. [CRX15]). Given an undirected graph $G$ and a terminal set $U\subseteq V(G)$, the element connectivity for two vertices $u,v\in U$, denoted by $\kappa^{\prime}_{G,U}(u,v)$, is the size of minimized set $C\subseteq E(G)\cup(V(G)\setminus U)$ whose removal will disconnect $u$ and $v$. An element connectivity Gomory-Hu tree for the graph $G$ and terminal set $U$ will capture $\kappa^{\prime}_{G,U}(u,v)$ for all pairs of $u,v\in U$. In [PSY22], they consider a variant called $k$-Gomory-Hu tree for element connectivity, which given an additional parameter $k$, will capture the value $\min\\{\kappa^{\prime}_{G,U}(u,v),k\\}$ for all $u,v\in U$ (namely we can query $\min\\{\kappa^{\prime}_{G,U}(u,v),k\\}$ for each given $u,v\in U$ in nearly constant time). By generalizing the $(1+\epsilon)$-approximate Gomory- Hu tree algorithm by [LP21] to the element connectivity setting, the following result was obtained by [PSY22]. ###### Theorem 2.4. Given an $n$-vertex $m$-edge undirected unweighted graph $G$, a terminal set $U\subseteq V(G)$ and a parameter $k$, there is a randomized algorithm to construct a $k$-Gomory-Hu tree for element connectivity in $\hat{O}(mk)$ time with high probability. Given the similarity of Gomory-Hu trees for edge connectivity and element connectivity, and the recent breakthrough of $\hat{O}(m)$-time construction algorithm for edge connectivity Gomroy-Hu tree, it seems reasonable to conjecture that element connectivity Gomory-Hu tree can also be constructed in $\hat{O}(m)$ time. ###### Conjecture 2.5. Given an $n$-vertex $m$-edge undirected unweighted graph $G$ and a terminal set $U\subseteq V(G)$, an element connectivity Gomory-Hu tree can be constructed in $\hat{O}(m)$ time. We leave this conjecture as a very interesting open problem. ## 3 The Lower Bound for the APVC Problem In this section, we will prove Theorem 3.1, a conditional lower bound of the APVC problem in undirected unweighted graphs conditioning on the 4-clique conjecture. Concretely, we will show a reduction from the 4-clique problem to the APVC problem. ###### Theorem 3.1. Assuming 2.1, for $n$-vertex undirected unweighted graphs, there is no combinatorial algorithm that solves the APVC problem in $O(n^{4-\epsilon})$ time for any constant $\epsilon>0$. Given an $n$-vertex 4-clique instance $G$, let $G_{4p}$ be the corresponding 4-partite graph defined in Definition 2.2 (where $V(G_{4p})$ is partitioned into 4 groups $A,B,C,D$). We start with some notations. For each vertex $a\in A$, we use $B_{a}=\\{b\in B|(a,b)\in E(G_{4p})\\}$ to denote the neighbors of $a$ in $B$ and let $\bar{B}_{a}=B\setminus B_{a}$. Analogously, $C_{a}$ is the set of vertices in $C$ adjacent to $a$ and $\bar{C}_{a}=C\setminus C_{a}$. For each $d\in D$, we define $B_{d},\bar{B}_{d},C_{d},\bar{C}_{d}$ in a similar way. As discussed in Section 1.2, our 4-clique instance $H$ will be constructed using two kinds of modules, the source-sink isolating gadgets and the set- intersection filters, which will be introduced in Section 3.1 and Section 3.2 respectively. After that, the final construction of $H$ and the proof of Theorem 3.1 will be completed in Section 3.3. ### 3.1 The Source-Sink Isolating Gadget We first introduce the _source-sink isolating gadget_. Basically, for an undirected graph $R$ and two disjoint groups of vertices $X,Y\subseteq V(R)$, a source-sink isolating gadget $Q(X,Y)$ (or just $Q$ for short) is a graph on vertices $X\cup Y$ with additional vertices outside $R$. Its formal guarantee is as follows. ###### Lemma 3.2. Given an undirected graph $R$ and two disjoint groups of vertices $X,Y\subseteq V(R)$, there is a graph $Q$ with $V(Q)\cap V(R)=X\cup Y$ and $|V(Q)|=O(|X|+|Y|)$ such that for any $x\in X,y\in Y$ with $(x,y)\notin E(R)$, $\kappa_{R\cup Q}(x,y)=\kappa_{R_{xy}}(x,y)+|X|+|Y|,$ where $R_{xy}=R\setminus((X\cup Y)\setminus\\{x,y\\})$. Such a graph $Q$ is called a source-sink isolating gadget, and moreover, the construction of $Q$ is independent from $R$. The reason we call the graph $Q$ a source-sink isolating gadget is that by adding $Q$ into the input graph $R$ the vertex connectivity between any pair of source $x\in X$ and sink $y\in Y$ in $R\cup Q$, i.e., $\kappa_{R\cup Q}(x,y)$, can be derived from their connectivity in $R_{xy}$, i.e., $\kappa_{R_{xy}}(x,y)$. But the graph $R_{xy}$, as defined in Lemma 3.2, is just the graph $R$ after removing all source and sink vertices in $X$ and $Y$ except $x$ and $y$. That is, the gadget “isolates” the pair $x$ and $y$ from the rest. We will use this gadget in Section 3.3. ###### Proof. We construct $Q$ in the following way. We create duplicated sets $X_{1},X_{2}$ of $X$, and also $Y_{1},Y_{2}$ of $Y$. For each vertex $x\in X$, we let $\hat{x}_{1}\in X_{1}$ and $\hat{x}_{2}\in X_{2}$ denote copies of $x$ in $X_{1}$ and $X_{2}$ respectively if there is no other specification (for each $y\in Y$, $\hat{y}_{1},\hat{y}_{2}$ are defined similarly). The vertex set of $Q$ is $V(Q)=X\cup X_{1}\cup X_{2}\cup Y\cup Y_{1}\cup Y_{2}$, and the edge set is $\displaystyle E(Q)=$ $\displaystyle\\{(x,\hat{x}_{1})\mid x\in X\\}\cup\\{(x_{1},y)\mid x_{1}\in X_{1},y\in Y\\}\cup$ $\displaystyle\\{(y,\hat{y}_{1})\mid y\in Y\\}\cup\\{(y_{1},x)\mid y_{1}\in Y_{1},x\in X\\}\cup$ $\displaystyle\\{(x,x_{2})\mid x\in X,x_{2}\in X_{2}\\}\cup$ $\displaystyle\\{(y,y_{2})\mid y\in Y,y_{2}\in Y_{2}\\}.$ The construction of $Q$ is illustrated in Figure 1. Figure 1: The source-sink isolating gadget $Q$ and the whole graph $R\cup Q$. There are bipartite cliques between $X$ and $Y_{1},X_{2}$, as well as $Y$ and $X_{1},Y_{2}$, and there are perfect matchings between $X$ and $X_{1}$, $Y$ and $Y_{1}$. Fixing some $x\in X$ and $y\in Y$, we let $\kappa=\kappa_{R_{xy}}(x,y)$ for short. We first show $\kappa_{R\cup Q}(x,y)\geq\kappa+|X|+|Y|$. From the flow view of vertex connectivity, there are $\kappa$ internal vertex-disjoint paths from $x$ to $y$ in $R_{xy}$. Combining 3.3 and $V(Q)\cap V(R_{xy})=\\{x,y\\}$, there are $\kappa+|X|+|Y|$ internal vertex-disjoint paths in $R\cup Q$. Therefore, $\kappa_{R\cup Q}(x,y)\geq\kappa+|X|+|Y|$. ###### Claim 3.3. There are $|X|+|Y|$ internal vertex-disjoint paths from $x$ to $y$ in $Q$. ###### Proof. The first path is $x\to\hat{x}_{1}\to y$. Each of the next $|X|-1$ paths corresponds to each $x^{\prime}\in X$ s.t. $x^{\prime}\neq x$, the path $x\to\hat{x}^{\prime}_{2}\to x^{\prime}\to\hat{x}^{\prime}_{1}\to y$ concretely, where $\hat{x}^{\prime}_{1}$ and $\hat{x}^{\prime}_{2}$ are copies of $x^{\prime}$ in $X_{1}$ and $X_{2}$. Symmetrically, there is a path $x\to y_{1}\to y$ and $|Y|-1$ paths, each of which corresponds to each $y^{\prime}\in Y$ s.t. $y^{\prime}\neq y$ (namely the path $x\to\hat{y}^{\prime}_{1}\to y^{\prime}\to\hat{y}^{\prime}_{2}\to y$, where $\hat{y}^{\prime}_{1}$ and $\hat{y}^{\prime}_{2}$ are copies of $y^{\prime}$ in $Y_{1}$ and $Y_{2}$). Observe that these $|X|+|Y|$ paths are internal vertex-disjoint. ∎ We then argue from the cut view that $\kappa_{R\cup Q}(x,y)\leq\kappa+|X|+|Y|$. Consider the vertex set $S_{Q}=\\{x^{\prime}\mid x^{\prime}\in X,x^{\prime}\neq x\\}\cup\\{y^{\prime}\mid y^{\prime}\in Y,y^{\prime}\neq y\\}\cup\\{\hat{x}_{1},\hat{y}_{1}\\}$. After removing $S_{Q}$ from $R\cup Q$, observe that vertices in both $R$ and $Q$ are only $x$ and $y$, so a simple path from $x$ to $y$ in graph $(R\cup Q)\setminus S_{Q}$ will be totally inside subgraphs $Q\setminus S_{Q}$ or $R\setminus S_{Q}$. Note that $x$ and $y$ are disconnected in $Q\setminus S_{Q}$. Moreover, subgraph $R\setminus S_{Q}$ is exactly $R_{xy}$, so removing $\kappa$ vertices can disconnect $x$ and $y$ in $R\setminus S_{Q}$. In conclusion, in graph $R\cup Q$, $x$ and $y$ can be disconnected by removing $|S_{Q}|+\kappa$ vertices, so $\kappa_{R\cup Q}(x,y)\leq\kappa+|X|+|Y|$. Finally, the size of $Q$ follows directly from the construction. ∎ ### 3.2 The Set-Intersection Filter We now introduce the set-intersection filter. For each $a\in A$, $d\in D$, the set-intersection filter $P_{ad}^{B}$ is a subgraph of the final $H$, which will “filter” the intersection $B_{a}\cap B_{d}$ from the whole set $B$ as Lemma 3.4 shows. It is constructed as follows. Let $V(P_{ad}^{B})=\\{a\\}\cup B\cup B^{\prime}\cup\\{d\\}$, where $B^{\prime}$ duplicates vertices in $B$. For each vertex $b\in B$, we use $\hat{b}^{\prime}$ to denote its copy in $B^{\prime}$, and for each (non-)neighbor sets $B_{a},\bar{B}_{a},B_{d},\bar{B}_{d}\subseteq B$, we use $B^{\prime}_{a},\bar{B}^{\prime}_{a},B^{\prime}_{d},\bar{B}^{\prime}_{d}\subseteq B^{\prime}$ to denote their counterparts respectively. The edge set of $P_{ad}^{B}$ is constructed by $\displaystyle E(P_{ad}^{B})=$ $\displaystyle\\{(a,b)\mid b\in B\\}\cup\\{(b,d)\mid b\in\bar{B}_{d}\\}\cup$ $\displaystyle\\{(a,\hat{b}^{\prime})\mid b\in B_{a}\\}\cup\\{(b^{\prime},d)\mid b^{\prime}\in B^{\prime}\\}\cup$ $\displaystyle\\{(b,\hat{b}^{\prime})\mid b\in B\\}.$ See Figure 2 for an illustration of $P^{B}_{ad}$. Figure 2: The set-intersection filter. The sets $P_{ad}^{B}$ and $R$ are the areas surrounded by dotted lines. The construction of $E(P^{B}_{ad})$ can be interpreted in the following intuitive way. * • First, the edges $\\{(a,b)\mid b\in B\\}$ and $\\{(b,d)\mid b\in\bar{B}_{d}\\}$ create vertex-disjoint paths of the format $a\to b\to d$ for all $b\in\bar{B}_{d}$, which implies $\bar{B}_{d}$ will be cut from $B$ in every vertex cut of $(a,d)$. * • Second, the edges $\\{(a,\hat{b}^{\prime})\mid b\in B_{a}\\}$ and $\\{(b^{\prime},d)\mid b^{\prime}\in B^{\prime}\\}$ create vertex-disjoint paths of the format $a\to\hat{b}^{\prime}\to d$ for all $b\in B_{a}$, which analogously implies that $B^{\prime}_{a}$ will be cut from $B^{\prime}$ in every $(a,d)$-vertex cut. * • Third, for every $(a,d)$-vertex cut, after $\bar{B}_{d}$ and $B^{\prime}_{a}$ are cut from $B$ and $B^{\prime}$ respectively from the above discussion, either $b$ or $\hat{b}^{\prime}$ should be cut for all $b\in B_{d}\cap\bar{B}_{a}$. The reason is that the edges $\\{(b,\hat{b}^{\prime})\mid b\in B\\}$ form a matching between $B$ and $B^{\prime}$, which will create vertex-disjoint paths of the format $a\to b\to\hat{b}^{\prime}\to d$ for all $b\in B_{d}\cap\bar{B}_{a}$. Therefore, suppose that in the third step we choose to cut vertex $b$ of all $b\in B_{d}\cap\bar{B}_{a}$. (In the formal proof of Lemma 3.4, we will see that, cutting $b$ rather than $\hat{b}^{\prime}$ for all $b\in B_{d}\cap\bar{B}_{a}$ is always a better choice when considering vertex min cut between $a$ and $d$.) Now $\bar{B}_{d}$ is cut in the first step and $B_{d}\cap\bar{B}_{a}$ is cut in the third step, so vertices in $B$ that survive are $B\setminus\bar{B}_{d}\setminus(B_{d}\cap\bar{B}_{a})=B_{a}\cap B_{d}$. Therefore, the set-intersection filter indeed obtain $B_{a}\cap B_{d}$ as desired. ###### Lemma 3.4. For each $a\in A,d\in D$, the set-intersection filter $P^{B}_{ad}$ has the following property. Let $R$ be an undirected graph such that $V(R)\cap V(P^{B}_{ad})=\\{a\\}\cup B\cup\\{d\\}$, then $\kappa_{R\cup P^{B}_{ad}}(a,d)=\kappa_{R^{B}_{ad}}(a,d)+|\bar{B}_{d}|+|B_{a}|+|B_{d}\cap\bar{B}_{a}|,$ where $R^{B}_{ad}=(R\cup P^{B}_{ad})\setminus((B\setminus(B_{a}\cap B_{d}))\cup B^{\prime})$ equivalent to $(R\setminus(B\setminus(B_{a}\cap B_{d})))\cup\\{(a,b)\mid b\in B_{a}\cap B_{d}\\}$, that is, the graph starting from $R$, removing non $B_{a}\cap B_{d}$ vertices and then adding edges connecting $a$ and each $b\in B_{a}\cap B_{d}$. ###### Proof. Let $\kappa=\kappa_{R^{B}_{ad}}(a,d)$ for short. We first show $\kappa_{R\cup P^{B}_{ad}}(a,d)\geq\kappa+|\bar{B}_{d}|+|B_{a}|+|B_{d}\cap\bar{B}_{a}|$. There are $\kappa$ internal vertex-disjoint paths from $a$ to $d$ in $R^{B}_{ad}$. Because $V(P^{B}_{ad})\cap V(R^{B}_{ad})=(B_{a}\cap B_{d})\cup\\{a,d\\}$, the paths from $a$ to $d$ in $P^{B}_{ad}\setminus(B_{a}\cap B_{d})$ are internal disjoint with those in $R^{B}_{ad}$, and there are $|\bar{B}_{d}|+|B_{a}|+|B_{d}\cap\bar{B}_{a}|$ of them by 3.5. Therefore, we have $\kappa+|\bar{B}_{d}|+|B_{a}|+|B_{d}\cap\bar{B}_{a}|$ internal vertex-disjoint paths from $a$ to $d$ in $R\cup P^{B}_{ad}$. ###### Claim 3.5. There are $|\bar{B}_{d}|+|B_{a}|+|B_{d}\cap\bar{B}_{a}|$ internal vertex- disjoint paths from $a$ to $d$ in $P^{B}_{ad}\setminus(B_{a}\cap B_{d})$. ###### Proof. The first $|\bar{B}_{d}|$ paths correspond to vertices $b\in\bar{B}_{d}$, each of which has a path $a\to b\to d$. The next $|B_{a}|$ paths correspond to vertices $b\in B_{a}$, each of which has a path $a\to\hat{b}^{\prime}\to d$. The last $|B_{d}\cap\bar{B}_{a}|$ paths correspond to vertices $b\in B_{d}\cap\bar{B}_{a}$, each of which has a path $a\to b\to\hat{b}^{\prime}\to d$. ∎ We then complete the proof by showing $\kappa_{R\cup P^{B}_{ad}}(a,d)\leq\kappa+|\bar{B}_{d}|+|B_{a}|+|B_{d}\cap\bar{B}_{a}|$. Let $S_{P}=\\{b\in B\mid b\in\bar{B}_{d}\\}\cup\\{\hat{b}^{\prime}\in B^{\prime}\mid b\in B_{a}\\}\cup\\{b\in B\mid b\in B_{d}\cap\bar{B}_{a}\\}$ be a vertex cut of $(a,d)$ in $P^{B}_{ad}$. Observe that in $(R\cup P^{B}_{ad})\setminus S_{P}$, a path from vertex $a$ to a vertex $b^{\prime}\in B^{\prime}\setminus S_{P}$ must go through vertex $d$, because each $b^{\prime}\in B^{\prime}\setminus S_{P}$ only connects to $d$ after removing $S_{P}$. Therefore, it is safe to ignore $B^{\prime}\setminus S_{P}$ when considering the vertex connectivity between $a$ and $d$ in graph $(R\cup P^{B}_{ad})\setminus S_{P}$, i.e. $\kappa_{(R\cup P^{B}_{ad})\setminus S_{P}}(a,d)=\kappa_{(R\cup P^{B}_{ad})\setminus(S_{P}\cup B^{\prime})}(a,d)=\kappa_{R^{B}_{ad}}(a,d)=\kappa,$ which means by further removing $\kappa$ vertices, we can disconnect $a$ and $d$ in $(R\cup P_{ad}^{B})\setminus S_{P}$. In conclusion, we can remove $|S_{P}|+\kappa$ vertices to disconnect $a$ and $d$ in $R\cup P^{B}_{ad}$, which implies $\kappa_{R\cup P^{B}_{ad}}(a,d)\leq\kappa+|\bar{B}_{d}|+|B_{a}|+|B_{d}\cap\bar{B}_{a}|$. ∎ For each $a\in A,d\in D$, we also define the set-intersection filter $P^{C}_{ad}$ similarly but symmetrically. Let $V(P^{C}_{ad})=\\{a\\}\cup C^{\prime}\cup C\cup\\{d\\}$, where $C^{\prime}$ duplicates vertices in $C$. For each $c\in C$, let $\hat{c}^{\prime}$ denote its copy in $C^{\prime}$. For each (non-)neighbor sets $C_{a},\bar{C}_{a},C_{d},\bar{C}_{d}\subseteq C$, let $C^{\prime}_{a},\bar{C}^{\prime}_{a},C^{\prime}_{d},\bar{C}^{\prime}_{d}\subseteq C^{\prime}$ denote their counterparts respectively. The edge set is $\displaystyle E(P^{C}_{ad})=$ $\displaystyle\\{(a,c)\mid c\in\bar{C}_{a}\\}\cup\\{(c,d)\mid c\in C\\}\cup$ $\displaystyle\\{(a,c^{\prime})\mid c^{\prime}\in C^{\prime}\\}\cup\\{(\hat{c}^{\prime},d)\mid c\in C_{d}\\}\cup$ $\displaystyle\\{(\hat{c}^{\prime},c)\mid c\in C\\}.$ The pattern $P^{C}_{ad}$ will also have similar properties as shown below. ###### Lemma 3.6. For each $a\in A,d\in D$, the set-intersection filter $P^{C}_{ad}$ has the following property. Let $R$ be an undirected graph such that $V(R)\cap V(P^{C}_{ad})=\\{a\\}\cup C\cup\\{d\\}$, then $\kappa_{R\cup P^{C}_{ad}}(a,d)=\kappa_{R^{C}_{ad}}(a,d)+|C_{d}|+|\bar{C}_{a}|+|C_{a}\cap\bar{C}_{d}|,$ where $R^{C}_{ad}=(R\cup P^{C}_{ad})\setminus((C\setminus(C_{a}\cap C_{d}))\cup C^{\prime})$ equivalent to $(R\setminus(C\setminus(C_{a}\cap C_{d})))\cup\\{(c,d)\mid c\in C_{a}\cap C_{d}\\}$, that is, the graph starting from $R$, removing non $C_{a}\cap C_{d}$ vertices and then adding edges connecting $d$ and each $c\in C_{a}\cap C_{d}$. ### 3.3 The Final Construction of the APVC Instance We are now ready to construct the final APVC instance $H$. For each $a\in A,d\in D$, we first construct a graph $H_{ad}$ as follows. Let $P^{B}_{ad}$ and $P^{C}_{ad}$ be the set-intersection filters defined in Section 3.2. Then the graph $H_{ad}$ will be defined by $H_{ad}=P^{B}_{ad}\cup P^{C}_{ad}\cup G_{4p}[B\cup C],$ which is the union of two set-intersection filters with edges in the graph $G_{4p}$ connecting $B$ and $C$. The final graph then will be constructed by $H=\bigcup_{a\in A,d\in D}H_{ad}\cup Q(A,D),$ where $Q(A,D)$ is the source-sink isolating gadget from Lemma 3.2 given $R=\bigcup_{a\in A,d\in D}H_{ad}$ and the sets $A,D$. See Figure 3 below for an illustration. Figure 3: The graph $H$. The edges connecting $B$ and $C$ (i.e. $G_{4p}[B\cup C]$) are omitted in the highlighted part. ###### Lemma 3.7. For each $a\in A$ and $d\in D$ of $G_{4p}$, we have $\kappa_{H}(a,d)\geq 4n+|N_{G}(a)\cap\bar{N}_{G}(d)|+|\bar{N}_{G}(a)\cap N_{G}(d)|.$ (1) Furthermore, without ambiguity let $a$ and $d$ also denote their original vertices in $G$. Then there is a 4-clique in $G$ containing $a$ and $d$ if and only if $a$ and $d$ are adjacent in $G_{4p}$ and $\kappa_{H}(a,d)\geq 4n+|N_{G}(a)\cap\bar{N}_{G}(d)|+|\bar{N}_{G}(a)\cap N_{G}(d)|+1.$ (2) ###### Proof. Fixing some $a\in A,d\in D$, we first apply Lemma 3.2 on $H$ and vertex sets $A,D$, which gives $\kappa_{H}(a,d)=\kappa_{R_{ad}}(a,d)+|A|+|D|$, where $R_{ad}=(\bigcup_{a\in A,d\in D}H_{ad})\setminus((A\cup D)\setminus\\{a,d\\})$. In fact, the graph $R_{ad}$ is exactly $H_{ad}$, because the induced subgraphs $H_{ad}[B\cup B^{\prime}\cup C\cup C^{\prime}]$ are the same for all $a\in A,d\in D$ by construction. Therefore, we have $\kappa_{H}(a,d)=\kappa_{H_{ad}}(a,d)+|A|+|D|.$ (3) Recall that $H_{ad}=P^{B}_{ad}\cup P^{C}_{ad}\cup G_{4p}[B\cup C]$. Let $R_{1}=P^{C}_{ad}\cup G_{4p}[B\cup C]$. We apply Lemma 3.4 on $P^{B}_{ad}$ and $R_{1}$, which gives $\kappa_{H_{ad}}(a,d)=\kappa_{R_{1}\cup P^{B}_{ad}}(a,d)=\kappa_{R^{\prime}_{1}}(a,d)+|\bar{B}_{d}|+|B_{a}|+|B_{d}\cap\bar{B}_{a}|,$ (4) where $\displaystyle R^{\prime}_{1}$ $\displaystyle=(R_{1}\setminus(B\setminus(B_{a}\cap B_{d})))\cup\\{(a,b)\mid b\in B_{a}\cap B_{d}\\}$ $\displaystyle=P^{C}_{ad}\cup G_{4p}[(B_{a}\cap B_{d})\cup C]\cup\\{(a,b)\mid b\in B_{a}\cap B_{d}\\}.$ Let $R_{2}=G_{4p}[(B_{a}\cap B_{d})\cup C]\cup\\{(a,b)\mid b\in B_{a}\cap B_{d}\\}$. We apply Lemma 3.6 on $P^{C}_{ad}$ and $R_{2}$, which gives $\kappa_{R^{\prime}_{1}}(a,d)=\kappa_{R_{2}\cup P^{C}_{ad}}(a,d)=\kappa_{R^{\prime}_{2}}(a,d)+|C_{d}|+|\bar{C}_{a}|+|C_{a}\cap\bar{C}_{d}|,$ (5) where $\displaystyle R^{\prime}_{2}=$ $\displaystyle(R_{2}\setminus(C\setminus(C_{a}\cap C_{d})))\cup\\{(c,d)\mid c\in C_{a}\cap C_{d}\\}$ $\displaystyle=$ $\displaystyle G_{4p}[(B_{a}\cap B_{d})\cup(C_{a}\cap C_{d})]\cup\\{(a,b)\mid b\in B_{a}\cap B_{d}\\}\cup$ $\displaystyle\\{(c,d)\mid c\in C_{a}\cap C_{d}\\}.$ We use $\hat{H}_{ad}$ to denote $R^{\prime}_{2}$ in the remaining proof and note that it is equivalent to the definition of $\hat{H}_{ad}$ in Section 1.2. Because $\kappa_{\hat{H}_{ad}}(a,d)\geq 0$, combining Equations 3, 4 and 5, we get $\kappa_{H}(a,d)\geq|A|+|D|+|\bar{B}_{d}|+|B_{a}|+|B_{d}\cap\bar{B}_{a}|+|C_{d}|+|\bar{C}_{a}|+|C_{a}\cap\bar{C}_{d}|.$ (6) We now prove the second part of the lemma. If $a$ and $d$ are not adjacent in $G_{4p}$, they are not adjacent in $G$ either, so there is no 4-clique in $G$ containing them. Otherwise, $a$ and $d$ are adjacent, we claim that there is a 4-clique containing $a$ and $d$ in $G_{4p}$ (which is equivalent to the existence of a 4-clique containing $a$ and $d$ in $G$ by Definition 2.2) if and only if $\kappa_{\hat{H}_{ad}}(a,d)\geq 1$. If the 4-clique exists, let $b\in B$ and $c\in C$ be the other vertices in the 4-clique. Then there is a path $a\to b\to c\to d$ in $\hat{H}_{ad}$ from the construction, which implies $\kappa_{\hat{H}_{ad}}(a,d)\geq 1$. On the other hand, if $\kappa_{\hat{H}_{ad}}(a,d)\geq 1$, there is a path $a\to b\to c\to d$, and $(a,b,c,d)$ forms a 4-clique in $G_{4p}$. Combining this claim with Equations 3, 4 and 5 gives that there is a 4-clique containing $a$ and $d$ in $G$ if and only if $\kappa_{H}(a,d)\geq|A|+|D|+|\bar{B}_{d}|+|B_{a}|+|B_{d}\cap\bar{B}_{a}|+|C_{d}|+|\bar{C}_{a}|+|C_{a}\cap\bar{C}_{d}|+1.$ (7) Finally, by the construction of $G_{4p}$, we have $|A|=|B|=|C|=|D|=n$, $|B_{a}|=|C_{a}|$, $|B_{d}|=|C_{d}|$, $|B_{d}\cap\bar{B}_{a}|=|N_{G}(d)\cap\bar{N}_{G}(a)|$ and $|C_{a}\cap\bar{C}_{d}|=|N_{G}(a)\cap\bar{N}_{G}(d)|$. Combining them with Inequalities (6) and (7) completes the proof. ∎ ###### Proof of Theorem 3.1. Assume for contradiction that there exists a combinatorial algorithm $\mathcal{A}$ for the APVC problem with running time $O(n^{4-\epsilon})$ for some constant $\epsilon>0$. Let $G$ be an arbitrary 4-clique instance. We first construct the 4-partite graph $G_{4p}$ and the graph $H$ following the construction in this section. Note that $V(H)=V(Q)\cup A\cup B\cup B^{\prime}\cup C\cup C^{\prime}\cup D$, so $|V(H)|=O(n)$ by the construction and Lemma 3.2. Also, $H$ can be constructed in $O(n^{2})$ time directly. By Equation 2, we can solve the 4-clique problem by first running $\mathcal{A}$ on graph $H$ and then checking for each adjacent $a\in A,d\in D$ whether $\kappa_{H}(a,d)$ reaches the threshold value $4n+|N_{G}(a)\cap\bar{N}_{G}(d)|+|\bar{N}_{G}(a)\cap N_{G}(d)|+1$. This takes $O(n^{4-\epsilon})+O(n^{3})=O(n^{4-\epsilon})$ time because computing the threshold value for each $a\in A,d\in D$ takes totally $O(n^{3})$ extra time. This contradicts 2.1. ∎ ###### Remark 3.8. The proof of Theorem 3.1 basically shows that if the APVC problem can be solved in time $T_{\mathrm{APVC}}(n)$, then the 4-clique problem can be solved in time $T_{\mathrm{4clique}}(n)=O(T_{\mathrm{APVC}}(n)+n^{3})$. We note that this relation can be improved to $T_{\mathrm{4clique}}(n)=O(T_{\mathrm{APVC}}(n)+n^{\omega})$ if we use fast matrix multiplication to speed up the reduction. For general algorithms, another version of the 4-clique conjecture (see e.g. [DW22]) states that solving the 4-clique problem requires $n^{\omega(1,2,1)-o(1)}$ time. Therefore, assuming this conjecture, solving the APVC problem also requires $n^{\omega(1,2,1)-o(1)}$ time for general algorithms. ### 3.4 Further Results In this section, we will show several corollaries from the lower bound of the APVC problem. The first corollary is a conditional lower bound of the SSVC problem. ###### Corollary 3.9. Assuming 2.1, for $n$-vertex undirected unweighted graphs, there is no combinatorial algorithm that solves the SSVC problem in $O(n^{3-\epsilon})$ time for any constant $\epsilon>0$. ###### Proof. Assume for the contradiction that the SSVC problem can be solved in $O(n^{3-\epsilon})$ time. Then for an APVC instance $G$, we may treat every vertex in $G$ as a source to get the correct output. It takes $O(n)$ SSVC calls and therefore the complexity is $O(n\cdot n^{3-\epsilon})=O(n^{4-\epsilon})$, which contradicts Theorem 3.1 assuming 2.1. ∎ The second corollary is a conditional lower bound of the APVC problem for graphs with general density. ###### Corollary 3.10. Given any constant $\delta\in[0,1]$, assuming 2.1, there is no combinatorial algorithm that solves the APVC problem for $n$-vertex $m$-edge unweighted graphs, where $m=\Theta(n^{1+\delta})$ with running time $O(m^{2-\epsilon})$ for any constant $\epsilon>0$. ###### Proof. Assume that for some constants $\delta$ and $\epsilon$, such $O(m^{2-\epsilon})$-time algorithm $\mathcal{A}$ exists. Let $H$ be an $\hat{n}$-vertex $\hat{m}$-edge APVC hard instance with $\hat{m}=\Theta(\hat{n}^{2})$, constructed as above for some 4-clique instance. Let $G$ be the union of $H$ and $\Theta(\hat{m}^{1/(1+\delta)})$ isolated vertices. Observe that $G$ now has $n=\hat{n}+\Theta(\hat{m}^{1/(1+\delta)})$ vertices and $m=\hat{m}$ edges, i.e. $m=\Theta(n^{1+\delta})$. By applying algorithm $\mathcal{A}$ on $G$, the all-pairs vertex connectivity of $H$ can be computed in $O(m^{2-\epsilon})$, i.e. $O(\hat{n}^{4-2\epsilon})$ time, contradicting 2.1 by the argument in the proof of Theorem 3.1. ∎ The last corollary is a conditional lower bound of the SSVC problem for graphs with general density. The proof is omitted since it is analogous to Corollary 3.10. ###### Corollary 3.11. Given any constant $\delta\in[0,1]$, assuming 2.1, there is no combinatorial algorithm that solves the SSVC problem for $n$-vertex $m$-edge unweighted graphs, where $m=\Theta(n^{1+\delta})$, with running time $O(m^{3/2-\epsilon})$ for any constant $\epsilon>0$. ## 4 The Lower Bound for Steiner Vertex Connectivity Problem In this section, we will prove Theorem 4.1, a conditional lower bound of the Steiner vertex connectivity problem in undirected unweighted graphs, conditioning on the edge-universal 4-clique problem. ###### Theorem 4.1. For $n$-vertex undirected unweighted graphs, assuming 2.3, there is no combinatorial algorithm that solves the Steiner vertex connectivity problem in $O(n^{4-\epsilon})$ time for any constant $\epsilon>0$. Given an $n$-vertex edge-universal 4-clique instance $G$ with a set $E_{\mathrm{dem}}$ of _demand_ edges, let $H$ be the hard APVC instance we construct in Section 3. We will strengthen $H$ to another graph $J$ such that the edge-universal 4-clique problem in $G$ can be reduced to a Steiner vertex connectivity problem in $J$ of the terminal set $A\cup D$. As mentioned in Section 1.2, the main idea is to add extra “flow” paths from $A$ to $D$ to make the additive deviations between $\kappa_{J}(a,d)$ and $\kappa_{\hat{H}_{ad}}(a,d)$ uniform for all pairs of $a\in A$ and $d\in D$. Furthermore, to exclude the interference of vertex connectivity of pairs $(a_{1},a_{2})$ s.t. $a_{1},a_{2}\in A$ or $(d_{1},d_{2})$ s.t. $d_{1},d_{2}\in D$, we artificially add large additive deviations for these pairs. The construction of $J$ is as follows and see Figure 4 for an illustration. The vertex set $V(J)=V(H)\cup Z\cup W\cup A^{\prime}\cup D^{\prime}$, where $Z,W,A^{\prime},D^{\prime}$ are disjoint groups of additional vertices to create extra “flow” paths. Concretely, $Z$ and $W$ will be copies of original vertex set $V(G)$, and $A^{\prime}$ and $D^{\prime}$ are additional sets of vertices of size $|A^{\prime}|=|D^{\prime}|=10n$. Let $\displaystyle E_{Z}=$ $\displaystyle\\{(a,z)|a\in A,z\in Z,(a,z)\in E(G)\\}\cup$ $\displaystyle\\{(d,z)|d\in D,z\in Z,(d,z)\in E(G)\\}$ and $\displaystyle E_{W}=$ $\displaystyle\\{(a,w)|a\in A,w\in W,(a,w)\not\in E(G)\\}\cup$ $\displaystyle\\{(d,w)|d\in D,w\in W,(d,w)\not\in E(G)\\}$ be the extra edges to “equalize the deviation” of all pairs $(a,d)$ between $A$ and $D$. Let $E_{A^{\prime}}=\\{(a,a^{\prime})\mid a\in A,a^{\prime}\in A^{\prime}\\}$ and $E_{D^{\prime}}=\\{(d,d^{\prime})\mid d\in D,d^{\prime}\in D^{\prime}\\}$ be extra edges that bring large deviations to pairs inside $A$ or $D$. Finally, we construct a set of extra edges $E_{AD}=\\{(a,d)\mid a\in A,d\in D,(a,d)\notin E_{\mathrm{dem}}\\}$ to prevent non-demand pairs $(V(G)\times V(G))\setminus E_{\mathrm{dem}}$ from affecting the Steiner connectivity value. The whole edge set $E(J)=V(H)\cup E_{Z}\cup E_{W}\cup E_{A^{\prime}}\cup E_{D^{\prime}}\cup E_{AD}$. Figure 4: The graph $J$. There are bipartite cliques between $A$ and $A^{\prime}$, $D$ and $D^{\prime}$, and the edges between $A,D$ and $Z,W$ are linked based on $E_{Z},E_{W}$. The correctness of the reduction will be established by the following lemmas. ###### Lemma 4.2. In graph $J$, for each $a_{1},a_{2}\in A$ with $a_{1}\neq a_{2}$ and $d_{1},d_{2}\in D$ with $d_{1}\neq d_{2}$, we have $\kappa_{J}(a_{1},a_{2})\geq 10n$ and $\kappa_{J}(d_{1},d_{2})\geq 10n$. ###### Proof. For each pair $a_{1},a_{2}\in A$ with $a_{1}\neq a_{2}$, we can construct at least $10n$ internally vertex-disjoint paths from $a_{1}$ to $a_{2}$ via vertices in $A^{\prime}$ and edges in $E_{A^{\prime}}$, so $\kappa_{J}(a_{1},a_{2})\geq 10n$. Similarly, $\kappa_{J}(d_{1},d_{2})\geq 10n$. ∎ ###### Lemma 4.3. In graph $J$, for each $a\in A$, $d\in D$ with $(a,d)\not\in E_{\mathrm{dem}}$, $\kappa_{J}(a,d)\geq 5n+1$. ###### Proof. From our construction, the number of nodes in $Z$ adjacent to both $a$ and $d$ is $|\\{z\in Z|(a,z)\in E_{Z},(d,z)\in E_{Z}\\}|=|N_{G}(a)\cap N_{G}(d)|,$ and the number of nodes in $W$ adjacent to both $a$ and $d$ is $|\\{w\in W|(a,w)\in E_{Z},(d,w)\in E_{Z}\\}|=|\bar{N}_{G}(a)\cap\bar{N}_{G}(d)|.$ Furthermore, there is an individual edge in $E_{AD}$ directly connecting $a$ and $d$ for each such $(a,d)\notin E_{\mathrm{dem}}$. Therefore, we have $|N_{G}(a)\cap N_{G}(d)|+|\bar{N}_{G}(a)\cap\bar{N}_{G}(d)|+1$ extra vertex-disjoint paths from $a$ to $d$. Furthermore, Inequality (1) from Equation 2 shows that $\kappa_{H}(a,d)\geq 4n+|N_{G}(a)\cap\bar{N}_{G}(d)|+|\bar{N}_{G}(a)\cap N_{G}(d)|$, so we can construct a vertex-disjoint path set with size at least $\displaystyle 4n$ $\displaystyle+|N_{G}(a)\cap\bar{N}_{G}(d)|+|\bar{N}_{G}(a)\cap N_{G}(d)|$ $\displaystyle+|N_{G}(a)\cap N_{G}(d)|+|\bar{N}_{G}(a)\cap\bar{N}_{G}(d)|+1=5n+1,$ which implies $\kappa_{J}(a,d)\geq 5n+1$. ∎ ###### Lemma 4.4. In graph $J$, for each $a\in A$, $d\in D$ with $(a,d)\in E_{\mathrm{dem}}$, there is a 4-clique containing $a$ and $d$ in $G$ if and only if $\kappa_{J}(a,d)\geq 5n+1$. ###### Proof. First, assume there is a 4-clique containing $a$ and $d$ in $G$. From Equation 2, we can construct $4n+|N_{G}(a)\cap\bar{N}_{G}(d)|+|\bar{N}_{G}(a)\cap N_{G}(d)|+1$ internally vertex-disjoint paths connecting $a$ and $d$ in $H$. An argument similar to the proof of Lemma 4.3 shows that we have extra $|N_{G}(a)\cap N_{G}(d)|+|\bar{N}_{G}(a)\cap\bar{N}_{G}(d)|$ internally vertex-disjoint paths via vertices in $Z,W$ and edges in $E_{Z},E_{W}$. Therefore, in graph $J$, we can find a set of internally vertex-disjoint paths from $a$ to $d$ with size $\displaystyle 4n$ $\displaystyle+|N_{G}(d)\cap\bar{N_{G}}(a)|+|N_{G}(a)\cap\bar{N_{G}}(d)|+1$ $\displaystyle+|N_{G}(a)\cap N_{G}(d)|+|\bar{N_{G}}(a)\cap\bar{N_{G}}(d)|=5n+1.$ Now assume there is no 4-clique containing $a$ and $d$ in $G$. From Lemma 3.2, we define $J_{ad}=J\setminus((A\cup D)\setminus\\{a,d\\})$ by removing vertices in $A$ and $D$ other than $a$ and $d$, and then we have $\kappa_{J}(a,d)=\kappa_{J_{ad}}(a,d)+|A|+|D|.$ (8) Note that although the source-sink isolating gadget $Q(A,D)$ is constructed by applying Lemma 3.2 on graph $H$ and vertex sets $A,D$, the construction of $Q(A,D)$ is independent from $H$, so we can also apply Lemma 3.2 on graph $J$ and vertex sets $A,D$. Next, we will show $\kappa_{J_{ad}}(a,d)\leq\kappa_{H_{ad}}(a,d)+|N_{G}(a)\cap N_{G}(d)|+|\bar{N}_{G}(a)\cap\bar{N}_{G}(d)|,$ (9) where $H_{ad}=H\setminus((A\cup D)\setminus\\{a,d\\})$ as defined in the proof of Equation 2. The reason is that, compared to $H_{ad}$, the extra vertices in $J_{ad}$ are in $A^{\prime},D^{\prime},Z,W$, which are only directly connected to $a$ or $d$. Let $C_{H}$ be the minimum vertex cut of size $|C_{H_{ad}}|=\kappa_{H_{ad}}(a,d)$ whose removal disconnects $a$ and $d$ in $H_{ad}$. Let $\displaystyle C_{J_{ad}}=$ $\displaystyle C_{H_{ad}}\cup\\{z\in Z\mid(a,z)\in E_{Z},(d,z)\in E_{Z}\\}\cup$ $\displaystyle\\{w\in W\mid(a,w)\in E_{Z},(d,w)\in E_{Z}\\}.$ We can easily observe that removing $C_{J_{ad}}$ will disconnect $a$ and $d$ in $J_{ad}$, which immediately implies Inequality (9). Finally, combining Equation (8) and Inequality (9), we have $\displaystyle\kappa_{J}(a,d)\leq$ $\displaystyle\kappa_{H_{ad}}(a,d)+|A|+|D|+|N_{G}(a)\cap N_{G}(d)|$ $\displaystyle+|\bar{N}_{G}(a)\cap\bar{N}_{G}(d)|.$ From Equation 2, we know that when there is no 4-clique containing $a$ and $d$ in $G$, $\kappa_{H}(a,d)=4n+|N_{G}(a)\cap\bar{N}_{G}(d)|+|\bar{N}_{G}(a)\cap N_{G}(d)|$. Combining $\kappa_{H}(a,d)=\kappa_{H_{ad}}(a,d)+|A|+|D|$ (by Lemma 3.2 again), $\displaystyle\kappa_{J}(a,d)\leq$ $\displaystyle\kappa_{H_{ad}}(a,d)+|A|+|D|$ $\displaystyle+|N_{G}(a)\cap N_{G}(d)|+|\bar{N}_{G}(a)\cap\bar{N}_{G}(d)|$ $\displaystyle=$ $\displaystyle\kappa_{H}(a,d)+|N_{G}(a)\cap N_{G}(d)|+|\bar{N}_{G}(a)\cap\bar{N}_{G}(d)|$ $\displaystyle=$ $\displaystyle 4n+|N_{G}(a)\cap\bar{N}_{G}(d)|+|\bar{N}_{G}(a)\cap N_{G}(d)|$ $\displaystyle+|N_{G}(a)\cap N_{G}(d)|+|\bar{N}_{G}(a)\cap\bar{N}_{G}(d)|$ $\displaystyle=$ $\displaystyle 5n<5n+1.$ ∎ We are now ready to prove Theorem 4.1. ###### Proof of Theorem 4.1. Given an edge-universal 4-clique instance $G$ with a set $E_{\mathrm{dem}}$ of demand edges. We first construct the Steiner vertex connectivity instance $J$ with terminal set $A\cup D$ as shown above. Then $G$ with $E_{\mathrm{dem}}$ is a yes instance of the edge-universal 4-clique problem if and only if the Steiner vertex connectivity of $A\cup D$ in $J$ is at least $5n+1$, by Lemmas 4.2, 4.3 and 4.4. For time analysis, note that $V(J)=O(n)$ by construction. Therefore, an $O(n^{4-\epsilon})$-time combinatorial algorithm for the Steiner vertex connectivity problem will imply an $O(n^{4-\epsilon})$ time combinatorial algorithm for the edge-universal 4-clique problem, which contradicts 2.3. ∎ ## 5 The Upper Bounds In this section, we will show the upper bounds of the APVC problem and SSVC problem in undirected unweighted sparse graphs (see Theorem 5.1 and Theorem 5.2 respectively). We only prove Theorem 5.1 in detail and briefly mention the proof of Theorem 5.2 since they are analogous. ###### Theorem 5.1. Given an undirected unweighted graph $G$ with $n$ vertices and $m$ edges, there is a randomized algorithm that solves the APVC problem in $\hat{O}(m^{\frac{11}{5}})$ time with high probability. Furthermore, assuming 2.5, the running time of this algorithm becomes $\hat{O}(m^{2})$. ###### Theorem 5.2. Given an undirected unweighted graph $G$ with $n$ vertices, $m$ edges and a source vertex, there is a randomized algorithm to solve the SSVC problem in $\hat{O}(m^{\frac{5}{3}})$ time with high probability. Furthermore, assuming 2.5, the running time becomes $\hat{O}(m^{1.5})$. A key subroutine to obtain the algorithm in Theorem 5.1 is the subroutine shown in Lemma 5.3, which can capture all vertex connectivity bounded by $k$. The proof of Lemma 5.3 has been shown in [IN12, PSY22] and we defer it to Section 5.1 for completeness. ###### Lemma 5.3. Given an $n$-vertex $m$-edge undirected unweighted graph $G$ and a threshold $k$, there is a randomized algorithm that computes the value of $\min\\{\kappa_{G}(u,v),k\\}$ for all vertex pairs $(u,v)$ in $\hat{O}(mk^{3}+n^{2})$ time with high probability. Furthermore, assuming 2.5, the running time of this algorithm becomes $\hat{O}(mk^{2}+n^{2})$. We now complete the proof of Theorem 5.1 using Lemma 5.3. ###### Proof of Theorem 5.1. As discussed in Section 1.2, our APVC algorithm will handle vertex connectivity between high-degree vertices and vertex connectivity involving low-degree vertices separately. Let $k$ be a degree threshold which will be chosen later. Given an input graph $G$, let $V_{h}=\\{v\in V(G)\mid\deg_{G}(v)>k\\}$ and $V_{l}=\\{v\in V(G)\mid\deg_{G}(v)\leq k\\}$. Note that for each vertex pair $(u,v)$ such that $u\in V_{l}$ or $v\in V_{l}$, the vertex connectivity $\kappa_{G}(u,v)$ is at most $k$. By Lemma 5.3, the vertex connectivity of all such pairs can be computed exactly in time $\hat{O}(mk^{3}+n^{2})$. Now consider all remaining pairs $(u,v)$ with $u\in V_{h}$ and $v\in V_{h}$. Since $|V_{h}|\leq O(m/k)$ by definition, there are $O(m^{2}/k^{2})$ remaining pairs. The vertex connectivity of each pair can be computed in $\hat{O}(m)$ time using one max flow call [CKL+22], so totally $\hat{O}(m^{3}/k^{2})$ time suffices for remaining pairs. The total running time of the above algorithm is then $\hat{O}(mk^{3}+n^{2}+m^{3}/k^{2})$, by choosing $k=\Theta(m^{2/5})$, the running time will be $\hat{O}(m^{11/5})$. Assuming 2.5, the running time will be improved to $\hat{O}(mk^{2}+n^{2}+m^{3}/k^{2})$, which is $\hat{O}(m^{2})$ by choosing $k=\Theta(m^{1/2})$. ∎ The proof of Theorem 5.2 is analogous and we sketch it below. ###### Proof of Theorem 5.2. Let $s\in V(G)$ be the source. Similarly, we partition $V(G)$ into two set $V_{h}=\\{v\in V(G)\mid\deg_{G}(v)>k\\}$ and $V_{l}=\\{v\in V(G)\mid\deg_{G}(v)\leq k\\}$ where $k$ is the degree threshold. The vertex connectivity $\kappa_{G}(s,v)$ for all pairs $(s,v)$ such that $v\in V_{l}$ can be computed in $\hat{O}(mk^{2}+n)$ time, using a subroutine analogous to Lemma 5.3. The vertex connectivity of remaining pairs can be computed in $\hat{O}(m^{2}/k)$ by trivially calling max flows. The total running time is $\hat{O}(mk^{2}+n+m^{2}/k)$, which will be $\hat{O}(m^{5/3})$ by choosing $k=\Theta(m^{1/3})$. Assuming 2.5, the running time is $\hat{O}(mk+n+m^{2}/k)$, which will be $\hat{O}(m^{3/2})$ by choosing $k=\Theta(m^{1/2})$. ∎ ### 5.1 Proof of Lemma 5.3 The algorithm and analysis follow the ideas in [IN12]. The algorithm is as follows. First, we create $t=O(k^{2}\log n)$ sample sets $U_{1},...,U_{t}$, each of which is generated by sampling each vertex in $V(G)$ independently with probability $1/k$. Moreover, for each set $U_{i}$, we construct a $k$-Gomory-Hu tree for element connectivity using Theorem 2.4. From Theorem 2.4, for each set $U_{i}$ and each pair $u,v\in U_{i}$, we can query $a_{i}(u,v)=\min\\{\kappa^{\prime}_{G,U_{i}}(u,v),k\\}$ in nearly constant time. Then for each $u,v\in V(G)$, we let the final output be $a(u,v)=\min\\{a_{i}(u,v)\mid u,v\in U_{i}\\}$. We then analyze the running time. The time to construct all $k$-Gomory-Hu trees is $\hat{O}(t\cdot mk)=\hat{O}(mk^{3})$. To compute all $a_{i}(u,v)$ and final output $a(u,v)$, each set $U_{i}$ will have size $\tilde{O}(n/k)$ w.h.p. by Chernoff bound, so the time is $\tilde{O}(t\cdot(n/t)^{2})=\tilde{O}(n^{2})$ (by aborting the algorithm when some $U_{i}$ has size not bounded by $\tilde{O}(n/k)$). Therefore, the total running time is $\hat{O}(mk^{3}+n^{2})$. The correctness is shown as follows. Consider a fixed $u,v\in V(G)$. First, $a(u,v)$ is well-defined with high probability, because for one sample set $U_{i}$, $u$ and $v$ are inside $U_{i}$ and $a_{i}(u,v)$ is well-defined with probability $1/k^{2}$ and there are $O(k^{2}\log n)$ sample sets. Given that $a(u,v)$ is well-defined, we know $a(u,v)\geq\min\\{\kappa_{G}(u,v),k\\}$ since $\kappa^{\prime}_{G,U_{i}}(u,v)\geq\kappa_{G}(u,v)$ for all $U_{i}$ by the definition of element connectivity. By the same reason, if $\kappa_{G}(u,v)\geq k$, we must have $a(u,v)=k=\min\\{\kappa_{G}(u,v),k\\}$ which is the correct answer. From now we suppose $\kappa_{G}(u,v)<k$ and we are going to show there exists $U_{i}$ such that $u,v\in U_{i}$ and $a_{i}(u,v)=\kappa_{G}(u,v)$ with high probability. Note that $\kappa^{\prime}_{G,U_{i}}(u,v)=\kappa$ if $U_{i}$ is disjoint with some minimum $u$-$v$ vertex cut $C_{u,v}$. From our sampling strategy, $U_{i}$ contains $u,v$ and is disjoint with $C_{u,v}$ with probability $\frac{1}{k^{2}}(1-\frac{1}{k})^{k}=\Omega(1/k^{2})$. Because there are $O(k^{2}\log n)$ sample set, $U_{i}$ exists with high probability. If we assuming 2.5, we simply subsitute $k$-Gomory-Hu tree for element connectivity with element connectivity Gomory-Hu tree and follow the same algorithm and analysis. We will obtain an algorithm with running time $\hat{O}(mk^{2}+n^{2})$. ## References * [ABHS22] Amir Abboud, Karl Bringmann, Danny Hermelin, and Dvir Shabtay. Scheduling lower bounds via AND subset sum. J. Comput. Syst. Sci., 127:29–40, 2022. * [AGI+19] Amir Abboud, Loukas Georgiadis, Giuseppe F. Italiano, Robert Krauthgamer, Nikos Parotsidis, Ohad Trabelsi, Przemyslaw Uznanski, and Daniel Wolleb-Graf. Faster algorithms for all-pairs bounded min-cuts. In Christel Baier, Ioannis Chatzigiannakis, Paola Flocchini, and Stefano Leonardi, editors, 46th International Colloquium on Automata, Languages, and Programming, ICALP 2019, July 9-12, 2019, Patras, Greece, volume 132 of LIPIcs, pages 7:1–7:15. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2019. * [AKL+22] Amir Abboud, Robert Krauthgamer, Jason Li, Debmalya Panigrahi, Thatchaphol Saranurak, and Ohad Trabelsi. Breaking the cubic barrier for all-pairs max-flow: Gomory-hu tree in nearly quadratic time. In 63rd IEEE Annual Symposium on Foundations of Computer Science, FOCS 2022, Denver, CO, USA, October 31 - November 3, 2022, pages 884–895. IEEE, 2022. * [AKT20] Amir Abboud, Robert Krauthgamer, and Ohad Trabelsi. New algorithms and lower bounds for all-pairs max-flow in undirected graphs. In Shuchi Chawla, editor, Proceedings of the 2020 ACM-SIAM Symposium on Discrete Algorithms, SODA 2020, Salt Lake City, UT, USA, January 5-8, 2020, pages 48–61. SIAM, 2020. * [AKT21a] Amir Abboud, Robert Krauthgamer, and Ohad Trabelsi. APMF < apsp? gomory-hu tree for unweighted graphs in almost-quadratic time. In 62nd IEEE Annual Symposium on Foundations of Computer Science, FOCS 2021, Denver, CO, USA, February 7-10, 2022, pages 1135–1146. IEEE, 2021. * [AKT21b] Amir Abboud, Robert Krauthgamer, and Ohad Trabelsi. Subcubic algorithms for gomory-hu tree in unweighted graphs. In Samir Khuller and Virginia Vassilevska Williams, editors, STOC ’21: 53rd Annual ACM SIGACT Symposium on Theory of Computing, Virtual Event, Italy, June 21-25, 2021, pages 1725–1737. ACM, 2021. * [AWW16] Amir Abboud, Virginia Vassilevska Williams, and Joshua R. Wang. Approximation and fixed parameter subquadratic algorithms for radius and diameter in sparse graphs. In Robert Krauthgamer, editor, Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2016, Arlington, VA, USA, January 10-12, 2016, pages 377–391. SIAM, 2016. * [Ben95] András A. Benczúr. Counterexamples for directed and node capacitated cut-trees. SIAM J. Comput., 24(3):505–510, 1995. * [BGL17] Karl Bringmann, Allan Grønlund, and Kasper Green Larsen. A dichotomy for regular expression membership testing. In Chris Umans, editor, 58th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2017, Berkeley, CA, USA, October 15-17, 2017, pages 307–318. IEEE Computer Society, 2017. * [CKL+22] Li Chen, Rasmus Kyng, Yang P. Liu, Richard Peng, Maximilian Probst Gutenberg, and Sushant Sachdeva. Maximum flow and minimum-cost flow in almost-linear time. In 63rd IEEE Annual Symposium on Foundations of Computer Science, FOCS 2022, Denver, CO, USA, October 31 - November 3, 2022, pages 612–623. IEEE, 2022. * [CKM+11] Paul F. Christiano, Jonathan A. Kelner, Aleksander Madry, Daniel A. Spielman, and Shang-Hua Teng. Electrical flows, laplacian systems, and faster approximation of maximum flow in undirected graphs. In Lance Fortnow and Salil P. Vadhan, editors, Proceedings of the 43rd ACM Symposium on Theory of Computing, STOC 2011, San Jose, CA, USA, 6-8 June 2011, pages 273–282. ACM, 2011. * [CRX15] Chandra Chekuri, Thapanapong Rukkanchanunt, and Chao Xu. On element-connectivity preserving graph simplification. In Nikhil Bansal and Irene Finocchi, editors, Algorithms - ESA 2015 - 23rd Annual European Symposium, Patras, Greece, September 14-16, 2015, Proceedings, volume 9294 of Lecture Notes in Computer Science, pages 313–324. Springer, 2015. * [DW22] Mina Dalirrooyfard and Virginia Vassilevska Williams. Induced cycles and paths are harder than you think. In 63rd IEEE Annual Symposium on Foundations of Computer Science, FOCS 2022, Denver, CO, USA, October 31 - November 3, 2022, pages 531–542. IEEE, 2022. * [FF56] L. R. Ford and D. R. Fulkerson. Maximal flow through a network. Canadian Journal of Mathematics, 8:399–404, 1956. * [GH61] R. E. Gomory and T. C. Hu. Multi-terminal network flows. Journal of the Society for Industrial and Applied Mathematics, 9(4):551–570, 1961. * [GIKW19] Jiawei Gao, Russell Impagliazzo, Antonina Kolokolova, and Ryan Williams. Completeness for first-order properties on sparse structures with algorithmic applications. ACM Trans. Algorithms, 15(2):23:1–23:35, 2019. * [GR98] Andrew V. Goldberg and Satish Rao. Beyond the flow decomposition barrier. J. ACM, 45(5):783–797, 1998. * [HRG00] Monika Rauch Henzinger, Satish Rao, and Harold N. Gabow. Computing vertex connectivity: New bounds from old techniques. J. Algorithms, 34(2):222–250, 2000. * [IN12] Rani Izsak and Zeev Nutov. A note on labeling schemes for graph connectivity. Inf. Process. Lett., 112(1-2):39–43, 2012. * [Kar00] David R. Karger. Minimum cuts in near-linear time. J. ACM, 47(1):46–76, 2000. * [Kle69] D. Kleitman. Methods for investigating connectivity of large graphs. IEEE Transactions on Circuit Theory, 16(2):232–233, 1969. * [KLOS14] Jonathan A. Kelner, Yin Tat Lee, Lorenzo Orecchia, and Aaron Sidford. An almost-linear-time algorithm for approximate max flow in undirected graphs, and its multicommodity generalizations. In Chandra Chekuri, editor, Proceedings of the Twenty-Fifth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2014, Portland, Oregon, USA, January 5-7, 2014, pages 217–226. SIAM, 2014. * [LLW88] Nathan Linial, László Lovász, and Avi Wigderson. Rubber bands, convex embeddings and graph connectivity. Comb., 8(1):91–102, 1988. * [LNP+21] Jason Li, Danupon Nanongkai, Debmalya Panigrahi, Thatchaphol Saranurak, and Sorrachai Yingchareonthawornchai. Vertex connectivity in poly-logarithmic max-flows. In Samir Khuller and Virginia Vassilevska Williams, editors, STOC ’21: 53rd Annual ACM SIGACT Symposium on Theory of Computing, Virtual Event, Italy, June 21-25, 2021, pages 317–329. ACM, 2021. * [LP20] Jason Li and Debmalya Panigrahi. Deterministic min-cut in poly-logarithmic max-flows. In Sandy Irani, editor, 61st IEEE Annual Symposium on Foundations of Computer Science, FOCS 2020, Durham, NC, USA, November 16-19, 2020, pages 85–92. IEEE, 2020. * [LP21] Jason Li and Debmalya Panigrahi. Approximate gomory-hu tree is faster than _n_ \- 1 max-flows. In Samir Khuller and Virginia Vassilevska Williams, editors, STOC ’21: 53rd Annual ACM SIGACT Symposium on Theory of Computing, Virtual Event, Italy, June 21-25, 2021, pages 1738–1748. ACM, 2021. * [LPS21] Jason Li, Debmalya Panigrahi, and Thatchaphol Saranurak. A nearly optimal all-pairs min-cuts algorithm in simple graphs. In 62nd IEEE Annual Symposium on Foundations of Computer Science, FOCS 2021, Denver, CO, USA, February 7-10, 2022, pages 1124–1134. IEEE, 2021. * [Mad16] Aleksander Madry. Computing maximum flow with augmenting electrical flows. In Irit Dinur, editor, IEEE 57th Annual Symposium on Foundations of Computer Science, FOCS 2016, 9-11 October 2016, Hyatt Regency, New Brunswick, New Jersey, USA, pages 593–602. IEEE Computer Society, 2016. * [PSY22] Seth Pettie, Thatchaphol Saranurak, and Longhui Yin. Optimal vertex connectivity oracles. In Stefano Leonardi and Anupam Gupta, editors, STOC ’22: 54th Annual ACM SIGACT Symposium on Theory of Computing, Rome, Italy, June 20 \- 24, 2022, pages 151–161. ACM, 2022. * [She13] Jonah Sherman. Nearly maximum flows in nearly linear time. In 54th Annual IEEE Symposium on Foundations of Computer Science, FOCS 2013, 26-29 October, 2013, Berkeley, CA, USA, pages 263–269. IEEE Computer Society, 2013. * [Vas09] Virginia Vassilevska. Efficient algorithms for clique problems. Inf. Process. Lett., 109(4):254–257, 2009. * [vdBLN+20] Jan van den Brand, Yin Tat Lee, Danupon Nanongkai, Richard Peng, Thatchaphol Saranurak, Aaron Sidford, Zhao Song, and Di Wang. Bipartite matching in nearly-linear time on moderately dense graphs. In Sandy Irani, editor, 61st IEEE Annual Symposium on Foundations of Computer Science, FOCS 2020, Durham, NC, USA, November 16-19, 2020, pages 919–930. IEEE, 2020. * [WW18] Virginia Vassilevska Williams and R. Ryan Williams. Subcubic equivalences between path, matrix, and triangle problems. J. ACM, 65(5):27:1–27:38, 2018.
# Extrinsic to intrinsic mechanism crossover of anomalous Hall effect in the Ir-doped MnPtSn Heusler system Sk Jamaluddin School of Physical Sciences, National Institute of Science Education and Research, An OCC of Homi Bhabha National Institute, Jatni-752050, India Roumita Roy School of Physical Sciences, Indian Institute of Technology Goa, Ponda-403401, Goa, India Amitabh Das Solid State Physics Division, Bhabha Atomic Research Centre, Trombay Mumbai 400085, India Homi Bhabha National Institute, Anushaktinagar, Mumbai 400094, India Sudipta Kanungo<EMAIL_ADDRESS>School of Physical Sciences, Indian Institute of Technology Goa, Ponda-403401, Goa, India Ajaya K. Nayak <EMAIL_ADDRESS>School of Physical Sciences, National Institute of Science Education and Research, An OCC of Homi Bhabha National Institute, Jatni-752050, India ###### Abstract Recent findings of large anomalous Hall signal in nonferromagnetic and nonferrimagnetic materials suggest that the magnetization of the system is not a critical component for the realization of the anomalous Hall effect (AHE). Here, we present a combined theoretical and experimental study demonstrating the evolution of different mechanisms of AHE in a cubic Heusler system MnPt1-xIrxSn. With the help of magnetization and neutron diffraction studies, we show that the substitution of nonmagnetic Ir in place of Pt significantly reduces the net magnetic moment from 4.17 $\mu_{B}$/f.u. in MnPtSn to 2.78 $\mu_{B}$/f.u. for MnPt0.5Ir0.5Sn. In contrast, the anomalous Hall resistivity is enhanced by nearly three times from 1.6 $\mu\Omega$ cm in MnPtSn to about 5 $\mu\Omega$ cm for MnPt0.5Ir0.5Sn. The power law analysis of the Hall resistivity data suggests that the extrinsic contribution of AHE that dominates in the case of the parent MnPtSn almost vanishes for MnPt0.5Ir0.5Sn, where the intrinsic mechanism plays the major role. The experimental results are well supported by our theoretical study, which shows a considerable enhancement of the spin-orbit coupling when Ir is introduced into the system. Our finding of a crossover of the anomalous Hall effect with chemical engineering is a major contribution toward the recent interest in controlling the band topology of topological materials, both in bulk and thin-film forms. Anomaous Hall Effect, Heusler compounds ## I INTRODUCTION In general, the anomalous Hall effect (AHE) in ferromagnetic (FM)/ferrimagnetic (FiM) materials scales with the magnetization of the samples. However, the recent observation of extremely large anomalous Hall conductivity in some of the moderate magnetization based FM materials [2, 1, 3] and nonmagnetic materials [4, 5] suggests that the spin-orbit coupling driven band topology is one of the primary microscopic mechanisms that governs the AHE. In the case of FM materials, the total Hall resistivity can be expressed as $\rho_{xy}=\rho^{o}_{xy}+\rho^{A}_{xy}=R_{0}H_{z}+R_{s}M_{z}$, where $R_{0}$ and $R_{S}$ are the ordinary and anomalous Hall coefficients, respectively. It is well established that mainly two different microscopic mechanisms, such as a scattering-dependent extrinsic contribution and scattering-independent intrinsic process, dictate the AHE in FM materials [6]. The extrinsic mechanism of AHE involves two types of scattering processes, the skew scattering [7] and the side jump [8]. Both of them are related to the scattering of charge carriers by the impurity sites and are the consequences of spin-orbit interaction. In contrast, the scattering-independent intrinsic mechanism proposed by Karplus and Luttinger arises due to the anomalous velocity of the charge carriers caused by the electronic band structure in the presence of spin-orbit coupling (SOC) [9]. Furthermore, the inclusion of the Berry phase to understand the intrinsic mechanism of AHE throws a deep insight regarding the band topology of the system [10, 12, 11]. In real systems, different microscopic mechanisms of AHE have been understood in terms of a power law relation between anomalous Hall resistivity ($\rho^{A}_{xy}$) and longitudinal resistivity ($\rho_{xx}$). For the intrinsic contribution, $\rho^{A}_{xy}$ scales quadratically with the longitudinal resistivity $\rho_{xx}$, i.e., $\rho^{A}_{xy}\propto\rho^{2}_{xx}$ [13, 14]. In the case of extrinsic skew scattering, the $\rho^{A}_{xy}$ linearly scales with $\rho_{xx}$, i.e., $\rho^{A}_{xy}\propto\rho_{xx}$, while for the side jump the relation is quadratic in nature, i.e., $\rho^{A}_{xy}$ $\propto$ $\rho^{2}_{xx}$. These relations can be further generalized to a scaling law $\rho^{A}_{xy}\propto\rho^{\alpha}_{xx}$ to identify the dominant mechanism of AHE [15], where $\alpha$ = 2 and 1 correspond to intrinsic / side jump and extrinsic skew scattering contributions, respectively [16, 17, 18]. Since the intrinsic and side jump mechanisms exhibit a similar relationship with the longitudinal resistivity, it is difficult to differentiate them experimentally. Recently, Tian et al [19] introduced a new scaling relation called the Tian-Ye-Jin ($TYJ$) model to separate the different contributions of AHE experimentally [20, 21, 22, 23]. It has been shown that the side jump contribution is negligibly small in most of the metallic ferromagnetic systems [19, 20], except in some of the multilayer systems where interfacial scattering plays an important role [24, 25]. The new scaling relation without considering the side jump mechanism can be expressed as $\rho^{A}_{xy}=a\rho_{xx0}+b\rho^{2}_{xx}$ (1) where $\rho_{xx0}$ is the residual resistivity. The first term corresponds to the extrinsic skew scattering and the second one deals with the intrinsic mechanism, where $b$ is the intrinsic parameter. Figure 1: (a) Temperature-dependent zero-field-cooled (ZFC; open symbols) and field cooled (FC; closed symbols) magnetization [M(T)] curves for MnPt1-xIrxSn (x=0, 0.1, 0.2, 0.3, 0.5). (b) Field-dependent isothermal magnetization loops [$M(H)$] measured at 5 K for the MnPt1-xIrxSn samples. (c) Valence electron count (e/a) dependent Curie temperature and saturation magnetization for these samples. Although AHE is a well studied phenomenon in several materials, a systematic manipulation of different mechanisms as discussed above is extremely important for its future implementation in real devices. In this regard, Heusler materials lay a great platform to willfully tune the material property by chemical doping [26, 27, 28, 29]. In some of the recent studies, it is demonstrated that these materials can exhibit extremely large AHE due to the presence of Weyl nodes near the Fermi energy [30, 31]. Therefore, expanding the material base in the Heusler family with respect to the AHE and understanding its governing mechanism is an important step to move forward. In this report, we carry out a combined theoretical and experimental study to show a systematic change in the governing mechanism of AHE in cubic Heusler compounds MnPt1-xIrxSn. First, we show that the magnetic moment of the system can be greatly modified by changing only the nonmagnetic element Pt and Ir. Then with the help of the scaling relation for the AHE and theoretical study, we present a detailed study on the changeover of the AHE mechanism from extrinsic to intrinsic by replacing Pt with Ir. ## II Methods Polycrystalline samples of MnPt1-xIrxSn (x=0, 0.1, 0.2, 0.3, 0.5) are prepared by arc-melting of ultrahighly pure constituent elements Mn, Pt, Ir, and Sn in argon atmosphere. For better chemical homogeneity, the ingots are melted four to five times by flipping the sides. As-prepared samples are annealed in an evacuated quartz tube at 850∘C for seven days, followed by quenching in ice water mixture. To check the phase purity, powder x-ray diffraction (XRD) measurement is performed on all the samples using a Rigaku SmartLab x-ray diffractometer with a Cu-Kα source. The neutron diffraction patterns are recorded using the PD2 powder diffractometer ($\lambda$ = 1.2443 ${\AA}$) at the Dhruva reactor, Bhabha Atomic Research Centre, Mumbai, India, at selected temperatures between 1.5 K and 300 K. The magnetic characterizations are carried out with the help of a Quantum Design MPMS3 (SQUID-VSM). The ac transport measurements are performed using a physical property measurement system (PPMS; Quantum Design). The density-functional theory calculations are performed using the plane-wave basis and pseudopotential framework as implemented in the Vienna ${ab-initio}$ simulation package (VASP) [32, 33]. The exchange-correlation functional is employed following the Perdew-Burke-Ernzerhof (PBE) prescription [34]. The experimentally determined lattice parameters are used in the calculations, while relaxing the atomic positions toward equilibrium until the Hellmann- Feynman force becomes less than 0.001 eV/$\AA$. In order to incorporate correlations beyond the scope of mean-field PBE, the Hubbard on-site $U$ is introduced by performing GGA+$U$ calculations [35, 36] with suitable values of Ueff (U-JH) of 5 eV at the Mn site and 2 eV at the Pt and Ir sites, respectively. The effect of SOC is introduced as a fully relativistic approach in the self-consistent calculations. The self-consistent electronic structure calculations are performed with a plane-wave cutoff of 500 eV, and an 8$\times 8\times 8$ k-mesh is used for the Brillouin zone integration. ## III RESULTS AND DISCUSSION Figure 2: Rietveld refinement of powder neutron diffraction (ND) patterns. (a) Room-temperature Rietveld refinement of the powder ND pattern with 10% atomic disorders between Mn/Ir site of MnPt0.5Ir0.5Sn. The region between 34 o and 36 o has been excluded due to contributions from the cryostat. (b) Low- temperature Rietveld refinement of powder ND pattern with 10 % atomic disorder for MnPt0.5Ir0.5Sn. (c) Low-temperature Rietveld refinement of the powder ND pattern for MnPtSn. ### III.1 Structural and Magnetic Characterization The structural characterization for all the samples is carried out by room- temperature XRD measurements combined with Rietveld analysis using the FullProf suite (see Fig. S1 in the Supplemental Material) [37]. All the samples crystallize in a single cubic structure with space group $F\bar{4}3m$. A negligible change in the lattice parameter from $a=$ 6.25 ${\AA}$ in MnPtIn to 6.23 ${\AA}$ is found with Ir doping for MnPt0.5Ir0.5Sn. The temperature- dependent magnetization, $M(T)$, curves measured in zero-field-cooled (ZFC) and field-cooled (FC) modes for all the samples are shown in Fig. 1(a). The magnetic ordering temperature, $T_{C}$, for the parent MnPtSn sample is found to be 326 K. The $T_{C}$ decreases substantially with increasing Ir concentration and falls to 226 K for MnPt0.5Ir0.5Sn. The ZFC and FC $M(T)$ curves do not display any noticeable difference, signifying the absence of any spin-glass-like phase or considerable magnetic anisotropy in the system. The isothermal magnetization, $M(H)$, loops measured at 5 K for all the samples are depicted in Fig. 1(b). As can be seen, the saturation magnetization, $M_{S}$, significantly reduces from 4.17 $\mu_{B}$/f.u. in the case of MnPtSn to about 2.78 $\mu_{B}$/f.u. for MnPt0.5Ir0.5Sn. It is well known that the cubic $L2_{1}$ based Heusler materials show a linear variation of the magnetic moment with the valence electrons concentration [38, 39, 40]. In the present case, the substitution of Ir in place of Pt decreases the number of valence electron in the system. This is expected to greatly affect the electronic and magnetic properties of the system. Figure 1(c) shows the variation of total magnetic moment and $T_{C}$ with the valence electron concentration (e/a). As can be seen, both $M_{S}$ and $T_{C}$ exhibit a nearly linear dependency with the e/a ratio. To further understand the microscopic origin of the reduction in the total magnetic moment with Ir doping, we carry out a neutron diffraction (ND) study on our well-characterized polycrystalline samples MnPtSn and MnPt0.5Ir0.5Sn. In the case of MnPt0.5Ir0.5Sn, the ND measurements are performed at 2 K and 300 K. Since the $T_{C}$ of MnPt0.5Ir0.5Sn is 226 K, we use the ND pattern at 300 K for this sample to extract the nuclear parameters. For this purpose, the Rietveld refinement using the FullProf suite [41] is performed with the Wyckoff positions of Sn, Mn, and Pt/Ir as 4a (0, 0, 0), 4b ($\frac{1}{2}$, $\frac{1}{2}$, $\frac{1}{2}$), and 4c ($\frac{1}{4}$, $\frac{1}{4}$, $\frac{1}{4}$), respectively. However, we are unable to achieve a good fitting for the 300 K ND using the above mentioned Wyckoff positions. It is worth mentioning that the atomic disorder is frequently encountered in Heusler compounds with different transition metals of comparable ionic radii. Although Ir is substituted in place of Pt, the higher electronegativity difference between Ir and Sn in comparison to that of Pt-Sn may result in some of the Ir atoms sitting along with the Sn atoms [26, 42]. To check this, we replace about 10% of Mn atoms with Ir in the Mn Wyckoff position (4b) and put these Mn atoms in the Ir position (4c), as shown in Fig. 2(a). By doing so, we achieve a better fitting in comparison to that of without disorder [see Supplemental Material Fig. S2 (a)] [37]. In the case of 0% disorder, the values of the $\chi^{2}$ and $R_{Bragg}$ factors are 6.40 and 13.74, respectively, whereas for 10% atomic disorder the values of the $\chi^{2}$ and $R_{Bragg}$ factors are reduced to be 5.96 and 11.63, respectively. By further increasing the disorder, both the $\chi^{2}$ and $R_{Bragg}$ factors increase [see the Supplemental Material table S1]. Hence, it can be concluded from our Rietveld refinement that MnPt0.5Ir0.5Sn consists of about 10% Mn-Ir atomic disorder. Figure 3: (a) Field-dependent Hall resistivity [$\rho_{xy}(H)$] measured at 5 K for different samples. (b) Variation of anomalous Hall resistivity [$\rho_{xy}^{A}(H)$; red curve] and saturation magnetization ($M_{S}$; blue curve) with iridium composition at 5 K for MnPt1-xIrxSn. (c) Compositional- dependent anomalous Hall conductivity ($\sigma^{A}_{xy}$) and anomalous Hall angle (AHA) at 5 K for MnPt1-xIrxSn. Next, we concentrate on the low-temperature powder ND data to find out the magnetic ground state of our samples. In the case of MnPt0.5Ir0.5Sn, the Mn atoms occupy both 4b and 4c sites due to the atomic disorder. First, we perform the refinement considering the magnetic moment at the 4b and 4c sites. The Mn moment at the 4b site comes to be 3.14 $\mu_{B}$, whereas the 4c site displays a very small moment. This may be due to the presence of a very small percentage of Mn atoms at the 4c site that makes it difficult to get the magnetic information. Then we carry out the refinement by giving the magnetic moment only at the 4b position. This scenario does not affect the fitting parameter much and we obtain a magnetic moment of 3.17 $\mu_{B}$ for the Mn atoms. This results in a total moment of 2.85 $\mu_{B}/f.u.$ which is in good agreement with the experimental magnetic moment observed from the magnetization data. In the case of the MnPtSn sample, the $T_{C}$ is about 326 K. We are unable to perform the ND experiment above room temperature; only 2 K ND data are collected for this sample. Here, we consider the regular structure without disorder and the refined magnetic moment of Mn at the 4b site comes out to be 3.52 $\mu_{B}$. In general for Heusler compounds, the magnetic atoms sitting at the 4b and 4c sites align antiferromagnetically. Hence, the reduction in the magnetic moment for MnPt0.5Ir0.5Sn might be arising from the lower Mn concentration at the 4b site as well as the antiferromagnetic alignment between Mn at the 4b and Mn at the 4c site. Hence, our ND study categorically establishes that the reduction of magnetic moment in the case of the Ir-doped samples arises mainly from the induced atomic disorder in the system. ### III.2 Anomalous Hall Effect In order to explore the effect of Ir substitution on the electrical transport, we focus on a detailed Hall resistivity measurement on our well-characterized samples. The field-dependent Hall resistivity [$\rho_{xy}(H)$] measured at 5 K for different Ir-doped samples is shown in Fig. 3(a). As is found, $\rho_{xy}(H)$ exhibits a monotonic increment from about 1.6 $\mu\Omega$ cm for MnPtSn to about 5 $\mu\Omega$ cm for MnPt0.5Ir0.5Sn. It is worth noting here that the enhancement of the $\rho_{xy}(H)$ is observed despite a significant reduction in the magnetic moment from about 4.2 $\mu_{B}$/f.u. for MnPtSn to 2.8 $\mu_{B}$/f.u. for MnPt0.5Ir0.5Sn [Fig. 3(b)]. Although there is a three fold increase in the Hall resistivity, we observe a small decrease in the anomalous Hall conductivity $\sigma^{A}_{xy}$ with Ir doping [Fig. 3(c), left y-axis]. The decrease in the $\sigma^{A}_{xy}$ might be arising due to the increase in residual resistivity as a result of atomic disorder due to Ir doping. In contrast, the anomalous Hall angle (AHE) considerably increases with increasing the Ir concentration. This suggests that the Ir doping actually helps in the overall conversion of the longitudinal resistivity to Hall resistivity. Figure 4: (a-e) Anomalous Hall resistivity ($\rho^{A}_{xy}$) as a function of square of longitudinal resistivity ($\rho^{2}_{xx}$) for MnPt1-xIrxSn (x= 0, 0.1, 0.2, 0.3, 0.5). Solid lines are the linear fitting to the equation $\rho^{A}_{xy}=a\rho_{xx0}+b\rho^{2}_{xx}$. Top axis indicates the corresponding measurement temperatures. (f) The extrinsic parameter ($a$; open squares) and intrinsic parameter ($b$; closed squares) as a function of iridium concentration $x$. The lines are guides to the eye. To understand the mechanism that governs the observed change in the AHE with Ir doping, we carry out a thorough analysis of the $\rho^{A}_{xy}$ data using $TYJ$ scaling relations [19] and the power law relation. In order to map out the magnitude of different contributions of AHE in the Ir-doped samples, we employ the scaling relation given in Eq. (1). For this purpose, we plot $\rho^{A}_{xy}$ vs $\rho^{2}_{xx}$ for different samples, as depicted in Figs. 4(a)-4(e). The $\rho_{xx}$ data are obtained from the temperature-dependent longitudinal resistivity measurements and $\rho^{A}_{xy}$ points are taken from the field-dependent Hall resistivity measurements at different temperatures as shown in the Supplemental Materials [37]. The fitting parameters obtained from the linear fit [as shown by black solid lines in Figs. 4(a)-4(e)] of the experimental data can be used to extract the information about different contributions of AHE. The slope of the linear fitting gives the intrinsic parameter $b$, whereas the extrinsic parameter $a$ can be calculated from the intercept. Figure 4(f) shows the variation of $a$ and $b$ with the iridium concentration. The extrinsic parameter $a$ decreases and the intrinsic parameter $b$ increases with increasing iridium concentration. This illustrates that the Ir doping enhances the intrinsic contribution to the AHE in the present system. It has been shown that the side jump contribution ($\sigma_{xx}^{SJ}$) is related to $\frac{e^{2}}{ha}(\frac{\epsilon_{so}}{E_{F}})$ [43, 44], where $\epsilon_{so}$ is the spin-orbit interaction energy, $E_{F}$ is the Fermi energy, and $a$ is the lattice constant. Using this expression, we have calculated the approximate side jump contribution using $\frac{\epsilon_{so}}{E_{F}}\sim 0.01$ and the experimental lattice constant $a$. We find that the side jump contribution for MnPtSn is almost negligible when compared with the extrinsic and intrinsic parameters [37]. For MnPt0.5Ir0.5Sn, the most dominant contribution arises from the intrinsic part, which is much larger than the side jump effect. However, the magnitudes of skew scattering and side jump contributions for this sample are comparable as both components are very small. Hence the extrinsic parameter ‘a’ mainly represents the skew scattering. We summarize the values of $a$, $b$, anomalous Hall resistivity arising from the extrinsic contribution $\rho^{ext}_{xy}$, the intrinsic anomalous Hall resistivity $\rho^{int}_{xy}$, and the total $\rho^{tot}_{xy}$ for all the compounds in table 1. Figure 5: Power law representation. (a-e) Plot of log ($\rho^{A}_{xy}$) vs log ($\rho_{xx}$) for MnPt1-xIrxSn (x= 0, 0.1, 0.2, 0.3, 0.5). Black solid lines are the fit using the relation ($\rho^{A}_{xy}\propto\rho^{\alpha}_{xx}$). Top axis indicates the corresponding measurement temperatures. (f) Scaling factor $\alpha$ as a function of iridium concentration $x$. Table 1: Values of extrinsic and intrinsic parameters and different contribution of AHE. —————————————————————————— x a b $\rho^{ext}_{xy}$ $\rho^{int}_{xy}$ $\rho^{tot}_{xy}$ $\times 10^{-3}$ $(\Omega$ $cm)^{-1}$ $(\mu\Omega$ $cm)$ 0.0 10.07 73.14 0.97 0.68 1.65 0.1 9.22 82.60 1.07 1.13 2.20 0.2 7.89 101.9 1.06 1.84 2.90 0.3 5.79 119.3 0.93 3.09 4.05 0.5 0.856 131.1 0.16 4.86 5.02 For the power law analysis, we plot log $\rho^{A}_{xy}$ vs log $\rho_{xx}$ as shown in Figs. 5(a)-5(e). The solid lines represent the linear fitting to the data and the slope gives the scaling factor $\alpha$. For each sample a good fit to the data can be obtained up to a certain temperature below which the magnetization is almost constant. As is found, $\alpha$ increases linearly with Ir doping from 0.96 for MnPtSn to 1.93 in MnPt0.5Ir0.5Sn [Fig. 5(f)]. The scaling factor $\alpha$ close to 1 suggests that the extrinsic contribution is the dominant mechanism that governs the AHE in MnPtSn, whereas $\alpha$ of nearly 2 in the case of MnPt0.5Ir0.5Sn points toward its intrinsic nature. The change of $\alpha$ from 0.96 to 1.93 implies a changeover in the mechanism of AHE from extrinsic to intrinsic with Ir doping. Figure 6: Projected density of states (DOS) with UMn = 5eV and UPt/Ir = 2eV for (a) MnPtSn, (b)MnPt0.5Ir0.5Sn, and (c) MnIrSn. The Mn-d, Pt-d, Ir-d, and Sn-p states are represented by red, green, blue, and cyan, respectively. Projected density of states with the variation of spin-orbit coupling (SOC) strength. (d), (e), and (f) represent the variation of DOS for Pt-d in MnPtSn, Pt-d in MnPt0.5Ir0.5Sn, and Ir-d in MnPt0.5Ir0.5Sn, respectively. The color scheme for various values of $\alpha$ is mentioned in the right panel. The Fermi level in the energy scale is set at zero. ### III.3 Electronic Band Structure and Effect of the Spin-Orbit Coupling (SOC) The electronic structure calculations are carried out for three compounds in the MnPt1-xIrxSn series with x=0.0, 0.5 and 1.0. Although MnIrSn could not be synthesized experimentally, the crystal structure is obtained by the structural optimization starting from the MnPtSn and included in the study for the sake of complete understanding of the role played by each component. Figure 6 shows the calculated GGA+$U$ comparative density of states (DOS) for the above-mentioned compounds. The DOS shows that the Mn-3d states are fully filled in the majority spin channel and completely empty in the minority spin channel, with a very small contribution at the Fermi energy. The major contribution at the Fermi level arises from the Pt-$5d$ and the Sn-$5p$ states as shown in Fig. 6(a). The DOS clearly shows the metallic character of the electronic structure, where the Pt-$5d$ states contribute at the Fermi energy for MnPtSn. In the case of x=0.5, the hybridized Ir-$5d$ and Pt-$5d$ states mostly appear at the Fermi energy. The calculated magnetic moment at the Mn site comes out to be about 4.52 $\mu_{B}$, while the induced moments at the Sn and Pt sites appear to be -0.12 $\mu_{B}$ and negligibly small, respectively. This results in a total moment of 4.53 $\mu_{B}$/f.u. After 50% Ir doping at the Pt site the DOS changes significantly as shown in Fig. 6(b). The Mn-$3d$ states remain almost unchanged, while the major renormalization happened at the Pt states due to incorporation of the Ir doping. Doped Ir-$5d$ states are pushed toward the Fermi level in MnPt0.5Ir0.5Sn. The magnetic moment at the Mn site is slightly reduced to 4.50 $\mu_{B}$, while the induced moment at the Ir site is around -0.05 $\mu_{B}$. The difference in the theoretical magnetic moment compared to the experimental moment obtained for the Ir doped sample might be arising due to the fact that the present theoretical calculations are performed without considering any disorder. For MnIrSn, the Mn-$3d$ states remain unchanged except slightly narrowing down the band width, and the Ir-$5d$ states are further pushed toward the Fermi level as shown in Fig. 6(c). Moreover, we calculate the magnetic transition temperatures in a mean-field method for the MnPtSn and MnPt0.5Ir0.5Sn. The calculated ratio of the transition temperature ($\frac{T^{MPS}_{c}}{T^{MPIS}_{c}}$) is around 1.52, which is in good agreement with the experimental value of 1.44 (see the Supplemental Material sec VII) [37]. Figure 7: Variation of magnetic moment with respect to (a)-(d) iridium doping concentration (x) and (e)-(h) scale factor of SOC strength $\alpha$. (a),(b),(e),(f) represent the variation of spin magnetic moment and (c),(d),(g),(h) represent the variation of orbital magnetic moment. Mn-d, Pt-d, Ir-d states are shown by circle, square, and triangle symbols, respectively. In (e)-(h) closed symbols represent MnPtSn and open symbols MnPt0.5Ir0.5Sn samples. In addition to the above-discussed effects, Ir doping may also modify the effective SOC strength in the material. To capture the effect of SOC-induced modifications in the system, we have done electronic structure calculations for MnPtSn and MnPt0.5Ir0.5Sn by varying the SOC strength manually. The calculated GGA+$U$+SOC electronic DOSs with varying SOC strength ($\alpha$) for Pt/Ir states are shown in Figs. 6(d)-6(f). As expected, the Mn-$3d$ state does not exhibit any substantial variation with the SOC strength [37]; however, the Pt/Ir-$5d$ states change significantly and move toward the Fermi energy with increasing SOC strength as shown in Figs. 6(d)-6(f). As discussed above, a similar kind of modification of the DOS is also observed with Ir doping in Figs. 6(a)-6(c). From these DOS calculations, it is very clear that the introduction of Ir into the system substantially changes the effective strength of the SOC as well as the Pt/Ir states. To understand the variation of the magnetic moment with the Ir doping, we also perform the calculations for intermediate concentrations (x= 0.25, 0.5, 0.75). The variation of spin and orbital magnetic moments with Ir concentration is shown in Figs. 7(a)-7(d). It is evident that the spin magnetic moment at the Mn sites gradually decrease with increasing Ir concentration, whereas the Pt/Ir spin moments, although very small, remain almost constant. The point to be noted here is that the sign of the induced moments at the Ir site is opposite to that of the Mn site. As expected, the Mn atom exhibits a very small orbital moment in comparison to its spin magnetic moment and the orbital moment at the Pt/Ir site. Interestingly, the orbital magnetic moment at the Ir site is larger than that of the Pt site with opposite sign. However, the spin and orbital magnetic moments at the Ir sites follow almost the same trend with the Ir doping , as expected due to the modification of the effective strength of the SOC in the system. Figure 8: Magnetization density for the MnPtSn and 50% Ir doped material for different SOC strength $\alpha$. (a) and (b) are the magnetization density of MnPtSn for $\alpha$=1.0 and 4.0 respectively, whereas (c) represents the same for MnPt0.5Ir0.5Sn with $\alpha$=1.0. Yellow and green represent the isosurfaces with + and - signs respectively. From our experimental and theoretical studies, it is very clear that Ir doping plays a crucial role in deciding the magnetic state of the material. Naively, the physical consequence of the Ir doping at the place of Pt can be mapped as reducing one electron per atom in the material. As we can see from the DOS calculation, Ir doping changes the SOC which influences the resultant magnetic moment of the system. Figures 7(e)-7(h) show the variation of spin and orbital magnetic moment as a function of the scale factor ($\alpha$) of the SOC strength. As the $\alpha$ increases, the spin magnetic moment at the Mn sites [Fig. 7(e)] decreases, whereas the orbital magnetic moment [Fig. 7(g)] increases. This trend can be understood from the transfer of spin contribution to the orbital contribution of magnetic moment with increasing SOC. In contrast, the variation of the spin and orbital moments as a function of $\alpha$ are very different for Pt and Ir. There is a marginal change in the spin and orbital magnetic moment for Pt as a function of $\alpha$, whereas the variation is quite substantial for Ir. The effect of SOC is more prominent in the case of Ir than Pt in both the MnPtSn and 50$\%$ Ir doped cases. Interestingly, the variations of magnetic moment at the Mn site as a function of the Ir doping and SOC strength ($\alpha$) are very similar, which suggests that the Ir doping at the Pt site plays a crucial role in dictating the effective SOC of the materials, apart from injecting an electron in the system. ### III.4 Magnetization Density The effect of the different roles played by the Pt and Ir atoms in terms of SOC strength can be conclusively visualized by the magnetization density plot shown in Fig. 8. Figures 8(a) and 8(b) show the magnetization density for MnPtSn with $\alpha$=1.0 and 4.0, respectively, whereas Fig. 8(c) represents the magnetization density for MnPt0.5Ir0.5Sn with $\alpha$=1.0. Since the Mn magnetic moment is much larger, the spin isosurface sitting at the Mn site is very large compared to the Pt/Ir sites. However, the small isosurface of the Pt or Ir carries very important information about the effective strength of the SOC. The Pt isosurface in Figs. 8(a) and 8(b), i.e. for $\alpha$=1.0 and 4.0, are very different, which is quite expected as the SOC strength modifies the shape of the magnetization density. A very interesting fact is that the shape of the Ir isosurface in Fig. 8(c) (i.e, for MnPt0.5Ir0.5Sn with $\alpha$=1.0) is similar to that of the Pt isosurface in Fig. 8(b) [i.e, for MnPtSn with $\alpha$=4.0]. This observation suggests that the effective strength of the SOC gets modified due to the Ir doping in the MnPtSn. Qualitatively, it can be said that the effective strength of the SOC in the case of the 50$\%$ Ir doped sample is greater than the parent MnPtSn. Hence, it is expected that the change in the SOC with Ir doping may significantly affect the electron transport in the system. The above discussed theoretical result nicely corresponds with the experimental findings. It is well established that the intrinsic contribution of the AHE is directly proportional to the relevant local Berry curvature of the material. To realize a non-vanishing Berry curvature by the breaking of the symmetry of the system, the SOC in the material plays a very crucial role [45]. From the electronic structure calculations, we show that Ir doping enhances the effective strength of SOC; as a result the intrinsic contribution to the AHE should also increase. The theoretical findings corroborate our experimental observation of enhanced AHE in case of the Ir-doped samples. It may be noted that the Ir-doped samples possess some amount of disorder as well as higher residual resistivity compared to the parent MnPtSn compound. Although the skew scattering contribution is sensitive to the impurity, it is larger in a clean regime [low resistivity, $\sigma_{xx}>10^{6}$ $(\Omega$ $cm)^{-1}$] and decays in a bad metal regime [high resistivity, $\sigma_{xx}<10^{4}$ $(\Omega$ $cm)^{-1})$]. This suggests that as the conductivity $(\sigma_{xx})$ decreases, the skew scattering contribution starts diminishing and the intrinsic mechanism begins to dominate [6, 46]. In the present case all the samples fall into moderate to bad metal regimes. Hence, our experimental and theoretical findings suggest that one can tune different contributions of AHE with changing the overall SOC strength in the system. In this regard, AHE driven by different underlying mechanisms is reported recently in several materials. In some cases, a large intrinsic anomalous Hall signal has been observed due to the presence of Weyl nodes [47, 48]. Additionally, nontrivial scalar spin chirality can also induce a large extrinsic anomalous Hall effect as recently found in MnGe [49]. Besides this, there are reports of modification of intrinsic and extrinsic AHE by tuning the Fermi level with the help of chemical doping [20, 50, 51]. Although SOC is the primary mechanism responsible for the AHE, it is important to understand the correlation among other mechanisms contributing its cause. In the present study, we demonstrate the manipulation of different contributions of AHE by introducing chemical doping. ## IV CONCLUSION In conclusion, we present a detailed study on the magnetic and electronic transport properties of the half-Heusler system MnPt1-xIrxSn. We find that the saturation magnetization systematically decreases while the anomalous Hall resistivity increases with increasing iridium concentration. The experimental Hall signal is analyzed using the scaling relation between anomalous Hall resistivity and longitudinal resistivity. We find that the scaling factor changes from close to 1 for the parent MnPtSn sample to quadratic for MnPt0.5Ir0.5Sn, signifying that Ir doping enhances the intrinsic contribution by suppressing the extrinsic mechanism. Our experimental results are well supported by the theoretical study that showed that the Ir doping significantly enhances the spin-orbit coupling in the system. The present study is an important contribution toward the basic understanding of different mechanisms of AHE, and thereby possesses a great importance in designing anomalous Hall sensor based spintronic devices. ## V Acknowledgment A.K.N. acknowledges support from the Department of Atomic Energy (DAE), the Department of Science and Technology (DST) Ramanujan Research Grant No. SB/S2/RJN-081/2016. R.R. acknowledges IIT Goa for her research fellowship and S.K. acknowledges DST INSPIRE for research funding. ## References * [1] I. Belopolski1, K. Manna, D. S. Sanchez, G.Chang,B. Ernst, J. Yin1, S. S. Zhang1, T. Cochran1, N. Shumiya1,H. Zheng1, B. Singh, G. Bian, D. Multer1, M. Litskevich1, X. , Shin-Ming Huang, B Wang, Tay-Rong Chang, Su-Yang Xu1,A. Bansi, C. Felser, H. Lin, M. Zahid Hasan, Science 365, 1278 (2019). * [2] Q. Wang, S. Sun, X. Zhang, F. Pang, and H. Lei, Phys. Rev. B 94, 075135 (2016). * [3] P. Li, J. Koo, W. Ning, J. Li, L. Miao, L. Min, Y. Zhu, Y. Wang, N. Alem, C.-X. Liu, Z. Mao, and B. Yan, Nat. Commun. 11, 3476 (2020). * [4] T. Liang, J. Lin, Q. Gibson, S. Kushwaha, M. Liu, W. Wang, H. Xiong, J. A. Sobota, M. Hashimoto, P. S. Kirchmann, Z.-X. Shen, R. J. Cava, N. P. Ong, Nat. Phys. 14, 451 (2018). * [5] S.-Y. Yang, Y. Wang, B. R. Ortiz, D. Liu, J. Gayles, E. Derunova, R. Gonzalez-Hernandez, L. Šmejkal, Y. Chen, S. S. P. Parkin, S. D. Wilson, E. S. Toberer, T. McQueen, and M. N. Ali, Sci. Adv. 6, eabb6003 (2020). * [6] N. Nagaosa, J. Sinova, S. Onoda, A. H. MacDonald, and N. P. Ong, Rev. Mod. Phys. 82, 1539 (2010). * [7] J. Smit, Physica 21, 877 (1955); 24, 39 (1958). * [8] L. Berger, Phys. Rev. B 2, 4559 (1970). * [9] R. Karplus and J. M. Luttinger, Phys. Rev. 95, 1154 (1954). * [10] T. Jungwirth, Q. Niu, and A. H. MacDonald, Phys. Rev. Lett. 88, 207208 (2002). * [11] S. Nakatsuji, N. Kiyohara, and T. Higo, Nature (London) 527, 212 (2015). * [12] A. K. Nayak, J. E. Fischer, Y. Sun, B. Yan, J. Karel, A. C.Komarek, C. Shekhar, N. Kumar, W. Schnelle, J. Kübler, C.Felser, and S. S. P. Parkin, Sci. Adv. 2, e1501870 (2016). * [13] W. L. Lee, S. Watauchi2, V. L. Miller, R. J. Cava, N. P. Ong, Science 303, 1647 (2004). * [14] Y. Pu, D. Chiba, F. Matsukura, H. Ohno, and J. Shi, Phys. Rev. Lett. 101, 117208 (2008). * [15] J. Shen, Q. Zeng, S. Zhang, H. Sun, Q. Yao, X. Xi, W. Wang, G. Wu, B. Shen, Q. Liu, and E. Liu, Adv. Funct. Mater. 30, 2000830, 2020. * [16] Q. Wang, Y. Xu, R. Lou, Z Liu, M. Li1, Y. Huang, D. Shen ,H. Weng, S. Wang1 and H. Lei, Nat. Commun. 9, 3681 (2018). * [17] H. Li, B. Zhang, J. Liang, B. Ding, J. Chen, J. Shen, Z. Li, E. Liu, X. Xi, G. Wu, Y. Yao, H. Yang, and W. W. Phys. Rev. B 101, 140409(R) (2020). * [18] Yu Liu , H. Tan, Z. Hu, B. Yan, and C. Petrovic, Phys. Rev. B 103, 045106 (2021). * [19] Y. Tian, L. Ye, and X. Jin, Phys. Rev. Lett. 103, 087206 (2009). * [20] D. Yu. Karpenkov, K. P. Skokov, I. A. Radulov, O. Gutfleisch, J. Weischenberg, and H. Zhang, Phys. Rev. B 100, 094445 (2019). * [21] K. K. Meng , L. J. Zhu, Z. H. Jin, E. K. Liu, X. P. Zhao, I. A. Malik, Z. G. Fu, Y. Wu, J. Miao, X. G. Xu, J. X. Zhang, J. H. Zhao, and Y. Jiang, Phys. Rev. B 100, 184410 (2019). * [22] L. J. Zhu, D. Pan and J. H. Zhao, Phys. Rev. B 89, 220406(R) (2014). * [23] J. Shen, Q. Yao, Q. Zeng, H. Sun, X. Xi, G. Wu, W. Wang, B. Shen, Q. Liu, and E. Liu, Phys. Rev. Lett. 125, 086602 (2020). * [24] Q. Zhang, P. Li, Y. Wen, C. Zhao, J. W. Zhang, A. Manchon, W. B. Mi, Y. Peng, and X. X. Zhang, Phys. Rev. B 94, 024428 (2016). * [25] Y. Zhang, Y. Yin, G. Dubuis, T. Butler, N. V. Medhekar and S. Granville, npj Quantum Mater, 6, 17 (2021). * [26] T. Graf, C. Felser, and S. S. P. Parkin, Prog. Solid State Chem. 39, 1 (2011). * [27] J. Kübler, G. H. Fecher, and C. Felser, Phys. Rev. B 76, 024414 (2007). * [28] E. Decolvenaere, M. Gordon, R. Seshadri, and Anton V. der Ven, Phys. Rev. B 96, 165109 (2017). * [29] S. D. Harrington, A. D. Rice, T. L. Brown-Heft, B. Bonef, A. Sharan, A. P. McFadden, J. A. Logan, M. Pendharkar, M. M. Feldman, O. Mercan, A. G. Petukhov, A. Janotti, L. Colakerol Arslan, and C. J. Palmstrøm, Phys. Rev. Mater. 2, 014406 (2018). * [30] A.A. Burkov, Phys. Rev. Lett. 113, 187202, (2014). * [31] J. Chen, X. Xu, H. Li, T. Guo, B. Ding, P. Chen, H. Zhang, X. Xi, and W. Wang, Phys. Rev. B 103, 144425,(2021). * [32] G. Kresse and J. Hafner, Phys. Rev. B $\bf{47}$, 558(R) (1993). * [33] G. Kresse and J. Furthmüller, Phys. Rev. B $\bf{54}$, 11169 (1996). * [34] J. P. Perdew, K. Burke, and M. Ernzerhof, Phys. Rev. Lett. $\bf{77}$ ,3865 (1996). * [35] V. I. Anisimov, I. V. Solovyev, M. A. Korotin, M. T. Czyzyk, and G. A. Sawatzky, Phys. Rev. B $\bf{48}$, 16929 (1993). * [36] S. L. Dudarev, G. A. Botton, S. Y. Savrasov, C. J. Humphreys, and A. P. Sutton, Phys. Rev. B $\bf{57}$, 1505 (1998). * [37] See Supplemental Material for details of structural, transport measurements, additional theoretical calculations and reference no. 50 to 51. * [38] S. Skaftouros, K. Özdoğan, E. Şaşıoğlu, and I. Galanakis, Phys. Rev. B 87, 024420 (2013). * [39] E. Şaşıoğlu, L. M. Sandratskii, and P. Bruno, Phys. Rev. B 77, 064417 (2008). * [40] L. Wollmann, S. Chadov, J. Kubler, and C Felser, Phys. Rev. B 90,214420 (2014). * [41] Juan Rodríguez-Carvajal, Physica B 192, 55 (1993). * [42] A. K. Nayak, C. Salazar Mejia, S. W. D’Souza, S. Chadov, Y. Skourski, C. Felser, and M. Nicklas, Phys. Rev. B 90, 220408(R) (2014). * [43] S. Onoda, N. Sugimoto, and N. Nagaosa, Phys. Rev. Lett. 97, 126602 (2006). * [44] P. Nozières. And C. Lewiner, J. Phys. France 34, 901 (1973). * [45] D. Xiao, M.-C. Chang, and Q. Niu, Rev. Mod. Phys. $\bf{82}$, 1959 (2010). * [46] Y. Shiomi, Y. Onose, and Y. Tokura, Phys. Rev. B 79, 100404 (2009). * [47] B. Ernst, R. Sahoo, Y. Sun, J. Nayak, L. Muchler, A K. Nayak, N. Kumar,J. Gayles, A. Markou, G. H. Fecher, and C. Felser, Phys. Rev. B 100, 054445 (2019). * [48] T. Suzuki, R. Chisnell, A. Devarakonda, Y.-T. Liu, W. Feng, D. Xiao, J. W. Lynn and J. G. Checkelsky, Nat. Phys. 12, 1119 (2016). * [49] Y. Fujishiro, N. Kanazawa, R. Kurihara, H. Ishizuka, T. Hori, F. S. Yasin, X. Yu, A. Tsukazaki, M. Ichikawa, M. Kawasaki, N. Nagaosa, M. Tokunaga and Y. Tokura. Nat. Commun. 12, 317, (2021). * [50] H.-Yu Yang, B. Singh, B. Lu, C.-Yi Huang, F. Bahrami, W.-Chi Chiu, D. Graf, S.M. Huang, B. Wang, H. Lin, D. Torchinsky, A. Bansil, and F. Tafti, APL Materials 8, 011111 (2020). * [51] K. M. Seemann, Y. Mokrousov, A. Aziz, J. Miguel, F. Kronast, W. Kuch, M. G. Blamire, A. T. Hindmarch, B. J. Hickey, I. Souza, and C. H. Marrows, Phys. Rev. Lett. 104, 076402, (2010). ## SUPPLEMENTARY INFORMATION ## VI X-ray Diffraction X-ray diffraction (XRD) data are collected at room temperature with a Cu-K source on a Rigaku SmartLab x-ray diffractometer to check the phase purity of all the samples. The Rietveld refinement of the room temperature powder XRD patterns for MnPt0.5Ir0.5Sn are shown in FIG. 9. The Rietveld refinement for (x=0, 0.1, 0.2, 0.3) is performed using the Wyckoff positions as Sn at (0, 0, 0), Mn at (1/2, 1/2, 1/2), and Pt/Ir at (1/4, 1/4, 1/4). For x=0.5, same Wyckoff positions are used with 10% atomic disorder between Mn/Ir sites (see section VII). The refinement confirms that all the samples are in single-phase and exhibit XRD patterns of cubic crystal structure of space group F$\bar{4}$3m. Figure 9: Room temperature powder XRD patterns and Rietveld refinements for MnPt1-xIrxSn (x=0, 0.1, 0.2, 0.3, 0.5 ## VII Neutron Diffraction (ND) The room-temperature Rietveld refinement of the powder ND data for MnPt0.5Ir0.5Sn is carried out using the FullProf software with different percentages of atomic disorder between Mn and Ir. Figs. 10 (a)-10 (d) shows the Rietveld refinement with 0%, 5%, 15%, 20% atomic disorders between the Mn and Ir atoms, respectively. The refined parameters with different percentages of atomic disorders are given in table II. We observe that 10% atomic disorder between Mn/Ir results in a minimum value of $\chi^{2}$ and $R_{Bragg}$. The magnitude of $\chi^{2}$ and $R_{Bragg}$ increases as the disorder increase or decreases. Our findings indicate that the best fit for MnPt0.5Ir0.5Sn is obtained with 10% atomic disorder between Mn/Ir sites. Figure 10: Rietveld refinement of the room-temperature ND pattern for MnPt0.5Ir0.5Sn with different atomic disorder between the Mn and Ir sites. The region between $34^{o}$ and $36^{o}$ is excluded due to contribution from the cryostat. Table 2: The percentage of disorder, site occupations, and other refined parameters for the room temperature powder neutron diffraction data of MnPt0.5Ir0.5Sn. Sl. No | | Wyckoff --- Position Site Occupations (%) | $\chi^{2}$ | $R_{Bragg}$ 0% Disorder | | 4b --- 4c | 100 Mn + 0 Ir --- 50 Pt +50 Ir 6.4 | 13.71 5% Disorder | | 4b --- 4c | 95 Mn + 5 Ir --- 50 Pt +45 Ir + 5 Mn 5.96 | 12.06 10% Disorder | | 4b --- 4c | 90 Mn + 10 Ir --- 50 Pt + 40 Ir + 10 Ir 5.96 | 11.62 15% Disorder | | 4b --- 4c | 85 Mn + 15 Ir --- 50 Pt +35 Ir + 15Ir 6.42 | 14.38 20% Disorder | | 4b --- 4c | 80 Mn + 20 Ir --- 50 Pt +30 +Ir +20 Ir 7.45 | 21.08 ## VIII Temperature dependence of longitudinal resistivity Longitudinal resistivity $\rho_{xx}$ as a function of temperature (T) for all the samples is shown in Fig. 11. Below the Curie temperature (Tc) resistivity decreases as the temperature decreases indicating that all the samples are in a metallic regime. At low temperature, $\rho_{xx}$ approaches the residual resistivity $\rho_{xx0}$ Figure 11: Temperature variation of longitudinal resistivity $\rho_{xx}$ (T) for the MnPt1-xIrxSn (with x= 0, 0.1, 0.2, 0.3, 0.5) samples. ## IX Field dependent Hall resistivity The Hall resistivity measurements for all the samples are performed at different temperatures with field sweep $\pm$ 5T, as shown in Fig. 12. For all the samples, the Hall resistivity increases with increasing temperature before it starts to decrease as the temperature approaches the Curie temperature. Figure 12: Field variation of Hall resistivity $\rho_{xy}(H)$ at different temperatures for the MnPt1-xIrxSn (with x= 0, 0.1, 0.2, 0.3, 0.5) samples. ## X Density of states with the variation of spin-orbit coupling (SOC) strength The density of states for Mn-d states of MnPtSn and MnPt0.5Ir0.5Sn with the variation of spin orbit-coupling (SOC) strength are calculated and shown in Figs. 13 (a)-13 (b), respectively. The calculated density of states with different SOC strengths does not show any significant changes for both the samples. Figure 13: Projected density of states of Mn-d with the variation of spin orbit coupling (SOC) strength for (a) MnPtSn and (b) MnPt0.5Ir0.5Sn ## XI Magnetic Exchange Interactions Figure 14: (a) Magnetic exchange interactions between different Mn sites as a function of the Mn-Mn distances for the MnPtSn and MnPt0.5Ir0.5Sn. Positive and negative signs represent ferromagnetic and antiferromagnetic type of the interactions, respectively. (b) Density of states (DOS) for MnPtSn. The Mn-3d, Pt-5d, Sn-5p, Sn-5s are represented by gray, green, blue and red respectively. Inset shows the Wannier function plot. With the Ir doping, substantial changes are observed in the magnetic properties of the MnPtSn half Heusler compound. In the experiments, it is observed that Mn spins are ferromagnetically ordered in the case of MnPtSn, however, there is some sort of antiferromagnetic correlation in the Ir doped cases. Through the electronic structure total energy method, we calculate the exchange interactions between different Mn sites as a function of the Mn-Mn distances as shown in Fig. 14 (a). We find that the magnetic exchange interactions of the MnPt0.5Ir0.5Sn are weaker than the MnPtSn, which is consistent with the smaller value of the TC for the Ir doped samples compared to that of the parent MnPtSn. We also find that the type of the exchange interactions are oscillatory in nature which ensures the conduction electron- driven RKKY type of ferromagnetic exchange interactions are dominating in these materials. In the case of MnPtSn, the Mn-Mn distance is large; thus the Mn-3d state does not overlap directly, and the direct exchange interactions can be neglected. On the contrary, the RKKY type of exchange interaction between two magnetic Mn ions is also mediated by the itinerant conduction electrons. The itinerant electrons close to the Fermi energy and having substantial hybridization with the Mn-d states will act as a mediator in the RKKY interactions. For the MnPtSn, the density of states (Fig Fig. 14 (b)) clearly shows that Sn-sp mixed states are taking part in the conduction electron near the Fermi energy and acting as a mediator for the RKKY type exchange interactions between two Mn-d ions. We also found a small presence of Pt/Ir-5d states in the Mn-Mn exchange interactions. The below DOS and Wannier function (inset figure Fig. 14 (b)) plot clearly shows that the finite tail (marked by red dotted circle in the inset) situated at the Sn site with a shape of sp hybridized state as a mediator between Mn-Mn interactions. We also found a very tiny tail at the Pt site, which also has little contribution as a mediator. ## XII Calculation of Magnetic transition temperature The magnetic transition temperatures in a mean-field approach for both the systems are calculated using the equation $T_{C}=\frac{2}{3}S(S+1)\sum_{i}J_{i}Z_{i}$ (2) The experimental ratio of $T_{C}$ is $\frac{T_{C}^{MPS}}{T_{C}^{MPIS}}$ =1.44 (MPS represents for MnPtSn and MPIS represents for MnPt0.5Ir0.5Sn). The calculated ratio of $T_{C}$ is $\frac{T_{C}^{MPS}}{T_{C}^{MPIS}}$ =1.52. ## XIII Side Jump Contribution Figure 15: Skew scattering $\rho_{xy}^{Skew}$, intrinsic $\rho_{xy}^{Int}$ and side-jump $\rho_{xy}^{SJ}$ contributions with respect to residual resistivity $\rho_{xx0}$ for MnPt$1-x$IrxSn (for x = 0, 0.1, 0.2, 0.3, 0.5). According to the existing theoretical model, the extrinsic side-jump contribution is in the order of $\sigma_{xy}^{SJ}\approx\frac{e^{2}}{ha}(\frac{\epsilon_{SO}}{E_{F}})$ , where $\epsilon_{SO}$ is the spin-orbit interaction energy, $E_{F}$ is the Fermi energy, a is the lattice constant. For metallic ferromagnet $\frac{\epsilon_{SO}}{E_{F}}\approx 0.01$. With the help of this expression, one can have the order of side jump contribution and compares it with the other mechanism. We calculated the approximate side jump contribution for our system using the above expression and taking the experimental lattice parameter. We plotted the different contributions with residual resistivity $\rho_{xx0}$ . It is clearly seen that the SJ contribution is minimal compared to other dominant contributions.
# Additive structure of non-monogenic simplest cubic fields Daniel Gil-Muñoz Department of Algebra, Faculty of Mathematics and Physics, Charles University, Sokolovská 83, 186 00 Praha 8, Czech Republic Magdaléna Tinková Faculty of Information Technology, Czech Technical University in Prague, Thákurova 9, 160 00 Praha 6, Czech Republic ###### Abstract We consider non-monogenic simplest cubic fields $K=\mathbb{Q}(\rho)$ in the family introduced by Shanks, and among these, we focus in the fields whose generalized module index $[\mathcal{O}_{K}:\mathbb{Z}[\rho]]$ is a prime number $p$. We prove that these fields arise exactly for $p=3$ or $p\equiv 1\,(\mathrm{mod}\,6)$ and we use the method in [16] to find the additive indecomposables of $\mathcal{O}_{K}$. We determine the whole structure of indecomposables for the family with $p=3$ and obtain that the behaviour is not uniform with respect to the indecomposables of $\mathbb{Z}[\rho]$. From the knowledge of the indecomposables we derive some arithmetical information on $K$, namely: the smallest and largest norms of indecomposables, the Pythagoras number of $\mathcal{O}_{K}$ and bounds for the minimal rank of universal quadratic forms over $K$. ## 1 Introduction The main motivation for the study of the additive structure of a totally real number field $K$ comes from the topic of universal quadratic forms. A quadratic form with coefficients in the ring of integers $\mathcal{O}_{K}$ is said to be universal (or universal over $K$) if it represents all totally positive elements in $\mathcal{O}_{K}$ (that is, those whose image by any embedding of $K$ is a positive real number). Some classical problems in arithmetic may be formulated in terms of these objects; for instance the well known Lagrange’s four-square theorem asserts that the quadratic form $x^{2}+y^{2}+z^{2}+w^{2}$ is universal over $\mathbb{Q}$. Although the sum of three squares $x^{2}+y^{2}+z^{2}$ is not universal over $\mathbb{Q}$, Maaß [24] proved in 1941 that it is universal over $\mathbb{Q}(\sqrt{5})$. Shortly after, in 1945, Siegel [31] proved that these are the only fields under which sums of squares can be universal. In his proof, indecomposables, which we study in this paper, appear under the name of extremal elements. All the examples of universal quadratic forms above are totally positive definite, that is, they only represent totally positive integers. In this paper we will restrict exclusively to such quadratic forms. From a result by Hsia, Kitaoka and Kneser [10] it follows that for every totally real number field there is some totally positive definite quadratic form which is universal over $K$. Let $m(K)$ be the minimal integer among all ranks of universal quadratic forms over $K$. The number $m(K)$ turns out to be rather unpredictable. In fact, Kitaoka conjectured that $m(K)=3$ only for finitely many totally real number fields $K$. Although a general solution is unknown, there are some partial results [6, 8, 22, 18]. The connection of $m(K)$ with additive indecomposables of $K$ came from the works by Blomer and Kala [1, 14], which is based on the suitability of these elements to construct quadratic lattices of large rank. For a totally real quadratic field $K=\mathbb{Q}(\sqrt{D})$, Perron [27], Dress and Scharlau [7] found a complete list of indecomposable integers in $K$ in terms of the continued fraction expansion of $\frac{-1+\sqrt{D}}{2}$ if $D\equiv 1\,(\mathrm{mod}\,4)$ and of $\sqrt{D}$ otherwise. Blomer and Kala used this description to show that for every positive integer $n$ there are infinitely many totally real quadratic fields $K$ such that $m(K)\geq n$. In other words, there are infinitely many quadratic fields without universal quadratic forms of rank $n$ arbitrarily large. In the following years an analogous result was proved for the families of multiquadratic and cubic fields [15, 36]. In fact, Kala [13] proved that it holds for each degree $d$ divisible by $2$ or $3$. On the other hand, Blomer and Kala [2] found concrete bounds for the number $m_{\mathrm{diag}}({\mathbb{Q}(\sqrt{D})})$, the minimal rank of diagonal universal quadratic forms over $\mathbb{Q}(\sqrt{D})$. This substantial progress on quadratic fields motivated the search of indecomposables in families of totally real number fields with higher degree. In this direction, Kala and the second author [16] proved that all indecomposables in a totally real number field $K$ are, up to multiplication by totally positive units, lattice points in a finite disjoint union of simplicial cones in the Minkowski space $\mathbb{R}^{d}$, where $d=[K:\mathbb{Q}]$. In the case that $d=3$, this description reduces to two specific parallelepipeds, whose nodes depend on a proper pair of totally positive units (according to the terminology in [32]). These considerations led to a geometric method consisting in finding all lattice points within those two parallelepipeds and identifying the indecomposables among them. A tool to identify the indecomposables among lattice points relies on their traces. The minimal trace of an algebraic integer $\alpha\in\mathcal{O}_{K}$ is the minimum among all numbers of the form $\mathrm{Tr}(\delta\alpha)$, for totally positive elements $\delta$ in the codifferent $\mathcal{O}_{K}^{\vee}=\\{\delta\in K\,|\,\mathrm{Tr}(\delta\alpha)\in\mathbb{Z}\text{ for all }\alpha\in{\mathcal{O}}_{K}\\}$ of $K$. Its importance in the search of indecomposables relies in the fact that every totally positive algebraic integer of minimal trace $1$ is indecomposable. While in the case of quadratic fields the converse holds, in the case of monogenic simplest cubic fields there is an indecomposable with minimal trace $2$. In fact, the second author [34, Theorem 1.1] proved that the minimal trace of indecomposables in totally real cubic orders can be arbitrarily large. The focus in [16] was in the family of simplest cubic fields. Those are the totally real cubic fields generated by a polynomial of the form $f(x)=x^{3}-ax^{2}-(a+3)x-1,\quad a\in\mathbb{Z}_{\geq-1},$ meaning that $K=\mathbb{Q}(\rho)$ for a root $\rho$ of $f$. These fields were originally studied in detail by Shanks [30], and among their nice arithmetic properties, there is the one that $\mathrm{disc}(f)=\Delta^{2}$, where $\Delta=a^{2}+3a+9$. Moreover, for a positive density of $a$, this is just the discriminant of $K$, and those are the cases in which $\mathcal{O}_{K}=\mathbb{Z}[\rho]$. In [16, Theorem 1.2] the complete list of the indecomposables in $\mathbb{Z}[\rho]$ is provided, which led to some bounds on the minimal rank of universal quadratic forms (see [16, Theorem 1.1]). Full lists of indecomposables of an analogous order of fields in some other families are given in [16, Section 8] and [34]. For some partial results on indecomposable integers in real biquadratic fields, see also [5, 22]. In this paper we shall consider the whole ring of integers $\mathcal{O}_{K}$ of the simplest cubic fields $K$ such that $\mathcal{O}_{K}\neq\mathbb{Z}[\rho]$. In Section 3 we shall review the characterization of the monogenity of $K$ found by Kashio and Sekigawa [19], and some consequences that will be needed later on. We have that $K$ is monogenic if and only if the generalized module index $\delta\coloneqq[\mathcal{O}_{K}:\mathbb{Z}[\rho]]$ is the cube of some integer number, unless $K$ is defined by more than one parameter $a$. In Section 4 we will focus on the simplest cubic fields $K$ with integral basis $B_{p}(k,l)\coloneqq\Big{\\{}1,\rho,\frac{k+l\rho+\rho^{2}}{p}\Big{\\}},$ for a prime number $p$ and integers $1\leq k,l\leq p-1$. Our main results concerning these can be summarized in: ###### Theorem 1.1. Let $K=\mathbb{Q}(\rho)$ be a simplest cubic field and call $\delta=[\mathcal{O}_{K}:\mathbb{Z}[\rho]]$. * 1. $\delta=3$ if and only if $K$ has integral basis $B_{3}(1,1)$. These simplest cubic fields arise exactly when $a\equiv 3$ or $21\,(\mathrm{mod}\,27)$, $a>12$ and $\frac{\Delta}{27}$ is square-free. * 2. $\delta=p$, $p>3$ prime, if and only if $K$ has integral basis $B_{p}(k,l)$ for some $1\leq k,l\leq p-1$. These simplest cubic fields arise exactly when $p\equiv 1\,(\mathrm{mod}\,6)$. In that case, the parameter $a$ may take two possible values $a_{1}$, $a_{2}$ modulo $p^{2}$, and when $a\equiv a_{i}\,(\mathrm{mod}\,p^{2})$ for $i\in\\{1,2\\}$, there are unique integers $k_{i}$, $l_{i}$ modulo $p$ such that $k\equiv k_{i}\,(\mathrm{mod}\,p)$ and $l\equiv l_{i}\,(\mathrm{mod}\,p)$. In order to find the elements of $K$ that may possibly be indecomposable, we will essentially follow the aforementioned method introduced in [16]. It turns out that they are contained in the same two parallelepipeds found by Kala and the second author for the case that $\mathcal{O}_{K}=\mathbb{Z}[\rho]$. While the first parallelepiped can be fully described without difficulty (see Proposition 4.10), the second one is much trickier. In Section 5, we will focus in the case that $p=3$, that is, the simplest cubic fields with integral basis $g_{1}=1,\quad g_{2}=\rho,\quad g_{3}=\frac{1+\rho+\rho^{2}}{3}.$ This family had already been considered in a different (but equivalent) form in [4, Theorem 3]. For these, we describe the lattice points in the second parallelepiped. The main result in this paper is a complete list of indecomposables for fields in this family, which we prove in Section 6. ###### Theorem 1.2. Let $K$ be a simplest cubic field with $a\equiv 3$ or $21\,(\mathrm{mod}\,27)$, $a>12$ and $\frac{\Delta}{27}$ square-free. Up to multiplication by totally positive units, the indecomposable integers in $K$ are 1, * (i) $g_{3}$, * (ii) $-g_{1}-(r+1)g_{2}+3g_{3}=-r\rho+\rho^{2}$ where $1\leq r\leq\frac{a}{3}$, * (iii) $-(2v+1)g_{1}-(v(a+3)+2)g_{2}+3(v+1)g_{3}=-v-(v(a+2)+1)\rho+(v+1)\rho^{2}$ where $\frac{2a}{3}+1\leq v\leq a$, * (iv) $-(2v+1)g_{1}-(v+1)(a+2)g_{2}+3(v+1)g_{3}=-v-(v(a+2)+a-v+1)\rho+(v+1)\rho^{2}$ where $0\leq v\leq\frac{a}{3}-1$, * (v) $-(2v+1)g_{1}-(v(a+3)+r+1)g_{2}+(3v+2)g_{3}=\\\ -v-(v(a+2)+r)\rho+(v+1)\rho^{2}-g_{3}$ where $0\leq v\leq\frac{a}{3}-1$ and $\frac{a}{3}+1\leq r\leq\frac{2a}{3}-v$, * (vi) $-(r+1)g_{2}+g_{3}=1-r\rho+\rho^{2}-2g_{3}$ where $0\leq r\leq\frac{a}{3}-1$, * (vii) $-(2v+2)g_{1}-(v(a+3)+\frac{2a}{3}+3)g_{2}+(3v+4)g_{3}=\\\ -v-(v(a+2)+\frac{2a}{3}+1)\rho+(v+2)\rho^{2}-2g_{3}$ where $0\leq v\leq\frac{a}{3}-1$, * (viii) $-(2v+2)g_{1}-(v(a+3)+\frac{4a}{3}-v+3)g_{2}+(3v+4)g_{3}=\\\ -v-(v(a+2)+\frac{4a}{3}-v+1)\rho+(v+2)\rho^{2}-2g_{3}$ where $\frac{a}{3}\leq v\leq\frac{2a}{3}-1$. The points from (ii) to (iv) have minimal trace $2$, while the remaining ones have minimal trace $1$. The total number of indecomposables in $K$ (up to multiplication by totally positive units) is $\frac{a^{2}+3a}{18}+2a+2$. The ordering of the indecomposables in the statement above will follow from the classification of lattice points carried out in Section 5. In Section 7, we shall study some applications of Theorem 1.2 for simplest cubic fields with integral basis $B_{3}(1,1)$. In [23], Lemmermeyer and Pethö proved that every $\gamma\in\mathbb{Z}[\rho]$ not associated with a rational integer, the norm $N(\gamma)$ of $\gamma$ satisfies $|N(\gamma)|\geq 2a+3$. In Proposition 7.1 we will see that except for some small values of $a$, this is also true for our subfamily with $p=3$ and $(k,l)=(1,1)$. On the other hand, we will also study the largest possible norm of the indecomposables, and concretely we will prove that $|N(\alpha)|\leq\frac{(a^{2}+3a+9)^{2}}{729}$, again with some exceptions (see Proposition 7.4). We will also show that the Pythagoras number of $K$ is $6$, just as in the monogenic case studied by Kala and the second author. We shall use these informations to find some bounds for $m(K)$, in the style of the stated ones in [16, Theorem 1.1]. For more details, see Section 7. Our results are in the line of the existing works for the additive structure of the totally real fields, which show evidence of its unpredictability. However, this work is pioneer in determining the whole structure of indecomposables in the ring of integers in a family of non-monogenic totally real fields. Remarkably, we have found this absence of pattern even within the same family of simplest cubic fields, for which the value of generalized module index $[\mathcal{O}_{K}:\mathbb{Z}[\rho]]$ turns out to be extremely influential. ## 2 Preliminaries Let $K$ be a totally real number field of degree $d$ and let $\sigma_{1},\dots,\sigma_{d}\colon K\hookrightarrow\mathbb{R}$ be its embeddings. By trace $\text{Tr}(\alpha)$ and norm $N(\alpha)$ of $\alpha\in K$, we will mean $\text{Tr}(\alpha)=\sum_{i=1}^{d}\sigma_{i}(\alpha)\hskip 28.45274pt\text{ and }\hskip 28.45274ptN(\alpha)=\prod_{i=1}^{d}\sigma_{i}(\alpha).$ An element $\alpha\in K$ is totally positive if $\sigma_{i}(\alpha)>0$ for every $1\leq i\leq d$, and we will denote it by $\alpha\succ 0$. The subset of the totally positive integers in $K$ is denoted by $\mathcal{O}_{K}^{+}$. Moreover, by $\alpha\succ\beta$, we will mean $\alpha-\beta\succ 0$. We will also use symbol $\succeq$ to include equality between our elements. A totally positive integer $\alpha\in\mathcal{O}_{K}$ is said to be indecomposable if it cannot be written as the sum of other two totally positive integers in $\mathcal{O}_{K}$. The codifferent of $\mathcal{O}_{K}$ is $\mathcal{O}_{K}^{\vee}=\\{\delta\in K\,|\,\text{Tr}(\delta\alpha)\in\mathbb{Z}\hbox{ for every }\alpha\in\mathcal{O}_{K}\\}.$ The subset of the totally positive integers in $\mathcal{O}_{K}^{\vee}$ is denoted by $\mathcal{O}_{K}^{\vee,+}$. For $\alpha\in\mathcal{O}_{K}^{+}$, the number $\mathrm{min}_{\delta\in\mathcal{O}_{K}^{\vee,+}}\mathrm{Tr}(\delta\alpha)$ will be referred to as the minimal trace of $\alpha$. It is immediate that the totally positive algebraic integers with minimal trace $1$ are indecomposable. We proceed to summarize the method introduced in [16] to find the indecomposables in $K$. Let us embed $K$ on the Minkowski space $\mathbb{R}^{d}$ by means of an integral basis of $K$. Following [32, Section 1, p. 2], we introduce the following terminology: ###### Definition 2.1. A pair $(\varepsilon_{1},\varepsilon_{2})$ of units in $K$ is proper if the following statements hold: * (1) The subgroup of $K^{*}$ generated by $\\{\varepsilon_{1},\varepsilon_{2}\\}$ has rank $2$. * (2) The set $\\{1,\varepsilon_{1},\varepsilon_{2}\\}$ is a $\mathbb{Q}$-basis of $K$. * (3) In the linear combination of $\varepsilon_{1}\varepsilon_{2}$ with respect to the above basis, the coefficient before $1$ is negative. On the other hand, the parallelepiped with nodes $\ell_{1},\dots,\ell_{e}\in\mathbb{Z}^{d}$ is $\mathcal{D}(\ell_{1},\dots,\ell_{e})=[0,1]\ell_{1}+\dots+[0,1]\ell_{e}.$ The key result is the following refinement of [32, Theorem 1] (see [16, Section 4, p. 11-12]): ###### Proposition 2.2. Let $K$ be a totally real cubic field and let $\\{\varepsilon_{1},\varepsilon_{2}\\}$ be a proper pair of totally positive units of $K$. The indecomposables of $K$ lie, up to multiplication by totally positive units in $K$, in either of the parallelepipeds $\mathcal{D}(1,\varepsilon_{1},\varepsilon_{2}),\quad\mathcal{D}(1,\varepsilon_{1},\varepsilon_{1}\varepsilon_{2}^{-1}).$ This method was applied successfully to the case that $K$ is a simplest cubic field (see Section 3 below) with $\mathcal{O}_{K}=\mathbb{Z}[\rho]$, leading to a list of lattice points among which the indecomposables were identified. ###### Theorem 2.3. [16, Theorem 1.2] Let $\rho$ be a root of the polynomial $x^{3}-ax^{2}-(a+3)x-1$ where $a\in{\mathbb{Z}}_{\geq-1}$. Up to multiplication by totally positive units, the indecomposables in the order $\mathbb{Z}[\rho]$ of ${\mathbb{Q}}(\rho)$ are $1$, * • the element $1+\rho+\rho^{2}$ with minimal trace $2$, * • the elements in the set $\blacktriangle=\\{-v-w\rho+(v+1)\rho^{2}\,|\,0\leq v\leq a,\,v(a+2)+1\leq w\leq(v+1)(a+1)\\},$ all of which have minimal trace $1$. ## 3 Simplest cubic fields and their monogenity A simplest cubic field is a number field $K=\mathbb{Q}(\rho)$, where $\rho$ is a root of the polynomial $f(x)=x^{3}-ax^{2}-(a+3)x-1,\quad a\in\mathbb{Z}_{\geq-1}.$ It is a cyclic cubic extension of $\mathbb{Q}$ (and then totally real). Let $\rho^{\prime}$ and $\rho^{\prime\prime}$ be the conjugates of $\rho$, that is, the other roots of $f$. When $a\geq 7$, reordering these roots if necessary, we can assume that [23] $a+1<\rho<a+1+\frac{2}{a},\quad-1-\frac{1}{a}<\rho^{\prime}<-1-\frac{1}{2a},\quad-\frac{1}{a+2}<\rho^{\prime\prime}<-\frac{1}{a+3}.$ We also have the explicit expressions $\rho^{\prime}=-1-\frac{1}{\rho}=a+2+a\rho-\rho^{2}\quad\text{and}\quad\rho^{\prime\prime}=-\frac{1}{1+\rho}=-2-(a+1)\rho+\rho^{2}.$ Let us denote $\Delta\coloneqq a^{2}+3a+9$. We know that $\mathrm{disc}(f)=\Delta^{2}$. If $\Delta$ is square-free, then $\mathcal{O}_{K}=\mathbb{Z}[\rho]$. The converse does not hold in general. We say that $K$ is monogenic if there is an element $\gamma\in\mathcal{O}_{K}$ such that $\mathcal{O}_{K}=\mathbb{Z}[\gamma]$. In that case, $\\{1,\gamma,\gamma^{2}\\}$ is a power integral basis of $K$. Denote by $\mathfrak{c}$ the conductor of $K$. The monogenity of $K$ is characterized in [19] in the following way: ###### Theorem 3.1. [19, Corollary 1.6] The following are equivalent: * 1. The field $K$ is monogenic. * 2. We have $a\in\\{-1,0,1,2,3,5,12,54,66,1259,2389\\}$ or $\frac{\Delta}{\mathfrak{c}}$ is a cube. * 3. We have $a\in\\{-1,0,1,2,3,5,12,54,66,1259,2389\\}$ or $a\not\equiv 3,21\,(\mathrm{mod}\,27)$ and $v_{p}(\Delta)\not\equiv 2\,(\mathrm{mod}\,3)$ for all $p\neq 3$. In such a case, a power integral basis of $K$ is generated by $\gamma=\frac{\rho-b}{n}$, where $n=\sqrt[3]{\frac{\Delta}{27}}$ and $b\in\mathbb{Z}$ satisfies $b\equiv\frac{a}{3}\,(\mathrm{mod}\,n)$ if $3\mid n$ and $3a\equiv b\,(\mathrm{mod}\,n)$ otherwise. ###### Remark 3.2. The list of possible values for $a$ in the second and the third statement in Theorem 3.1 are the ones for which the corresponding simplest cubic field $K$ is also defined by some other value $a^{\prime}$ in the same list. Hence, its presence is due to the fact that for such a simplest cubic field $K$, one of the values satisfies the conditions of monogenity and the other one does not, resulting that $K$ is monogenic. We proceed to reinterpret the relevant number $\frac{\Delta}{\mathfrak{c}}$ in the second statement of Theorem 3.1. Since $K$ is an abelian extension of $\mathbb{Q}$, from Kronecker-Weber theorem we know that it is contained in some cyclotomic field. Since in addition $K$ is totally real, $\mathfrak{c}$ is the smallest integer $n\in\mathbb{Z}_{>0}$ such that $K$ is contained in the $n$-th cyclotomic field. In particular, $\mathfrak{c}$ is positive. Moreover, it is related to the discriminant by the so called conductor- discriminant formula, $\mathfrak{c}^{2}=\mathrm{disc}(K)$. Together with the well known formula $\mathrm{disc}(f)=[\mathcal{O}_{K}:\mathbb{Z}[\rho]]^{2}\mathrm{disc}(K)$, we deduce that $\mathfrak{c}$ divides $\Delta$ and $[\mathcal{O}_{K}:\mathbb{Z}[\rho]]=\frac{\Delta}{\mathfrak{c}}.$ ###### Remark 3.3. Note that $K$ may be monogenic even if $\mathcal{O}_{K}\neq\mathbb{Z}[\rho]$, namely whenever $[\mathcal{O}_{K}:\mathbb{Z}[\rho]]$ is a non-trivial cube. This means that $\mathcal{O}_{K}=\mathbb{Z}[\gamma]$ for some $\gamma\neq\rho$. However, we will usually refer to the case $\mathcal{O}_{K}=\mathbb{Z}[\rho]$ as the monogenic case. In order to compute effectively the generalized module index $[\mathcal{O}_{K}:\mathbb{Z}[\rho]]$, we denote $\Delta=bc^{3}$ with $b,c>0$ coprime integers and $b$ cube-free. Then: ###### Proposition 3.4. [19, Remark 1.5] The conductor can be described as follows: $\mathfrak{c}=\begin{cases}\prod_{p\mid b}p&\hbox{if }3\nmid a\hbox{ or }a\equiv 12\,(\mathrm{mod}\,27),\\\ 3^{2}\prod_{p\mid b,p\neq 3}p&\hbox{otherwise}.\end{cases}$ ###### Remark 3.5. If $3\nmid a$, we have that $\Delta=a^{2}+3a+9\equiv 1\,(\mathrm{mod}\,3)$, and if $a\equiv 12\,(\mathrm{mod}\,27)$, then $\Delta\equiv 0\,(\mathrm{mod}\,27)$. In both cases, we see that $3\nmid\mathfrak{c}$. We deduce that the conductor of a simplest cubic field $K$ satisfies either that $3\nmid\mathfrak{c}$ or $v_{3}(\mathfrak{c})=2$. ###### Example 3.6. Let $K$ be the simplest cubic field with $a=21$. Then $\Delta=3^{3}\cdot 19$. Therefore, $c=3$ and $b=19$. Since $3$ divides $a$ and $a\not\equiv 12\,(\mathrm{mod}\,27)$, we find that $\mathfrak{c}=19\cdot 3^{2}$. Then $\frac{\Delta}{\mathfrak{c}}=3$, which is not a cube, and then $K$ is not monogenic. Actually, the integral basis of $K$ is $\Big{\\{}1,\rho,\frac{1+\rho+\rho^{2}}{3}\Big{\\}}.$ ###### Example 3.7. Let $K$ be the simplest cubic field with $a=41$. Then $\Delta=7^{2}\cdot 39$ and $\mathfrak{c}=7\cdot 39$, so $\frac{\Delta}{\mathfrak{c}}=7$ is not a cube and $K$ is not monogenic. In this case, the integral basis of $K$ is $\Big{\\{}1,\rho,\frac{4+3\rho+\rho^{2}}{7}\Big{\\}}.$ ### 3.1 The parallelepipeds of lattice points We focus on the problem of finding the indecomposable integers in a simplest cubic field $K$ for the case when $\mathcal{O}_{K}\neq\mathbb{Z}[\rho]$. That is, calling $\delta=[\mathcal{O}_{K}:\mathbb{Z}[\rho]]$, we remove the restriction that $\delta=1$ and we aim to find all indecomposables in $K$. In order to find candidates, we will follow a slight variation of the method in [16], based on the well known fact that $\delta\mathcal{O}_{K}\subseteq\mathbb{Z}[\rho]$. Let us fix the power basis $\\{1,\rho,\rho^{2}\\}$, which is not integral unless $\mathcal{O}_{K}=\mathbb{Z}[\rho]$. We have that $\mathcal{O}_{K}$ embeds in $\frac{1}{\delta}\mathbb{Z}^{3}$ through $\begin{array}[]{rccl}\tau\colon&K&\longrightarrow&\mathbb{R}^{3},\\\ &\alpha_{1}+\alpha_{2}\rho+\alpha_{3}\rho^{2}&\longmapsto&(\alpha_{1},\alpha_{2},\alpha_{3}).\end{array}$ Now, let us call $\varepsilon_{1}=\rho^{2}$ and $\varepsilon_{2}=(\rho^{\prime\prime})^{-2}$. From [32, Corollary 2] we know that the pairs $(1,\varepsilon_{1},\varepsilon_{2})$ and $(1,\varepsilon_{1},\varepsilon_{1}\varepsilon_{2}^{-1})$ are proper (note that conditions (1)-(3) in Definition 2.1 do not depend on whether we have $\mathcal{O}_{K}=\mathbb{Z}[\rho]$). Then, the problem now turns to find the intersection of $\frac{1}{\delta}\mathbb{Z}^{3}$ with the parallelepipeds $\mathcal{D}(1,\rho^{2},1+2\rho+\rho^{2}),\quad\mathcal{D}(1,\rho^{2},-1-a-(a^{2}+3a+3)\rho+(a+2)\rho^{2}).$ (1) From now on, these will be referred to as the first parallelepiped and the second parallelepiped, respectively. ###### Remark 3.8. There might be elements in $\frac{1}{\delta}\mathbb{Z}[\rho]$ that do not lie in $\mathcal{O}_{K}$. In other words, there might be some lattice points that are not algebraic integers. We shall find the lattice points in the first of the above parallelepipeds. The second one will be completely determined for a specific family of simplest cubic fields $K$ in Section 5, by using an integral basis of $K$. We look for $t_{1},t_{2},t_{3}\in[0,1]$ such that for some $(m,n,o)\in\mathbb{Z}^{3}$, $t_{1}+t_{2}\rho^{2}+t_{3}(1+2\rho+\rho^{2})=\frac{1}{\delta}(m+n\rho+o\rho^{2}).$ This gives rise to the system of equations $\displaystyle t_{1}+t_{3}$ $\displaystyle=\frac{1}{\delta}m,$ $\displaystyle 2t_{3}$ $\displaystyle=\frac{1}{\delta}n,$ $\displaystyle t_{2}+t_{3}$ $\displaystyle=\frac{1}{\delta}o.$ The second equation gives that $t_{3}=\frac{n}{2\delta}$ with $0\leq n\leq 2\delta$. Carrying this to the first equation, $t_{1}=\frac{2m-n}{2\delta}$, with $\frac{n}{2}\leq m\leq\frac{n}{2}+\delta$. As for the third one, $t_{2}=\frac{2o-n}{2\delta}$, with $\frac{n}{2}\leq o\leq\frac{n}{2}+\delta$. We plug these values in the parametric equation, obtaining the points $\frac{1}{\delta}(m+n\rho+o\rho^{2}),\quad 0\leq n\leq 2\delta,\quad\frac{n}{2}\leq m,o\leq\frac{n}{2}+\delta,$ which are candidates to be indecomposables of $K$. Note that choosing $m=n=o=\delta$ yields the point $1+\rho+\rho^{2}$, which was identified as indecomposable of the $\mathbb{Z}$-order $\mathbb{Z}[\rho]$ in [16]. On the other hand, unlike in that case, the trivial choices $n=0$ or $n=2\delta$ give non-trivial candidates for $\frac{n}{2}<m,o<\frac{n}{2}+\delta$. ###### Proposition 3.9. Let $K$ be a simplest cubic field and let $\delta=[\mathcal{O}_{K}:\mathbb{Z}[\rho]]$. The candidates (i.e, lattice points that are possibly indecomposables of $K$) from the first parallelepiped in (1) are the points in the set $\left\\{\frac{m}{\delta}+\frac{n}{\delta}\rho+\frac{o}{\delta}\rho^{2};\;0\leq n\leq 2\delta,\;\frac{n}{2}\leq m,o\leq\frac{n}{2}+\delta\right\\}\cap{\mathcal{O}}_{K}^{+}.$ ## 4 Non-monogenic simplest cubic fields with prime index In this section we introduce the simplest cubic fields $K$ with an integral basis of the form $B_{p}(k,l)\coloneqq\Big{\\{}1,\rho,\frac{k+l\rho+\rho^{2}}{p}\Big{\\}},$ (2) where $p$ is an odd prime, and $k,l\in\mathbb{Z}$ are not divisible by $p$. Note that we may assume without loss of generality that $1\leq k,l\leq p-1$. When $(k,l)$ is implicit in the context, we just denote $B_{p}=B_{p}(k,l)$. The simplest cubic fields in Examples 3.6 and 3.7 are of this form: in the former we have $(k,l)=(1,1)$ and $p=3$, while in the latter it is $(k,l)=(4,3)$ and $p=7$. We begin with the following important lemma. ###### Lemma 4.1. Let $K$ be a simplest cubic field with integral basis $B_{p}(k,l)$ for $1\leq k,l\leq p-1$. Then, $K$ does not admit an integral basis of the same form for another pair $(k^{\prime},l^{\prime})\neq(k,l)$, $1\leq k^{\prime},l^{\prime}\leq p-1$. ###### Proof. We start with the assumption that $K$ has integral basis as in the statement, so that $\frac{k+l\rho+\rho^{2}}{p}\in\mathcal{O}_{K}$. Let $(k^{\prime},l^{\prime})$ be a pair of integer numbers such that $\frac{k^{\prime}+l^{\prime}\rho+\rho^{2}}{p}$ belongs to $\mathcal{O}_{K}$ as well. Then, their difference $\frac{k^{\prime}-k+(l^{\prime}-l)\rho}{p}\in\mathcal{O}_{K}.$ Let us call $c=k^{\prime}-k$ and $d=l^{\prime}-l$. Then there are unique $x_{1},x_{2},x_{3}\in\mathbb{Z}$ such that $\frac{c+d\rho}{p}=x_{1}+x_{2}\rho+x_{3}\frac{k+l\rho+\rho^{2}}{p}.$ Using the uniqueness of coordinates with respect to the power basis, we obtain that $x_{1}=\frac{c}{p}$, $x_{2}=\frac{d}{p}$ and $x_{3}=0$. Hence $c$ and $d$ are multiples of $p$, meaning that $k\equiv k^{\prime}\,(\mathrm{mod}\,p)$ and $l\equiv l^{\prime}\,(\mathrm{mod}\,p)$. ∎ In order to find the pair $(k,l)$ for which $\frac{k+l\rho+\rho^{2}}{p}$ is the third element of the integral basis, we follow this procedure. The minimal polynomial of $\frac{k+l\rho+\rho^{2}}{p}$ is $x^{3}-\frac{h_{1}}{p}x^{2}+\frac{h_{2}}{p^{2}}x-\frac{h_{3}}{p^{3}},$ where $\begin{split}h_{1}=h_{1}(k,l)&=a^{2}+(l+2)a+3k+6,\\\ h_{2}=h_{2}(k,l)&=(2k-l+1)a^{2}+(-l^{2}+2kl+4k-3l+4)a+3k^{2}-3l^{2}+12k-3l+9,\\\ h_{3}=h_{3}(k,l)&=(k^{2}-kl+k)a^{2}+(k^{2}l-kl^{2}+2k^{2}+l^{2}-3kl+4k-l)a+k^{3}+l^{3}\\\ &\hskip 28.45274pt-3kl^{2}+6k^{2}-3kl+9k-3l+1.\end{split}$ Then, we are just looking for the pair $(k,l)$, $1\leq k,l\leq p-1$, such that $h_{i}(k,l)\equiv 0\,(\mathrm{mod}\,p^{i}),\quad i\in\\{1,2,3\\}.$ (3) Eventually we will prove that such a pair actually exists. Let $K$ be a simplest cubic field with integral basis $B_{p}(k,l)$ and let us compute the module index $\delta=\frac{\Delta}{\mathfrak{c}}$. It may be checked, for instance using mathematical software, that $\mathrm{disc}\Big{(}1,\rho,\frac{k+l\rho+\rho^{2}}{p}\Big{)}=\mathrm{det}\begin{pmatrix}1&\rho&\frac{k+l\rho+\rho^{2}}{p}\\\ 1&\rho^{\prime}&\frac{k+l\rho^{\prime}+\rho^{\prime 2}}{p}\\\ 1&\rho^{\prime\prime}&\frac{k+l\rho^{\prime\prime}+\rho^{\prime\prime 2}}{p}\end{pmatrix}^{2}=\Big{(}\frac{a^{2}+3a+9}{p}\Big{)}^{2},$ and recall that the conductor $\mathfrak{c}$ must be positive, so $\mathfrak{c}=\frac{a^{2}+3a+9}{p}$. It follows that $\delta=p$, so that $\Delta=p\mathfrak{c}$. In particular, $K$ is not monogenic. Recall from Remark 3.2 that the simplest cubic fields defined by more than one value of $a$ (i.e. $a\in\\{-1,0,1,2,3,5,12,54,66,1259,2389\\}$) are monogenic. Since $K$ is not monogenic, this is not the case. Then, Theorem 3.1 (3) gives that $a\equiv 3$ or $21$ $(\mathrm{mod}\,27)$ or $v_{q}(\Delta)\equiv 2\,(\mathrm{mod}\,3)$ for some prime $q\neq 3$. Keeping the assumption that $K$ has integral basis $B_{p}(k,l)$, we explore each of these two possibilities. #### Case $a\equiv 3$ or $21$ $(\mathrm{mod}\,27)$ Suppose that $a\equiv 3$ or $21$ $(\mathrm{mod}\,27)$. A direct computation shows that $\Delta=a^{2}+3a+9$ is divisible by $27$. But since $K$ has integral basis $B_{p}$, we have that $\Delta=p\mathfrak{c}$, and from Proposition 3.4 we see that $v_{3}(\mathfrak{c})=2$, so it must be $p=3$. That is, $K$ has integral basis $B_{3}(k,l)$ for some $1\leq k,l\leq 2$ (which are unique by Lemma 4.1). If we choose $(k,l)=(1,1)$, we have $\begin{split}h_{1}&=a^{2}+3a+9=\Delta\equiv 0\,(\mathrm{mod}\,3),\\\ h_{2}&=2a^{2}+6a+18=2\Delta\equiv 0\,(\mathrm{mod}\,9),\\\ h_{3}&=a^{2}+3a+9=\Delta\equiv 0\,(\mathrm{mod}\,27).\end{split}$ Then the only pair $(k,l)$ with $1\leq k,l\leq 2$ for which $\frac{k+l\rho+\rho^{2}}{3}$ is the third element of $B_{3}$ is $(k,l)=(1,1)$. In fact, $a\equiv 3$ or $21$ $(\mathrm{mod}\,27)$ only when $K$ has integral basis $B_{3}$. In the next section, we will completely characterize the simplest cubic fields with integral basis of this form. #### Case $v_{q}(\Delta)\equiv 2\,(\mathrm{mod}\,3)$, $q\neq 3$ prime Next, assume that $K$ satisfies that $v_{q}(\Delta)\equiv 2\,(\mathrm{mod}\,3)$ for some prime $q\neq 3$, which is equivalent to $q^{2}\mid\Delta$. Since $\Delta=p\mathfrak{c}$ and $q\neq 3$, from Proposition 3.4 it follows that $q=p$. Hence we obtain that $p\neq 3$ and $p^{2}\mid\Delta$. In particular, if $K$ has integral basis $B_{p}$ for $p>3$, then $p^{2}\mid\Delta$. There is a partial converse for this implication. ###### Proposition 4.2. Let $K=\mathbb{Q}(\rho)$ be a simplest cubic field such that $p^{2}\mid\Delta$ for a prime $p>3$. Then there are integers $1\leq k,l\leq p-1$ such that $\frac{k+l\rho+\rho^{2}}{p}\in\mathcal{O}_{K}$. ###### Proof. We will check that under the hypothesis that $p^{2}\mid\Delta$, there exist $1\leq k,l\leq p-1$ such that $h_{i}(k,l)\equiv 0\,(\mathrm{mod}\,p^{i})$ for all $i\in\\{1,2,3\\}$. Since $p>3$, $3$ has a modular inverse $3^{-1}$ modulo $p$. If the equations (3) have some solution $(k,l)$ as stated, from the first one we obtain that $k=-3^{-1}(al+a^{2}+2a+6)+Up,$ where $U\in\mathbb{Z}$ is such that $1\leq k\leq p-1$. We will check that this is satisfied for some value of $l$ modulo $p$. Let us take $U\in\mathbb{Z}$. Since $p^{2}\mid\Delta$, we have that $3h_{2}(-3^{-1}(al+a^{2}+2a+6)+Up,l)=-(a^{2}+a+1+(2a+1)l+l^{2})\Delta+9p^{2}U^{2}\equiv 0\,(\mathrm{mod}\,p^{2}).$ Thus, if $p\mid h_{1}(k,l)$ for some $(k,l)\in\mathbb{Z}^{2}$, then $p^{2}\mid h_{2}(k,l)$. On the other hand, $27h_{3}(-3^{-1}(al+a^{2}+2a+6)+Up,l)=g(l)-9p\Delta U(a^{2}+a+1+(2a+1)l+l^{2})+27p^{3}U^{3},$ where $g(l)=\Delta((2a^{2}-3)\Delta+8(2a+3)+((6a-3)\Delta-12a+36)l+(6\Delta-6a-36)l^{2}+(2a+3)l^{3}).$ Now, using again that $p^{2}\mid\Delta$ gives $27h_{3}(-3^{-1}(al+a^{2}+2a+6)+Up,l)\equiv g(l)\,(\mathrm{mod}\,p^{3})$. The proof will be finished as soon as we check that $g(l)$ has some root modulo $p^{3}$, and since $\Delta|g(l)$ for every $l$, it is enough to check that $\frac{g(l)}{\Delta}$ has a root modulo $p$. Let $q(l)=\frac{g(l)}{\Delta}\in\mathbb{Z}[l]$. We have that $\begin{split}q^{\prime}(l)&=(6a-3)\Delta-12a+36+2(6\Delta-6a-36)l+3(2a+3)l^{2}\\\ &=3(2a^{3}+5a^{2}+11a+3)+6(2a^{2}+4a+6)l+3(2a+3)l^{2}.\end{split}$ If $2a+3\equiv 0\,(\mathrm{mod}\,p)$, then $a\equiv-3\cdot 2^{-1}\,(\mathrm{mod}\,p)$, and replacing in the congruence $a^{2}+3a+9\equiv 0\,(\mathrm{mod}\,p)$ gives that $27\equiv 0\,(\mathrm{mod}\,p)$, which is a contradiction because $p>3$. Then $2a+3\not\equiv 0\,(\mathrm{mod}\,p)$; let $(2a+3)^{-1}$ be its modular inverse modulo $p$. Since $\mathrm{disc}(q^{\prime})=2^{2}3^{3}\Delta\equiv 0\,(\mathrm{mod}\,p)$, $q(l)$ and $q^{\prime}(l)$ share a root modulo $p$, which is $l\equiv-(2a^{2}+4a+6)(2a+3)^{-1}\,(\mathrm{mod}\,p)$. In particular this is a root of $q(l)$ modulo $p$, and since $p^{2}\mid\Delta$, it is a root of $g(l)$ modulo $p^{3}$. ∎ From the previous proof we can extract a simple relation between $k$ and $l$. ###### Corollary 4.3. Let $K=\mathbb{Q}(\rho)$ be a simplest cubic field such that $p^{2}\mid\Delta$. The integers $k$, $l$ in the proof of Proposition 4.2 satisfy $2k-l\equiv-2\,(\mathrm{mod}\,p)$. In particular, this holds if $K$ has integral basis $B_{p}(k,l)$ with $p>3$. ###### Proof. We have that $l\equiv-(2a^{2}+4a+6)(2a+3)^{-1}\,(\mathrm{mod}\,p)$ and $k\equiv-3^{-1}(al+a^{2}+2a+6)\,(\mathrm{mod}\,p)$. Then $\begin{split}2k-l&\equiv-2\cdot 3^{-1}(al+a^{2}+2a+6)-l\\\ &\equiv-3^{-1}((2a+3)l+2(a^{2}+2a+6))\\\ &\equiv-3^{-1}(-2a^{2}-4a-6+2a^{2}+4a+12)\equiv-2\,(\mathrm{mod}\,p)\end{split}$ The last part of the statement holds because of the uniqueness of $k$ and $l$ given by Lemma 4.1. ∎ ###### Corollary 4.4. Let $p>3$ be a prime number and let $K=\mathbb{Q}(\rho)$ be a simplest cubic field. Then $K$ has integral basis $B_{p}$ if and only if $p^{2}\mid\Delta$ and $\Delta=p^{2}d$ or $\Delta=9p^{2}d$ for some $d\in\mathbb{Z}$ square-free and coprime to $3$. ###### Proof. We already know that if $K$ has integral basis $B_{p}$, then $p^{2}\mid\Delta$ and $\Delta=p\mathfrak{c}$. Looking at the description of the conductor in Proposition 3.4 and Remark 3.5, we see that $\Delta=p^{2}d$ or $9p^{2}d$ for some square-free integer $d$ coprime to $3$, depending on which of the two possible description of $\mathfrak{c}$ occurs. Conversely, if $p^{2}\mid\Delta$ and $\Delta=p^{2}d$ or $9p^{2}d$ with $d$ square-free and coprime to $3$, by Proposition 4.2 we have that $\frac{k+l\rho+\rho^{2}}{p}\in\mathcal{O}_{K}$ for unique integers $1\leq k,l\leq p-1$. Now, $\mathrm{disc}(1,\rho,\frac{k+l\rho+\rho^{2}}{p})=\Big{(}\frac{\Delta}{p}\Big{)}^{2}$. Recall from Remark 3.5 that either $3\nmid\mathfrak{c}$ or $v_{3}(\mathfrak{c})=2$. Thus the hypothesis gives that $\frac{\Delta}{p}=\mathfrak{c}$, so the discriminant of $B_{p}$ is just the discriminant of $K$. Then $B_{p}$ is an integral basis of $K$. ∎ ###### Remark 4.5. If $K$ is a simplest cubic field with $[\mathcal{O}_{K}:\mathbb{Z}[\rho]]=p$ for $p>3$, then $\Delta=p\mathfrak{c}$, so $p^{2}\mid\Delta$ and $\Delta=p^{2}d$ or $9p^{2}d$ for some square-free integer $d$. Then, we can argue as in the second part of the previous proof to conclude that $B_{p}$ is an integral basis of $K$. ###### Example 4.6. The simplest cubic field $K$ defined by $a=90$ has integral basis $B_{7}(4,3)$ and $\Delta=9\cdot 7^{2}\cdot 19$. In this case, we have that $3\mid a$ and $a\equiv 41\,(\mathrm{mod}\,49)$. We may interpret the condition that $p^{2}\mid\Delta$ as an equation $a^{2}+3a+9\equiv 0\,(\mathrm{mod}\,p^{2})$ in $a$. Then, in order to have a simplest cubic field with integral basis $B_{p}$ we need this equation to be solvable. If we call $h(x)=x^{2}+3x+9$, we know by Hensel’s lemma that any factorization of $h$ modulo $p$ lifts to a factorization modulo $p^{2}$. Now, if the equation is solvable, $h$ has two simple roots modulo $p$, so they lift to two simple roots modulo $p^{2}$. We conclude that for a prime $p$ such that the equation above is solvable, there is a simplest cubic field with integral basis $B_{p}$ for exactly two values of $a$ modulo $p^{2}$ (and each such an integer $a$ produces a simplest cubic field of this type unless $a\in\\{-1,0,1,2,3,5,12,54,66,1259,2389\\}$). In the next result, we find a nice characterization of the condition that $p^{2}\mid\Delta$. ###### Lemma 4.7. Let $p>3$ be a prime number. The equation $a^{2}+3a+9\equiv 0\,(\mathrm{mod}\,p^{2})$ is solvable if and only if $p\equiv 1\,(\mathrm{mod}\,6)$. ###### Proof. Let us call $h(x)=x^{2}+3x+9$. We have that $\mathrm{disc}(h)=-27$, so the equation in the statement is solvable only if $-27$ is a square modulo $p$. If we consider the Legendre symbol, we obtain $\genfrac{(}{)}{}{0}{-27}{p}=\genfrac{(}{)}{}{0}{-1}{p}\genfrac{(}{)}{}{0}{3}{p}^{3}.$ Then the result follows easily by determining the value of each factor in terms of $p$. ∎ ###### Remark 4.8. If $a\in\mathbb{Z}$ is such that $a^{2}+3a+9\equiv 0\,(\mathrm{mod}\,p^{2})$, the other solution of this equation in congruences is the class of $a^{2}+2a+6$ modulo $p^{2}$, that is, $(a^{2}+2a+6)^{2}+3(a^{2}+2a+6)+9\equiv 0\,(\mathrm{mod}\,p^{2})$. ###### Proposition 4.9. For each prime number $p>3$, there is a simplest cubic field with integral basis $B_{p}$ if and only if $p\equiv 1\,(\mathrm{mod}\,6)$. ###### Proof. Assume that $K$ is a simplest cubic field with integral basis $B_{p}$. Then we know that $p^{2}\mid\Delta$, so that the equation $a^{2}+3a+9\equiv 0\,(\mathrm{mod}\,p^{2})$ is solvable. By Lemma 4.7, we deduce that $p\equiv 1\,(\mathrm{mod}\,6)$. Conversely, assume that $p\equiv 1\,(\mathrm{mod}\,6)$. Then the equation $a^{2}+3a+9\equiv 0\,(\mathrm{mod}\,p^{2})$ has two solutions, namely it is satisfied for two values of $a$ modulo $p^{2}$. Among the values of $a\in\mathbb{Z}$ satisfying the equation above modulo $p^{2}$, we claim that there is some for which $\frac{a^{2}+3a+9}{p^{2}}$ is square- free and $a\notin\\{-1,0,1,2,3,5,12,54,66,1259,2389\\}$. Indeed, if for $a_{0}\in\mathbb{Z}$ we have $\Delta_{0}\coloneqq a_{0}^{2}+3a_{0}+9\equiv 0\,(\mathrm{mod}\,p^{2})$, the integers $a_{k}=a_{0}+kp^{2}$, $k\in\mathbb{Z}$, also satisfy this equation. Now, for $\Delta_{k}\coloneqq a_{k}^{2}+3a_{k}+9$, we have $\frac{\Delta_{k}}{p^{2}}=p^{2}k^{2}+(2a_{0}+3)k+\frac{\Delta_{0}}{p^{2}}.$ This is a quadratic polynomial in $k$, so it is well known that it takes infinitely many square-free values, see [25]. By Corollary 4.4, the simplest cubic field $K$ with some such a parameter $a$ has integral basis $B_{p}$. ∎ Thus, given a prime $p>3$ with $p\equiv 1\,(\mathrm{mod}\,6)$, we can determine the family of simplest cubic fields with integral basis $B_{p}$ as follows: * • We find the two values of $a$ modulo $p^{2}$ for which $a^{2}+3a+9\equiv 0\,(\mathrm{mod}\,p^{2})$. * • For each of those values, we find the pair $(k,l)\in{\mathbb{Z}}_{p}^{2}$ so that $\frac{k+l\rho+\rho^{2}}{p}\in{\mathcal{O}}_{K}$. In Table 1, we show the two values of $a$ for which there is a simplest cubic field with integral basis $B_{p}$, for some low values of $p>3$, $p\equiv 1\,(\mathrm{mod}\,6)$. It is implicitly assumed that $a\notin\\{-1,0,1,2,3,5,12,54,66,1259,2389\\}$ and $\Delta=p^{2}d$ or $9p^{2}d$ for $d\in\mathbb{Z}$ square-free. $p$ | Values of $a$ mod $p^{2}$ | $(k,l)$ ---|---|--- $7$ | $5$ | $(2,6)$ $41$ | $(4,3)$ $13$ | $66$ | $(3,8)$ $100$ | $(9,7)$ $19$ | $154$ | $(11,5)$ $204$ | $(7,16)$ $31$ | $356$ | $(25,21)$ $602$ | $(5,12)$ $37$ | $374$ | $(10,22)$ $992$ | $(26,17)$ $43$ | $577$ | $(36,31)$ $1269$ | $(6,14)$ $p$ | Values of $a$ mod $p^{2}$ | $(k,l)$ ---|---|--- $61$ | $1259$ | $(47,35)$ $2459$ | $(13,28)$ $67$ | $2097$ | $(37,9)$ $2389$ | $(29,60)$ $73$ | $1265$ | $(64,57)$ $4061$ | $(8,18)$ $79$ | $1096$ | $(55,33)$ $5142$ | $(23,48)$ $97$ | $4451$ | $(35,72)$ $4955$ | $(61,27)$ $103$ | $271$ | $(46,94)$ $10335$ | $(56,11)$ Table 1: Families of simplest cubic fields with integral basis $B_{p}$ for $7\leq p\leq 103$ ### 4.1 Lattice points in the first parallelepiped Let $K$ be a simplest cubic field with integral basis $B_{p}(k,l)$ for $1\leq k,l\leq p-1$. We describe the lattice points in the first parallelepiped following the description obtained in the general case. Namely, we have the points $\frac{1}{p}(m+n\rho+o\rho^{2}),\quad 0\leq n\leq 2p,\quad\frac{n}{2}\leq m,o\leq\frac{n}{2}+p.$ Let us assume that $p$ divides $n$. Then $n=n^{\prime}p$, $0\leq n^{\prime}\leq 2$. Therefore, the lattice point above is $\frac{m}{p}+n^{\prime}\rho+\frac{o}{p}\rho^{2}=\frac{m-ko}{p}+\Big{(}n^{\prime}-\frac{lo}{p}\Big{)}\rho+o\frac{k+l\rho+\rho^{2}}{p}.$ Since $k$ and $l$ are not divisible by $p$, this belongs to $\mathcal{O}_{K}$ if and only if $p\mid m$ and $p\mid o$. From this, we obtain the candidate $1+\rho+\rho^{2}$; the other choices give sums of totally positive elements. Now assume that $p$ does not divide $n$, so that $n\in\\{1,\dots,p-1\\}\cup\\{p+1,\dots,2p-1\\}$. Then, we have $\frac{m}{p}+\frac{n}{p}\rho+\frac{o}{p}\rho^{2}=\frac{m-ko}{p}+\frac{n-lo}{p}\rho+o\frac{k+l\rho+\rho^{2}}{p},$ which belongs to $\mathcal{O}_{K}$ if and only if $m\equiv ko\,(\mathrm{mod}\,p)$ and $n\equiv lo\,(\mathrm{mod}\,p)$. Let $n\in\\{1,\dots,p-1\\}\cup\\{p+1,\dots,2p-1\\}$. We claim that there are unique $\frac{n}{2}\leq m,o\leq\frac{n}{2}+p$ such that the congruences above are satisfied. Since $K$ has integral basis $B_{p}(k,l)$, Corollary 4.3 gives that $2k-l\equiv-2\,(\mathrm{mod}\,p)$. Now, since the interval depending on $n$ has length $p$, the only exception to uniqueness would be that $m$ or $o\equiv\frac{n}{2}\,(\mathrm{mod}\,p)$. If $m\equiv\frac{n}{2}\,(\mathrm{mod}\,p)$, since also $n\equiv lo\,(\mathrm{mod}\,p)$, we obtain that $2ko\equiv lo\,(\mathrm{mod}\,p)$. If it was $o\equiv 0\,(\mathrm{mod}\,p)$, then $p$ divides $n$, which is against the assumption on $n$. Then $o\not\equiv 0\,(\mathrm{mod}\,p)$ and $2k\equiv l\,(\mathrm{mod}\,p)$, which contradicts that $2k-l\equiv-2\,(\mathrm{mod}\,p)$. If $o\equiv\frac{n}{2}\,(\mathrm{mod}\,p)$, the congruence $n\equiv lo\,(\mathrm{mod}\,p)$ gives that $l\equiv 2\,(\mathrm{mod}\,p)$. Now, since $2k-l\equiv-2\,(\mathrm{mod}\,p)$, we have that $k$ is divisible by $p$, which is again a contradiction. This proves the claim. Therefore, we have $2p-2$ lattice points arising from these choices. ###### Proposition 4.10. Let $K$ be a simplest cubic field with integral basis of the form (2). Then the lattice points in the first parallelepiped which are algebraic integers are $\frac{1}{p}(m+n\rho+o\rho^{2}),$ where, for each $1\leq n\leq 2p-1$, $m$ and $o$ are the unique integer numbers in the interval $[\frac{n}{2},\frac{n}{2}+p]$ satisfying the congruences $m\equiv ko\,(\mathrm{mod}\,p)$ and $n\equiv lo\,(\mathrm{mod}\,p)$. ## 5 The family with $p=3$ From now on, we focus on the simplest cubic fields $K$ with integral basis $B_{p}$ for $p=3$ and $(k,l)=(1,1)$, that is, $B_{3}=\Big{\\{}1,\rho,\frac{1+\rho+\rho^{2}}{3}\Big{\\}}.$ We first describe them in terms of the parameter $a$. ###### Proposition 5.1. The simplest cubic fields $K$ with integral basis $\Big{\\{}1,\rho,\frac{1+\rho+\rho^{2}}{3}\Big{\\}}$ are the ones for which $a\equiv 3,21\,(\mathrm{mod}\,27)$, $a>12$ and $\frac{\Delta}{27}$ is square- free. ###### Proof. Assume that $K$ has an integral basis as in the statement. Then $\mathrm{disc}(K)=\mathrm{disc}(1,\rho,\frac{1+\rho+\rho^{2}}{3})=\Big{(}\frac{\Delta}{3}\Big{)}^{2}$, so $\Delta=3\mathfrak{c}$ and $K$ is not monogenic. In particular, $3$ divides $\Delta=a^{2}+3a+9$, so $3\mid a$ and $9\mid\Delta$. Thus $3\mid\mathfrak{c}$, so Remark 3.5 gives that $9\mid\mathfrak{c}$, whence $27\mid\Delta$. Since $p^{2}$ does not divide $\mathfrak{c}$ for every prime $p\neq 3$, we obtain that $\frac{\Delta}{27}$ is square-free. On the other hand, by Theorem 3.1 (3), the non-monogenity of $K$ gives that $a\equiv 3$ or $21$ $(\mathrm{mod}\,27)$. Now, since $K$ is not monogenic, $K$ is defined by a single value of $a$. That is, $a\notin\\{-1,0,1,2,3,5,12,54,66,1259,2389\\}$. But in this list no element greater than $12$ is congruent to $3$ or $21$ mod $27$, so $a>12$. Conversely, let $K$ be a simplest cubic field defined by $a$ such that $a\equiv 3,21\,(\mathrm{mod}\,27)$ (so that $\Delta$ is divisible by $27$), $a>12$ and $\frac{\Delta}{27}$ is square-free. Then, we can write $\Delta=27d$ with $d$ square-free and from Proposition 3.4 we see that $\mathfrak{c}=9d$, so $\delta=3$. In particular, $K$ is not monogenic. On the other hand, $\frac{1+\rho+\rho^{2}}{3}$ is an algebraic integer because the coefficients of its minimal polynomial over $\mathbb{Q}$ are integers. Now, $\mathrm{disc}(1,\rho,\frac{1+\rho+\rho^{2}}{3})=\Big{(}\frac{\Delta}{3}\Big{)}^{2}=\mathfrak{c}^{2}$, so $K$ has integral basis $\Big{\\{}1,\rho,\frac{1+\rho+\rho^{2}}{3}\Big{\\}}$. ∎ ###### Remark 5.2. In the first part of the previous proof, we only have used that $\Delta=3\mathfrak{c}$. We deduce that the only simplest cubic fields with $[\mathcal{O}_{K}:\mathbb{Z}[\rho]]=3$ are the ones with integral basis $B_{3}$. ###### Remark 5.3. Not all simplest cubic fields with $a\equiv 3,21\,(\mathrm{mod}\,27)$ satisfy that $\frac{\Delta}{27}$ is square-free. For instance, if $a=678\equiv 3\,(\mathrm{mod}\,27)$, then $v_{7}(\Delta)=2$. In that case, the integral basis of $K$ is $\Big{\\{}1,\rho,\frac{4+10\rho+\rho^{2}}{21}\Big{\\}}.$ Let us write $a=27t+r$, where $r\in\\{3,21\\}$. Then, $\frac{\Delta}{27}=\begin{cases}27t^{2}+9t+1&\hbox{ if }a\equiv 3\,(\mathrm{mod}\,27),\\\ 27t^{2}+45t+19&\hbox{ if }a\equiv 21\,(\mathrm{mod}\,27).\end{cases}$ These are quadratic polynomials in $t$, so they take infinitely many square- free values. Then, we have infinitely many simplest cubic fields within this family. Our family of simplest cubic fields is just the family considered by Cánovas Orvay in [4, Theorem 3]. ###### Lemma 5.4. The simplest cubic fields $K$ with integral basis $B_{3}(1,1)$ are just the fields of the form $K=\mathbb{Q}(\theta)$ with $\theta^{3}-p\theta+pq=0$, where $p=9d$ with $d\in\mathbb{Z}$ square-free and $q\in\mathbb{Z}$ is such that $q>2$, $3\nmid q$ and $4p-27q^{2}=9$. ###### Proof. Let $K=\mathbb{Q}(\rho)$ be a simplest cubic field as in the statement. Setting $\theta=-\rho+\frac{a}{3}$, we have $K=\mathbb{Q}(\theta)$. Let $p=\frac{a^{2}+3a+9}{3}$ and $q=\frac{2a+3}{9}$. Since $\Delta=27d$, we have $\begin{split}\theta^{3}-p\theta+pq&=-\rho^{3}+\rho^{2}a-\rho\frac{a^{2}}{3}+\frac{a^{3}}{27}+\Big{(}\frac{a^{2}}{3}+a+3\Big{)}\rho-\frac{(a^{2}+3a+9)(a-3)}{27}\\\ &=-\rho^{3}+a\rho^{2}+(a+3)\rho+\frac{a^{3}}{27}-\frac{a^{3}-27}{27}\\\ &=-\rho^{3}+a\rho^{2}+(a+3)\rho+1=0,\end{split}$ proving that $\theta$ has minimal polynomial $x^{3}-px+pq$ over $\mathbb{Q}$. Moreover, we have that $4p-27q^{2}=4\Big{(}\frac{a^{2}}{3}+a+3\Big{)}-\frac{1}{3}\Big{(}4a^{2}+12a+9\Big{)}=9.$ Conversely, let $p=9d$ with $d\in\mathbb{Z}$ square-free, $q\in\mathbb{Z}$ with $q>2$ and $3\nmid q$ and $4p-27q^{2}=9$. Let $K=\mathbb{Q}(\theta)$ with $\theta^{3}-p\theta+pq=0$, and define $a=\frac{9q-3}{2}$. From the equality with $p$ and $q$ we deduce that $q\equiv 1$ or $5$ $(\mathrm{mod}\,6)$, so $a\in\mathbb{Z}$ and $a\equiv 3$ or $21$ $(\mathrm{mod}\,27)$. Let us call $\rho=-\theta+\frac{a}{3}$. Going backwards in the chain of equalities above, we find that $\rho$ has irreducible polynomial $x^{3}-ax^{2}-(a+3)x-1$ over $\mathbb{Q}$. ∎ ###### Corollary 5.5. Let $K=\mathbb{Q}(\rho)$ be a simplest cubic field with integral basis $B_{3}(1,1)$. Then $\\{\rho,\rho^{\prime}\\}$ is a system of fundamental units of $K$. ###### Proof. Let $\theta_{1}=4\frac{p}{3}-3q\theta-2\theta^{2}$, $\sigma=\frac{\theta_{1}}{3}$, $\tau=\frac{1}{q}\Big{(}\frac{q-1}{2}+\frac{q-1}{2}\sigma+\sigma^{2}\Big{)}$ and $\mu=\sigma+\tau$. In [4, Theorem 3 (iii)], it is stated that $\\{\mu,\mu^{\prime}\\}$ is a system of fundamental units of $K$. Now, we may check directly that $\mu=\rho^{\prime}+1$. Since Galois automorphisms preserve systems of fundamental units, we deduce that $\\{\rho+1,\rho^{\prime}+1\\}$ is a system of fundamental units. Finally, since $\rho+1=-\rho\rho^{\prime}$ and $\rho^{\prime}+1=-\frac{1}{\rho}$, $\\{\rho,\rho^{\prime}\\}$ is a system of fundamental units of $K$. ∎ ### 5.1 Algebraic integers in parallelepipeds Following Proposition 4.10, in the first parallelepiped we have the lattice points $\frac{1}{3}(m+n\rho+o\rho^{2}),\quad 1\leq n\leq 5,\quad\frac{n}{2}\leq m,o\leq\frac{n}{2}+3.$ In this case, the congruences therein become $m\equiv n\equiv o\,(\mathrm{mod}\,3)$, and since $\frac{n}{2}\leq m,n,o\leq\frac{n}{2}+3$, we have $m=n=o$. Then we obtain the points $n\frac{1+\rho+\rho^{2}}{3}$ for $1\leq n\leq 5$. The choices $n>1$ give sums of the choice for $n=1$, so the former gives the unique candidate of this parallelepiped $\frac{1+\rho+\rho^{2}}{3}.$ For the second parallelepiped, we change the strategy and work with the integral basis of a simplest cubic field in this family $\Big{\\{}1,\rho,\frac{1+\rho+\rho^{2}}{3}\Big{\\}}.$ We just rewrite the nodes of the second parallelepiped with respect to this integral basis. Call $g_{1}=1$, $g_{2}=\rho$, $g_{3}=\frac{1+\rho+\rho^{2}}{3}$. Then, the second parallelepiped becomes $\mathcal{D}(g_{1},-g_{1}-g_{2}+3g_{3},-(2a+3)g_{1}-(a^{2}+4a+5)g_{2}+3(a+2)g_{3}),$ and we want to determine its lattice points. That is, we look for $t_{1},t_{2},t_{3}\in[0,1]$ and integer numbers $m,n,o\in\mathbb{Z}$ such that $t_{1}g_{1}+t_{2}(-g_{1}-g_{2}+3g_{3})+t_{3}(-(2a+3)g_{1}-(a^{2}+4a+5)g_{2}+3(a+2)g_{3})=mg_{1}+ng_{2}+og_{3}.$ This is equivalent to the following system of equations $\displaystyle t_{1}-t_{2}-(2a+3)t_{3}$ $\displaystyle=m,$ $\displaystyle- t_{2}-(a^{2}+4a+5)t_{3}$ $\displaystyle=n,$ (4) $\displaystyle 3t_{2}+3(a+2)t_{3}$ $\displaystyle=o.$ From the second and the third equation we obtain $-3(a^{2}+3a+3)t_{3}=o+3n$. As in the monogenic case, the choices $t_{3}=0$ or $t_{3}=3(a^{2}+3a+3)$ give either totally positive units or their sums. Therefore, $t_{3}=\frac{u}{3(a^{2}+3a+3)},\quad 1\leq u\leq 3(a^{2}+3a+3)-1.$ (5) From the first and the second equation we obtain that $t_{1}+(a^{2}+2a+2)t_{3}=m-n$, that is, $t_{1}=m-n-u\Big{(}\frac{1}{3}-\frac{a+1}{3(a^{2}+3a+3)}\Big{)}.$ Write $u=3w+s$, where $0\leq w\leq a^{2}+3a+2$, $0\leq s\leq 2$ and $s>0$ if $w=0$. Then $t_{1}=m-n-w+\frac{w(a+1)}{a^{2}+3a+3}-s\Big{(}\frac{1}{3}-\frac{a+1}{3(a^{2}+3a+3)}\Big{)}.$ Write $w=v(a+2)+r$, where $0\leq v,r\leq a+1$ and $r=0$ if $v=a+1$. Then $t_{1}=m-n-w+v+\frac{r(a+1)-v}{a^{2}+3a+3}-s\frac{a^{2}+2a+2}{3(a^{2}+3a+3)}.$ Let us define $e_{1}=\Big{\lceil}-\Big{(}\frac{r(a+1)-v}{a^{2}+3a+3}-s\frac{a^{2}+2a+2}{3(a^{2}+3a+3)}\Big{)}\Big{\rceil}.$ Hence, $t_{1}\in[0,1]$ if and only if $m-n=w-v+e_{1}$. We can easily write $e_{1}=\Big{\lceil}\frac{s(a^{2}+2a+2)-3(r(a+1)-v)}{3(a^{2}+3a+3)}\Big{\rceil}.$ Then, we have that * • For $s=0$, $e_{1}=1$ if $r=0$, and $e_{1}=0$ otherwise. * • For $s=1$, $e_{1}=1$ if $3(r(a+1)-v)\leq a^{2}+2a$, and $e_{1}=0$ otherwise. * • For $s=2$, $e_{1}=1$ if $3(r(a+1)-v)\leq 2a^{2}+4a+3=2a(a+2)+3$, and $e_{1}=0$ otherwise. ###### Remark 5.6. In the case $s=0$, $e_{1}=0$ would be also possible if $v=0$ and $r=0$, but this corresponds to the zero point, which is not totally positive. On the other hand, replacing the expression (5) of $t_{3}$ in the second equation of (4) gives $t_{2}=-n-u\Big{(}\frac{1}{3}+\frac{a+2}{3(a^{2}+3a+3)}\Big{)}.$ Now, replacing $u=3w+s$, $t_{2}=-n-w-\frac{w(a+2)}{a^{2}+3a+3}-s\Big{(}\frac{1}{3}+\frac{a+2}{3(a^{2}+3a+3)}\Big{)}.$ Write $w=t(a+1)+l$, with $0\leq t\leq a+2$, $0\leq l\leq a$ and $l=0$ if $t=a+2$. Now, $t_{2}=-n-w-t+\frac{t-l(a+2)}{a^{2}+3a+3}-s\frac{a^{2}+4a+5}{3(a^{2}+3a+3)}.$ Let us call $e_{2}=\Big{\lceil}-\Big{(}\frac{t-l(a+2)}{a^{2}+3a+3}-s\frac{a^{2}+4a+5}{3(a^{2}+3a+3)}\Big{)}\Big{\rceil}.$ Then $t_{2}\in[0,1]$ if and only if $n=-w-t-e_{2}$. An easy calculation shows that $e_{2}=\Big{\lceil}\frac{s(a^{2}+4a+5)+3(l(a+2)-t)}{3(a^{2}+3a+3)}\Big{\rceil}.$ Thus, in this case: * • For $s=0$, $e_{2}=0$ if $l=0$, and $e_{2}=1$ otherwise. * • For $s=1$, $e_{2}=1$ if $3(l(a+2)-t)\leq 2a^{2}+5a+3=(2a+3)(a+1)$, and $e_{2}=2$ otherwise. * • For $s=2$, $e_{2}=1$ if $3(l(a+2)-t)\leq a^{2}+a-3=a(a+1)-3$, and $e_{2}=2$ otherwise. ###### Remark 5.7. In the case $s=0$, $e_{2}=0$ would be also possible if $l=1$ and $t=a+2$, but we have excluded that combination by taking $l=0$ if $t=a+2$. Moreover, since $-u=o+3n$, we have that $o=-u+3(w+t+e_{2})$. This gives the lattice points $-(v+t-e_{1}+e_{2})g_{1}-(w+t+e_{2})g_{2}+(-u+3(w+t+e_{2}))g_{3}.$ Since $u=3w+s$, we may rewrite them as $-(v+t-e_{1}+e_{2})g_{1}-(w+t+e_{2})g_{2}+(-s+3(t+e_{2}))g_{3}.$ With respect to the power (non-integral) basis, these become $-v-\frac{s}{3}+e_{1}-\Big{(}w+\frac{s}{3}\Big{)}\rho+\Big{(}t-\frac{s}{3}+e_{2}\Big{)}\rho^{2}.$ ###### Proposition 5.8. Let $K$ be a simplest cubic field with $a\equiv 3$ or $21\,(\mathrm{mod}\,27)$, $a>12$ and $\frac{\Delta}{27}$ square-free. The lattice points of $K$ (and therefore non-unit candidates to be indecomposables of $K$; up to multiplication by totally positive units) are the point $\frac{1+\rho+\rho^{2}}{3}$ from the first parallelepiped, and the points $-(v+t-e_{1}+e_{2})g_{1}-(w+t+e_{2})g_{2}+(-s+3(t+e_{2}))g_{3}$ $=-(v-e_{1})-w\rho+(t+e_{2})\rho^{2}-sg_{3}$ from the second one, where: * • $0\leq w\leq a^{2}+3a+2$, * • $0\leq s\leq 2$, and $s>0$ if $w=0$, * • $v$ and $r$ are defined from $w=v(a+2)+r$, $0\leq v,r\leq a+1$ and $r=0$ if $v=a+1$, * • $t$ and $l$ are defined from $w=t(a+1)+l$, $0\leq t\leq a+2$, $0\leq l\leq a$, and $l=0$ if $t=a+2$, * • $e_{1}=\Big{\lceil}\frac{s(a^{2}+2a+2)+3(v-r(a+1))}{3(a^{2}+3a+3)}\Big{\rceil}$, * • $e_{2}=\Big{\lceil}\frac{s(a^{2}+4a+5)+3(l(a+2)-t)}{3(a^{2}+3a+3)}\Big{\rceil}$. Note that each lattice point $\alpha$ in the second parallelepiped is completely determined from $s$ and the integers $v$ and $r$ as in Proposition 5.8 such that $\alpha=-(v-e_{1})-(v(a+2)+r)\rho+(t+e_{2})\rho^{2}-sg_{3}$, because such a choice determines completely the pair $(t,l)$. From now on, we will denote $\alpha=\alpha_{s}(v,r)$ or, for $w=v(a+2)+r$, $\alpha=\alpha_{s}(w)$. ### 5.2 Codifferent In order to decide whether a given candidate is indecomposable or not, the codifferent appears as a very useful tool. Recall that $\mathcal{O}_{K}^{\vee}=\\{\delta\in K\,|\,\mathrm{Tr}(\alpha\delta)\in\mathbb{Z}\,\forall\alpha\in\mathcal{O}_{K}\\}.$ To determine an ${\mathbb{Z}}$-basis of $\mathcal{O}_{K}^{\vee}$, we use [26, Proposition 4.14]. Recall the notation $g_{1}=1$, $g_{2}=\rho$, $g_{3}=\frac{1+\rho+\rho^{2}}{3}$. Since $\\{g_{i}\\}_{i=1}^{3}$ is an integral basis of $K$, a collection $\\{\varphi_{i}\\}_{i=1}^{3}$ of elements in $K$ satisfying $\mathrm{Tr}(g_{i}\varphi_{j})=\delta_{ij},\quad i,j=1,2,3,$ is an ${\mathbb{Z}}$-basis of $\mathcal{O}_{K}^{\vee}$. At the same time, we may write $\varphi_{j}=\sum_{k=1}^{3}a_{jk}g_{k}$, $a_{jk}\in\mathbb{Q}$. Then $\delta_{ij}=\mathrm{Tr}(g_{i}\varphi_{j})=\sum_{k=1}^{3}a_{jk}\mathrm{Tr}(g_{i}g_{k}).$ We can express these equalities as $\begin{pmatrix}\mathrm{Tr}(g_{1}^{2})&\mathrm{Tr}(g_{1}g_{2})&\mathrm{Tr}(g_{1}g_{3})\\\ \mathrm{Tr}(g_{1}g_{2})&\mathrm{Tr}(g_{2}^{2})&\mathrm{Tr}(g_{2}g_{3})\\\ \mathrm{Tr}(g_{1}g_{3})&\mathrm{Tr}(g_{2}g_{3})&\mathrm{Tr}(g_{3}^{2})\end{pmatrix}\begin{pmatrix}a_{11}&a_{21}&a_{31}\\\ a_{12}&a_{22}&a_{32}\\\ a_{13}&a_{23}&a_{33}\end{pmatrix}=\begin{pmatrix}1&0&0\\\ 0&1&0\\\ 0&0&1\end{pmatrix}.$ Let $M=(\mathrm{Tr}(g_{i}g_{k}))_{i,k=1}^{3}$. Then the coordinates of $\varphi_{j}$ with respect to the basis $\\{g_{i}\\}_{i=1}^{3}$ is given by the $j$-th column of the matrix $M^{-1}$. Carrying out the product traces, we have $M=\begin{pmatrix}3&a&\frac{a^{2}+3a+9}{3}\\\ a&a^{2}+2a+6&\frac{(a+1)(a^{2}+3a+9)}{3}\\\ \frac{a^{2}+3a+9}{3}&\frac{(a+1)(a^{2}+3a+9)}{3}&\frac{(a^{2}+3a+5)(a^{2}+3a+9)}{9}\end{pmatrix}.$ From this, we find $M^{-1}=\frac{1}{a^{2}+3a+9}\begin{pmatrix}a^{2}+7a+21&a^{2}+7a+9&-3(a+6)\\\ a^{2}+7a+9&2(a^{2}+3a+3)&-3(2a+3)\\\ -3(a+6)&-3(2a+3)&18\end{pmatrix}.$ Then, we conclude that $\begin{split}\varphi_{1}&=\frac{1}{a^{2}+3a+9}\Big{(}a^{2}+7a+21+(a^{2}+7a+9)\rho-3(a+6)\frac{1+\rho+\rho^{2}}{3}\Big{)},\\\ \varphi_{2}&=\frac{1}{a^{2}+3a+9}\Big{(}a^{2}+7a+9+2(a^{2}+3a+3)\rho-3(2a+3)\frac{1+\rho+\rho^{2}}{3}\Big{)},\\\ \varphi_{3}&=\frac{1}{a^{2}+3a+9}\Big{(}-3(a+6)-3(2a+3)\rho+18\frac{1+\rho+\rho^{2}}{3}\Big{)}.\end{split}$ When computing codifferent traces of specific elements, we have the following important remark: If $\alpha=\sum_{i=1}^{3}a_{i}g_{i}\in\mathcal{O}_{K}$ and $\varphi=\sum_{j=1}^{3}b_{j}\varphi_{j}\in\mathcal{O}_{K}^{\vee}$, then $\mathrm{Tr}(\varphi\alpha)=\sum_{i,j=1}^{3}a_{i}b_{j}\mathrm{Tr}(g_{i}\varphi_{j})=\sum_{i,j=1}^{3}a_{i}b_{j}\delta_{ij}=\sum_{i=1}^{3}a_{i}b_{i}.$ ## 6 Indecomposables In this part we will prove Theorem 1.2. In the previous sections, we found all the lattice points in the two parallelepipeds, and now we determine which ones are indecomposable, either using the codifferent or other means. We identify the lattice points for the case $w=0$. In this case, $s=0$ would give the zero point, which is excluded. Otherwise, since we have $v=t=r=l=0$, then $e_{1}=e_{2}=1$. We obtain the points $\alpha_{2}(0,0)=-g_{2}+g_{3},\quad\alpha_{1}(0,0)=-g_{2}+2g_{3}.$ (6) The case $w=(a+1)(a+2)$ is easy as well. This value of $w$ corresponds to $t=a+2$, $v=a+1$ and $r=l=0$. For $s=0$, we see that $e_{1}=1$ and $e_{2}=0$. We obtain the point $\alpha_{0}(a+1,0)=-(2a+2)-(a+2)^{2}g_{2}+3(a+2)g_{3}=-a-(a+1)(a+2)\rho+(a+2)\rho^{2}.$ For $s\neq 0$, we have $e_{1}=e_{2}=1$, giving the point $\begin{split}\alpha_{s}(a+1,0)&=-(2a+3)-((a+2)^{2}+1)g_{2}+(-s+3(a+3))g_{3}\\\ &=-a-(a+1)(a+2)\rho+(a+3)\rho^{2}-sg_{3}.\end{split}$ (7) Fixed a value of $s$, we will distribute the lattice points $\alpha_{s}(v,r)$ in four cases depending on the value of $w=v(a+2)+r$. * (a) $x(a+2)<w<(x+1)(a+1)$, $0\leq x\leq a-1$, which corresponds to $1\leq r\leq a-v$. In this case, $v=t=x$ and $l=r+v$. * (b) $w=(x+1)(a+1)$, $0\leq x\leq a$. In this case, $v=x$, $t=v+1$, $r=a-v+1$ and $l=0$. * (c) $(x+1)(a+1)<w<(x+1)(a+2)$, $1\leq x\leq a$, which corresponds to $v=x$, $a-v+2\leq r\leq a+1$, $t=v+1$ and $l=r-(a-v+1)$. * (d) $w=(x+1)(a+2)$, $0\leq x\leq a-1$. Then $v=x+1$, $r=0$ and $t=l=v$. It may be checked easily that these cases, together with $\alpha_{s}(0,0)$, $s\in\\{1,2\\}$ and $\alpha_{s}(a+1,0)$, $s\in\\{0,1,2\\}$, cover exactly once all points in Proposition 5.8. ### 6.1 The case $s=0$ If in Proposition 5.8 we take $s=0$, this corresponds to the intersection of the second parallelepiped with $\mathbb{Z}[\rho]$. Then, we are left just with the lattice points in $\mathbb{Z}[\rho]$, which have been already determined in [16]. In other words, the points $\alpha_{s}(v,r)$ with $s=0$ are exactly the ones that have integer coordinates with respect to the power basis $\\{1,\rho,\rho^{2}\\}$. With our notation, these are $\begin{array}[]{ll}-v-w\rho+(v+1)\rho^{2},&\hbox{if }v(a+2)<w\leq(v+1)(a+1)\hbox{ and }0\leq v\leq a,\\\ -v-w\rho+(v+2)\rho^{2},&\hbox{if }(v+1)(a+1)<w<(v+1)(a+2)\hbox{ and }0\leq v\leq a,\\\ -(v-1)-w\rho+(v+1)\rho^{2},&\hbox{if }w=v(a+2)\hbox{ and }0\leq v\leq a+1.\\\ \end{array}$ While the first line corresponds to situations (a) and (b), the second (resp. third) one corresponds to situation (c) (resp. (d)). ### 6.2 The case $s=1$ Assume that $s=1$. Recall that $w=v(a+2)+r=t(a+1)+l$. Now, we have that $e_{1}=\begin{cases}1&\hbox{if }3(r(a+1)-v)\leq a(a+2),\\\ 0&\hbox{otherwise},\end{cases}$ $e_{2}=\begin{cases}1&\hbox{if }3(l(a+2)-t)\leq(2a+3)(a+1),\\\ 0&\hbox{otherwise}.\end{cases}$ #### Situation (a) We first take $x(a+2)<w<(x+1)(a+1)$, $0\leq x\leq a-1$. We know that $x=v$. Then: $\begin{split}&3(r(a+1)-v)\leq a(a+2)\\\ \Longleftrightarrow\;&r(a+1)\leq\frac{a}{3}(a+2)+v\\\ \Longleftrightarrow\;&r\leq\Big{\lfloor}\frac{a(a+2)}{3(a+1)}+\frac{v}{a+1}\Big{\rfloor}=\Big{\lfloor}\frac{a}{3}\Big{(}1+\frac{1}{a+1}\Big{)}+\frac{v}{a+1}\Big{\rfloor}=\frac{a}{3}+\Big{\lfloor}\frac{3v+a}{3(a+1)}\Big{\rfloor}.\end{split}$ Now, $3v+a<3(a+1)$ if and only if $v<\frac{2a}{3}+1$. We deduce that $e_{1}=\begin{cases}1&\hbox{if }0\leq v\leq\frac{2a}{3}\hbox{ and }1\leq r\leq\frac{a}{3}\\\ &\hbox{or }\frac{2a}{3}+1\leq v\leq a-1,\\\ 0&\hbox{if }0\leq v\leq\frac{2a}{3}-1\hbox{ and }\frac{a}{3}+1\leq r\leq a-v.\end{cases}$ On the other hand, $\begin{split}&3(l(a+2)-t)=3(v(a+1)+r(a+2))\leq(2a+3)(a+1)\\\ \Longleftrightarrow\;&r(a+2)\leq\Big{(}\frac{2a}{3}+1-v\Big{)}(a+1)\\\ \Longleftrightarrow\;&r\leq\Big{\lfloor}\Big{(}\frac{2a}{3}+1-v\Big{)}\frac{a+1}{a+2}\Big{\rfloor}=\frac{2a}{3}+1-v+\Big{\lfloor}\frac{1}{a+2}\Big{(}v-\Big{(}\frac{2a}{3}+1\Big{)}\Big{)}\Big{\rfloor}.\end{split}$ Note that the right-side member needs to be positive, i.e. $v<\frac{2a}{3}+1$. Under this assumption, we have that $e_{2}=1$ if and only if $1\leq r\leq\frac{2a}{3}-v$ (since $\frac{2a}{3}+1-v<a+2$). We deduce that $\begin{cases}e_{2}=1&\hbox{if }0\leq v\leq\frac{2a}{3}\hbox{ and }1\leq r\leq\frac{2a}{3}-v,\\\ e_{2}=2&\hbox{if }0\leq v\leq\frac{2a}{3}\hbox{ and }\frac{2a}{3}+1-v\leq r\leq a-v\\\ &\hbox{or }\frac{2a}{3}+1\leq v\leq a.\\\ \end{cases}$ Joining all the information together, we have that: * • $e_{1}=e_{2}=1$ exactly in the cases: * – $0\leq v\leq\frac{a}{3}\hbox{ and }1\leq r\leq\frac{a}{3}$, * – $\frac{a}{3}+1\leq v\leq\frac{2a}{3}-1\hbox{ and }1\leq r\leq\frac{2a}{3}-v$. These correspond to the points $-2vg_{1}-(w+v+1)g_{2}+(3v+2)g_{3}=-(v-1)-w\rho+(v+1)\rho^{2}-g_{3}.$ * • $e_{1}=0$ and $e_{2}=1$ if and only if $0\leq v\leq\frac{a}{3}-1$ and $\frac{a}{3}+1\leq r\leq\frac{2a}{3}-v$. These correspond to the points $-(2v+1)g_{1}-(w+v+1)+(3v+2)g_{3}=-v-w\rho+(v+1)\rho^{2}-g_{3}.$ * • $e_{1}=1$ and $e_{2}=2$ exactly in the cases: * – $\frac{a}{3}+1\leq v\leq\frac{2a}{3}$ and $\frac{2a}{3}-v+1\leq r\leq\frac{a}{3}$, * – $\frac{2a}{3}+1\leq v\leq a-1$. These correspond to the points $-(2v+1)g_{1}-(w+v+2)g_{2}+(3v+5)g_{3}=-(v-1)-w\rho+(v+2)\rho^{2}-g_{3}.$ * • $e_{1}=0$ and $e_{2}=2$ exactly in the cases: * – $0\leq v\leq\frac{a}{3}$ and $\frac{2a}{3}-v+1\leq r\leq a-v$, * – $\frac{a}{3}+1\leq v\leq\frac{2a}{3}-1$ and $\frac{a}{3}+1\leq r\leq a-v$. These correspond to the points $-(2v+2)g_{1}-(w+v+2)g_{2}+(3v+5)g_{3}=-v-w\rho+(v+2)\rho^{2}-g_{3}.$ #### Situation (b) Suppose that $w=(x+1)(a+1)$ with $0\leq x\leq a$ and $x=v$. Now, we have $\begin{split}&3(r(a+1)-v)=3((a-v)(a+2)+1)\leq a(a+2)\\\ \Longleftrightarrow\;&(a-v)(a+2)+1\leq\frac{a}{3}(a+2)\\\ \Longleftrightarrow\;&a-v+\frac{1}{a+2}\leq\frac{a}{3}\\\ \Longleftrightarrow\;&a-v+1\leq\frac{a}{3}\\\ \Longleftrightarrow\;&v\geq\frac{2a}{3}+1.\end{split}$ Therefore, $e_{1}=1$ if $v\geq\frac{2a}{3}+1$, and $e_{1}=0$ otherwise. On the other hand, since $l=0$ we always have that $e_{2}=1$. Then: * • $e_{1}=1$ and $e_{2}=1$ if and only if $\frac{2a}{3}+1\leq v\leq a$. This corresponds to the points $-(2v+1)g_{1}-(w+v+2)g_{2}+(3v+5)g_{3}=-(v-1)-w\rho+(v+2)\rho^{2}-g_{3}.$ * • $e_{1}=0$ and $e_{2}=1$ if and only if $0\leq v\leq\frac{2a}{3}$. This corresponds to the points $-(2v+2)g_{1}-(w+v+2)g_{2}+(3v+5)g_{3}=-v-w\rho+(v+2)\rho^{2}-g_{3}.$ #### Situation (c) Next, assume that $(x+1)(a+1)+1\leq w<(x+1)(a+2)$ with $1\leq x\leq a$ and $x=v$. Arguing as in the situation (a), we see that $3(r(a+1)-v)\leq a(a+2)$ if and only if $r\leq\frac{a}{3}+\Big{\lfloor}\frac{3v+a}{3(a+1)}\Big{\rfloor}$. Since $\frac{a}{3}+1\geq a-v+2$ if and only if $v\geq\frac{2a}{3}+1$, we deduce that $e_{1}=\begin{cases}1&\hbox{if }\frac{2a}{3}+1\leq v\leq a\hbox{ and }a-v+2\leq r\leq\frac{a}{3}+1,\\\ 0&\hbox{if }\frac{2a}{3}+1\leq v\leq a\hbox{ and }\frac{a}{3}+2\leq r\leq a+1\\\ &\hbox{or }1\leq v\leq\frac{2a}{3}.\\\ \end{cases}$ On the other hand, $\begin{split}&3(l(a+2)-t)=3(l(a+2)-(v+1))\leq(2a+3)(a+1)\\\ \Longleftrightarrow\;&l(a+2)\leq\Big{(}\frac{2a}{3}+1\Big{)}(a+1)+v+1\\\ \Longleftrightarrow\;&l\leq\Big{(}\frac{2a}{3}+1\Big{)}\Big{(}1-\frac{1}{a+2}\Big{)}+\frac{v+1}{a+2}=\frac{2a}{3}+1+\frac{v-\frac{2a}{3}}{a+2}\\\ \Longleftrightarrow\;&r\leq\frac{5a}{3}-v+2+\frac{v-\frac{2a}{3}}{a+2},\end{split}$ as $l=r-(a-v+1)$. Now, we see that $\Big{\lfloor}\frac{v-\frac{2a}{3}}{a+2}\Big{\rfloor}=\begin{cases}-1&\hbox{if }v<\frac{2a}{3},\\\ 0&\hbox{if }v\geq\frac{2a}{3}.\end{cases}$ In the first case, we have that $e_{2}=1$ when $a-v+2\leq r\leq\mathrm{min}(a+1,\frac{5a}{3}-v+1)=a+1$, so there are no further restrictions on $r$. As for the second one, we have that $e_{2}=1$ when $a-v+2\leq r\leq\mathrm{min}(a+1,\frac{5a}{3}-v+2)$. This equals $\frac{5a}{3}-v+2$ unless $v=\frac{2a}{3}$, in which case the minimum is $a+1$. We conclude that $e_{2}=\begin{cases}1&\hbox{if }1\leq v\leq\frac{2a}{3},\\\ &\hbox{or }\frac{2a}{3}+1\leq v\leq a\hbox{ and }a-v+2\leq r\leq\frac{5a}{3}-v+2,\\\ 2&\hbox{if }\frac{2a}{3}+1\leq v\leq a\hbox{ and }\frac{5a}{3}-v+3\leq r\leq a+1.\end{cases}$ We note that $e_{1}=1$ implies that $e_{2}=1$, and consequently the case that $e_{1}=1$ and $e_{2}=2$ is not possible. * • $e_{1}=e_{2}=1$ if and only if $\frac{2a}{3}+1\leq v\leq a\hbox{ and }a-v+2\leq r\leq\frac{a}{3}+1$. This corresponds to the points $-(2v+1)g_{1}-(w+v+2)g_{2}+(3v+5)g_{3}=-(v-1)-w\rho+(v+2)\rho^{2}-g_{3}.$ * • $e_{1}=0$ and $e_{2}=1$ exactly in the cases: * – $1\leq v\leq\frac{2a}{3}$. * – $\frac{2a}{3}+1\leq v\leq a$ and $\frac{a}{3}+2\leq r\leq\frac{5a}{3}-v+2$. This corresponds to the points $-(2v+2)g_{1}-(w+v+2)g_{2}+(3v+5)g_{3}=-v-w\rho+(v+2)\rho^{2}-g_{3}.$ * • $e_{1}=0$ and $e_{2}=2$ if and only if $\frac{2a}{3}+2\leq v\leq a$ and $\frac{5a}{3}-v+3\leq r\leq a+1$. This corresponds to the points $-(2v+3)g_{1}-(w+v+3)g_{2}+(3v+8)g_{3}=-v-w\rho+(v+3)\rho^{2}-g_{3}.$ #### Situation (d) Let us take $w=(x+1)(a+2)$ with $0\leq x\leq a-1$. Now, we have $v=x+1$. Since $r=0$, we always have $3(r(a+1)-v)<a(a+2)$, so $e_{1}=1$. Moreover, $3(l(a+2)-t)=3v(a+1)\leq(2a+3)(a+1)$ if and only if $v\leq\frac{2a}{3}+1$. Therefore, $e_{2}=1$ if $v\leq\frac{2a}{3}+1$ and $e_{2}=2$ otherwise. Then: * • $e_{1}=1$ and $e_{2}=1$ if and only if $1\leq v\leq\frac{2a}{3}+1$. Together with the point $-g_{2}+2g_{3}$ from (6) (so that $0\leq v\leq\frac{2a}{3}+1$), this corresponds to the points $-2vg_{1}-(w+v+1)g_{2}+(3v+2)g_{3}=-(v-1)-w\rho+(v+1)\rho^{2}-g_{3}.$ * • $e_{1}=1$ and $e_{2}=2$ if and only if $\frac{2a}{3}+2\leq v\leq a$. Together with the point from (7) with $s=1$ (so that $\frac{2a}{3}+2\leq v\leq a+1$), this corresponds to the points $-(2v+1)g_{1}-(w+v+2)g_{2}+(3v+5)g_{3}=-(v-1)-w\rho+(v+2)\rho^{2}-g_{3}.$ ### 6.3 The case $s=2$ Assume that $s=2$. In this case, $e_{1}=\begin{cases}1&\hbox{if }3(r(a+1)-v)\leq 2a(a+2)+3,\\\ 0&\hbox{otherwise},\end{cases}$ $e_{2}=\begin{cases}1&\hbox{if }3(l(a+2)-t)\leq a(a+1)-3,\\\ 2&\hbox{otherwise}.\end{cases}$ #### Situation (a) We first take $x(a+2)<w<(x+1)(a+1)$, where $x=v$. Then: $\begin{split}&3(r(a+1)-v)\leq 2a(a+2)+3\\\ \Longleftrightarrow\;&r(a+1)\leq\frac{2a}{3}(a+2)+v+1\\\ \Longleftrightarrow\;&r\leq\frac{2a}{3}\Big{(}1+\frac{1}{a+1}\Big{)}+\frac{v+1}{a+1}=\frac{2a}{3}+\frac{v+1+\frac{2a}{3}}{a+1}.\end{split}$ We deduce that $e_{1}=\begin{cases}1&\hbox{if }0\leq v\leq\frac{a}{3}-1\hbox{ and }1\leq r\leq\frac{2a}{3}\\\ &\hbox{or }\frac{a}{3}\leq v\leq a-1,\\\ 0&\hbox{if }0\leq v\leq\frac{a}{3}-1\hbox{ and }\frac{2a}{3}+1\leq r\leq a-v.\end{cases}$ On the other hand, $\begin{split}&3(l(a+2)-t)=3(v(a+1)+r(a+2))\leq a(a+1)-3\\\ \Longleftrightarrow\;&r(a+2)\leq\frac{a}{3}(a+1)-1-v(a+1)=\Big{(}\frac{a}{3}-v\Big{)}(a+1)-1\\\ \Longleftrightarrow;&r\leq\Big{(}\frac{a}{3}-v\Big{)}\Big{(}1-\frac{1}{a+2}\Big{)}-\frac{1}{a+2}=\frac{a}{3}-v+\frac{v-\frac{a}{3}-1}{a+2}.\end{split}$ Then, $e_{2}=\begin{cases}1&\hbox{if }0\leq v\leq\frac{a}{3}-2\hbox{ and }1\leq r\leq\frac{a}{3}-v-1,\\\ 2&\hbox{if }0\leq v\leq\frac{a}{3}-2\hbox{ and }\frac{a}{3}-v\leq r\leq a-v\\\ &\hbox{or }\frac{a}{3}-1\leq v\leq a-1.\end{cases}$ * • $e_{1}=e_{2}=1$ if and only if $0\leq v\leq\frac{a}{3}-2$ and $1\leq r\leq\frac{a}{3}-v-1$. This corresponds to the points $-2vg_{1}-(w+v+1)g_{2}+(3v+1)g_{3}=-(v-1)-w\rho+(v+1)\rho^{2}-2g_{3}.$ * • Since $e_{2}=1$ implies that $e_{1}=1$, the case that $e_{1}=0$ and $e_{2}=1$ is not possible. * • $e_{1}=1$ and $e_{2}=2$ exactly in the cases: * – $0\leq v\leq\frac{a}{3}-1$ and $\frac{a}{3}-v\leq r\leq\frac{2a}{3}$. * – $\frac{a}{3}\leq v\leq a-1$. These correspond to the points $-(2v+1)g_{1}-(w+v+2)g_{2}+(3v+4)g_{3}=-(v-1)-w\rho+(v+2)\rho^{2}-2g_{3}.$ * • $e_{1}=0$ and $e_{2}=2$ if and only if $0\leq v\leq\frac{a}{3}-1$ and $\frac{2a}{3}+1\leq r\leq a-v$. In this case, we have the points $-(2v+2)g_{1}-(w+v+2)g_{2}+(3v+4)g_{3}=-v-w\rho+(v+2)\rho^{2}-2g_{3}.$ #### Situation (b) Suppose that $w=(x+1)(a+1)$ with $0\leq x\leq a$ and $x=v$. Now, we have $\begin{split}&3(r(a+1)-v)=3((a-v)(a+2)+1)\leq 2a(a+2)+3\\\ \Longleftrightarrow\;&a(a+2)-v(a+2)+1\leq\frac{2a}{3}(a+2)+1\\\ \Longleftrightarrow\;&v(a+2)\geq\frac{a}{3}(a+2)\\\ \Longleftrightarrow\;&v\geq\frac{a}{3}.\end{split}$ We deduce that $e_{1}=1$ if $v\geq\frac{a}{3}$ and $e_{1}=0$ otherwise. On the other hand, we always have $e_{2}=1$ because $l=0$. * • $e_{1}=e_{2}=1$ if and only if $\frac{a}{3}\leq v\leq a$, giving the points $-(2v+1)g_{1}-(w+v+2)g_{2}+(3v+4)g_{3}=-(v-1)-w\rho+(v+2)\rho^{2}-2g_{3}.$ * • $e_{1}=0$ and $e_{2}=1$ if and only if $0\leq v\leq\frac{a}{3}-1$, which corresponds to $-(2v+2)g_{1}-(w+v+2)g_{2}+(3v+4)g_{3}=-v-w\rho+(v+2)\rho^{2}-2g_{3}.$ #### Situation (c) Next, assume that $(x+1)(a+1)+1\leq w<(x+1)(a+2)$ with $1\leq x\leq a$ and $x=v$. Arguing as in the situation (a), $3(r(a+1)-v)\leq 2a(a+2)+3$ if and only if $r\leq\frac{2a}{3}+\frac{v+1+\frac{2a}{3}}{a+1}$. Assume that $0\leq v\leq\frac{a}{3}-1$. If it was $e_{1}=1$, we would have that $r\leq\frac{2a}{3}$ but $a-v+2>\frac{2a}{3}$, which is a contradiction. For $v=\frac{a}{3}$, if it was $e_{1}=1$ we would have that $r\leq\frac{2a}{3}+1$, but $a-v+2=\frac{2a}{3}+2$, giving again a contradiction. Thus, when $v\leq\frac{a}{3}$ we always have $e_{1}=0$. Now, suppose that $v\geq\frac{a}{3}+1$. Again, $e_{1}=1$ if and only if $r\leq\frac{2a}{3}+1$, and in this case $a-v+2\geq\frac{2a}{3}+1$. Thus, $e_{1}=\begin{cases}1&\hbox{if }\frac{a}{3}+1\leq v\leq a\hbox{ and }a-v+2\leq r\leq\frac{2a}{3}+1,\\\ 0&\hbox{if }\frac{a}{3}+1\leq v\leq a\hbox{ and }\frac{2a}{3}+2\leq r\leq a+1\\\ &\hbox{or }1\leq v\leq\frac{a}{3}.\end{cases}$ On the other hand, $\begin{split}&3(l(a+2)-t)=3(l(a+2)-(v+1))\leq a(a+1)-3\\\ \Longleftrightarrow\;&l(a+2)\leq\frac{a}{3}(a+1)+v\\\ \Longleftrightarrow\;&l\leq\frac{a}{3}\frac{a+1}{a+2}+\frac{v}{a+2}\\\ \Longleftrightarrow\;&l\leq\frac{a}{3}+\frac{v-\frac{a}{3}}{a+2}\\\ \Longleftrightarrow\;&r\leq\frac{4a}{3}-v+1+\frac{v-\frac{a}{3}}{a+2}.\end{split}$ If $1\leq v\leq\frac{a}{3}-1$, we have that $e_{2}=1$ if and only if $a-v+2\leq r\leq\mathrm{min}(a+1,\frac{4a}{3}-v)=a+1$, so there are no further restrictions on $r$. Otherwise, if $\frac{a}{3}\leq v\leq a$, $e_{2}=1$ if and only if $a-v+2\leq r\leq\min(a+1,\frac{4a}{3}-v+1)=\frac{4a}{3}-v+1$. We conclude that $e_{2}=\begin{cases}1&\hbox{if }1\leq v\leq\frac{a}{3}-1\\\ &\hbox{or }\frac{a}{3}\leq v\leq a\hbox{ and }a-v+2\leq r\leq\frac{4a}{3}-v+1,\\\ 2&\hbox{if }\frac{a}{3}+1\leq v\leq a\hbox{ and }\frac{4a}{3}-v+2\leq r\leq a+1.\end{cases}$ * • $e_{1}=e_{2}=1$ exactly in the cases: * – $\frac{a}{3}+1\leq v\leq\frac{2a}{3}$ and $a-v+2\leq r\leq\frac{2a}{3}+1$, * – $\frac{2a}{3}+1\leq v\leq a$ and $a-v+2\leq r\leq\frac{4a}{3}-v+1$. These correspond to the points $-(2v+1)g_{1}-(w+v+2)g_{2}+(3v+4)g_{3}=-(v-1)-w\rho+(v+2)\rho^{2}-2g_{3}.$ * • $e_{1}=0$ and $e_{2}=1$ exactly in the cases: * – $1\leq v\leq\frac{a}{3}-1$, * – $\frac{a}{3}\leq v\leq\frac{2a}{3}-1$ and $\frac{2a}{3}+2\leq r\leq\frac{4a}{3}-v+1$. These correspond to the points $-(2v+2)g_{1}-(w+v+2)g_{2}+(3v+4)g_{3}=-v-w\rho+(v+2)\rho^{2}-2g_{3}.$ * • $e_{1}=1$ and $e_{2}=2$ if and only if $\frac{2a}{3}+1\leq v\leq a$ and $\frac{4a}{3}-v+2\leq r\leq\frac{2a}{3}+1$. This corresponds to the points $-(2v+2)g_{1}-(w+v+3)g_{2}+(3v+7)g_{3}=-(v-1)-w\rho+(v+3)\rho^{2}-2g_{3}.$ * • $e_{1}=0$ and $e_{2}=2$ exactly in the cases: * – $\frac{a}{3}+1\leq v\leq\frac{2a}{3}$ and $\frac{4a}{3}-v+2\leq r\leq a+1$, * – $\frac{2a}{3}+1\leq v\leq a$ and $\frac{2a}{3}+2\leq r\leq a+1$. These correspond to the points $-(2v+3)g_{1}-(w+v+3)g_{2}+(3v+7)g_{3}=-v-w\rho+(v+3)\rho^{2}-2g_{3}.$ #### Situation (d) Let us take $w=(x+1)(a+2)$ with $0\leq x\leq a-1$. In this case, $x=v+1$. First, we always have $3(r(a+1)-v)<2a(a+2)+3$, so $e_{1}=1$. Moreover, $3(l(a+2)-t)=3v(a+1)\leq a(a+1)-3$ if and only if $v\leq\frac{a}{3}-1$. Therefore, $e_{2}=1$ if $v\leq\frac{a}{3}-1$ and $e_{2}=2$ otherwise. Then: * • $e_{1}=1$ and $e_{2}=1$ if and only if $1\leq v\leq\frac{a}{3}-1$. Together with the point $-g_{2}+g_{3}$ from (6) (so that $0\leq v\leq\frac{a}{3}-1$), this corresponds to the points $-2vg_{1}-(w+v+1)g_{2}+(3v+1)g_{3}=-(v-1)-w\rho+(v+1)\rho^{2}-2g_{3}.$ * • $e_{1}=1$ and $e_{2}=2$ if and only if $\frac{a}{3}\leq v\leq a$. Together with the point in (7) with $s=2$ (so that $\frac{a}{3}\leq v\leq a+1$), this corresponds to the points $-(2v+1)g_{1}-(w+v+2)g_{2}+(3v+4)g_{3}=-(v-1)-w\rho+(v+2)\rho^{2}-2g_{3}.$ ### 6.4 Representing the lattice points We list all lattice points in the second parallelepiped in Tables 2, 3 and 4 for the cases $s=0$, $s=1$ and $s=2$, respectively. We call $P_{i}$ (resp. $R_{i}$, resp. $S_{i}$) the set of points in the $i$-th row of Table 2 (resp. 3, resp. 4). $P_{i}$ | Point | Situation | Region ---|---|---|--- $P_{1}$ | $-v-w\rho+(v+1)\rho^{2}$ | (a) | $0\leq v\leq a-1$, $1\leq r\leq a-v$ $P_{2}$ | (b) | $0\leq v\leq a$, $r=a-v+1$ $P_{3}$ | $-v-w\rho+(v+2)\rho^{2}$ | (c) | $0\leq v\leq a$, $a-v+2\leq r\leq a+1$ $P_{4}$ | $-(v-1)-w\rho+(v+1)\rho^{2}$ | (d) | $0\leq v\leq a+1$, $r=0$ Table 2: Lattice points in the second parallelepiped with $s=0$ $R_{i}$ | Point | Situation | Region ---|---|---|--- $R_{1}$ | $-(v-1)-w\rho+(v+1)\rho^{2}-g_{3}$ | (a) | $0\leq v\leq\frac{a}{3}$, $1\leq r\leq\frac{a}{3}$ $R_{2}$ | (a) | $\frac{a}{3}+1\leq v\leq\frac{2a}{3}-1$, $1\leq r\leq\frac{2a}{3}-v$ $R_{3}$ | (d) | $0\leq v\leq\frac{2a}{3}+1$, $r=0$ $R_{4}$ | $-v-w\rho+(v+1)\rho^{2}-g_{3}$ | (a) | $0\leq v\leq\frac{a}{3}-1$, $\frac{a}{3}+1\leq r\leq\frac{2a}{3}-v$ $R_{5}$ | $-(v-1)-w\rho+(v+2)\rho^{2}-g_{3}$ | (a) | $\frac{a}{3}+1\leq v\leq\frac{2a}{3}$, $\frac{2a}{3}-v+1\leq r\leq\frac{a}{3}$ $R_{6}$ | (a) | $\frac{2a}{3}+1\leq v\leq a-1$, $1\leq r\leq a-v$ $R_{7}$ | (b) | $\frac{2a}{3}+1\leq v\leq a$, $r=a-v+1$ $R_{8}$ | (c) | $\frac{2a}{3}+1\leq v\leq a$, $a-v+2\leq r\leq\frac{a}{3}+1$ $R_{9}$ | (d) | $\frac{2a}{3}+2\leq v\leq a+1$, $r=0$ $R_{10}$ | $-v-w\rho+(v+2)\rho^{2}-g_{3}$ | (a) | $0\leq v\leq\frac{a}{3}$, $\frac{2a}{3}-v+1\leq r\leq a-v$ $R_{11}$ | (a) | $\frac{a}{3}+1\leq v\leq\frac{2a}{3}-1$, $\frac{a}{3}+1\leq r\leq a-v$ $R_{12}$ | (b) | $0\leq v\leq\frac{2a}{3}$, $r=a-v+1$ $R_{13}$ | (c) | $1\leq v\leq\frac{2a}{3}$, $a-v+2\leq r\leq a+1$ $R_{14}$ | (c) | $\frac{2a}{3}+1\leq v\leq a$, $\frac{a}{3}+2\leq r\leq\frac{5a}{3}-v+2$ $R_{15}$ | $-v-w\rho+(v+3)\rho^{2}-g_{3}$ | (c) | $\frac{2a}{3}+2\leq v\leq a$, $\frac{5a}{3}-v+3\leq r\leq a+1$ Table 3: Lattice points in the second parallelepiped with $s=1$ $S_{i}$ | Point | Situation | Region ---|---|---|--- $S_{1}$ | $-(v-1)-w\rho+(v+1)\rho^{2}-2g_{3}$ | (a) | $0\leq v\leq\frac{a}{3}-2$, $1\leq r\leq\frac{a}{3}-v-1$ $S_{2}$ | (d) | $0\leq v\leq\frac{a}{3}-1$, $r=0$ $S_{3}$ | $-(v-1)-w\rho+(v+2)\rho^{2}-2g_{3}$ | (a) | $0\leq v\leq\frac{a}{3}-1$, $\frac{a}{3}-v\leq r\leq\frac{2a}{3}$ $S_{4}$ | (a) | $\frac{a}{3}\leq v\leq a-1$, $1\leq r\leq a-v$ $S_{5}$ | (b) | $\frac{a}{3}\leq v\leq a$, $r=a-v+1$ $S_{6}$ | (c) | $\frac{a}{3}+1\leq v\leq\frac{2a}{3}$, $a-v+2\leq r\leq\frac{2a}{3}+1$ $S_{7}$ | (c) | $\frac{2a}{3}+1\leq v\leq a$, $a-v+2\leq r\leq\frac{4a}{3}-v+1$ $S_{8}$ | (d) | $\frac{a}{3}\leq v\leq a+1$, $r=0$ $S_{9}$ | $-v-w\rho+(v+2)\rho^{2}-2g_{3}$ | (a) | $0\leq v\leq\frac{a}{3}-1$, $\frac{2a}{3}+1\leq r\leq a-v$ $S_{10}$ | (b) | $0\leq v\leq\frac{a}{3}-1$, $r=a-v+1$ $S_{11}$ | (c) | $1\leq v\leq\frac{a}{3}-1$, $a-v+2\leq r\leq a+1$ $S_{12}$ | (c) | $\frac{a}{3}\leq v\leq\frac{2a}{3}-1$, $\frac{2a}{3}+2\leq r\leq\frac{4a}{3}-v+1$ $S_{13}$ | $-(v-1)-w\rho+(v+3)\rho^{2}-2g_{3}$ | (c) | $\frac{2a}{3}+1\leq v\leq a$, $\frac{4a}{3}-v+2\leq r\leq\frac{2a}{3}+1$ $S_{14}$ | $-v-w\rho+(v+3)\rho^{2}-2g_{3}$ | (c) | $\frac{a}{3}+1\leq v\leq\frac{2a}{3}$, $\frac{4a}{3}-v+2\leq r\leq a+1$ $S_{15}$ | (c) | $\frac{2a}{3}+1\leq v\leq a$, $\frac{2a}{3}+2\leq r\leq a+1$ Table 4: Lattice points in the second parallelepiped with $s=2$ For a fixed value of $s$, we identify a lattice point $\alpha_{s}(v,r)$ with a point $(v,r)\in\mathbb{Z}^{2}$. In Figures 1, 3 and 3, we represent the previous regions in two dimensional planes, where the X-axis (resp. Y-axis) corresponds to the values of $v$ (resp. $r$). In the statement of Theorem 1.2, the element $g_{3}$ in (i) is the only candidate from the first parallelepiped. The elements in (ii), (iii) and (iv) come from the case $s=0$, the ones in (v) have $s=1$, and the remaining ones correspond to $s=2$. In the already mentioned figures, we highlight them in red colour. $0$$v$$r$$a+1$$\frac{2a}{3}+1$$\frac{a}{3}+1$$a+1$$\frac{2a}{3}+1$$\frac{a}{3}+1$$1$$P_{1}$$P_{2}$$P_{3}$$P_{4}$ Figure 1: Lattice points with $s=0$ $0$$v$$r$$a+1$$\frac{2a}{3}+1$$\frac{a}{3}+1$$a+1$$\frac{2a}{3}+1$$\frac{a}{3}+1$$1$$R_{1}$$R_{2}$$R_{3}$$R_{4}$$R_{5}$$R_{6}$$R_{7}$$R_{8}$$R_{9}$$R_{10}$$R_{11}$$R_{12}$$R_{13}$$R_{14}$$R_{15}$ Figure 2: Lattice points with $s=1$ $0$$v$$r$$a+1$$\frac{2a}{3}+1$$\frac{a}{3}+1$$a+1$$\frac{2a}{3}+1$$\frac{a}{3}+1$$1$$S_{1}$$S_{2}$$S_{3}$$S_{4}$$S_{5}$$S_{6}$$S_{7}$$S_{8}$$S_{9}$$S_{10}$$S_{11}$$S_{12}$$S_{13}$$S_{14}$$S_{15}$ Figure 3: Lattice points with $s=2$ ### 6.5 Proof of the main theorem We proceed to validate Theorem 1.2. We need to prove that the elements listed therein are indecomposable and that all the other lattice points in the parallelepipeds are decomposable. #### 6.5.1 Transformations of lattice points For later use, we consider the transformations $T_{1}$ and $T_{2}$ introduced in [16, Section 5.1] and extend them to all elements of $K$. Namely, for $\alpha\in K$, we define $T_{1}(\alpha)=\alpha^{\prime}(-1-a-(a^{2}+3a+3)\rho+(a+2)\rho^{2}),\quad T_{2}(\alpha)=\alpha^{\prime\prime}\rho^{2}.$ The choice $s=0$ corresponds to the ones studied in [16]. Our points $\alpha_{0}(v,r)$ are represented therein as $\alpha(v,W)$ with $W=r-1$. With our notation, the definitions of $T_{1}$ and $T_{2}$ in $\blacktriangle$ are: $\begin{split}&T_{1}(\alpha_{0}(v,r))=(\alpha_{0}(v,r))^{\prime}(-1-a-(a^{2}+3a+3)\rho+(a+2)\rho^{2})=\alpha_{0}(r-1,a+2-v-r),\\\ &T_{2}(\alpha_{0}(v,r))=(\alpha_{0}(v,r))^{\prime\prime}\rho^{2}=\alpha_{0}(a+1-v-r,v+1).\end{split}$ Let us study the behaviour of transformations $T_{1}$ and $T_{2}$ over lattice points with $s\neq 0$. It turns out that their behaviour over these points is far more unpredictable than in the case $s=0$. For the sake of simplicity, we will just work with the transformation $T_{1}$. Let us assume that $s=1$ and let us identify a lattice point $\alpha_{1}(v,r)$ with a point $(v,r)\in\mathbb{Z}^{2}$. Using some mathematical software, we can easily check the following: * • If $(v,r)\in\bigcup_{i=1}^{3}R_{i}$, $T_{1}(v,r)=\begin{cases}\Big{(}\frac{2a}{3}+1,0\Big{)}&\hbox{if }(v,r)=(0,0),\\\ \Big{(}r+\frac{2a}{3},a+2-v-r\Big{)}&\hbox{otherwise}.\end{cases}$ * • If $(v,r)\in R_{4}$, $T_{1}(v,r)=\Big{(}r-\frac{a}{3}-1,a+1-v-r\Big{)}.$ * • If $(v,r)\in\bigcup_{i=5}^{9}R_{i}$, $T_{1}(v,r)=\begin{cases}\Big{(}r+\frac{2a}{3},a+2-v-r\Big{)}&\hbox{if }v+r\leq a+2,\\\ \Big{(}r+\frac{2a}{3}-1,2a+4-v-r\Big{)}&\hbox{if }v+r>a+2.\end{cases}$ * • If $(v,r)\in\bigcup_{i=10}^{14}R_{i}$, $T_{1}(v,r)=\begin{cases}\Big{(}r-\frac{a}{3}-1,a+1-v-r\Big{)}&\hbox{if }v+r\leq a+1,\\\ \Big{(}r-\frac{a}{3}-2,2a+3-v-r\Big{)}&\hbox{if }v+r>a+1.\end{cases}$ * • If $(v,r)\in R_{15}$, $T_{1}(v,r)=\Big{(}r-\frac{a}{3}-2,2a+3-v-r\Big{)}.$ We see that at each case $T_{1}$ behaves as a composition of symmetries and translations. In the following, we list the images of some regions $R_{i}$ by $T_{1}$. ###### Lemma 6.1. Under the previous notations, we have: * • $T_{1}(R_{1}\cup R_{2})=R_{14}$, * • $T_{1}(R_{3}-\\{(0,0)\\})=\\{v=\frac{2a}{3},\,\frac{a}{3}+1\leq r\leq a+1\\}\subset R_{12}\cup R_{13}$, * • $T_{1}(0,0)=(\frac{2a}{3}+1,0)$, * • $T_{1}(R_{10}\cup R_{11})=R_{1}\cup R_{2}$, * • $T_{1}(R_{12})=R_{3}-\\{(\frac{2a}{3}+1,0)\\}$, * • $T_{1}(R_{5})=R_{8}$, * • $T_{1}(R_{9}\cup R_{15})=R_{5}$. ###### Proof. We may check directly each inequality. Let us do it explicitly for the first one. Given $(v,r)\in R_{1}\cup R_{2}$, we know that $T_{1}(v,r)=(v^{\prime},r^{\prime})$, where $\begin{cases}v^{\prime}=r+\frac{2a}{3},\\\ r^{\prime}=a+2-v-r.\end{cases}$ If $(v,r)\in R_{1}$, we have that $0\leq v\leq\frac{a}{3}$ and $1\leq r\leq\frac{a}{3}$. We see directly that $\frac{2a}{3}+1\leq v^{\prime}\leq a$. On the other hand, $\frac{2a}{3}+2-r\leq r^{\prime}\leq a+2-r$. Now, $\frac{2a}{3}+2-r=\frac{4a}{3}+2-v^{\prime}$ and $a+2-r=\frac{5a}{3}-v^{\prime}+2$, so $\frac{4a}{3}-v^{\prime}+2\leq r^{\prime}\leq\frac{5a}{3}-v^{\prime}+2$. If $(v,r)\in R_{2}$, we have that $\frac{a}{3}+1\leq v\leq\frac{2a}{3}-1$ and $1\leq r\leq\frac{2a}{3}-v$. It is direct that $\frac{2a}{3}+1\leq v^{\prime}\leq\frac{4a}{3}-v+1\leq a$. On the other hand, since $v+r\leq\frac{2a}{3}$, we have that $r^{\prime}\geq\frac{2a}{3}+2$. Finally, $v^{\prime}+r^{\prime}=\frac{5a}{3}+2-v\leq\frac{4a}{3}+1$, so $r^{\prime}\leq\frac{4a}{3}-v^{\prime}+1$. Joining the two cases, we see that $\frac{2a}{3}+1\leq v^{\prime}\leq a$ and $\frac{2a}{3}+2\leq r^{\prime}\leq\frac{5a}{3}-v^{\prime}+2$. Since all these steps are reversible, we conclude that $T_{1}(R_{1}\cup R_{2})=R_{14}$. ∎ Now, let us assume that $s=2$ and study the behaviour of $T_{1}$ over the points $\alpha_{2}(v,r)$. As in the previous case, we identify a lattice point $\alpha_{2}(v,r)$ with a point $(v,r)\in\mathbb{Z}^{2}$. * • If $(v,r)\in S_{1}\cup S_{2}$, $T_{1}(v,r)=\Big{(}r+\frac{a}{3},a+1-r-v\Big{)}.$ * • If $(v,r)\in S_{3}\cup S_{4}\cup S_{5}\cup S_{6}\cup S_{7}\cup S_{8}$, $T_{1}(v,r)=\begin{cases}\Big{(}r+\frac{a}{3},a+1-v-r\Big{)}&\hbox{if }v+r\leq a+1,\\\ \Big{(}r+\frac{a}{3}-1,2a+3-v-r\Big{)}&\hbox{if }v+r>a+1.\end{cases}$ * • If $(v,r)\in S_{9}\cup S_{10}\cup S_{11}\cup S_{12}$, $T_{1}(v,r)=\begin{cases}\Big{(}r-\frac{2a}{3}-1,a-r-v\Big{)}&\hbox{if }v+r\leq a,\\\ \Big{(}r-\frac{2a}{3}-2,2a-r-v+2\Big{)}&\hbox{if }v+r>a.\end{cases}$ * • If $(v,r)\in S_{13}$, $T_{1}(v,r)=(r+\frac{a}{3}-1,2a+3-v-r).$ * • If $(v,r)\in S_{14}\cup S_{15}$, $T_{1}(v,r)=\Big{(}r-\frac{2a}{3}-2,2a+2-v-r\Big{)}.$ Again, we can use these expressions to identify the images of regions $S_{i}$ by $T_{1}$. ###### Lemma 6.2. Let us identify a lattice point $\alpha_{2}(v,r)$ with a point $(v,r)\in\mathbb{Z}^{2}$. We have: * • $T_{1}(S_{1}\cup S_{2})=S_{12}$, * • $T_{1}(S_{3})=S_{6}\cup S_{7}$, * • $T_{1}(S_{6}\cup S_{7})=S_{14}\cup S_{15}$, * • $T_{1}(S_{12})=S_{9}$. #### 6.5.2 Proving the indecomposability First, we prove the indecomposability of the elements in the statement of Theorem 1.2. We begin with the elements of minimal trace 1. ###### Proposition 6.3. The elements in Theorem 1.2 (i), (v), (vi), (vii) and (viii) have minimal trace $1$, and consequently they are indecomposables. ###### Proof. For (i), i.e. the point $g_{3}$, we just note that $\varphi_{1}+\varphi_{3}\in\mathcal{O}_{K}^{\vee}$ is a root of the polynomial $x^{3}-x^{2}+\frac{a+3}{a^{2}+3a+9}x-\frac{2a+3}{(a^{2}+3a+9)^{2}},$ so it is totally positive, and that $\mathrm{Tr}((\varphi_{1}+\varphi_{3})g_{3})=1.$ For (v), let $0\leq v\leq\frac{a}{3}-1$ and $\frac{a}{3}+1\leq r\leq\frac{2a}{3}-v$. The element $3\varphi_{1}+2\varphi_{3}\in\mathcal{O}_{K}^{\vee}$ is a root of the polynomial $x^{3}-3x^{2}+\frac{18}{a^{2}+3a+9}x-\frac{27}{(a^{2}+3a+9)^{2}},$ so it is totally positive. Now, we have $\begin{split}&\mathrm{Tr}((3\varphi_{1}+2\varphi_{3})(-(2v+1)g_{1}-(v(a+3)+r+1)g_{2}+(3v+2)g_{3})=\\\ &-6v-3+6v+4=1.\end{split}$ Since the elements in (vii) and (viii) are associated with conjugates of elements in (vi), it is enough to work with the last ones. Let $0\leq r\leq\frac{a}{3}-1$. We already know that $\varphi_{1}+\varphi_{3}\in\mathcal{O}_{K}^{\vee,+}$. Now, we have $\mathrm{Tr}((\varphi_{1}+\varphi_{3})(-(r+1)g_{2}+g_{3}))=1.$ ∎ On the other hand, the remaining indecomposables have minimal trace $2$. ###### Proposition 6.4. The elements in Theorem 1.2 (ii), (iii) and (iv) have minimal trace $2$. ###### Proof. Let us recover the notation in Section 6.1. We have that $T_{1}$ maps bijectively the points from (ii) to (iv), from (iv) to (iii), and from (iii) to (ii). This means that the elements in (iii) and (iv) are associated with conjugates of elements in (ii). Moreover, the points in (ii) lie in $\blacktriangle_{0}(a)$. Therefore, it is enough to prove the statement of the points in (ii). Let $\alpha=-g_{1}-(r+1)g_{2}+3g_{3}$, $1\leq r\leq\frac{a}{3}$, be a lattice point as in (ii). Since $\mathrm{Tr}((\varphi_{1}+\varphi_{3})\alpha)=2$, the minimal trace of $\alpha$ is upper bounded by $2$. Let us assume that there exists a totally positive element of the codifferent $\delta=u_{1}\varphi_{1}+u_{2}\varphi_{2}+u_{3}\varphi_{3}$ such that $\mathrm{Tr}(\alpha\delta)=1$, i.e., $\mathrm{Tr}(\alpha\delta)=-u_{1}-(r+1)u_{2}+3u_{3}=1.$ It implies that $u_{1}=-(r+1)u_{2}+3u_{3}-1$. Now we will consider several totally positive elements, whose product with $\delta$ gives a totally positive element with a positive trace. We will show that these traces cannot be all positive at the same time. First of all, let us consider the element $-g_{1}-rg_{2}+3g_{3}$. This element is totally positive for $1\leq r\leq\frac{a}{3}$. Note that it is also true for $r=1$ as $-g_{1}-g_{2}+3g_{3}=\rho^{2}$. Thus, we obtain $\mathrm{Tr}(\delta(-g_{1}-rg_{2}+3g_{3}))=(r+1)u_{2}-3u_{3}+1-ru_{2}+3u_{3}=u_{2}+1.$ This trace is positive only if $u_{2}\geq 0$. Similarly, when we consider the totally positive element $-g_{1}-(r+2)g_{2}+3g_{3}$, we get $\mathrm{Tr}(\delta(-g_{1}-(r+2)g_{2}+3g_{3}))=(r+1)u_{2}-3u_{3}+1-(r+2)u_{2}+3u_{3}=-u_{2}+1.$ This trace is positive only if $u_{2}\leq 0$, which, together with the previous part, gives $u_{2}=0$. When we consider the element $g_{3}$, we can conclude that $\mathrm{Tr}(\delta g_{3})=u_{3},$ which implies $u_{3}>0$. On the other hand, for $-g_{1}-(\frac{a}{3}+2)g_{2}+2g_{3}$, we see that $\mathrm{Tr}\Big{(}\delta\Big{(}-g_{1}-\Big{(}\frac{a}{3}+2\Big{)}g_{2}+2g_{3}\Big{)}\Big{)}=-3u_{3}+1+2u_{3}=-u_{3}+1,$ which leads to $u_{3}\geq 0$. Thus, there is no $u_{3}\in\mathbb{Z}$ for which $\delta$ would be totally positive. ∎ Next, we prove that these are indecomposable. ###### Proposition 6.5. The elements in Theorem 1.2 (ii), (iii) and (iv) are indecomposable. ###### Proof. As in Proposition 6.4, it is enough to prove the indecomposability for points in (ii). Let $\alpha=-g_{1}-(r+1)g_{2}+3g_{3}$, $1\leq r\leq\frac{a}{3}$, be a lattice point as in (ii). Since it has minimal trace $2$, its only possible decomposition is as a sum of elements whose trace after multiplication by $\varphi_{1}+\varphi_{3}$ is equal to $1$, i.e. $\alpha=(-v_{1}g_{1}-w_{1}g_{2}+(v_{1}+1)g_{3})+(-v_{2}g_{1}-w_{2}g_{2}+(v_{2}+1)g_{3}).$ First of all, let us discuss the case when $v_{1}<0$, and firstly assume $v_{1}\leq-2$. If $w_{1}>0$, then $-v_{1}g_{1}-w_{1}g_{2}+(v_{1}+1)g_{3}<0,$ since $a+1<\rho<a+1+\frac{2}{a}$ and $g_{3}\geq 3$ for $a\geq 21$. Thus, in this case we must have $w_{1}\leq 0$. Moreover, we can conclude that $\begin{split}\mathrm{Tr}(-v_{1}g_{1}-w_{1}g_{2}+(v_{1}+1)g_{3})&=-3v_{1}-aw_{1}+(v_{1}+1)\frac{a^{2}+3a+9}{3}\\\ &=\frac{a^{2}+3a}{3}v_{1}-aw_{1}+\frac{a^{2}+3a+9}{3}.\end{split}$ This trace is negative or zero if $\frac{(a+3)(v_{1}+1)}{3}+\frac{3}{a}\leq w_{1},$ in which case our element is not totally positive. Thus, let us assume that $\frac{(a+3)(v_{1}+1)}{3}+\frac{3}{a}>w_{1}$. We have $-v_{1}g_{1}-w_{1}g_{2}+(v_{1}+1)g_{3}=\frac{-2v_{1}+1}{3}-\Big{(}w_{1}-\frac{v_{1}+1}{3}\Big{)}\rho+\frac{v_{1}+1}{3}\rho^{2}.$ Under our assumption on $v_{1}$ and $w_{1}$, we see that the coefficient before $\rho$ is positive, and the coefficient before $\rho^{2}$ is negative. Since $-1-\frac{1}{a}<\rho^{\prime}<-1-\frac{1}{2a}$, we can conclude that $\begin{split}\frac{-2v_{1}+1}{3}-\Big{(}w_{1}-\frac{v_{1}+1}{3}\Big{)}\rho^{\prime}+\frac{v_{1}+1}{3}\rho^{\prime 2}<\frac{-2v_{1}+1}{3}+\Big{(}w_{1}-\frac{v_{1}+1}{3}\Big{)}+\frac{v_{1}+1}{3}\\\ \leq\frac{-2v_{1}+1}{3}+\frac{(a+3)(v_{1}+1)}{3}+\frac{3}{a}-\frac{v_{1}+1}{3}+\frac{v_{1}+1}{3}=\frac{(a+1)(v_{1}+1)}{3}+1+\frac{3}{a}\leq 0\end{split}$ for $a\geq 21$. It implies that if $v_{1}\leq-2$, then the element of this form is never totally positive. If $v_{1}=-1$, we obtain the element $1-w_{1}\rho$. If $w_{1}>0$, then $1-w_{1}\rho<0$, and if $w_{1}<0$, then $1-w_{1}\rho^{\prime}<0$. We obtain a totally positive element only if $w_{1}=0$, i.e., the element $1$. However, we know that $\alpha$ is smaller than $1$ for some embedding, so $1$ cannot appear in its decomposition. Therefore, we must have $v_{1}\geq 0$, and this is also true for $v_{2}$. Without loss of generality, we thus must have $v_{1}=1$ and $v_{2}=0$. If $w_{2}<0$, we get $-w_{2}\rho^{\prime}+\frac{1+\rho^{\prime}+\rho^{\prime 2}}{3}=\frac{1}{3}-\Big{(}w_{2}-\frac{1}{3}\Big{)}\rho^{\prime}+\frac{1}{3}\rho^{\prime 2}<\frac{1}{3}+w_{2}-\frac{1}{3}+\frac{\Big{(}1+\frac{1}{a}\Big{)}^{2}}{3}=w_{2}+\frac{\Big{(}1+\frac{1}{a}\Big{)}^{2}}{3}<0$ as, clearly, $\frac{\Big{(}1+\frac{1}{a}\Big{)}^{2}}{3}<1$ for $a\geq 21$. Therefore, $w_{2}\geq 0$. Similarly, if $w_{1}<0$, we obtain $-1+\frac{2}{3}-\Big{(}w_{1}-\frac{2}{3}\Big{)}\rho^{\prime}+\frac{2}{3}\rho^{\prime 2}<-1+\frac{2}{3}+w_{1}-\frac{2}{3}+\frac{2}{3}\Big{(}1+\frac{1}{a}\Big{)}^{2}<0.$ Thus, we have $0\leq w_{1},w_{2}\leq r+1\leq\frac{a}{3}+1$. Now, we will use the remaining root $-\frac{1}{a+2}<\rho^{\prime\prime}<-\frac{1}{a+3}$. If $w_{1}=0$, we have $-1+2\frac{1+\rho^{\prime\prime}+\rho^{\prime\prime 2}}{3}<-\frac{1}{3}-\frac{2}{3}\frac{1}{a+3}+\frac{2}{3}\frac{1}{(a+2)^{2}}<0$ for $a\geq 21$. Similarly, if $w_{1}>0$, we obtain $-1-w_{1}\rho^{\prime\prime}+2\frac{1+\rho^{\prime\prime}+\rho^{\prime\prime 2}}{3}=-\frac{1}{3}-\left(w_{1}-\frac{2}{3}\right)\rho^{\prime\prime}+\frac{2}{3}\rho^{\prime\prime}<-\frac{1}{3}+\left(w_{1}-\frac{2}{3}\right)\frac{1}{a+2}+\frac{2}{3}\frac{1}{(a+2)^{2}}$ The expression on the right side is positive only if $w_{1}>\frac{a}{3}+\frac{4}{3}-\frac{2}{3(a+2)}>\frac{a}{3}+1$ for $a\geq 21$. This is impossible as we must have $w_{1}\leq\frac{a}{3}+1$. Therefore, $\alpha$ is indecomposable. ∎ #### 6.5.3 Finding decompositions of lattice points We have proved that the elements in Theorem 1.2 are indecomposable, and we have given their minimal traces. Now, we prove that all the other lattice points in the second parallelepiped are decomposable. Again, we have to distinguish different cases depending on the value of $s$. Let us assume that $s=0$. We already know that the points from $P_{3}$ and $P_{4}$ are decomposable (because they are indecomposables in $\mathbb{Z}[\rho]$, see [16, Lemma 4.2]). Thus, we are left with the regions $P_{1}$ and $P_{2}$. Their union correspond to the triangle $\blacktriangle$ (note that since the points in situation (d) are discarded, we can denote $v=x$). Recall the definition of $T_{1}$ and $T_{2}$ in Section 6.5.1. Since the factors multiplying the conjugate of $\alpha_{0}(v,r)$ at each line are totally positive units, we have that for each $i\in\\{1,2\\}$, $\alpha_{0}(v,r)$ is indecomposable if and only if so is $T_{i}(\alpha_{0}(v,r))$. Moreover, for each $\alpha\in\blacktriangle$, there is $i\in\\{1,2\\}$ such that $T_{i}(\alpha)$ lies in the region $\blacktriangle_{0}(a)=\Big{\\{}\alpha_{0}(v,r)\,|\,0\leq v\leq\frac{a}{3}-1,\,v+1\leq r\leq a-2v\Big{\\}}\cup\Big{\\{}\alpha_{0}\Big{(}\frac{a}{3},\frac{a}{3}+1\Big{)}\Big{\\}}.$ Therefore, it is enough to study the decomposability of the lattice points in $\blacktriangle_{0}(a)$. ###### Proposition 6.6. The elements with $s=0$ other than the ones in Theorem 1.2 (ii), (iii) and (iv) are decomposable. ###### Proof. We know that it is enough to look at the points in $\blacktriangle_{0}(a)$. Thus, let us consider $\alpha\in\blacktriangle_{0}(a)$, i.e. we have $\begin{split}\alpha=-v-(v(a+2)+r)\rho+(v+1)\rho^{2},\end{split}$ (8) where $0\leq v\leq\frac{a}{3}-1$ and $v+1\leq r\leq a-2v$, or $\alpha=\alpha_{0}(\frac{a}{3},\frac{a}{3}+1)$. Among these, the points that are known to be indecomposable (from Proposition 6.5) are the ones in Theorem 1.2 (ii), i.e. for $v=0$ and $1\leq r\leq\frac{a}{3}$. Let us assume that $0\leq v\leq\frac{a}{3}-1$ and $\frac{a}{3}+1\leq r\leq\frac{2a}{3}-v$. Then $\alpha-g_{3}$ is one of the points in Theorem 1.2 (v), so $\alpha$ is decomposable. If $v=0$ and $\frac{2a}{3}+1\leq r\leq a$, we have that $\begin{split}\alpha&=\Big{[}-\frac{2a}{3}\rho+\rho^{2}-g_{3}\Big{]}\\\ &+\Big{[}1-\Big{(}r-\frac{2a}{3}-1\Big{)}\rho+\rho^{2}-2g_{3}\Big{]}.\end{split}$ The first summand is $\alpha_{1}(0,\frac{2a}{3})\in R_{4}$. On the other hand, $0\leq r^{\prime}\leq\frac{a}{3}-1$ for $r^{\prime}=r-\frac{2a}{3}-1$, so the second summand is the point $\alpha_{2}(0,r^{\prime})$ that belongs to $S_{1}$ if $1\leq r^{\prime}\leq\frac{a}{3}-1$ and to $S_{2}$ if $r^{\prime}=0$. Therefore, $\alpha$ is decomposable for this case. Assume that $1\leq v\leq\frac{a}{3}-1$ and $v+1\leq r\leq\frac{a}{3}$. We have $\begin{split}\alpha&=\Big{[}-\Big{(}\frac{a}{3}+1\Big{)}\rho+\rho^{2}-g_{3}\Big{]}\\\ &+\Big{[}-(v-1)-(v(a+2)+r-\frac{a}{3}-2)\rho+(v+1)\rho^{2}-2g_{3}\Big{]}.\end{split}$ The first summand is $\alpha_{1}(0,\frac{a}{3}+1)\in R_{4}$. The second one can be rewritten as $-(v-1)-\Big{(}(v-1)(a+2)+r+\frac{2a}{3}\Big{)}\rho+(v+1)\rho^{2}-2g_{3}.$ For $v^{\prime}=v-1$ and $r^{\prime}=r+\frac{2a}{3}$, this is of the form $-v^{\prime}-(v^{\prime}(a+2)+r^{\prime})\rho+(v^{\prime}+2)\rho^{2}-2g_{3},$ with $0\leq v^{\prime}\leq\frac{a}{3}-2$. * • If $r\leq\frac{a}{3}-v+1$, then $\frac{2a}{3}+1+v\leq r^{\prime}\leq a-v+1=a-v^{\prime}$ and the point above is $\alpha_{2}(v^{\prime},r^{\prime})\in S_{9}$. * • If $r=\frac{a}{3}-v+2$, then $r^{\prime}=a-v^{\prime}+1$ and we obtain $\alpha_{2}(v^{\prime},r^{\prime})\in S_{10}$. * • In the remaining cases, $a-v^{\prime}+2\leq r^{\prime}\leq a$ and we obtain $\alpha_{2}(v^{\prime},r^{\prime})\in S_{11}$. Now, consider the same decomposition for $\alpha=\alpha_{0}(\frac{a}{3},\frac{a}{3}+1)$, i.e. $v=\frac{a}{3}$ and $r=\frac{a}{3}+1$. Then $v^{\prime}=\frac{a}{3}-1$ and $r^{\prime}=a+1$, so that $\alpha_{2}(v^{\prime},r^{\prime})\in S_{11}$. Assume that $1\leq v\leq\frac{a}{3}-1$ and $\frac{2a}{3}-v+1\leq r\leq a-2v$. We have $\begin{split}\alpha&=\Big{[}-\frac{2a}{3}\rho+\rho^{2}-g_{3}\Big{]}\\\ &+\Big{[}-(v-1)-\Big{(}v(a+2)+r-\frac{2a}{3}-1\Big{)}\rho+(v+1)\rho^{2}-2g_{3}\Big{]}.\end{split}$ We already know that the first summand is totally positive. Let us assume that $\frac{2a}{3}-v+1\leq r\leq\frac{2a}{3}$ (for simplicity, even when $a-2v\leq\frac{2a}{3}$, which happens if and only if $v\geq\lceil\frac{a}{6}\rceil$ and in which case $\alpha$ is as in (8)). We rewrite the second summand as $-(v-1)-\Big{(}(v-1)(a+2)+r+\frac{a}{3}+1\Big{)}\rho+(v+1)\rho^{2}-2g_{3},$ which is of the form $-v^{\prime}-(v^{\prime}(a+2)+r^{\prime})\rho+(v^{\prime}+2)\rho^{2}-2g_{3}$ for $v^{\prime}=v-1$ and $r^{\prime}=r+\frac{a}{3}+1$, so that $a-v^{\prime}+1\leq r^{\prime}\leq a+1$. If $v=1$, then $v^{\prime}=0$ and $r^{\prime}=a+1$, and we obtain the lattice point $\alpha_{2}(0,a+1)\in S_{10}$. Otherwise, if $2\leq v\leq\frac{a}{3}-1$, then $1\leq v^{\prime}\leq\frac{a}{3}-2$, and we obtain a lattice point $\alpha_{2}(v^{\prime},r^{\prime})$ in $S_{10}$ if $r^{\prime}=a-v^{\prime}+1$ and in $S_{11}$ if $a-v^{\prime}+2\leq r^{\prime}\leq a+1$. Finally, let us assume that $\frac{2a}{3}+1\leq r\leq a-2v$, which in particular implies that $1\leq v\leq\lfloor\frac{a-3}{6}\rfloor<\frac{a}{3}-1$. Let $r^{\prime}=r-\frac{2a}{3}-1$. If $r=\frac{2a}{3}+1$, then $r^{\prime}=0$ and the second summand above equals $\alpha_{2}(v,0)\in S_{2}$. Otherwise, if $\frac{2a}{3}+2\leq r\leq a-2v$, then $1\leq r^{\prime}\leq\frac{a}{3}-2v-1$, and we obtain $\alpha_{2}(v,r^{\prime})\in S_{1}$. We have seen that all points in $\blacktriangle_{0}(a)$ other than the ones in Theorem 1.2 (ii) can be written as a sum of totally positive elements, so they are decomposable. ∎ In the cases with $s\neq 0$, $T_{1}$ also preserves decompositions in both directions. Hence, we can use it to reduce the list of lattice points for which we have to find a decomposition. ###### Proposition 6.7. The elements with $s=1$ other than the ones in Theorem 1.2 (v) are decomposable. ###### Proof. Note that the points in Theorem 1.2 (v) are just the ones in $R_{4}$. By Lemma 6.1, it is enough to prove the decomposability of points in $R_{1}$, $R_{2}$, $R_{3}$, $R_{5}$, $R_{6}$, $R_{7}$ and $R_{13}-\\{v=\frac{2a}{3},\,\frac{a}{3}+2\leq r\leq a+1\\}$. We first consider the points of the form $\alpha=-(v-1)-(v(a+2)+r)\rho+(v+1)\rho^{2}-g_{3},$ which correspond to the regions $R_{1}$, $R_{2}$ and $R_{3}$. We have the decompositions $\begin{split}\alpha&=[1+\rho^{2}-2g_{3}]+[-(v-1)-(v(a+2)+r-1)\rho+(v+1)\rho^{2}-2g_{3}]=\alpha_{2}(0,0)+P,\\\ \alpha&=\Big{[}-\frac{2a}{3}\rho+\rho^{2}-g_{3}\Big{]}+\Big{[}-(v-1)-\Big{(}(v-1)(a+2)+r+\frac{a}{3}+2\Big{)}\rho+v\rho^{2}\Big{]}\\\ &=\alpha_{1}\Big{(}0,\frac{2a}{3}\Big{)}+Q.\end{split}$ We will prove that either $P$ or $Q$ are totally positive depending on the values of $v$ and $r$. Let us suppose that $\alpha\in R_{1}\cup R_{2}$, so $0\leq v\leq\frac{2a}{3}-1$ and $1\leq r\leq\mathrm{min}(\frac{a}{3},\frac{2a}{3}-v)$. If $v=0$ and we call $r^{\prime}=r-1$, then $P=1-r^{\prime}\rho+\rho^{2}-2g_{3}$ with $0\leq r^{\prime}\leq\frac{a}{3}-1$. For $r=1$, we obtain $P=\alpha_{2}(0,0)\in S_{2}$, with $(e_{1},e_{2})=(1,1)$, while for $2\leq r\leq\frac{a}{3}$, we obtain $P=\alpha_{2}(0,r^{\prime})\in S_{1}$. Next, suppose that $1\leq v\leq\frac{2a}{3}-1$. Let us call $v^{\prime}=v-1$ and $r^{\prime}=r+\frac{a}{3}+2$. Then $Q=-v^{\prime}-(v^{\prime}(a+2)+r^{\prime})\rho+(v^{\prime}+1)\rho^{2}$, with $0\leq v^{\prime}\leq\frac{2a}{3}-2$ and $\frac{a}{3}+3\leq r^{\prime}\leq\mathrm{min}(\frac{2a}{3},a-v^{\prime}+1)$. Then $Q=\alpha_{0}(v^{\prime},r^{\prime})\in\blacktriangle$ is totally positive. Finally, we assume that $\alpha\in R_{3}$, so that $0\leq v\leq\frac{2a}{3}+1$ and $r=0$. For $v=0$, we have $\alpha=1+\rho^{2}-g_{3}=\alpha_{1}(0,0)$, which we already know is decomposable, and $T_{1}(\alpha_{1}(0,0))=\alpha_{1}(\frac{2a}{3}+1,0)$, so we may assume that $1\leq v\leq\frac{2a}{3}$. For $v^{\prime}=v-1$ and $r^{\prime}=\frac{a}{3}+2$, we have $Q=-v^{\prime}-(v^{\prime}(a+2)+r^{\prime})\rho+(v^{\prime}+1)\rho^{2}$, with $0\leq v^{\prime}\leq\frac{2a}{3}-1$ and $0<r^{\prime}<a-v^{\prime}+1$. Then $Q=\alpha_{0}(v^{\prime},r^{\prime})\in\blacktriangle$ is totally positive. Next, let us take $\alpha\in R_{5}\cup R_{6}\cup R_{7}$, so that $\alpha=-(v-1)-(v(a+2)+r)\rho+(v+2)\rho^{2}-g_{3}$ with $\frac{a}{3}+1\leq v\leq\frac{2a}{3}$ and $\frac{2a}{3}-v+1\leq r\leq\frac{a}{3}$ or $\frac{2a}{3}+1\leq v\leq a$ and $1\leq r\leq a-v+1$. Let us write $\alpha=\alpha_{2}(0,0)+[-(v-1)-(v(a+2)+r-1)\rho+(v+2)\rho^{2}-2g_{3}].$ On the other hand, our hypotheses on $v$ and $r$ imply that in particular $\frac{a}{3}\leq v\leq a$ and $1\leq r\leq a-v+1$, so the second summand is just $\alpha_{2}(v,r)\in S_{4}\cup S_{5}$, and hence it is totally positive. Therefore, $\alpha$ is decomposable for this case. It remains to study the points in the region $R_{13}-\\{v=\frac{2a}{3},\,\frac{a}{3}+2\leq r\leq a+1\\}$. That is, we consider $\alpha=-v-(v(a+2)+r)\rho+(v+2)\rho^{2}-g_{3},$ with $1\leq v\leq\frac{2a}{3}-1$ and $a-v+2\leq r\leq a+1$. Let us write $\begin{split}\alpha&=\Big{[}-\Big{(}\frac{a}{3}+1\Big{)}\rho+\rho^{2}-g_{3}\Big{]}+[-v-\Big{(}v(a+2)+r-\frac{a}{3}-1\Big{)}\rho+(v+1)\rho^{2}]\end{split}$ The first summand is $\alpha_{1}(0,\frac{a}{3}+1)$, which is totally positive. As for the second one, let $r^{\prime}=r-\frac{a}{3}-1$ and assume that $r\leq\frac{4a}{3}-v+1$ (note that this does not impose any restriction if $1\leq v\leq\frac{a}{3}$). Then $\frac{2a}{3}-v+1\leq r^{\prime}\leq\mathrm{min}(\frac{2a}{3},a-v)$, so we obtain $\alpha_{0}(v,r^{\prime})\in\blacktriangle$, which is totally positive. Hence, $\alpha$ is decomposable. The remaining case corresponds to the points for which $\frac{a}{3}+1\leq v\leq\frac{2a}{3}-1$ and $\frac{4a}{3}-v+2\leq r\leq a+1$. But those are the images by $T_{1}$ of the points in $R_{13}$ for which $1\leq v\leq\frac{a}{3}$, so they are decomposable. ∎ ###### Proposition 6.8. The elements with $s=2$ other than the ones in Theorem 1.2 (vi), (vii) and (viii) are decomposable. ###### Proof. Among these lattice points, the indecomposables correspond to: * • the line $\\{v=0,\,0\leq r\leq\frac{a}{3}-1\\}$, contained in $S_{1}\cup S_{2}$, * • the line $\\{0\leq v\leq\frac{a}{3}-1,\,r=\frac{2a}{3}+1\\}$, contained in $S_{9}$, * • the line $\\{\frac{a}{3}\leq v\leq\frac{2a}{3}-1,\,r=\frac{4a}{3}-v+1\\}$, contained in $S_{12}$. and we know that they are indecomposable from Proposition 6.3. Moreover, by Lemma 6.2, it is enough to find decompositions for all the other points in the regions $S_{1}$, $S_{2}$, $S_{3}$, $S_{4}$, $S_{5}$, $S_{8}$, $S_{10}$, $S_{11}$ and $S_{13}$. We start with the regions $S_{1}$ and $S_{2}$. Among their points, the line $\\{\alpha_{2}(0,r)\,|\,0\leq r\leq\frac{a}{3}-1\\}$ are the elements in Theorem 1.2 (vi), and we know that they are indecomposable from Proposition 6.3. The remaining points are of the form $\begin{split}\alpha&=-(v-1)-(v(a+2)+r)\rho+(v+1)\rho^{2}-2g_{3}\\\ &=\alpha_{1}\Big{(}0,\frac{2a}{3}\Big{)}+\Big{[}-(v-1)-\Big{(}v(a+2)+r-\frac{2a}{3}\Big{)}\rho+v\rho^{2}-g_{3}\Big{]}\end{split}$ with $1\leq v\leq\frac{a}{3}-1$ and $0\leq r\leq\frac{a}{3}-v-1$. We rewrite the second summand as $-v^{\prime}-\Big{(}v^{\prime}(a+2)+r^{\prime}\Big{)}\rho+(v^{\prime}+1)\rho^{2}-g_{3},$ where $v^{\prime}=v-1$ and $r^{\prime}=r+\frac{a}{3}+2$. Since $0\leq v^{\prime}\leq\frac{a}{3}-2$ and $\frac{a}{3}+2\leq r^{\prime}\leq\frac{2a}{3}-v^{\prime}$, this is one of the indecomposables in Theorem 1.2 (v). Next, we consider the points from the regions $S_{3}$, $S_{4}$, $S_{5}$ and $S_{8}$, which are of the form $\alpha=-(v-1)-(v(a+2)+r)\rho+(v+2)\rho^{2}-2g_{3}$ with $0\leq v\leq\frac{a}{3}-1$ and $\frac{a}{3}-v\leq r\leq\frac{2a}{3}$ or $\frac{a}{3}\leq v\leq a+1$ and $0\leq r\leq a+1-v$. First we assume that $r>0$, so that $v\leq a$. Let us write $\alpha=[1+\rho^{2}-2g_{3}]+[-v-(v(a+2)+r)\rho+(v+1)\rho^{2}].$ Under these hypotheses we have $1\leq r\leq a-v+1$, so the second summand belongs to $\blacktriangle$ and hence it is totally positive. Now, let us choose $r=0$ and write $\alpha=\Big{[}-\Big{(}\frac{a}{3}+1\Big{)}\rho+\rho^{2}-g_{3}\Big{]}+\Big{[}-(v-1)-\Big{(}(v-1)(a+2)+\frac{2a}{3}+1\Big{)}\rho+(v+1)\rho^{2}-g_{3}\Big{]}.$ The first summand is $\alpha_{1}(0,\frac{a}{3}+1)$. The second one can be rewritten as $-v^{\prime}-\Big{(}v^{\prime}(a+2)+r^{\prime}\Big{)}\rho+(v^{\prime}+2)\rho^{2}-g_{3},$ where $v^{\prime}=v-1$ and $r^{\prime}=\frac{2a}{3}+1$. This belongs to: * • $R_{10}$, if $v=\frac{a}{3}$; * • $R_{12}$, if $v=\frac{a}{3}+1$; * • $R_{13}$, if $\frac{a}{3}+2\leq v\leq\frac{2a}{3}+1$; * • $R_{14}$, if $\frac{2a}{3}+2\leq v\leq a+1$. Hence, the points in these regions are decomposable. Now, we consider the points from the regions $S_{10}$ and $S_{11}$, which are of the form $\alpha=-v-(v(a+2)+r)\rho+(v+2)\rho^{2}-2g_{3}$ with $0\leq v\leq\frac{a}{3}-1$ and $a-v+1\leq r\leq a+1$. Let us write $\alpha=\Big{[}-\Big{(}v+r-\frac{2a}{3}\Big{)}\rho+\rho^{2}-g_{3}\Big{]}+\Big{[}-v-\Big{(}v(a+2)+\frac{2a}{3}-v\Big{)}\rho+(v+1)\rho^{2}-g_{3}\Big{]}.$ The second summand is clearly the point $\alpha_{1}(v,\frac{2a}{3}-v)\in R_{4}$, and consequently totally positive. As for the first one, if we define $r^{\prime}=v+r-\frac{2a}{3}$, from $a-v+1\leq r\leq a+1$ we obtain that $\frac{a}{3}+1\leq r^{\prime}\leq\frac{a}{3}+1+v$, and since $v\leq\frac{a}{3}-1$ we have actually that $\frac{a}{3}+1\leq r^{\prime}\leq\frac{2a}{3}$. Therefore the first summand is $\alpha_{1}(0,r^{\prime})\in R_{4}$. Hence it is totally positive, and $\alpha$ is decomposable. Finally, we consider the points of the region $S_{13}$, which are of the form $\alpha=-(v-1)-(v(a+2)+r)\rho+(v+3)\rho^{2}-2g_{3},$ where $\frac{2a}{3}+1\leq v\leq a$ and $\frac{4a}{3}-v+2\leq r\leq\frac{2a}{3}+1$. We have that $\alpha=[1+\rho^{2}-2g_{3}]+[-v-(v(a+2)+r)\rho+(v+2)\rho^{2}].$ Since $r>a-v+2$, the second summand is a lattice point $\alpha_{0}(v,r)\in P_{3}$, so it is totally positive, and $\alpha$ is decomposable. ∎ We conclude that Theorem 1.2 is established from Propositions 6.3, 6.4, 6.5, 6.6, 6.7 and 6.8. ### 6.6 Other families $p$ | $a$ mod $p^{2}$ | $(k,l)$ | Indecomposables in the first parallelepiped ---|---|---|--- $7$ | $5$ | $(2,6)$ | none $41$ | $(4,3)$ | none $13$ | $66$ | $(3,8)$ | $-g_{2}+2g_{3}$, $-g_{1}-3g_{2}+6g_{3}$, $-g_{1}-3g_{2}+5g_{3}$ $100$ | $(9,7)$ | $-g_{1}-g_{2}+2g_{3}$, $-3g_{1}-2g_{2}+5g_{3}$, $-4g_{1}-3g_{2}+6g_{3}$ $19$ | $154$ | $(11,5)$ | $-2g_{1}-g_{2}+4g_{3}$, $-3g_{1}-g_{2}+6g_{3}$, $-5g_{1}-2g_{2}+9g_{3}$ $204$ | $(7,16)$ | $-g_{1}-3g_{2}+4g_{3}$, $-2g_{1}-5g_{2}+6g_{3}$, $-3g_{1}-7g_{2}+9g_{3}$ $31$ | $356$ | $(25,21)$ | $-2g_{1}-2g_{2}+3g_{3}$, $-10g_{1}-8g_{2}+13g_{3}$, $-12g_{1}-10g_{2}+15g_{3}$, $-6g_{1}-5g_{2}+8g_{3}$, $-7g_{1}-6g_{2}+9g_{3}$, $-11g_{1}-9g_{2}+14g_{3}$ $602$ | $(5,12)$ | $-g_{2}+3g_{3}$, $-2g_{1}-5g_{2}+15g_{3}$, $-2g_{1}-5g_{2}+13g_{3}$, $-g_{1}-3g_{2}+9g_{3}$, $-g_{1}-3g_{2}+8g_{3}$, $-2g_{1}-5g_{2}+14g_{3}$ Table 5: Indecomposables in the first parallelepiped for low values of $p>3$ Now we will briefly discuss indecomposables for families for other primes $p$. It seems that the behaviour of indecomposables is different in comparison to $p=3$, which can be seen even from some partial results. For example, in Table 5, we can find all indecomposable integers in the first parallelepiped formed by units $1$, $\rho^{2}$ and $\rho^{2}\rho^{\prime 2}$. Recall that for $p=3$, we have one indecomposable in this parallelepiped, namely $g_{3}=\frac{1+\rho+\rho^{2}}{3}$. For $p=7$, we do not obtain any indecomposable integer here. On the other hand, for families with $p=13,19$, we get three indecomposables, and for $p=31$, we obtain six such elements. Note that each of these indecomposables has minimal trace $1$. Moreover, similarly as in the second parallelepiped, these elements form triples $\alpha_{1}$, $\alpha_{2}$ and $\alpha_{3}$ so that $\alpha_{1}=\alpha$, $\alpha_{2}=\alpha^{\prime}\varepsilon_{1}$ and $\alpha_{3}=\alpha^{\prime\prime}\varepsilon_{2}$ where $\varepsilon_{1}$ and $\varepsilon_{2}$ are concrete units in ${\mathcal{O}}_{K}$. In the following, similarly as for $p=3$, we use the notation $g_{1}=1$, $g_{2}=\rho$ and $g_{3}=\frac{k+l\rho+\rho^{2}}{p}$. Moreover, the fact that the situation significantly differs for primes $p>3$ can also be seen in the following proposition, which gives the complete characterization of indecomposable integers in ${\mathcal{O}}_{K}$ for $a=41$. ###### Proposition 6.9. Let $K={\mathbb{Q}}(\rho)$ where $\rho$ is a root of the polynomial $x^{3}-41x^{2}-44x-1$. Then, up to multiplication of totally positive units, all the indecomposable integers in ${\mathcal{O}}_{K}$ are $1$, 1. 1. $-(5v+2)g_{1}-wg_{2}+(7v+3)g_{3}$ where $0\leq v\leq 5$ and $46v+14\leq w\leq 45v+19$, 2. 2. $-vg_{2}+g_{3}$ where $1\leq v\leq 6$, 3. 3. $-g_{1}-vg_{2}+2g_{3}$ where $1\leq v\leq 12$, 4. 4. $-(5v+125)g_{1}-(45v+1132)g_{2}+(7v+176)g_{3}$ where $0\leq v\leq 5$, 5. 5. $-(5v+31)g_{1}-(45v+283)g_{2}+(7v+44)g_{3}$ where $0\leq v\leq 11$, 6. 6. $-(5v+5)g_{1}-(46v+22)g_{2}+(7v+8)g_{3}$ where $0\leq v\leq 5$, 7. 7. $-(5v+6)g_{1}-(46v+41)g_{2}+(7v+9)g_{3}$ where $0\leq v\leq 11$, which are of minimal trace $1$, 1. 8. $-3g_{1}-vg_{2}+5g_{3}$ where $9\leq v\leq 14$, 2. 9. $-(5v+3)g_{1}-(45v+32)g_{2}+(7v+5)g_{3}$ where $0\leq v\leq 5$, 3. 10. $-(5v+93)g_{1}-(46v+837)g_{2}+(7v+131)g_{3}$ where $0\leq v\leq 5$, 4. 11. $-4g_{1}-4g_{2}+7g_{3}$, $-4g_{1}-45g_{2}+7g_{3}$ and $-209g_{1}-1890g_{2}+294g_{3}$, which are of minimal trace $2$, 1. 12. $-4g_{1}-vg_{2}+7g_{3}$ where $5\leq v\leq 9$, 2. 13. $-(5v+4)g_{1}-(45v+45)g_{2}+(7v+7)g_{3}$ where $1\leq v\leq 5$, 3. 14. $-(5v+184)g_{1}-(46v+1660)g_{2}+(7v+259)g_{3}$ where $0\leq v\leq 4$, which are of minimal trace $3$. The distribution of elements in Proposition 6.9 is not random, and it is highly influenced by elements of the codifferent, which we use in the determination of the minimal trace. In particular, for almost all elements, we can use one of four concrete elements of the codifferent as follows: first one for the triangle of elements in (1), second one for (2), (3), (8) and (12), third one for (4), (5), (9) and (13), and fourth one for (6), (7), (10) and (14). Only elements from (11) need different elements of the codifferent to get the minimal trace $2$ (and not $3$, which we would get by applying one of four elements of the codifferent mentioned above). We can also highlight that in comparison to the family with $p=3$, we obtain indecomposable integers with minimal trace $3$. Note that these elements are also indecomposable in the order ${\mathbb{Z}}[\rho]$. ## 7 Consequences In this section, we will use our knowledge of indecomposable integers for $p=3$ to derive several other results for these fields. First, we will focus on the minimal norm of algebraic integers in $K$ not associated with rational integers. Such a minimal norm for monogenic simplest cubic fields is $2a+3$ as was shown by Lemmermeyer and Pethö [23]. We will prove that except for a few cases of $a$, $2a+3$ is still minimal for our family with $p=3$. Then, we will find an upper bound on the norm of indecomposable integers in ${\mathcal{O}}_{K}$. Recall that in any totally real number field, the norm of indecomposable integers is bounded [3, 18]. So far, this upper bound was intensively studied for real quadratic fields [7, 11, 12, 35] and several families of monogenic totally real cubic number fields [34]. Let us highlight that our bound differs from the bound given in [34] for the simplest cubic fields with ${\mathcal{O}}_{K}={\mathbb{Z}}[\rho]$. Moreover, we will use indecomposable integers to show that the Pythagoras number of ${\mathcal{O}}_{K}$ is always $6$ in our subfamily. A similar result was derived by Tinková [33], who proved that the Pythagoras number of ${\mathbb{Z}}[\rho]$ is $6$ whenever $a\geq 3$. Moreover, in the final subsection, we will find both upper and lower bounds on the minimal number of variables of universal quadratic forms over ${\mathcal{O}}_{K}$. For that, we will follow the procedure developed in [16], which is based on the knowledge of indecomposable integers and their minimal traces after multiplication by elements of the codifferent. ### 7.1 The smallest norm Now, we will find the smallest norm of elements which are not associated with rational integers. To do that, we use the knowledge of the structure of indecomposables stated in Theorem 1.2. ###### Proposition 7.1. Let $\alpha\in{\mathcal{O}}_{K}$. Then either $|N(\alpha)|\geq\left\\{\begin{array}[]{ll}\frac{\Delta}{27}&\text{ if }a=21,30,48,\\\ 2a+3&\text{ if }a>48,\end{array}\right.$ or $\alpha$ is associated with a rational integer. Moreover, this lower bound is attained by some $\alpha\in{\mathcal{O}}_{K}$. ###### Proof. The smallest such norm can be attained either by an indecomposable integer or by a sum of two totally positive units. Let us first discuss the second case. Recall that even in this case, a system of fundamental units is formed by the pair $\rho$ and $\rho^{\prime}$, see Corollary 5.5. Every sum of two units is associated with an element of the form $1+\varepsilon$ where $\varepsilon$ is a totally positive unit. Moreover, we have $\varepsilon\neq 1$ since otherwise, $1+\varepsilon$ would be associated with the rational integer $2$. Therefore, by [16, Lemma 6.2], $\varepsilon$ is greater than $a^{2}$ in some embedding. Thus, $N(1+\varepsilon)=N(1)+N(\varepsilon)+\text{Tr}(\varepsilon)+\text{Tr}(\varepsilon\varepsilon^{\prime})>a^{2}.$ However, $a^{2}>2a+3$ for $a\geq 21$, and $2a+3$ is the norm of one of the indecomposable integers in ${\mathcal{O}}_{K}$. Thus, a sum of two units cannot have the smallest norm. Since the elements in Theorem 1.2 (iii) and (iv) are associated with conjugates of elements in 1.2 (ii), and the same is true for (vi)–(viii), it is enough to consider elements in (i), (ii), (v) and (vi). The norm of $\frac{1+\rho+\rho^{2}}{3}$ is equal to $\frac{a^{2}+3a+9}{27}$. The smallest norm among elements in (ii) is $2a+3$. Regarding elements in (vi), we have $N\left(-(r+1)g_{2}+g_{3}\right)=-r^{3}-3r^{2}+\frac{a^{2}+3a-18}{9}r+\frac{4a^{2}+12a+9}{27}$ where $0\leq r\leq\frac{a}{3}-1$. This norm is a cubic polynomial in $r$ with a negative leading coefficient. Moreover, for $r=-2$, it is equal to $-\frac{2a^{2}+6a-9}{27}<0$ for $a\geq 21$. It means that in the interval $\left[0,\frac{a}{3}-1\right]$, it attains its smallest value at $0$ or at $\frac{a}{3}-1$. For $r=0$, we obtain $\frac{4a^{2}+12a+9}{27}>2a+3$ for $a\geq 21$. Similarly, for $r=\frac{a}{3}-1$, we again obtain the norm $\frac{4a^{2}+12a+9}{27}$. Thus, the smallest norm is not attained by elements in (vi). Let us now focus on elements in (v). We can deduce that $N\left(-(2v+1)g_{1}-(v(a+3)+r+1)g_{2}+(3v+2)g_{3}\right)\\\ =-r^{3}-(av+3v-a)r^{2}-\left(av^{2}-a^{2}v-3av-3v+\frac{2a^{2}}{9}-\frac{a}{3}-1\right)r\\\ +v^{3}+\frac{a^{2}}{3}v^{2}-\frac{2a^{3}+7a^{2}+12a+9}{9}v-\frac{4a^{2}+12a+9}{27}$ where $0\leq v\leq\frac{a}{3}-1$ and $\frac{a}{3}+1\leq r\leq\frac{2a}{3}-v$. This is again a cubic polynomial in $r$ with a negative leading coefficient. If $r=\frac{a}{3}$, we obtain $v^{3}-\frac{a^{3}+3a+9}{9}v-\frac{a^{2}+3a+9}{27},$ which is a cubic polynomial in $v$ with a positive leading coefficient. Moreover, for $v=-1$, it is $\frac{2a^{2}+6a-9}{27}>0$; for $v=0$, it is equal to $-\frac{a^{2}+3a+9}{27}<0$, and for $v=\frac{a}{3}-1$, it equals $-\frac{10a^{2}-24a+9}{27}<0$ for $a\geq 21$. It gives that whenever $r=\frac{a}{3}$ and $0\leq v\leq\frac{a}{3}-1$, then our norm is negative. It implies that for a fixed $v$, the smallest norm is attained either for $r=\frac{a}{3}+1$, or for $r=\frac{2a}{3}-v$. Let us now discuss the case when $r=\frac{a}{3}+1$. Here, we get the norm $v^{3}-av^{2}+\frac{2a^{2}-3a-9}{9}v+\frac{2a^{2}+6a-9}{27}.$ For $v=\frac{a}{3}$, it is equal to $-\frac{a^{2}+3a+9}{27}<0$. Thus, it again attains its smallest value in one of the border points of our interval. For both of them, it is equal to $\frac{2a^{2}+6a-9}{27}>2a+3$ for $a\geq 26$. If $a=21$, then $\frac{2a^{2}+6a-9}{27}=37>19=\frac{a^{2}+3a+9}{27}$, so we can exclude this norm even in this case. Similarly, for $r=\frac{2a}{3}-v$, we obtain $-v^{3}-3v^{2}+\frac{a^{2}+3a-18}{9}v+\frac{2a^{2}+6a-9}{27}.$ For $v=-1$, it is equal to $-\frac{a^{2}+3a+9}{27}$; thus, the smallest norm is attained in a border point. For both of them, we again obtain the value $\frac{2a^{2}+6a-9}{27}$ which we have excluded before. Thus, we are left with the norms $\frac{a^{2}+3a+9}{27}$ and $2a+3$, and $\frac{a^{2}+3a+9}{27}>2a+3$ if $a\geq 53$. Between $21$ and $52$, only the cases $a=21,30,48$ belong to our family, and for them, the norm $\frac{a^{2}+3a+9}{27}$ is the smallest one. ∎ ### 7.2 The largest norm In this part, we will find a sharp upper bound on the norm of indecomposable integers in ${\mathcal{O}}_{K}$. As in the case of the smallest norm, it is enough to discuss elements in (i), (ii), (v) and (vi). We will start with elements in (ii) and (vi). ###### Lemma 7.2. Let $\alpha\in{\mathcal{O}}_{K}$ be an indecomposable integer. Then: 1. 1. If $\alpha$ is as in (ii), then $N(\alpha)\leq\frac{2a^{3}+9a^{2}+27a+27}{27}=\frac{(2a+3)\Delta}{27}$. 2. 2. If $\alpha$ is as in (vi), then $N(\alpha)<\frac{(2a+3)\Delta}{27}$. ###### Proof. To prove (1), it is not difficult to see that the largest norm among these elements is attained when $r=\frac{a}{3}$, for which $N(\alpha)=\frac{2a^{3}+9a^{2}+27a+27}{27}$. On the other hand, if $\alpha=-(r+1)g_{2}+g_{3}$, we have $g(r)=\frac{2a^{3}+9a^{2}+27a+27}{27}-N(\alpha)=r^{3}+3r^{2}-\frac{a^{2}+3a-18}{9}r+\frac{2a^{3}+5a^{2}+15a+18}{27}.$ This is a cubic polynomial in $r$ with a positive leading coefficient. Moreover, the discriminant of $g$ is negative, in which case $g$ has only one real root. This together with $g(0)=\frac{2a^{3}+5a^{2}+15a+18}{27}>0$ implies that $g(r)>0$ for all $0\leq r\leq\frac{a}{3}-1$, completing the proof. ∎ Note that $N\left(\frac{1+\rho+\rho^{2}}{3}\right)=\frac{a^{2}+3a+9}{27}<\frac{2a^{3}+9a^{2}+27a+27}{27}$ for $a\geq 21$. Therefore, we can also exclude the element $\frac{1+\rho+\rho^{2}}{3}$ from the consideration. It remains to discuss the elements in (v), which form the triangle $R_{4}$ in Section 6.4. Let us denote these integers as $\alpha(v,r)=-(2v+1)g_{1}-(v(a+3)+r+1)g_{2}+(3v+2)g_{3}$ where $0\leq v\leq\frac{a}{3}-1$ and $\frac{a}{3}+1\leq r\leq\frac{2a}{3}-v$. Furthermore, for the transformation $T_{1}$ in Section 6.5.3 we have $T_{1}(R_{4})=R_{4}$. In other words, $R_{4}$ also contains $\alpha^{\prime}\varepsilon_{1}$ and $\alpha^{\prime\prime}\varepsilon_{2}$ where $\varepsilon_{1}$ and $\varepsilon_{2}$ are concrete totally positive units. Thus, it suffices to consider only one third of this triangle, in particular, $R_{4}^{0}=\left\\{\alpha(v,r);0\leq v\leq\frac{a-3}{9}-1,\frac{a}{3}+1+v\leq r\leq\frac{2a}{3}-2v-1\right\\}\cup\left\\{\alpha\left(\frac{a-3}{9},\frac{a}{3}+1+\frac{a-3}{9}\right)\right\\}.$ The following lemma compares norms of elements contained in $R_{4}^{0}$. Note that the proof is almost the same as in [34, Lemma 3.2]. ###### Lemma 7.3. We have 1. 1. $N(\alpha(v,r))<N(\alpha(v+1,r))$ for all $0\leq v\leq\frac{a-3}{9}-1$ and $\frac{a}{3}+2+v\leq r\leq\frac{2a}{3}-2v-3$, 2. 2. $N\big{(}\alpha\big{(}v,\frac{a}{3}+1+v\big{)}\big{)}<N\big{(}\alpha\big{(}v+1,\frac{a}{3}+1+v+1\big{)}\big{)}$ for all $0\leq v\leq\frac{a-3}{9}-1$, 3. 3. $N\big{(}\alpha\big{(}v,\frac{2a}{3}-2v-1\big{)}\big{)}<N\big{(}\alpha\big{(}v+1,\frac{2a}{3}-2(v+1)-1\big{)}\big{)}$ for all $0\leq v\leq\frac{a-3}{9}-2$, 4. 4. $N\big{(}\alpha\big{(}v,\frac{2a}{3}-2v-2\big{)}\big{)}<N\big{(}\alpha\big{(}v+1,\frac{2a}{3}-2(v+1)-2\big{)}\big{)}$ for all $0\leq v\leq\frac{a-3}{9}-2$. Now we have everything we need to determine the upper bound on the norm of indecomposable integers in ${\mathcal{O}}_{K}$. ###### Proposition 7.4. If $\alpha$ is indecomposable in ${\mathcal{O}}_{K}$, then $|N(\alpha)|\leq\left\\{\begin{array}[]{ll}\frac{2a^{3}+9a^{2}+27a+27}{27}\text{ if }a=21,30,48,\\\ \frac{(a^{2}+3a+9)^{2}}{729}\text{ if }a>48,\end{array}\right.$ ###### Proof. As we have seen, the largest norm is either $\frac{2a^{3}+9a^{2}+27a+27}{27}$ found in Proposition 7.2 or is attained by some elements from $R_{4}^{0}$. Moreover, applying Lemma 7.3, we are left only with two candidates from $R_{4}^{0}$: $\alpha\left(\frac{a-3}{9}-1,\frac{a}{3}+2+\frac{a-3}{9}\right)$ and $\alpha\left(\frac{a-3}{9},\frac{a}{3}+1+\frac{a-3}{9}\right)$. Furthermore, $N\left(\alpha\left(\frac{a-3}{9}-1,\frac{a}{3}+2+\frac{a-3}{9}\right)\right)=\frac{a^{4}+6a^{3}-54a^{2}-189a+81}{729}\\\ <\frac{(a^{2}+3a+9)^{2}}{729}=N\left(\alpha\left(\frac{a-3}{9},\frac{a}{3}+1+\frac{a-3}{9}\right)\right).$ Moreover, $\frac{2a^{3}+9a^{2}+27a+27}{27}<\frac{(a^{2}+3a+9)^{2}}{729}$ for $a\geq 53$, which gives a few exceptional cases listed in the statement of the proposition. ∎ ### 7.3 Pythagoras number In this part, we will show that the Pythagoras number of ${\mathcal{O}}_{K}$ is $6$. Before that, let us briefly overview basic facts and results on this number. Consider a commutative ring ${\mathcal{O}}$, for which we define the following two sets: $\sum{\mathcal{O}}_{K}^{2}=\Bigg{\\{}\sum_{i=1}^{n}\alpha_{i}^{2};\alpha_{i}\in{\mathcal{O}}_{K}\text{ and }n\in{\mathbb{N}}\Bigg{\\}}$ and for each $m\in\mathbb{N}$ $\sum^{m}{\mathcal{O}}_{K}^{2}=\Bigg{\\{}\sum_{i=1}^{m}\alpha_{i}^{2};\alpha_{i}\in{\mathcal{O}}_{K}\Bigg{\\}}.$ Then by the Pythagoras number ${\mathcal{P}}({\mathcal{O}})$ of ${\mathcal{O}}$, we will mean the following infimum: ${\mathcal{P}}({\mathcal{O}})=\inf\Bigg{\\{}m\in{\mathbb{N}}\cup\\{\infty\\};\sum{\mathcal{O}}_{K}^{2}=\sum^{m}{\mathcal{O}}_{K}^{2}\Bigg{\\}}.$ For example, ${\mathcal{P}}({\mathbb{C}})={\mathcal{P}}({\mathbb{R}})=1$ and ${\mathcal{P}}({\mathbb{Z}})={\mathcal{P}}({\mathbb{Q}})=4$. Although there are many results on the Pythagoras number of fields, we will only summarize results in the case when ${\mathcal{O}}$ is an order of a totally real number field $K$. For them, we know that the value of ${\mathcal{P}}({\mathcal{O}})$ is finite but can attain arbitrarily large values [29]. Moreover, in this case, the Pythagoras number can be bounded by a function depending on the degree $d$ of a field $K$, and, furthermore, ${\mathcal{P}}({\mathcal{O}})\leq d+3$ for $2\leq d\leq 5$ [17]. So far, the Pythagoras number was fully determined for orders in real quadratic fields [28], and we have some partial results for real biquadratic fields [21, 9] and totally real cubic fields [20, 33]. In particular, we closely follow [33], where the second author showed that ${\mathcal{P}}({\mathbb{Z}}[\rho])=6$ for $\rho$ being a root of the polynomial $x^{3}-ax^{2}-(a+3)x-1$ with $a\geq 3$, i.e., for a concrete order of the simplest cubic fields. In cubic fields, we know that ${\mathcal{P}}({\mathcal{O}}_{K})\leq 6$ by the result of Kala and Yatsyna [17], and we will prove that ${\mathcal{P}}({\mathcal{O}}_{K})\geq 6$ for our subfamily of non-monogenic simplest cubic fields with the base $B_{3}(1,1)$. For that, we will use the method developed in [33], and our determination of the structure of indecomposable integers in ${\mathcal{O}}_{K}$. In particular, we will do the following: 1. 1. We will choose a suitable element $\gamma\in{\mathcal{O}}_{K}$ such that $\gamma$ can be written as a sum of six non-zero squares of elements in ${\mathcal{O}}_{K}$. 2. 2. We will determine all elements $\alpha$ from Theorem 1.2 and all units $\varepsilon\in{\mathcal{O}}_{K}$ such that $\gamma\succeq(\varepsilon\alpha)^{2}$. 3. 3. Using elements in (2), we will find all elements $\omega\in{\mathcal{O}}_{K}$ such that $\gamma\succeq\omega^{2}$. 4. 4. Considering elements in (3), we will discuss all possible decompositions of $\gamma$ as a sum of squares and show that for every such decomposition, we need at least $6$ squares. Note that in fact, elements $\varepsilon\alpha$ in (2) are indecomposable integers in other signatures. Since the simplest cubic fields have units of all signatures, they can be expressed as presented here, i.e., as $\varepsilon\alpha$ where $\varepsilon$ is a (not necessarily totally positive) unit and $\alpha$ is one of totally positive indecomposables listed in Theorem 1.2. For more details, see, for example, [33]. In our proof, we will fix $\gamma$ as $\displaystyle\gamma$ $\displaystyle=1+1+1+4+\left(\frac{a+6}{3}g_{1}+\frac{a}{3}g_{2}-g_{3}\right)^{2}+\left(\frac{5a+3}{9}g_{1}+\frac{2a+3}{3}g_{2}-2g_{3}\right)^{2}$ $\displaystyle=\frac{34a^{2}+15a+783}{81}+\frac{11a^{2}-29a-39}{27}\rho-\frac{11a-33}{27}\rho^{2}.$ Using estimates on $\rho,\rho^{\prime}$, and $\rho^{\prime\prime}$ (see Section 2; note that for $\rho^{\prime}$, we use bounds $-1-\frac{1}{a+1}<\rho^{\prime}<-1-\frac{1}{a+2}$ found in [33]), we obtain $\displaystyle\gamma$ $\displaystyle<\frac{13a^{2}}{81}+\frac{14a}{27}+\frac{197}{27}-\frac{26}{9a}<a^{2},$ $\displaystyle\gamma^{\prime}$ $\displaystyle<\frac{a^{4}+40a^{3}+1234a^{2}+4596a+4725}{81(a+2)^{2}}<a^{2},$ $\displaystyle\gamma^{\prime\prime}$ $\displaystyle<\frac{34a^{4}+186a^{3}+1167a^{2}+5178a+7497}{81(a+3)^{2}}<a^{2}$ for $a\geq 21$. We see that if $\omega\in{\mathbb{Z}}$ is such that $\gamma\succeq\omega^{2}$, then $|\omega|<a$. In contrast with [33], we will not determine the precise set of elements $\omega$ satisfying $\gamma\succeq\omega^{2}$, but we will significantly restrict the set of such elements. As we will see, this restriction will be enough to show that we need at least $6$ squares to express $\gamma$. First of all, in a series of lemmas, we will find all $\omega\in{\mathcal{O}}_{K}$ such that $\gamma\succeq\omega^{2}$. We will start with units and the indecomposable integer $g_{3}=\frac{1+\rho+\rho^{2}}{3}$. Recall that in the simplest cubic fields, all totally positive units are squares. Moreover, conjugates of $g_{3}$ are associated with $g_{3}$. ###### Lemma 7.5. We have 1. 1. if $\gamma\succeq\varepsilon^{2}$ where $\varepsilon$ is a unit in ${\mathcal{O}}_{K}$, then $\varepsilon^{2}=1$, 2. 2. if $\gamma\succeq\left(\varepsilon\frac{1+\rho+\rho^{2}}{3}\right)^{2}$ where $\varepsilon$ is a unit in ${\mathcal{O}}_{K}$, then $\varepsilon^{2}=\rho^{\prime 2}\rho^{\prime\prime 2}$. ###### Proof. Recall from Corollary 5.5 that a system of fundamental units is formed by the pair $\rho$ and $\rho^{\prime}$. Thus, we can use [16, Lemma 6.2], which says that if $\varepsilon^{2}\neq 1$ is a unit in ${\mathcal{O}}_{K}$, then $\varepsilon^{2}$ has a conjugate greater than $a^{2}$. Since $\gamma$ has all conjugates smaller than $a^{2}$, no totally positive unit except for $1$ can be totally smaller than $\gamma$. Considering the second part of the statement, we see that $\displaystyle\left(\frac{1+\rho+\rho^{2}}{3}\right)^{2}$ $\displaystyle>\frac{(a^{2}+3a+3)^{2}}{9}>a^{2},$ $\displaystyle\left(\frac{1+\rho^{\prime}+\rho^{\prime 2}}{3}\right)^{2}$ $\displaystyle>\frac{1}{9}\left(\frac{(a+3)^{2}}{(a+2)^{2}}-\frac{1}{a+1}\right)^{2},$ $\displaystyle\left(\frac{1+\rho^{\prime\prime}+\rho^{\prime\prime 2}}{3}\right)^{2}$ $\displaystyle>\frac{1}{9}\left(1-\frac{1}{a+2}+\frac{1}{(a+3)^{2}}\right)^{2}.$ Thus, we can immediately exclude that $\varepsilon^{2}=1$. Moreover, it can be easily shown that $a^{2}\left(\frac{1+\rho^{\prime}+\rho^{\prime 2}}{3}\right)^{2}>\gamma^{\prime}\qquad\text{ and }\qquad a^{4}\left(\frac{1+\rho^{\prime\prime}+\rho^{\prime\prime 2}}{3}\right)^{2}>\gamma^{\prime\prime}.$ Here, we use [16, Lemma 6.3]. It says that if $\varepsilon>a^{2}$ is a totally positive unit such that $\varepsilon\neq\rho^{2},\rho^{2}\rho^{\prime 2}$, then either $\varepsilon>a^{4}$, or one of $\varepsilon^{\prime}$ and $\varepsilon^{\prime\prime}$ is greater than $a^{2}$. Together with previous estimates, it implies $\varepsilon^{\prime\prime 2}=\rho^{2},\rho^{2}\rho^{\prime 2}$ in our case, from which we can exclude $\varepsilon^{\prime\prime 2}=\rho^{2}$, which does not have $\varepsilon^{2}<1$. ∎ We will proceed with indecomposables from ${\mathbb{Z}}[\rho]$. ###### Lemma 7.6. Let $\alpha$ be as in (ii), (iii) or (iv) of Theorem 1.2. Then there is no unit $\varepsilon\in{\mathcal{O}}_{K}$ such that $\gamma\succeq(\varepsilon\alpha)^{2}$. ###### Proof. Let $\alpha=-r\rho+\rho^{2}$ where $1\leq r\leq\frac{a}{3}$, i.e., $\alpha$ is one of elements in (ii). In this case, we have $\displaystyle(-r\rho+\rho^{2})^{2}$ $\displaystyle>((a+1)^{2}-r(a+2))^{2}\geq\frac{(2a^{2}+4a+3)^{2}}{9}>a^{2},$ $\displaystyle(-r\rho^{\prime}+\rho^{\prime 2})^{2}$ $\displaystyle>(r+1)^{2},$ $\displaystyle(-r\rho^{\prime\prime}+\rho^{\prime\prime 2})^{2}$ $\displaystyle>\frac{r^{2}}{(a+3)^{2}}.$ Therefore, we must have $\varepsilon^{2}<1$, $\varepsilon^{\prime 2}<a^{2}$ and $\varepsilon^{\prime\prime 2}>a^{2}$, which again follows from [16, Lemma 6.2]. Moreover, easily, $\frac{r^{2}a^{4}}{(a+3)^{2}}>\gamma,\gamma^{\prime},\gamma^{\prime\prime}$. Therefore, again by [16, Lemma 6.3], we can conclude that $\varepsilon^{\prime\prime 2}=\rho^{2},\rho^{2}\rho^{\prime 2}$. However, for both of these units, we have $(\varepsilon\alpha)^{2}>\gamma,\gamma^{\prime},\gamma^{\prime\prime}$. Note that by this, we have also excluded Cases (iii) and (iv). It follows from the fact that elements in (iii) and (iv) are associated with conjugates in (ii), and, thus, $(\varepsilon\alpha)^{2}>\gamma,\gamma^{\prime},\gamma^{\prime\prime}$ also covers elements originating from indecomposables in (iii) and (iv). ∎ Now we will discuss the other three lines of indecomposables. ###### Lemma 7.7. Let $a\geq 55$. Let $\alpha$ be as in (vi), (vii) or (viii) of Theorem 1.2, and let $\gamma\succeq(\varepsilon\alpha)^{2}$ for some unit $\varepsilon$. Then $\alpha$ is as in (vi) and $\varepsilon^{2}=\rho^{\prime 2}\rho^{\prime\prime 2}$. ###### Proof. Let $\alpha=-(r+1)g_{2}+g_{3}$ where $0\leq r\leq\frac{a}{3}-1$, i.e., $\alpha$ is one of elements in (vi). Then we have $\displaystyle\alpha^{2}$ $\displaystyle>\left(\frac{a^{2}}{3}-\frac{4}{3a}-r\Big{(}a+1+\frac{2}{a}\Big{)}\right)^{2}>\left(\frac{2a}{3}+\frac{1}{3}+\frac{2}{3a}\right)^{2}>\gamma,\gamma^{\prime},\gamma^{\prime\prime},$ $\displaystyle\alpha^{\prime 2}$ $\displaystyle>\left(\frac{4}{3}+\frac{4}{3(a+2)}+\frac{1}{3(a+2)^{2}}+r\Big{(}1+\frac{1}{a+2}\Big{)}\right)^{2}>1,$ $\displaystyle\alpha^{\prime\prime 2}$ $\displaystyle>\left(\frac{1}{3}+\frac{2}{3(a+3)}+\frac{1}{3(a+3)^{2}}+\frac{r}{a+3}\right)^{2}.$ Moreover, $a^{4}\alpha^{\prime\prime 2}>\gamma,\gamma^{\prime},\gamma^{\prime\prime}$. It implies that again $\varepsilon^{\prime\prime 2}=\rho^{2},\rho^{2}\rho^{\prime 2}$ by [16, Lemma 6.3]. We can exclude $\varepsilon^{\prime\prime 2}=\rho^{2}$ as $\varepsilon^{2}>1$. Now we will discuss all three conjugates of $(\rho^{\prime}\rho^{\prime\prime}\alpha)^{2}$. Since $(\rho\rho^{\prime}\alpha^{\prime\prime})^{2}>\gamma^{\prime}$, the conjugate $(\rho^{\prime\prime}\rho\alpha^{\prime})^{2}$ cannot be totally smaller than $\gamma$. If $r\geq\frac{a}{6}$, then $(\rho\rho^{\prime}\alpha^{\prime\prime})^{2}>\gamma$, and if $r<\frac{a}{6}$, then $(\rho^{\prime}\rho^{\prime\prime}\alpha)^{2}>\gamma^{\prime}$ for $a\geq 55$. This excludes the conjugate $(\rho\rho^{\prime}\alpha^{\prime\prime})^{2}$. Thus, we are left only with $(\rho^{\prime}\rho^{\prime\prime}\alpha)^{2}$, which states the lemma. Note that similarly as in the proof of Lemma 7.6, conjugates of elements in (vi) are associated with elements in (vii) and (viii), and in this proof, we have discussed all conjugates of $\alpha$ at the same time. ∎ In the end, we will look at the triangle of indecomposables in (v). For that, let us denote $\alpha(v,r)=-(2v+1)g_{1}-(v(a+3)+r+1)g_{2}+(3v+2)g_{3}$ where $0\leq v\leq\frac{a}{3}-1$ and $\frac{a}{3}+1\leq r\leq\frac{2a}{3}-v$. ###### Lemma 7.8. Let $a\geq 63$. Moreover, let $\gamma\succeq(\varepsilon\alpha(v,r))^{2}$ for some unit $\varepsilon\in{\mathcal{O}}_{K}$, $0\leq v\leq\frac{a}{3}-1$ and $\frac{a}{3}+1\leq r\leq\frac{2a}{3}-v$. Then either 1. 1. $(\varepsilon\alpha(v,r))^{2}=\rho^{\prime\prime 2}\rho^{2}\alpha(0,r)^{\prime 2}$ for some $\frac{a}{3}+1\leq r<\frac{5a-6}{12}$, or 2. 2. $(\varepsilon\alpha(v,r))^{2}=\rho^{2}\rho^{\prime 2}\alpha(0,r)^{\prime\prime 2}$ for some $\frac{5(a-3)}{9}\leq r\leq\frac{2a}{3}-1$. ###### Proof. As before, it suffices to restrict to a subset of the triangle of indecomposables in (v). In particular, we can discuss only elements in $R_{4}^{0}$. Thus, let $\alpha(v,r)\in R_{4}^{0}$, i.e., $0\leq v\leq\frac{a-3}{9}-1$ and $\frac{a}{3}+1+v\leq r\leq\frac{2a}{3}-2v-1$, or $v=\frac{a-3}{9}$ and $r=\frac{a}{3}+1+\frac{a-3}{9}$. Then $\displaystyle\alpha(v,r)^{2}$ $\displaystyle>\left(\frac{2a^{2}}{3}+a-\frac{2}{3a}-v\Big{(}a+4+\frac{4}{a}\Big{)}-r\Big{(}a+1+\frac{2}{a}\Big{)}\right)^{2}\geq\left(\frac{4a}{3}-\frac{1}{3}+\frac{4}{3a}\right)^{2}>a^{2},$ $\displaystyle\alpha(v,r)^{\prime 2}$ $\displaystyle>\left(\frac{2}{3}+\frac{5}{3(a+2)}+\frac{2}{3(a+2)^{2}}+v\Bigg{(}a+2+\frac{a+4}{a+2}+\frac{1}{(a+2)^{2}}\Bigg{)}+r\Bigg{(}1+\frac{1}{a+2}\Bigg{)}\right)^{2}$ $\displaystyle\hskip 142.26378pt\geq\left(\frac{a+5}{3}+\frac{a+8}{3(a+2)}+\frac{2}{3(a+2)^{2}}\right)^{2}>1,$ $\displaystyle\alpha(v,r)^{\prime\prime 2}$ $\displaystyle>\left(-\frac{1}{3}+\frac{1}{3(a+3)}+\frac{2}{3(a+3)^{2}}+v\Bigg{(}-1+\frac{a+2}{a+3}+\frac{1}{(a+3)^{2}}\Bigg{)}+\frac{r}{a+3}\right)^{2}\geq\frac{(a+5)^{2}}{9(a+3)^{4}}.$ Furthermore, $a^{6}\alpha(v,r)^{\prime\prime 2}>\gamma,\gamma^{\prime},\gamma^{\prime\prime}$. That means that our unit has to satisfy $\varepsilon^{2}<1$, $\varepsilon^{\prime 2}<a^{2}$ and $a^{2}<\varepsilon^{\prime\prime 2}<a^{6}$. This is fulfilled by units $\varepsilon^{\prime\prime 2}=\rho^{2}\rho^{\prime 2},\rho^{2}\rho^{\prime 4},\rho^{4}\rho^{\prime 2},\rho^{4}\rho^{\prime 4},\rho^{4}\rho^{\prime 6}$. Now, we will use the fact that if $\gamma\succeq\omega^{2}$ for some $\omega\in{\mathcal{O}}_{K}$, then necessarily $\text{Tr}(\gamma)\geq\text{Tr}(\omega^{2})$. It can be computed that $\text{Tr}(\gamma)=\frac{16a^{2}-24a+981}{27}$. Moreover, we have $\displaystyle\text{Tr}(\rho^{\prime 2}\rho^{\prime\prime 4}\alpha(v,r)^{2})$ $\displaystyle\geq\frac{a^{4}+12a^{3}+56a^{2}+132a+153}{9}>\text{Tr}(\gamma),$ $\displaystyle\text{Tr}(\rho^{\prime 4}\rho^{\prime\prime 6}\alpha(v,r)^{2})$ $\displaystyle\geq\frac{a^{4}+10a^{3}+41a^{2}+96a+126}{9}>\text{Tr}(\gamma),$ so we can exclude $\varepsilon^{2}=\rho^{\prime 2}\rho^{\prime\prime 4},\rho^{\prime 4}\rho^{\prime\prime 6}$. Moreover, if $v\geq 1$, then $\displaystyle\text{Tr}(\rho^{\prime 2}\rho^{\prime\prime 2}\alpha(v,r)^{2})$ $\displaystyle\geq\frac{17a^{2}+96a+369}{9}>\text{Tr}(\gamma),$ $\displaystyle\text{Tr}(\rho^{\prime 4}\rho^{\prime\prime 2}\alpha(v,r)^{2})$ $\displaystyle\geq\frac{26a^{2}+6a+153}{9}>\text{Tr}(\gamma),$ $\displaystyle\text{Tr}(\rho^{\prime 4}\rho^{\prime\prime 4}\alpha(v,r)^{2})$ $\displaystyle\geq\frac{41a^{2}+150a+234}{9}>\text{Tr}(\gamma),$ by which we have excluded all cases with $v\geq 1$. Let us now focus on the case when $v=0$. If $r\geq\frac{a}{3}+2$, then $\displaystyle\text{Tr}(\rho^{\prime 4}\rho^{\prime\prime 2}\alpha(0,r)^{2})$ $\displaystyle\geq\frac{26a^{2}+42a+126}{9}>\text{Tr}(\gamma),$ $\displaystyle\text{Tr}(\rho^{\prime 4}\rho^{\prime\prime 4}\alpha(0,r)^{2})$ $\displaystyle\geq\frac{26a^{2}+114a+234}{9}>\text{Tr}(\gamma).$ If $r=\frac{a}{3}+1$, we obtain $(\gamma-\beta)(\gamma^{\prime}-\beta^{\prime})+(\gamma-\beta)(\gamma^{\prime\prime}-\beta^{\prime\prime})+(\gamma^{\prime}-\beta^{\prime})(\gamma^{\prime\prime}-\beta^{\prime\prime})<0$ (9) where $\beta$ is any conjugate of $\rho^{\prime 4}\rho^{\prime\prime 2}\alpha\big{(}0,\frac{a}{3}+1\big{)}^{2}$ or $\rho^{\prime 4}\rho^{\prime\prime 4}\alpha\big{(}0,\frac{a}{3}+1\big{)}^{2}$. Note that (9) gives the second coefficient in the minimal polynomial of $\gamma-\beta$, which has to be nonnegative if $\gamma-\beta\succeq 0$. By this, we have excluded the units $\varepsilon^{2}=\rho^{\prime 4}\rho^{\prime\prime 2},\rho^{\prime 4}\rho^{\prime\prime 4}$. Thus, it remains to discuss the unit $\varepsilon^{2}=\rho^{\prime 2}\rho^{\prime\prime 2}$. We have $\rho^{\prime\prime 2}\rho^{2}\alpha(0,r)^{\prime 2}>\gamma^{\prime}$ for all $\frac{a}{3}+1\leq r\leq\frac{2a}{3}-1$, which excludes the conjugate $\rho^{\prime 2}\rho^{\prime\prime 2}\alpha(0,r)^{2}$. Let us now focus on the conjugate $\rho^{\prime\prime 2}\rho^{2}\alpha(0,r)^{\prime 2}$. If $r\geq\frac{5a-6}{12}$, we have $\rho^{\prime\prime 2}\rho^{2}\alpha(0,r)^{\prime 2}>\frac{1}{(a+3)^{2}}(a+1)^{2}\frac{(5a^{3}+27a^{2}+52a+44)^{2}}{144(a+2)^{4}}>\gamma$ for $a\geq 63$. Therefore, we must have $\frac{a}{3}+1\leq r<\frac{5a-6}{12}$, which gives the first case in our statement. For the conjugate $\rho^{2}\rho^{\prime 2}\alpha(0,r)^{\prime\prime 2}$, we similarly see that for $\frac{a}{3}+1\leq r\leq\frac{5(a-3)}{9}-1$, we obtain $\rho^{\prime 2}\rho^{\prime\prime 2}\alpha(0,r)^{2}>\left(1+\frac{1}{a+2}\right)^{2}\frac{1}{(a+3)^{2}}\frac{(a^{3}+28a^{2}+14a+42)^{2}}{81a^{2}}>\gamma^{\prime}$ for $a\geq 34$. This implies $\frac{5(a-3)}{9}\leq r\leq\frac{2a}{3}-1$, i.e., the second case in our statement. ∎ In Lemmas 7.5, 7.6, 7.7 and 7.8, we have found some of the squares totally smaller than $\gamma$. There can exist other squares $\omega^{2}\in{\mathcal{O}}_{K}$ satisfying $\gamma\succeq\omega^{2}$; however, these squares can be determined from elements derived so far. In particular, every such $\omega$ can be written as $\omega=\sum_{i=1}^{n}\beta_{i}$ where elements $\beta_{i}$ are square roots of squares found in above lemmas. Moreover, we can assume that all $\beta_{i}$ in this decomposition have the same signature. By the signature of an element $\alpha\in K$, we mean the triple $(\text{sgn}(\alpha),\text{sgn}(\alpha^{\prime}),\text{sgn}(\alpha^{\prime\prime}))$, where sgn is the signum function. In the following, we will also replace $1$ and $-1$ in this triple by symbols $+$ and $-$. Therefore, $\gamma\succeq\omega^{2}=\left(\sum_{i=1}^{n}\beta_{i}\right)^{2}\succeq(\beta_{i}+\beta_{j})^{2}\succeq\beta_{i}^{2}$ for every $1\leq i,j\leq n$. Thus, to determine all the remaining squares totally smaller than $\gamma$, it is enough for each possible signature 1. (1) to take all square roots of squares from Lemmas 7.5, 7.6, 7.7 and 7.8 with this signature, 2. (2) to study the element $(\beta_{i}+\beta_{j})^{2}$ for all pairs of elements $\beta_{i}$ and $\beta_{j}$ from 1. * • If $(\beta_{i}+\beta_{j})^{2}\succ\gamma$ for all such pairs, then we do not get any additional square totally smaller than $\gamma$. In the following proof, we will sometimes show that $\beta_{i}^{2},\beta_{j}^{2}\succ\frac{1}{2}\gamma$, from which $(\beta_{i}+\beta_{j})^{2}\succ\gamma$ follows. * • Otherwise, we can get more squares. However, we will see that for $\gamma$, this can occur only in one trivial case. Moreover, $\alpha$ and $-\alpha$ produce the same square, so it suffices to study only one of every pair of (mutually opposite) signatures; e.g., instead of studying both elements with signatures $(+,+,+)$ and $(-,-,-)$, it is enough to examine those with the signature $(+,+,+)$. This problem is related to the study of indecomposability of algebraic integers which are not totally positive. The above summary is not detailed, and the full description can be found in [33]. ###### Lemma 7.9. Let $a\geq 63$ and let $\gamma\succeq\omega^{2}$ where $\omega^{2}\in{\mathcal{O}}_{K}$ is not as in Lemmas 7.5, 7.6, 7.7 and 7.8. Then $\omega\in{\mathbb{Z}}$. ###### Proof. Recall that the unit $\rho$ has the signature $(+,-,-)$. Thus, the so far found elements, whose squares are totally smaller than $\gamma$, have the following signatures: 1. 1. $1$ has the signature $(+,+,+)$, 2. 2. $\rho^{\prime}\rho^{\prime\prime}\frac{1+\rho+\rho^{2}}{3}$ has the signature $(+,-,-)$, 3. 3. $\rho^{\prime}\rho^{\prime\prime}\left(-(r+1)g_{2}+g_{3}\right)$ where $0\leq r\leq\frac{a}{3}-1$ have the signature $(+,-,-)$, 4. 4. $\rho^{\prime\prime}\rho\alpha(0,r)^{\prime}$ where $\frac{a}{3}+1\leq r<\frac{5a-6}{12}$ have the signature $(-,-,+)$, 5. 5. $\rho\rho^{\prime}\alpha(0,r)^{\prime\prime}$ where $\frac{5(a-3)}{9}\leq r\leq\frac{2a}{3}-1$ have the signature $(-,+,-)$. From elements in (1), we can get only squares $\omega^{2}$ such that $\omega\in{\mathbb{Z}}$, which are included in the statement of the lemma. We will proceed with elements in (2) and (3), which have the same signature. Let $\alpha_{1}=-(r_{1}+1)g_{2}+g_{3}$ and $\alpha_{2}=-(r_{2}+1)g_{2}+g_{3}$ where $-1\leq r_{1},r_{2}\leq\frac{a}{3}-1$. Note for $r_{i}=-1$, we obtain the element $\frac{1+\rho+\rho^{2}}{3}$ from (2). Then $(\rho\rho^{\prime}\alpha_{1}^{\prime\prime}+\rho\rho^{\prime}\alpha_{2}^{\prime\prime})^{2}=\rho^{2}\rho^{\prime 2}(\alpha_{1}^{\prime\prime 2}+2\alpha_{1}^{\prime\prime}\alpha_{2}^{\prime\prime}+\alpha_{2}^{\prime\prime 2})\\\ >4(a+1)^{2}\left(1+\frac{1}{a+2}\right)^{2}\frac{1}{9}\left(1-\frac{1}{a+2}+\frac{1}{(a+3)^{2}}\right)^{2}>\gamma^{\prime\prime}.$ Thus, the square of a sum of elements in (2) and (3) cannot be totally smaller than $\gamma$. Let us now focus on elements in (4). For them, we have $\rho^{\prime\prime 2}\rho^{2}\alpha(0,r)^{\prime 2}>\frac{1}{(a+3)^{2}}(a+1)^{2}\left(\frac{a+5}{3}+\frac{a+8}{3(a+2)}+\frac{2}{3(a+2)^{2}}\right)^{2}>\frac{1}{2}\gamma$ for $\frac{a}{3}+1\leq r<\frac{5a-6}{12}$. That means that the square of a sum of two elements in (4) cannot be totally smaller than $\gamma$. Likewise, if we consider elements in (5), we obtain $\rho^{\prime\prime 2}\rho^{2}\alpha(0,r)^{\prime 2}>\frac{1}{(a+3)^{2}}(a+1)^{2}\left(\frac{5a}{9}-1+\frac{5a}{9(a+2)}+\frac{2}{3(a+2)^{2}}\right)^{2}>\frac{1}{2}\gamma^{\prime\prime}$ for $\frac{5(a-3)}{9}\leq r\leq\frac{2a}{3}-1$. That completes the proof. ∎ Now, we are able to prove that the Pythagoras number of ${\mathcal{O}}_{K}$ is $6$ as for ${\mathbb{Z}}[\rho]$. ###### Proposition 7.10. We have ${\mathcal{P}}({\mathcal{O}}_{K})=6$. ###### Proof. If $a<63$, i.e, for $a=21,30,48,57$, we can use a computer program to check that we need at least $6$ squares to express $\gamma$. Mathematica notebook with our code is available at https://sites.google.com/view/tinkovamagdalena/research. Thus, let us assume that $a\geq 63$. In Lemmas 7.5, 7.6, 7.7, 7.8 and 7.9, we have determined all squares which can be totally smaller than $\gamma$. We have obtained the following elements: 1. 1. $n^{2}\in{\mathbb{Z}}$ where $n^{2}<a^{2}$, 2. 2. $\rho^{\prime 2}\rho^{\prime\prime 2}\left(\frac{1+\rho+\rho^{2}}{3}\right)^{2}=\frac{a^{2}+4a+6}{9}g_{1}+\frac{a^{2}+2a+3}{9}g_{2}-\frac{a}{3}g_{3}$, 3. 3. $\rho^{\prime 2}\rho^{\prime\prime 2}\left(-(r+1)g_{2}+g_{3}\right)^{2}=\frac{a^{2}+10a+33+(6a+36)r+9r^{2}}{9}g_{1}+\frac{a^{2}+8a+3+6ar}{9}g_{2}-\frac{a+6+6r}{3}g_{3}$ where $0\leq r\leq\frac{a}{3}-1$, 4. 4. $\rho^{\prime\prime 2}\rho^{2}\left(-g_{1}^{\prime}-(r+1)g_{2}^{\prime}+2g_{3}^{\prime}\right)^{2}=\frac{4a^{2}+10a+6-(12a+18)r+9r^{2}}{9}g_{1}+\frac{3a^{2}+2a+3-6ar}{9}g_{2}-(a-2r)g_{3}$ where $\frac{a}{3}+1\leq r<\frac{5a-6}{12}$, 5. 5. $\rho^{2}\rho^{\prime 2}\left(-g_{1}^{\prime\prime}-(r+1)g_{2}^{\prime\prime}+2g_{3}^{\prime\prime}\right)^{2}=\frac{15-8a+36r+9r^{2}}{9}g_{1}+\frac{3-4a-4a^{2}+(12a+18)r}{9}g_{2}+\frac{4a-3-12r}{3}g_{3}$ where $\frac{5(a-3)}{9}\leq r\leq\frac{2a}{3}-1$. Recall that only these elements can appear in a decomposition of $\gamma$ to a sum of squares. First of all, let us note that for all these elements, the coefficients before $g_{2}$ are nonnegative, and the coefficients before $g_{3}$ are negative or zero. This is clear for Cases (1), (2) and (3). For (4), we see that $\frac{3a^{2}+2a+3-6ar}{9}>\frac{a^{2}+10a+6}{18}>0$ and $-(a-2r)<-1-\frac{a}{6}<0.$ Similarly, for (5), we obtain $\frac{3-4a-4a^{2}+(12a+18)r}{9}\geq\frac{8a^{2}-42a-81}{27}>0$ and $\frac{4a-3-12r}{3}\leq-\frac{8a-51}{9}<0.$ Recall that $\displaystyle\gamma$ $\displaystyle=\frac{34a^{2}+15a+783}{81}+\frac{11a^{2}-29a-39}{27}\rho-\frac{11a-33}{27}\rho^{2}$ $\displaystyle=\frac{2(17a^{2}+24a+342)}{81}g_{1}+\frac{11a^{2}-18a-72}{27}g_{2}-\frac{11(a-3)}{9}g_{3}.$ If $a$ is even, then $\frac{11(a-3)}{9}$ is odd. If we consider the coefficient before $g_{3}$ for elements in (1)–(5), we see that this coefficient is odd only for squares in (5) under the assumption that $a$ is odd. Thus, at least one element from (5) must appear in every decomposition of $\gamma$ to a sum of squares. Similarly, let $a$ be odd. Then $\frac{11a^{2}-18a-72}{27}$, i.e., the coefficient before $g_{2}$ of $\gamma$, is odd. If we again consider the squares in (1)–(5), only elements in (5) has an odd coefficient before $g_{2}$ for odd $a$. That implies that for all possible $a\geq 63$, at least one element from (5) appear in every decomposition of $\gamma$ to a sum of squares. Considering elements from (5), we can conclude that $-\frac{11(a-3)}{9}=\frac{4a-3-12r}{3}$ only if $r=\frac{23a-42}{36}$. For this concrete $r$, the coefficient before $g_{2}$ is $\frac{3-4a-4a^{2}+(12a+18)r}{9}=\frac{11a^{2}}{27}-\frac{13a}{18}-2.$ Moreover, $\frac{11a^{2}}{27}-\frac{13a}{18}-2=\frac{11a^{2}-18a-72}{27}$ only if $a=12$, which does not give a simplest cubic field from our subfamily. Thus, if this concrete element $r=\frac{23a-42}{36}$ were in the decomposition of $\gamma$, then some other summand would have to have * • a non-zero coefficient before $g_{2}$, * • a zero coefficient before $g_{3}$. However, these conditions are not both satisfied for any element from (1)–(5). Thus, the element from (5), which appears in the decomposition of $\gamma$, satisfies $-\frac{11(a-3)}{9}\neq\frac{4a-3-12r}{3}.$ That means that we need at least one more element from (2)–(5) to express $\gamma$. We know that for all elements from (5), the coefficient before $g_{3}$ satisfies $\frac{4a-3-12r}{3}\leq-\frac{8a-51}{9}.$ Thus, the other summand from (2)–(5) has this coefficient between zero and $-\frac{11(a-3)}{9}+\frac{8a-51}{9}=-\frac{a}{3}-2.$ That is true for 1. (a) the element $\frac{a^{2}+4a+6}{9}g_{1}+\frac{a^{2}+2a+3}{9}g_{2}-\frac{a}{3}g_{3}$ from (2), 2. (b) the element with $r=0$ from (3), 3. (c) elements from (4). Let us now discuss the case when at least two elements from a–c appeared in the decomposition of $\gamma$. In this case, it suffices to realize that a sum of two elements from a–c never has the coefficient before $g_{3}$ between $-\frac{a}{3}-2$ and $0$, which we have derived to be a necessary condition in this part of the proof. This follows from the fact that the largest coefficient before $g_{3}$ is achieved by the element in c with the largest $r$, where we have $-(a-2r)<-1-\frac{a}{6}.$ So its twice is $-2(a-2r)<-\frac{a}{3}-2.$ It implies that in every decomposition of $\gamma$, there must appear exactly one element from (5), and exactly one element from a–c. The remaining summands belong to (1). Let us now focus on elements from c, i.e., elements from (4), and let us assume that in a decomposition of $\gamma$, there is one element from (4) and one from (5). If we consider coefficients before $g_{2}$ and $g_{3}$, we can obtain the following system of equations: $\frac{11a^{2}-18a-72}{27}=\frac{3a^{2}+2a+3-6ar_{1}}{9}+\frac{3-4a-4a^{2}+(12a+18)r_{2}}{9}$ and $-\frac{11(a-3)}{9}=-(a-2r_{1})+\frac{4a-3-12r_{2}}{3}$ where $\frac{a}{3}+1\leq r_{1}<\frac{5a-6}{12}$ and $\frac{5(a-3)}{9}\leq r_{2}\leq\frac{2a}{3}-1$. This system has the solution $r_{1}=\frac{a}{3}-1$ and $r_{2}=\frac{5(a-3)}{9}$. However, $r_{1}=\frac{a}{3}-1$ does not belong to our interval. Thus, the elements from c cannot appear in any decomposition of $\gamma$ to a sum of squares. We will proceed with a. In this case, the coefficient before $g_{3}$ leads to the equation $-\frac{11(a-3)}{9}=-\frac{a}{3}+\frac{4a-3-12r}{3}$ with the solution $r=\frac{10a-21}{18}$. However, regarding the coefficient before $g_{2}$, we obtain $\frac{a^{2}+2a+3}{9}+\frac{3-4a-4a^{2}+(12a+18)r}{9}=\frac{11a^{2}-18a-45}{27}\neq\frac{11a^{2}-18a-72}{27}$ for every $a$. Thus, only the element in b can appear in every decomposition of $\gamma$. Moreover, $r=\frac{5(a-3)}{9}$ for the element from (5) in this decomposition, and, thus, $\gamma=\left(\frac{a^{2}+10a+33}{9}g_{1}+\frac{a^{2}+8a+3}{9}g_{2}-\frac{a+6}{3}g_{3}\right)\\\ +\left(\frac{25a^{2}-42a-180}{81}g_{1}+\frac{8a^{2}-42a-81}{27}g_{2}-\frac{8a-51}{9}g_{3}\right)+7.$ To express $7$, we need at least $4$ squares. Therefore, we need at least $6$ squares to express $\gamma$, which gives ${\mathcal{P}}({\mathcal{O}}_{K})=6$ as desired. ∎ ### 7.4 Universal quadratic forms In this part, we will use the results on indecomposables in ${\mathcal{O}}_{K}$ to get information on universal quadratic forms over ${\mathcal{O}}_{K}$. Before that, we will summarize several basic facts on quadratic forms over number fields. Let us consider quadratic form of the form $Q(x_{1},x_{2},\ldots,x_{n})=\sum_{1\leq i\leq j\leq n}a_{ij}x_{i}x_{j}$ where $a_{ij}\in{\mathcal{O}}_{K}$. We say that $Q$ is totally positive definite if $Q(\gamma_{1},\gamma_{2},\ldots,\gamma_{n})\in{\mathcal{O}}_{K}^{+}$ for every $n$-tuple $\gamma_{1},\ldots,\gamma_{n}\in{\mathcal{O}}_{K}$ such that elements $\gamma_{i}$ are not all equal to zero. The quadratic form $Q$ is classical if $2$ divides $a_{ij}$ for all $i\neq j$. The quadratic form $Q$ is called diagonal if $a_{ij}=0$ whenever $i\neq j$. And finally, $Q$ is universal if it represents all totally positive algebraic integers in $K$. In this part, we will assume that all quadratic forms are totally positive definite. To bound the minimal number of variables of universal quadratic forms over our subfamily of the non-monogenic simplest cubic fields, we will use the method developed in [16, Section 7]. This approach is based on the knowledge of indecomposable integers in $K$. In this part, we will use the following notation. Let $S$ denote the set of indecomposable integers in ${\mathcal{O}}_{K}$ up to multiplication by squares of units in ${\mathcal{O}}_{K}$. Moreover, let us assume that there exist $\delta\in{\mathcal{O}}_{K}^{\vee,+}$ and $\alpha_{1},\alpha_{2},\ldots,\alpha_{n}\in{\mathcal{O}}_{K}^{+}$ such that $\text{Tr}(\alpha_{i}\delta)=1$ for all $1\leq i\leq n$. Then, a more general result for totally real number fields derived in [16, Section 7] implies the following theorem for cubic fields: ###### Theorem 7.11 ([16, Section 7]). Let $K$ be a totally real cubic field. 1. 1. There exist a diagonal universal quadratic form over ${\mathcal{O}}_{K}$ with ${\mathcal{P}}({\mathcal{O}}_{K})\\#S$ variables. 2. 2. Every classical universal quadratic form over ${\mathcal{O}}_{K}$ has at least $\frac{n}{3}$ variables. 3. 3. If $n\geq 240$, then every (non-classical) universal quadratic form over ${\mathcal{O}}_{K}$ has at least $\frac{\sqrt{n}}{3}$ variables. Now we will use Theorem 7.11 and results on indecomposable integers to derive bounds on the minimal number of variables of universal quadratic forms over our nice subfamily of the simplest cubic fields. ###### Proposition 7.12. Let $K$ be a simplest cubic field such that $a\equiv 3,21\;(\textup{mod 27})$, $a>12$ and $\frac{\Delta}{27}$ is square-free. Then: 1. 1. There exists a diagonal universal quadratic form over ${\mathcal{O}}_{K}$ with $\frac{a^{2}+3a}{3}+12a+12$ variables. 2. 2. Every classical universal quadratic form over ${\mathcal{O}}_{K}$ has at least $\frac{a^{2}+3a}{54}$ variables. 3. 3. If $a>64$, then every (non-classical) universal quadratic form over ${\mathcal{O}}_{K}$ has at least $\frac{\sqrt{a^{2}+3a}}{9\sqrt{2}}$ variables. ###### Proof. Regarding the first part of the statement, we know that ${\mathcal{P}}({\mathcal{O}}_{K})=6$ by Proposition 7.10. Moreover, in $K$, every totally positive unit is a square. Thus $\\#S=\frac{a^{2}+3a}{18}+2a+2$, which is the number of indecomposable integers up to multiplication by totally positive units. To get the largest possible value in (2) and (3), it is convenient to consider the triangle of indecomposables $\alpha$ in (v) of Theorem 1.2, for which we have found one totally positive element $\delta$ of the codifferent satisfying $\text{Tr}(\alpha\delta)=1$. In this case, $n=\frac{a^{2}+3a}{18}$, which gives parts (2) and (3) of the proposition. ∎ ## Acknowledgements The authors want to thank Vítězslav Kala and Giacomo Cherubini for enriching discussions and helpful suggestions on the contents of this work. The authors also wish to express their thanks to László Remete and Stéphane Louboutin for their quick answers. The first author was supported by Czech Science Foundation GAČR, grant 21-00420M, and by Charles University Research Centre program UNCE/SCI/022. The second author was supported by Czech Science Foundation GAČR, grant 22-11563O. ## References * [1] V. Blomer and V. Kala “Number fields without universal n-ary quadratic forms” In _Math. Proc. Cambridge Philos. Soc._ 159, 2015, pp. 239–252 * [2] V. Blomer and V. Kala “On the rank of universal quadratic forms over real quadratic fields” In _Doc. Math._ 23, 2018, pp. 15–34 * [3] H. Brunotte “Zur Zerlegung totalpositiver Zahlen in Ordnungen totalreeller algebraischer Zahlkörper” In _Arch. Math. (Basel)_ 41 (6), 1983, pp. 502–503 * [4] F. Cánovas Orvay “On cyclic cubic fields.” In _Extracta Mathematicae_ 6.1, 1991, pp. 28–30 URL: http://eudml.org/doc/39912 * [5] M. Čech et al. “Universal quadratic forms and indecomposables over biquadratic fields” In _Math. Nachr._ 292, 2019, pp. 540–555 * [6] W.. Chan, M.-H. Kim and S. Raghavan “Ternary universal integral quadratic forms over real quadratic fields” In _Japan. J. Math._ 22, 1996, pp. 263–273 * [7] A. Dress and R. Scharlau “Indecomposable Totally Positive Numbers in Real Quadratic Orders” In _J. Number Theory_ 14, 1982, pp. 292–306 * [8] A.. Earnest and A. Khosravani “Universal positive quaternary quadratic lattices over totally real number fields” In _Mathematika_ 44, 1997, pp. 342–347 * [9] Z. He and Y. Hu “Pythagoras number of quartic orders containing $\sqrt{2}$” Preprint., 2022 * [10] J.. Hsia, Y. Kitaoka and M. Kneser “Representations of positive definite quadratic forms” In _J. Reine Angew. Math._ 301, 1978, pp. 132–141 * [11] S.. Jang and B.. Kim “A refinement of the Dress-Scharlau theorem” In _J. Number Theory_ 158, 2016, pp. 234–243 * [12] V. Kala “Norms of indecomposable integers in real quadratic fields” In _J. Number Theory_ 166, 2016, pp. 193–207 * [13] V. Kala “Number fields without universal quadratic forms of small rank exist in most degrees” To appear in Math. Proc. Cambridge Philos. Soc., 2022 * [14] V. Kala “Universal quadratic forms and elements of small norm in real quadratic fields” In _Bull. Aust. Math. Soc._ 94, 2016, pp. 7–14 * [15] V. Kala and J. Svoboda “Universal quadratic forms over multiquadratic fields” In _Ramanujan J._ 48, 2019, pp. 151–157 * [16] V. Kala and M. Tinková “Universal quadratic forms, small norms and traces in families of number fields” To appear in Int. Math. Res. Not. IMRN, 2022 * [17] V. Kala and P. Yatsyna “Lifting problem for universal quadratic forms” In _Adv. Math._ 337, 2021, pp. 107497 * [18] V. Kala and P. Yatsyna “On Kitaoka’s conjecture and lifting problem for universal quadratic forms” To appear in Bull. Lond. Math. Soc., 2022 * [19] T. Kashio and R. Sekigawa “The characterization of cyclic cubic fields with power integral bases” In _Kodai Mathematical Journal_ 44, 2021, pp. 290–306 * [20] J. Krásenský “A cubic ring of integers with the smallest Pythagoras number” In _Arch. Math (Basel)_ 118, 2022, pp. 39–48 * [21] J. Krásenský, M. Raška and E. Sgallová “Pythagoras numbers of orders in biquadratic fields” To appear in Expo. Math., 2022 * [22] J. Krásenský, M. Tinková and K. Zemková “There are no universal ternary quadratic forms over biquadratic fields” In _Proc. Edinb. Math. Soc._ 63 (3), 2020, pp. 861–912 * [23] F. Lemmermeyer and A. Pethö “Simplest Cubic Fields” In _Manuscripta Math._ 88, 1995, pp. 53–58 * [24] H. Maaß “Über die Darstellung total positiver Zahlen des Körpers $R(\sqrt{5})$ als Summe von drei Quadraten” In _Abh. Math. Sem. Univ. Hamburg_ 14, 1941, pp. 185–191 * [25] T. Nagell “Zur Arithmetik der Polynome” In _Abhandl. Math. Sem. Hamburg_ 1, 1922, pp. 178–193 * [26] W. Narkiewicz “Elementary and Analytic Theory of Algebraic Numbers” PWN/POLISH SCIENTIFIC PUBLISHE, 1974 * [27] O. Perron “Die Lehre von den Kettenbrüchen” B. G. Teubner, 1913 * [28] M. Peters “Quadratische Formen über Zahlringen” In _Acta Arith._ 24, 1973, pp. 157–164 * [29] R. Scharlau “On the Pythagoras number of orders in totally real number fields” In _J. Reine Angew. Math._ 316, 1980, pp. 208–210 * [30] D. Shanks “The Simplest Cubic Fields” In _Math. Comput._ 28, 1974, pp. 1137–1152 * [31] C.. Siegel “Sums of $m$-th powers of algebraic integers” In _Annals of Mathematics_ 46, 1945, pp. 313–339 * [32] E. Thomas and A.. Vasquez “On the Resolution of Cusp Singularities and the Shintani Decomposition in Totally Real Cubic Number Fields” In _Math. Ann._ 247, 1980, pp. 1–20 * [33] M. Tinková “On the Pythagoras number of the simplest cubic fields” Preprint, 2022 * [34] M. Tinková “Trace and norm of indecomposable integers in cubic orders” To appear in Ramanujan J., 2022 * [35] M. Tinková and P. Voutier “Indecomposable integers in real quadratic fields” In _J. Number Theory_ 212, 2020, pp. 458–482 * [36] P. Yatsyna “A lower bound for the rank of a universal quadratic form with integer coefficients in a totally real field” In _Comment. Math. Helvet._ 94, 2019, pp. 221–239
# Electronic spectrum and optical properties of Y-shaped Kekulé-patterned graphene: Band nesting resonance as an optical signature Yawar Mohammadi E-mail address<EMAIL_ADDRESS> ###### Abstract Employing tight-binding model we investigate the effects of a uniform Y-shaped Kekulé lattice distortion on the electronic spectrum and optical conductivity of graphene. We derive a low-energy effective Hamiltonian which is found to be in excellent agreement with one calculated from a diagonalization of the full tight-binding Hamiltonian. Then using the low-energy Hamiltonian and Kubo formula we obtain an analytical expression for the real part of the optical conductivity used to explore the effects of chemical potential, temperature and on-site and hopping energy deviations in details. In particular we find that Y-shaped Kekué-patterned graphene at finite chemical potential displays a large optical response called band nesting resonance. This effect is shown to be robust against increasing temperature, facilitating its detection as an optical signature for the Y-shaped Kekulé distortion even at room temperature. Department of Physics, Farhangian University, Tehran, Iran _Keywords_ : Kekulé-patterned graphene; Kubo Formula; Optical conductivity; Band nesting resonance. ## 1 Introduction Recently Gutiérrez et. al [1] have realized experimentally a Y-shaped periodic alternation of weak and strong C-C bonds in a superlattice of graphene grown epitaxially onto copper substrate, called Y-shaped Kekulé lattice distortion with reference to the Kekulé dimerization in a benzene ring. Using density functional theory calculations they showed [1] that removing copper atoms from the topmost surface allows the system to relax created an alternating network of C-C bonds similar to the Y-shaped Kekulé lattice distortion. Gamayun et.al introduced a tight-binding Hamiltonian to explore the effects of a uniform Y-shaped Kekulé lattice on the properties of graphene, in which this effect has been taken into account by adding a nearest-neighbor hopping energy deviation to the minimal tight-binding Hamiltonian of graphene. According to this Hamiltonian the uniform Y-shaped Kekulé distortion locks the valley degree of freedom of the charge carriers to their direction of motion, resulting in the breaking of the valley degeneracy of graphene and emerging two species of massless Dirac fermions[2]. Some researchers, using this Hamiltonian, have investigated different properties of Y-shaped Kekulé- patterned graphene. Wu et. al[3] proposed a type of valley field-effect transistors for Y-shaped Kekulé-patterned graphene and explored tuning valley pseudomagnetoresistance via an electric field. Andrade et. al[4] studied the effects of uniaxial strain on the band structure of all types of Kekulé- distorted graphene. In other works the effect of the Y-shaped Kekulé distortion on the electronic transport properties[5, 6], dynamical polarization[7], magneto-optical conductivity[8] and quantum Hall effect[9] in graphene has been investigated. In this paper we consider the electronic spectrum and the optical conductivity of such a system with particular emphasis on the effects of the on-site energy deviation, which in the previous researches has been neglected[3, 4, 5, 6, 7, 8, 9, 10, 11] while as shown here it leads to fascinating optical properties. The structure of our paper is organized as follows. In section 2, we introduce our tight-binding model in which the effects of the Y-shaped Kekulé distortion is taken into account by including both on-site and hopping energy deviations in the minimal tight-binding Hamiltonian of graphene. Then we derive a low- energy effective Hamiltonian in subsection 2.1 which is found to be in excellent agreement with one calculated from a diagonalization of the full tight-binding Hamiltonian. The main effect of the on-site energy deviation on the band structure is that a set of bands gains an effective mass and a shift in energy, thus lifting the degeneracy of the conduction bands at the Dirac point. In the next subsection we obtain an analytical expression for the real part of the optical conductivity using Kubo formula, which is used to explore the effects of temperature, chemical potential and on-site and nearest- neighbor hopping energy deviations on the optical conductivity. Then we present our results for the optical conductivity as a function of the photon energy in next section. In particular we find that in the limit of zero chemical potential the optical conductivity displays a dip-peak structure located at the photon energy corresponding to 2 times the effective on-site energy deviation. Furthermore, it is shown that at finite chemical potential Y-shaped Kekué-patterned graphene exhibits a large optical response caused by nesting of the conduction or valance bands. This effect is shown to be robust with respect to increasing temperature, facilitating its detection even at room temperature. In this section we also discuss the effects of next-nearest- neighbor hopping energy deviation. Finally we end the paper by presenting summary and conclusions in section 4. ## 2 Model and Formulation A schematic representation of Y-shaped Kekulé-patterned graphene superlattice has been displayed in Fig. 1(a), in which the three thick lines in each unit cell of graphene superlattice denote the bonds that connect the carbon atom located on the copper-atom vacancies in substrate to its nearest neighbors. These bonds acquire a shorter nearest-neighbor bond [1], leading to elongation of the other nearest-neighbor bonds, which is shown by thin lines. Also the on-site energy of carbon atoms located on the cooper-atom vacancies is expected to be changed with respect to other carbon atoms. Therefore, by taking into account deviations in the hopping and on-site energies [12], the tight-binding Hamiltonian for Y-shaped Kekulé-patterned graphene in the nearest-neighbor approximation is given by $\displaystyle H$ $\displaystyle=$ $\displaystyle u\sum_{i=1}^{N/3}a_{\vec{\mathbf{R}}_{i}}^{{\dagger}}a_{\vec{\mathbf{R}}_{i}}-(t_{0}+\delta t)\sum_{i=1}^{N/3}\sum_{n=1}^{3}(b_{\vec{\mathbf{R}}_{i}+\vec{\mathbf{\delta}}_{n}}^{{\dagger}}a_{\vec{\mathbf{R}}_{i}}+h.c.)$ (1) $\displaystyle-$ $\displaystyle(t_{0}-\delta t)\sum_{i=1}^{N/3}\sum_{n=1}^{3}(b_{\vec{\mathbf{R}}_{i}+\vec{\mathbf{\delta}}_{3}-\vec{\mathbf{\delta}}_{1}+\vec{\mathbf{\delta}}_{n}}^{{\dagger}}a_{\vec{\mathbf{R}}_{i}+\vec{\mathbf{\delta}}_{3}-\vec{\mathbf{\delta}}_{1}}+b_{\vec{\mathbf{R}}_{i}+\vec{\mathbf{\delta}}_{3}-\vec{\mathbf{\delta}}_{2}+\vec{\mathbf{\delta}}_{n}}^{{\dagger}}a_{\vec{\mathbf{R}}_{i}+\vec{\mathbf{\delta}}_{3}-\vec{\mathbf{\delta}}_{2}}+h.c.)$ where $t_{0}$ and $N$ are the nearest-neighbor hopping energy and the number of the unit cells in pristine graphene respectively (The number of the nuit cells of the Y-shaped Kekulé-patterned graphene is $N/3$.). The operator $a_{\vec{\mathbf{R}}_{i}}^{{\dagger}}$ ($b_{\vec{\mathbf{R}}_{i}}^{{\dagger}}$) creates an electron in the carbon atoms on sublattice $A$ ($B$) located at $i^{th}$ unit cell. $\vec{\mathbf{R}}_{i}=n_{i}\vec{\mathbf{A}}_{1}+m_{i}\vec{\mathbf{A}}_{2}$ are the translation vectors of Y-shaped Kekulé-patterned graphene with $\vec{\mathbf{A}}_{1}=(3\sqrt{3}a/2,3a/2)$ and $\vec{\mathbf{A}}_{2}=(-3\sqrt{3}a/2,3a/2)$ its primitive translation vectors, and $n_{i}$ and $m_{i}$ are integer numbers. The vectors $\vec{\mathbf{\delta}}_{1}=(\sqrt{3}a/2,a/2)$, $\vec{\mathbf{\delta}}_{2}=(\sqrt{3}a/2,a/2)$ and $\vec{\mathbf{\delta}}_{3}=(0,-a)$ are drown from each A carbon atom to its nearest neighbors and $a$ is the shortest carbon-carbon distance. The on-site energy deviation, $u$, is assumed to be nonzero only for the carbon atoms located on the cooper-atom vacancies in the substrate. The hopping energy deviation, which is due to the change in the bond length, is assumed to be $+\delta t$ and $-\delta t$ for carbon-carbon bonds shown by thick and thin lines, respectively, in Fig. 1(a). This Hamiltonian can be diagonalized by selecting the appropriate unit cell, red hexagonal in 1(a), in space k. The corresponding band structure for different values of the hopping and on-site energy deviations, (a) $\delta t=0.1t_{0}$ and $u=0$, (b) $\delta t=0$ and $u=0.1t_{0}$, (c) $\delta t=0.1t_{0}$ and $u=0.1t_{0}$ has been shown in Figure 1 by solid black lines. Figure 1: (a) A schematic representation of a single layer graphene with a uniform Y-shaped Kekulé lattice distortion in which primitive cells of pristine and Y-shaped Kekulé-patterned has been shown by green and red hexagonal respectively. $\vec{a}_{i}$ ($\vec{A}_{i}$) indicates primitive translation vectors of pristine (Y-shaped Kekulé-patterned) graphene. (b) Reciprocal lattices for pristine(in green) and Y-shaped Kekulé-patterned (in black) graphene. ### 2.1 Low-energy effective Hamiltonian To build a low-energy effective Hamiltonian, we must first express the Hamiltonian of Eq. 1 in the momentum space. As it is clear from Fig. 1(b), the first Brillouin zone of pristine graphene can be represented through three copies of that of graphene superlattice which are set to be centered at $\mathbf{K}_{j}$ (j=1,2 and 3), $\mathbf{K}_{1}=\mathbf{K}=(+4\pi/3\sqrt{3},0)$, $\mathbf{K}_{2}=\mathbf{\Gamma}=(0,0)$ and $\mathbf{K}_{2}=\mathbf{K}^{{}^{\prime}}=(-4\pi/3\sqrt{3},0)$. Therefore, the annihilation (creation) operators can be expanded in the momentum space as [13] $a_{i}=\frac{1}{\sqrt{N}}\sum_{\mathbf{k}}\sum_{j}e^{i(\mathbf{K}_{j}+\mathbf{k}).\mathbf{R}_{i}}a_{\mathbf{K}_{j}+\mathbf{k}}$ ($a_{i}^{{\dagger}}=\frac{1}{\sqrt{N}}\sum_{\mathbf{k}}\sum_{j}e^{-i(\mathbf{K}_{j}+\mathbf{k}).\mathbf{R}_{i}}a^{{\dagger}}_{\mathbf{K}_{j}+\mathbf{k}}$) where N is the number of unit cells of pristine graphene, and $\mathbf{k}$ runs over the first Brillouin of graphene superlattice, red hexagonal in 1(b). The Fourier transform of Eq. (1) is then, $\displaystyle H(\mathbf{k})$ $\displaystyle=$ $\displaystyle\sum_{\mathbf{k}}\sum_{j}(\epsilon^{\ast}(\mathbf{K}_{j}+\mathbf{k})b_{\mathbf{K}_{j}+\mathbf{k}}^{{\dagger}}a_{\mathbf{K}_{j}+\mathbf{k}}+h.c.)+\frac{u}{3}\sum_{\mathbf{k}}\sum_{j,j^{{}^{\prime}}}a_{\mathbf{K}_{j}+\mathbf{k}}^{{\dagger}}a_{\mathbf{K}_{j^{{}^{\prime}}}+\mathbf{k}}$ (2) $\displaystyle+$ $\displaystyle\Delta\sum_{\mathbf{k}}\sum_{j,j^{{}^{\prime}}}([1-2\cos(2(j-j^{{}^{\prime}})\pi/3)]\epsilon^{\ast}(\mathbf{K}_{j}+\mathbf{k})b_{\mathbf{K}_{j}+\mathbf{k}}^{{\dagger}}a_{\mathbf{K}_{j^{{}^{\prime}}}+\mathbf{k}}+h.c.),$ where $\epsilon^{\ast}(\mathbf{K}_{j}+\mathbf{k})=-t_{0}\sum_{n=1}^{3}e^{-i(\mathbf{K}_{j}+\mathbf{k}).\mathbf{\delta}_{n}}$ and $\Delta=\frac{\delta t}{3t_{0}}$. By defining $\hat{\psi}_{\mathbf{k}}=(a_{\mathbf{\Gamma}+\mathbf{k}},b_{\mathbf{\Gamma}+\mathbf{k}},a_{\mathbf{K}+\mathbf{k}},b_{\mathbf{K}+\mathbf{k}},a_{\mathbf{K}^{{}^{\prime}}+\mathbf{k}},b_{\mathbf{K}^{{}^{\prime}}+\mathbf{k}})^{T}$ we can rewrite the Hamiltonian as $\displaystyle H(\mathbf{k})=\hat{\psi}^{{\dagger}}_{\mathbf{k}}\left(\begin{array}[]{cc}\hat{H}_{H}&\hat{T}^{{\dagger}}\\\ \hat{T}&\hat{H}_{L}\end{array}\right)\hat{\psi}_{\mathbf{k}},$ (5) where $\displaystyle\hat{H}_{H}=\left(\begin{array}[]{cc}\frac{u}{3}&(1-\Delta)\epsilon(\mathbf{\Gamma}+\mathbf{k})\\\ (1-\Delta)\epsilon^{\ast}(\mathbf{\Gamma}+\mathbf{k})&0\end{array}\right),$ (8) and $\displaystyle\hat{H}_{L}=\left(\begin{array}[]{cccc}\frac{u}{3}&(1-\Delta)\epsilon(\mathbf{K}+\mathbf{k})&\frac{u}{3}&2\Delta\epsilon(\mathbf{K}^{{}^{\prime}}+\mathbf{k})\\\ (1-\Delta)\epsilon^{\ast}(\mathbf{K}+\mathbf{k})&0&2\Delta\epsilon^{\ast}(\mathbf{K}+\mathbf{k})&0\\\ \frac{u}{3}&2\Delta\epsilon(\mathbf{K}+\mathbf{k})&\frac{u}{3}&(1-\Delta)\epsilon(\mathbf{K}^{{}^{\prime}}+\mathbf{k})\\\ 2\Delta\epsilon^{\ast}(\mathbf{K}^{{}^{\prime}}+\mathbf{k})&0&(1-\Delta)\epsilon^{\ast}(\mathbf{K}^{{}^{\prime}}+\mathbf{k})&0\end{array}\right),$ (13) are the high and low energy sectors of $\hat{H}(\mathbf{k})$, respectively, and $\displaystyle\hat{T}=\left(\begin{array}[]{cccc}\frac{u}{3}&2\Delta\epsilon(\mathbf{K}+\mathbf{k})&\frac{u}{3}&2\Delta\epsilon(\mathbf{K}^{{}^{\prime}}+\mathbf{k})\\\ 2\Delta\epsilon^{\ast}(\mathbf{\Gamma}+\mathbf{k})&0&2\Delta\epsilon^{\ast}(\mathbf{\Gamma}+\mathbf{k})&0\end{array}\right),$ (16) is the coupling between them. Then, we consider the Schrodinger equations applied to Eq. (5), $\displaystyle\hat{H}_{H}\hat{\Psi}_{H}+\hat{T}\hat{\Psi}_{L}=E\hat{\Psi}_{H},$ (17) and $\displaystyle\hat{T}^{{\dagger}}\hat{\Psi}_{H}+\hat{H}_{L}\hat{\Psi}_{L}=E\hat{\Psi}_{L},$ (18) on each space, where $\hat{\Psi}_{H}$ and $\hat{\Psi}_{L}$ are the components of the eigenfunction, $\hat{\Psi}=(\hat{\Psi}_{H},\hat{\Psi}_{L})$. One can obtain $\hat{\Psi}_{H}$ from Eq. (17) and use it on Eq. (18) to obtain $\displaystyle\hat{H}_{eff}=\hat{H}_{L}+\hat{T}^{{\dagger}}(E\hat{1}-\hat{H}_{L})^{-1}\hat{T},$ (19) which is the effective Hamiltonian for the $\hat{\Psi}_{L}$ component. Then, we expand $\epsilon(\mathbf{K}_{j}+\mathbf{k})$ and their complex conjugates in Eqs. (8-16) up to the first order in $\mathbf{k}$, $\epsilon(\mathbf{\Gamma}+\mathbf{k})\approx-3t_{0}$, $\epsilon(\mathbf{K}+\mathbf{k})\approx\frac{3}{2}t_{0}a(k_{x}+ik_{y})$ and $\epsilon(\mathbf{K}^{{}^{\prime}}+\mathbf{k})\approx-\frac{3}{2}t_{0}a(k_{x}-ik_{y})$, and insert them into Eq. (19). Therefore, we arrive at $\displaystyle\hat{H}_{eff}=\left(\begin{array}[]{cccc}U&v_{B}(k_{x}+ik_{y})&U&-v_{C}(k_{x}-ik_{y})\\\ v_{B}(k_{x}-ik_{y})&0&v_{C}(k_{x}-ik_{y})&0\\\ U&v_{C}(k_{x}+ik_{y})&U&-v_{B}(k_{x}-ik_{y})\\\ -v_{C}(k_{x}+ik_{y})&0&-v_{B}(k_{x}+ik_{y})&0\end{array}\right),$ (24) where $U=\frac{u}{3}(1-\frac{4\Delta}{1-\Delta}+\frac{4\Delta^{2}}{(1-\Delta)^{2}})$, $v_{B}=\frac{3}{2}t_{0}a(1-\Delta-\frac{4\Delta^{2}}{1-\Delta})$, and $v_{C}=\frac{3}{2}t_{0}a(2\Delta-\frac{4\Delta^{2}}{1-\Delta})$ (Notice that although Eq. (24) is convenient to obtain the paramagnetic current operator, to get the diamagnetic part of the current operator we must keep the terms up to second order in $\mathbf{k}$). The energy bands of the low-energy effective Hamiltonian are $\displaystyle E_{1\mathbf{k}}$ $\displaystyle=$ $\displaystyle- E_{2\mathbf{k}}=-(v_{B}-v_{C})k,$ $\displaystyle E_{3\mathbf{k}}$ $\displaystyle=$ $\displaystyle U-\sqrt{(v_{B}+v_{C})^{2}k^{2}+U^{2}},$ $\displaystyle E_{4\mathbf{k}}$ $\displaystyle=$ $\displaystyle U+\sqrt{(v_{B}+v_{C})^{2}k^{2}+U^{2}},$ (25) where $k=\sqrt{k^{2}_{x}+k^{2}_{y}}$. Figure 2 displays the exact band structure resulting from Eq. (1) or (2) in comparison with that obtained from our low-energy effective Hamiltonian for different values of $u$ and $\delta t$, (a) $u=0$ and $\delta t=0.1t_{0}$, (b) $u=0.1t_{0}$ and $\delta t=0$, (c) $u=0.1t_{0}$ and $\delta t=0.1t_{0}$. An excellent agreement between the energy bands of the effective Hamiltonian and the exact ones in the low-energy regime is clear. One can see that, the Y-shaped Kekulé distortion breaks the valley degeneracy and couples the energy bands at $\mathbf{K}$ and $\mathbf{K}^{{}^{\prime}}$ in the firs brillouin zone, resulting in two concentric energy bands with different velocities at $\mathbf{\Gamma}$ point, but contrary to what has been reported in previous works [2, 14], their deviation from initial velocity is not symmetric. According to our results, different velocities are $(1-3\Delta)v_{F}$ and $(1+\Delta-\frac{8\Delta^{2}}{1-\Delta})v_{F}$ with $v_{F}=\frac{3}{2}t_{0}a$ being the initial Fermi velocity. On the other hand, the on-site potential deviation, in addition two breaking valley symmetry, has two other effects. First, it causes two energy bands gain effective mass. Second, it shifts the massive energy bands, resulting in a particle-hole symmetry breaking and lifting the four fold degeneracy at Dirac point. These effects lead to the appearance of attractive optical signatures, which are discussed in the next section. Figure 2: Comparison of the exact energy bands resulting from Eq. (1) or (2) with the effective low-energy bands, Eq. (2.1), for different values of the hopping and on-site energy deviations, (a) $\delta t=0.1t_{0}$ and $u=0$, (b) $\delta t=0$ and $u=0.1t_{0}$ and (c) $\delta t=0.1t_{0}$ and $u=0.1t_{0}$. $\mathbf{M^{\ast}}=(+2\pi/3\sqrt{3},0)$, $\mathbf{\Gamma}=(0,0)$ and $\mathbf{K^{\ast}}=(+2\pi/3\sqrt{3},+2\pi/9)$ are high-symmetry points in the first brillouin zone of graphene superlattice. See red hexagonal in Fig. 1(b) . ### 2.2 Current operator and Kubo Formula The finite frequency conductivity calculated by Kubo Formula is given by[15] $\displaystyle\sigma_{\alpha\alpha}(\Omega)=\frac{\Pi_{\alpha\alpha}(\Omega+i0^{+})}{i\hbar(\Omega+i0^{+})S}+\frac{<j_{\alpha}^{D}>}{i(\Omega+i0^{+})S},$ (26) where $S$ is the area of the sample, $\alpha$ denotes $x$ or $y$, $\Omega$ is the frequency, $\Pi_{\alpha\alpha}(\Omega+i0^{+})$ is the correlation function of the paramagnetic current operator and $<j_{\alpha}^{D}>$ is the expectation value of the diamagnetic part of current operator. The function $\Pi_{\alpha\alpha}(\Omega+i0^{+})$ is obtained by analytical continuum of the corresponding Matsubara current-current correlation function defined as[16, 17, 18] $\displaystyle\Pi_{\alpha\alpha}(i\omega_{n})=\int_{0}^{\hbar\beta}d\tau e^{i\omega_{n}\tau}<T_{\tau}j_{\alpha}^{P}(\tau)j_{\alpha}^{P}(0)>,$ (27) where $i\omega_{n}$ are the bosonic Matsubara frequencies, $\beta=1/k_{B}T$ with $k_{B}$ the Boltzmann constant and T the temperature, and $T_{\tau}$ is time ordering operator. One can show that the paramagnetic current operator can be written as [17] $j_{\alpha}^{P}=-g_{s}e\Sigma_{\mathbf{k}}\hat{\psi}^{{\dagger}}_{\mathbf{k},L}\hat{v}_{\alpha}\hat{\psi}_{\mathbf{k},L}$ where $g_{s}=2$ is the spin degeneracy, $e$ is the electron charge, $\hat{\psi}_{\mathbf{k},L}$ is the low-energy component of $\hat{\psi}_{\mathbf{k}}=(\hat{\psi}_{\mathbf{k},H},\hat{\psi}_{\mathbf{k},L})^{T}$, and $\hat{v}_{\alpha}=\partial\hat{H}_{eff}/\hbar\partial k_{\alpha}$. Consequently, the current-current correlation function can then be written in the usual bubble approximation as[16, 17] $\displaystyle\Pi_{\alpha\alpha}(i\nu_{m})=\frac{g_{s}e^{2}}{\beta\hbar}\sum_{i\nu_{m}}\sum_{\mathbf{k}}Tr[\hat{v}_{\alpha}\hat{G}(i\nu_{m}+i\omega_{n},\mathbf{k})\hat{v}_{\alpha}\hat{G}(i\nu_{m},\mathbf{k})],$ (28) where $\hat{G}(i\nu_{m},\mathbf{k})=(i\nu_{m}\hat{1}-\hat{H}_{eff}/\hbar)^{-1}$, $Tr[\hat{A}]$ denotes the trace of $\hat{A}$ matrix, and $i\nu_{m}$ are the fermionic Matsubara frequencies. Using the spectral representation of the Green’s function matrix,[17, 19] $\displaystyle\hat{G}(z,\mathbf{k})=\int_{-\infty}^{+\infty}\frac{d\omega^{{}^{\prime}}}{2\pi}\frac{\hat{A}(\omega^{{}^{\prime}},\mathbf{k})}{z-\omega^{{}^{\prime}}},$ (29) the imaginary and real part of the current-current correlation function can be written as $\displaystyle\Im\Pi_{xx}(\Omega+i0^{+})=-\frac{\pi g_{s}e^{2}}{\hbar^{2}}\sum_{\mathbf{k}}\sum_{i,j}\chi(E_{j\mathbf{k}},E_{i\mathbf{k}})[n_{F}(E_{j\mathbf{k}})-n_{F}(E_{i\mathbf{k}})]\delta(\Omega+E_{j\mathbf{k}}/\hbar- E_{i\mathbf{k}}/\hbar),$ (30) and $\displaystyle\Re\Pi_{xx}(\Omega+i0^{+})=\frac{g_{s}e^{2}}{\hbar^{2}}\sum_{\mathbf{k}}\sum_{i,j}\chi(E_{i\mathbf{k}},E_{j\mathbf{k}})\frac{n_{F}(E_{j\mathbf{k}})-n_{F}(E_{i\mathbf{k}})}{\Omega+E_{j\mathbf{k}}/\hbar- E_{i\mathbf{k}}/\hbar},$ (31) where $n_{F}(x)=1/(1+exp[(x-\mu)/k_{B}T])$ is Fermi-Dirac distribution function and $\displaystyle\chi(E_{i\mathbf{k}},E_{j\mathbf{k}})$ $\displaystyle=$ $\displaystyle 4(v^{2}_{B}\cos(2\varphi_{\mathbf{k}})+v^{2}_{C})\mathcal{M}_{12,E_{i\mathbf{k}}}\mathcal{M}_{12,E_{j\mathbf{k}}}+4(v^{2}_{B}+v^{2}_{C}\cos(2\varphi_{\mathbf{k}}))\mathcal{M}_{12,E_{i\mathbf{k}}}\mathcal{M}_{12,E_{j\mathbf{k}}}$ (32) $\displaystyle+$ $\displaystyle 4v_{B}v_{C}(1+\cos(2\varphi_{\mathbf{k}}))[\mathcal{M}_{12,E_{i\mathbf{k}}}\mathcal{M}_{13,E_{j\mathbf{k}}}+\mathcal{M}_{13,E_{i\mathbf{k}}}\mathcal{M}_{12,E_{j\mathbf{k}}}]$ $\displaystyle+$ $\displaystyle 2(v^{2}_{B}+v^{2}_{C})[\mathcal{M}_{11,E_{i\mathbf{k}}}\mathcal{M}_{22,E_{j\mathbf{k}}}+\mathcal{M}_{22,E_{i\mathbf{k}}}\mathcal{M}_{11,E_{j\mathbf{k}}}]$ $\displaystyle+$ $\displaystyle 4v_{B}v_{C}[\mathcal{M}_{11,E_{i\mathbf{k}}}\mathcal{M}_{23,E_{j\mathbf{k}}}+\mathcal{M}_{23,E_{i\mathbf{k}}}\mathcal{M}_{11,E_{j\mathbf{k}}}]$ $\displaystyle-$ $\displaystyle 4v_{B}v_{C}\cos(2\varphi_{\mathbf{k}})[\mathcal{M}_{14,E_{i\mathbf{k}}}\mathcal{M}_{22,E_{j\mathbf{k}}}+\mathcal{M}_{22,E_{i\mathbf{k}}}\mathcal{M}_{14,E_{j\mathbf{k}}}]$ $\displaystyle-$ $\displaystyle 2(v^{2}_{B}+v^{2}_{C})[\mathcal{M}_{14,E_{i\mathbf{k}}}\mathcal{M}_{23,E_{j\mathbf{k}}}+\mathcal{M}_{23,E_{i\mathbf{k}}}\mathcal{M}_{14,E_{j\mathbf{k}}}],$ in which $\varphi_{\mathbf{k}}=\tan^{-1}(k_{y}/k_{x})$ and $\mathcal{M}_{mn,E_{i\mathbf{k}}}$ are the coefficients of delta functions $\delta(\omega-E_{i\mathbf{k}}/\hbar)$ in the matrix components of the spectral function, $A_{mn}(\omega,\mathbf{k})=2\pi\sum_{i}\mathcal{M}_{mn,E_{i\mathbf{k}}}\delta(\omega- E_{i\mathbf{k}}/\hbar)$. To obtain $<j_{\alpha}^{D}>$ we keep the terms up to second order in $\mathbf{k}$ in the expansion of $\epsilon(\mathbf{K}_{j}+\mathbf{k})$ and its complex conjugate in Eq. (19), so we can write the diamagnetic current operator as[21] $j^{D}_{\alpha}=-g_{s}e^{2}\Sigma_{\mathbf{k}}\hat{\psi}^{{\dagger}}_{\mathbf{k},L}\hat{w}_{\alpha}\hat{\psi}_{\mathbf{k},L}$, where $\displaystyle\hat{w}_{x}=\frac{1}{\hbar^{2}}\frac{\partial^{2}\hat{H}_{eff}}{\partial k^{2}_{x}}=\frac{a}{2\hbar^{2}}\left(\begin{array}[]{cccc}0&-v_{B}&0&v_{C}\\\ -v_{B}&0&-v_{C}&0\\\ 0&-v_{C}&0&v_{B}\\\ v_{C}&0&v_{B}&0\end{array}\right).$ (37) After inserting above equations into $<j_{\alpha}^{D}>$ and performing some straightforward calculations[20], we reach the following statement $\displaystyle<j_{x}^{D}>=\frac{g_{s}ae^{2}}{2\hbar^{{}^{2}}}\sum_{\mathbf{k}}\sum_{i}\Lambda(E_{i\mathbf{k}})n_{F}(E_{i\mathbf{k}}),$ (38) where $\displaystyle\Lambda(E_{i\mathbf{k}})=4v_{B}\mathcal{M}_{12,E_{i\mathbf{k}}}\cos(2\varphi_{\mathbf{k}})+2v_{C}\mathcal{M}_{14,E_{i\mathbf{k}}}\cos(4\varphi_{\mathbf{k}})+2v_{C}\mathcal{M}_{23,E_{i\mathbf{k}}}.$ (39) The real part of the finite frequency conductivity, which is related to the optical quantities such optical absorption and reflectivity, is given by $\displaystyle\Re\sigma_{\alpha\alpha}(\Omega)=D\delta(\Omega)+\frac{\Im\Pi_{\alpha\alpha}(\Omega+i0^{+})}{\hbar\Omega S},$ (40) with $\displaystyle D=-\pi\frac{<j_{\alpha}^{D}>}{S}-\pi\frac{\Re\Pi_{\alpha\alpha}(\Omega+i0^{+})}{\hbar S}.$ (41) called Drude weight. Finally, by inserting Eqs. (30), (31) and (38) into (40), we arrive at the following equation for the real part of the optical conductivity $\displaystyle\Re\sigma_{\alpha\alpha}(\Omega)/\sigma_{0}=$ $\displaystyle-$ $\displaystyle\frac{2\pi ag_{s}}{\hbar}\frac{1}{S}\sum_{\mathbf{k}}\sum_{i}\Lambda(E_{i\mathbf{k}})n_{F}(E_{i\mathbf{k}})\delta(\Omega)$ $\displaystyle-$ $\displaystyle\frac{4\pi g_{s}}{\hbar^{2}}\frac{1}{S}\sum_{\mathbf{k}}\sum_{i,j}\chi(E_{i\mathbf{k}},E_{j\mathbf{k}})\frac{n_{F}(E_{j\mathbf{k}})-n_{F}(E_{i\mathbf{k}})}{\Omega+E_{j\mathbf{k}}/\hbar- E_{i\mathbf{k}}/\hbar}\delta(\Omega)$ $\displaystyle-$ $\displaystyle\frac{4\pi g_{s}}{\hbar^{2}}\frac{1}{S}\sum_{\mathbf{k}}\sum_{i,j}\chi(E_{i\mathbf{k}},E_{j\mathbf{k}})\frac{[n_{F}(E_{j\mathbf{k}})-n_{F}(E_{i\mathbf{k}})]}{\Omega}\delta(\Omega+E_{j\mathbf{k}}/\hbar- E_{i\mathbf{k}}/\hbar),$ where $\sigma_{0}=\frac{e^{2}}{4\hbar}$ is the conductivity of the pristine graphene. Eqs. (32), (39), and (2.2) along with the low-energy effective Hamiltonian and its energy bands are our main results, which are solved numerically in the next section to investigate the effects of the kekulé distortion on the optical conductivity of graphene at low energy regime. ## 3 Results and discussion Here we present our results exploring the effects of Y-shaped Kekulé distortion on the optical conductivity of graphene, which is obtained by evaluating Eq. (2.2) numerically. Although it is expected that the Y-shaped Kekulé distortion leads to a deviation in both hopping and on-site energies, but in order to clarify the role of each of them separately, in addition to the $\delta t\neq 0$ and $u\neq 0$ case which is expected to be more realistic, we also investigate the cases in which $\delta t\neq 0$ or $u\neq 0$. For the numerical evaluation of Eq. (2.2), we use the Lorentzian representation of the delta function, $\delta(x)=(\eta/\pi)/(x^{2}+\eta^{2})$ with broadening $\eta$ reflecting the effect of electron scattering from disorder, $1/\tau_{imp}=2\eta$, phenomenologically. In all calculations we set $\eta=0.00002t_{0}$. Furthermore, we put $t_{0}$ as the energy scale and the optical conductivity curves are scaled by $\sigma_{0}$. So, $\sigma_{xx}/\sigma_{0}$ is calculated and plotted as a function of $\hbar\Omega/t_{0}$ with $\hbar\Omega$ being the photon energy. Zero chemical potential: Our results for the optical conductivity at zero chemical potential and temperature are shown in Fig. 3, in which the hopping and on-site energy deviations are (a) $\delta t=0.10t_{0}$ and $u=0$, (b) $\delta t=0$ and $u=0.10t_{0}$ and (c) $\delta t=0.10t_{0}$ and $u=0.10t_{0}$. The results for Kekulé-patterned (pristine) graphene are shown by solid magenta (dashed black) curves. All possible optical transitions are displayed in the lower insets of Fig. 3 and the resulting structures in the optical conductivity are identified by arrows with the same color on the curves. The upper insets exhibit the contribution of each transition in optical conductivity separately identified by the curve with same color. In the case of $\delta t=0.10t_{0}$ and $u=0$, the optical conductivity displays a flat interband response at all frequencies which also is found in undoped pristine graphene[22, 23], reflecting the linear band structure of Kekulé-patterned graphene near $\mathbf{\Gamma}$ point. This is to be contrast with the case of $u\neq 0$ shown in the middle panels Fig. 3, in which a dip-peak structure is also seen around $2U$. The appearance of this structure in the optical conductivity, which can be detected as a optical signature for the Y-shaped Kekulé distortion in single layer graphene at the charge neutrality, is explained by examining the contribution of each of the transitions in the optical conductivity. In the presence of the on-site energy deviation $E_{3\mathbf{k}}\rightarrow E_{2\mathbf{k}}$ transition is possible for all $\hbar\Omega>0$, but $E_{1\mathbf{k}}\rightarrow E_{4\mathbf{k}}$ transition, due to the presence of the energy gap, is restricted to $\hbar\Omega>2U$. The $E_{3\mathbf{k}}\rightarrow E_{2\mathbf{k}}$ ($E_{1\mathbf{k}}\rightarrow E_{4\mathbf{k}}$) transition, as shown in the upper inset of the middle frame of Fig. 3, starts at zero ($2U$) photon energy with a spectral weight that decreases (increases) by increasing the photon energy and finally approaches $\sigma_{0}/2$ at high photon energies. These facts manifest themselves as a dip in the curve of the total optical conductivity. This spectral weight lost, for $\hbar\Omega>2U$, is compensate by the appearance of a new transition which starts at $\hbar\Omega=2U$, $E_{3\mathbf{k}}\rightarrow E_{4\mathbf{k}}$, and its large spectral weight also leads to the appearance of the peak structure. At high photon energies, where the effect of the energy gap becomes insignificant, $E_{2\mathbf{k}}$ and $E_{4\mathbf{k}}$ bands inherit the chirality of the energy bands in $\mathbf{K}$ and $\mathbf{K}^{{}^{\prime}}$ valleys in pristine graphene and become completely chiral and anti-chiral respectively. So, the optical transitions between them becomes forbidden[24]. Therefore, at high photon energies only $E_{1\mathbf{k}}\rightarrow E_{4\mathbf{k}}$ and $E_{2\mathbf{k}}\rightarrow E_{4\mathbf{k}}$ transitions contribute to the optical conductivity and the usual flat interband response is recovered. Returning to the case of $\delta t=0.10t_{0}$ and $u=0.10t_{0}$ (Fig. 3(c)), one can see that the only additional change caused by the simultaneous presence of the hopping and the on-site energy deviations is that the dip-peak structure of the optical conductivity is shifted to lower energies originating from the dependence of $U$ on the hopping energy deviation, $U=\frac{u}{3}(1-\frac{4\Delta}{1-\Delta}+\frac{4\Delta^{2}}{(1-\Delta)^{2}})$. Figure 3: The normalized optical conductivity, $\sigma_{xx}/\sigma_{0}$, of undoped Kekulé-patterned graphene (solid magenta curves) as function of $\hbar\Omega/t_{0}$ at zero temperature for (a) $\delta t=0.1t_{0}$ and $u=0$, (b) $\delta t=0$ and $u=0.1t_{0}$, and (c) $\delta t=0.1t_{0}$ and $u=0.1t_{0}$, compared with that of pristine graphene (dashed black curves). The lower insets show the band structures around $\mathbf{\Gamma}$ point for each case, and allowed optical transitions indicated by arrows which give rise to structure in the conductivity curves shown by arrows with same color on plots. The upper insets in each panel displays the contribution of each of the optical transitions in the optical conductivity separately. Finite chemical potential: Figure 4 shows our numerical results for the optical conductivity of Y-shaped Kekulé-patterned graphene in comparison with that of pristine graphene for $\mu>0$. The upper (lower) panels display the result for $\mu=+0.05t_{0}$ ($\mu=+0.10t_{0}$), in which the left, middle and panels is for $\delta t=0.10t_{0}\leavevmode\nobreak\ \leavevmode\nobreak\ u=0$, $\delta t=0.0\leavevmode\nobreak\ \leavevmode\nobreak\ u=0.10t_{0}$ and $\delta t=0.10t_{0}\leavevmode\nobreak\ \leavevmode\nobreak\ u=0.10t_{0}$ respectively. As in Fig. 3, the solid magenta (dashed black) curves display the optical conductivity of Kekulé-patterned (pristine) graphene, and in each panel, the inset shows the corresponding band structure and the allowed inter- band optical transitions. The colored arrows in the band structure show the allowed optical transitions which give rise to some structures in the optical conductivity identified by arrows with same color on the plots. Figure 5 shows the same plots as Fig. 4, but for other different values of the chemical potentials, $\mu=-0.05t_{0}$ (upper panels) and $\mu=-0.10t_{0}$ (lower panels). For $\mu\neq 0$, as can be seen in all panels of Figs. 4 and 5, there is the usual Drude conductivity arising from the intra-band transitions. Starting from the case of $\delta t=0.10t_{0}$ and $u=0$, shown in Figs. 4(a) and 3(d), one can see that the interband absorption edge that is at $2\mu$ for pristine graphene[21] is now splitting into two edges moving to lower and higher photon energies. So, the optical conductivity exhibits a two-steps absorption which onset at $2\mu\frac{v_{B}}{v_{B}+v_{C}}$ and $2\mu\frac{v_{B}}{v_{B}-v_{C}}$ photon energies coming from the interband transitions between different valleys, $E_{1\mathbf{k}}\rightarrow E_{4\mathbf{k}}$ and $E_{3\mathbf{k}}\rightarrow E_{2\mathbf{k}}$. The presence of the Kekule distortion, which couples different valleys, also leads to the appearance of a sharp peak in the optical conductivity which occurs in the range of $2\mu\frac{v_{C}}{v_{B}+v_{C}}$ to $2\mu\frac{v_{C}}{v_{B}-v_{C}}$ and is due to the inter-valley transitions in the conduction band, shown by red arrow in the insets. See also (a), (d), (g) and (j) panels of Fig. 6 in which the contribution of all transitions of Figs. 4 and 5 has been displayed. From Fig. 4 (d) one can also see that increasing the chemical potential enhances the splitting of the interband absorption edges and moves them further to higher and lower energies. It also shifts the sharp peak to higher energies and increase its spectral weight. The effects of nonzero on-site energy deviation on the optical conductivity has been addressed in (b) and (e) panels of Fig. 4. One can see that due to the gapped nature of $E_{4\mathbf{k}}$ band, which increases the energy required for $E_{3\mathbf{k}}\rightarrow E_{4\mathbf{k}}$ transition, the sharp peak in optical conductivity is enhanced and moved to higher energies, with a nonzero spectral weight in the energy range $U+\sqrt{U^{2}+[(v_{B}+v_{C})/(v_{B}-v_{C})]^{2}\mu^{2}}<\hbar\Omega<2U$. If the chemical potential increases and cuts $E_{4\mathbf{k}}$ band, due to the occurrence of a phenomenon called band nesting resonance[25, 26, 27], the spectral weight of the peak, as can be seen in Fig. 4(b), increases sharply and its broadening becomes limited to a small range of photon energies, facilitating its observation as an optical signature for the Y-shaped Kekulé distortion with nonzero on-site energy deviation. In order to understand the origin of band nesting resonance, we rewrite the contribution of $E_{i\mathbf{k}}\rightarrow E_{j\mathbf{k}}$ transition in the optical conductivity in Eq. 2.2, $-\frac{4\pi g_{s}}{\hbar^{2}}\frac{1}{S}\sum_{\mathbf{k}}\chi(E_{j\mathbf{k}},E_{i\mathbf{k}})\frac{[n_{F}(E_{i\mathbf{k}})-n_{F}(E_{j\mathbf{k}})]}{\Omega}\delta(\Omega+E_{i\mathbf{k}}/\hbar- E_{j\mathbf{k}}/\hbar)$, in a more illustrative form. If we consider cuts $S(E)$ of constant energy $E$, $E=E_{j\mathbf{k}}-E_{i\mathbf{k}}$, in the band structure, we can write $d^{2}\mathbf{k}=dS[d(E_{j\mathbf{k}}-E_{i\mathbf{k}})/|\nabla_{\mathbf{k}}(E_{j\mathbf{k}}-E_{i\mathbf{k}})|]$. So, we arrive at $\displaystyle\Re\sigma_{xx}^{E_{i}\rightarrow E_{j}}(\Omega)/\sigma_{0}=-\frac{g_{s}}{\pi\hbar\Omega}\int_{S(E)}\frac{\chi(E_{j\mathbf{k}},E_{i\mathbf{k}})}{|\nabla_{\mathbf{k}}(E_{j\mathbf{k}}-E_{i\mathbf{k}})|}dS,$ (43) for the contribution of $E_{i\mathbf{k}}\rightarrow E_{j\mathbf{k}}$ transition in optical conductivity at zero temperature indicating that strong peaks in the optical conductivity will come from regions in the spectrum where $|\nabla_{\mathbf{k}}(E_{j\mathbf{k}}-E_{i\mathbf{k}})|\approx 0$. One of the ways to fulfill the condition $|\nabla_{\mathbf{k}}(E_{j\mathbf{k}}-E_{i\mathbf{k}})|\approx 0$ is that the condition $|\nabla_{\mathbf{k}}E_{i\mathbf{k}}|\approx|\nabla_{\mathbf{k}}E_{j\mathbf{k}}|>0$, which occurs when two energy bands are approximately equispaced over regions in the Brillouin zone and called band nesting, is satisfied. Band nesting results in a strong peak in the optical conductivity which is called band nesting resonance. One can easily check that the band nesting condition is satisfied for $E_{2\mathbf{k}}$ and $E_{4\mathbf{k}}$ bands in the case of $u\neq 0$ when the chemical potential become larger than $2U$, as can be seen in Fig. 4(e), resulting in the appearance of band nesting resonance in case of $\delta t=0$ and $u=0.1t_{0}$. Notice that the band nesting resonance also takes place for $E_{1\mathbf{k}}\rightarrow E_{3\mathbf{k}}$ transition but with a spectral weight more dependent on the chemical potential and restricted to limited ranges of the chemical potential. The simultaneous presence of the hopping and on-site energy deviations makes the band nesting condition better fulfilled. Therefore, as Fig. 4(f) shows, for the case of $\delta t=0.1t_{0}$ and $u=0.1t_{0}$ the intensity of the peak increases. Furthermore, since the nonzero hopping and on-site energy deviations decreases $U$ and shifts $E_{4\mathbf{k}}$ band to lower energies, the band nesting condition is also satisfied for a wide range of $\mathbf{k}$ in $E_{2\mathbf{k}}\rightarrow E_{4\mathbf{k}}$ transition and as a result, we see a large increase in the intensity of the peak in Fig. 4(c) compared to that in Fig. 4(b). Figure 4: The normalized optical conductivity, $\sigma_{xx}/\sigma_{0}$, of undoped Kekulé-patterned graphene (solid magenta curves) as function of $\hbar\Omega/t_{0}$ at zero temperature for different values of the chemical potential, $\mu=0.05t_{0}$ (upper panels) and $\mu=0.10t_{0}$ (lower panels), and different values of the on-site and hopping energy deviations, $\delta t=0.1t_{0}$ and $u=0$ (left panels), $\delta t=0$ and $u=0.1t_{0}$ (middle panels) and $\delta t=0.1t_{0}$ and $u=0.1t_{0}$ (right panels), compared with that of pristine graphehe. The insets show the corresponding band structures around $\mathbf{\Gamma}$ point. The colored arrows in the band structure show the allowed optical transitions which give rise to some structures in the optical conductivity curves identified by arrows with same color on the plots. Returning to the cases of $u\neq 0$ and $\mu=0.05t_{0}<2U$, shown in Figs. 4(b) and 4(c), one can see that the optical conductivity displays a similar two-steps interband absorbtion but with a interband absorbtion edges that, with respect to that in case of $\delta t=0.1t_{0}$ and $u=0$, moves to lower energies and onsets at $2U$ and $\mu+\sqrt{U^{2}+[(v_{B}+v_{C})/(v_{B}-v_{C})]^{2}\mu^{2}}-U$. A nonzero hopping energy deviation, as shown in Fig. 4(c), enhances the splitting of the interband absorbtion edge. Moreover, the shape of the steps, in cases of $u\neq 0$ and $\mu<2U$, which is typical of a gapped band structure [28, 29] shows a peak at the absorbtion edge arising from a discontinuity in the electronic density of states. As the photon energy increases the usual flat interband response of pristine single layer graphene is recovered. By increasing the chemical potential, as can be seen in Figs. 4(e) and 4(f), $E_{2\mathbf{k}}\rightarrow E_{4\mathbf{k}}$ transitions become similar to that between linear energy bands leading to disappearance of the peak at the absorbtion edge. One can also see that in this case the optical conductivity displays a three-steps absorbtion arising from the splitting of the absorbtion edge of $E_{1\mathbf{k}}\rightarrow E_{4\mathbf{k}}$ and $E_{3\mathbf{k}}\rightarrow E_{4\mathbf{k}}$ transitions. Figure 5 shows the same plots as Fig. 4, but for $\mu=-0.05t_{0}$ (upper panels) and $\mu=-0.10t_{0}$ (lower panels). As expected, due to the presence of particle-hole symmetry, the optical conductivity in the case of $u=0$ does not change with changing the sign of the chemical potential, leading to similar plots as Figs. 4(a) and 4(d). This is in contrast to the case of nonzero on-site energy deviation, middle and right panels of Fig. 5, in which the particle-hole symmetry is absent and the allowed optical transition due to the gapped nature of $E_{4\mathbf{k}}$ is shifted to higher energies. In these cases the optical conductivity, similar to that in the middle and right panels of Fig. 4, displays a three-steps interband absorbtion but with the edges that move to higher energies. For $\mu<0$ the band nesting condition is satisfied for the case of $\delta t=0$ and $u=0.1t_{0}$ leading to appearance of band nesting resonance, but when both hopping and on-site energy deviation are nonzero the energy difference between $E_{1\mathbf{k}}$ and $E_{3\mathbf{k}}$ which determine the location and intensity of the intra-valley transition in the valance band (that turns into in band nesting resonance in the appropriate conditions[25]) is strongly dependent on the chemical potentials. Figure 5: Same plots as Fig. 4, but for $\mu=-0.05t_{0}$ (upper panels) and $\mu=-0.10t_{0}$ (lower panels). Figure 6: Calculated normalized optical conductivity, $\sigma_{xx}/\sigma_{0}$, as function of $\hbar\Omega/t_{0}$ for each of the optical transitions shown in the insets of Figs. 4 and 4 separately. The main optical feature of the case of $\delta t\neq 0$ and $u\neq 0$, which is expected to be more realistic, is the occurrence of band nesting resonance. To show that the band nesting resonance is robust with respect to increasing temperature, the optical conductivity of Y-shaped Kekul-patterned graphene at different temperatures, $T=0$ (solid magenta curves) and $T=300\leavevmode\nobreak\ K$ (dashed black curves), has been displayed in Figure 7. One can see that increasing temperature only broadens the interband absorbtion edge. It is also clear that the band nesting resonance which occurs in both electron- and hole-doped cases is resistant against increasing temperature except when the chemical potential approaches the intersection of $E_{1\mathbf{k}}$ and $E_{3\mathbf{k}}$ energy bands ((d) panel). Therefore the occurrence of band nesting resonance is a significant optical signature to detect the Y-shaped Kekulé distortion in single layer graphene. Figure 7: The optical conductivity of Kekulé-patterned graphene for the case of $\delta t=0.1t_{0}$ and $u=0.1t_{0}$ at different temperatures, $T=0\leavevmode\nobreak\ K$ (magenta solid curves) and $T=300\leavevmode\nobreak\ K$ (black dashed-dotted curves), and for different chemical potentials (a) $\mu=+0.05\leavevmode\nobreak\ t_{0}$, (b) $\mu=+0.05\leavevmode\nobreak\ t_{0}$, (c) $\mu=+0.05\leavevmode\nobreak\ t_{0}$ and (d) $\mu=+0.05\leavevmode\nobreak\ t_{0}$. It is evident that the band nesting resonance identicated by red arrow, except when the chemical potential approaches the intersection of $E_{1\mathbf{k}}$ and $E_{3\mathbf{k}}$ energy bands ((d) panel), is robust with respect to increasing temperature. Finally it is desirable to mention to the effect of the next-nearest hopping energy deviations on the optical conductivity of Y-shaped Kekulé-patterned graphene. In Ref. [21] the effect of next-nearest hopping energy on the optical conductivity of pristine graphene was investigated and it was found that its effect is negligible at low energies. Moreover, recently the effect of the next nearest neighbor hopping energy deviation on the electronic structure of kekule-patterned graphene has been investigated analytically[14]. Comparing equation 14c of this article with our result for the effective low energy Hamiltonian, Eq. 24, shows that including this effect in our calculations only leads to the normalization of $U$. Therefore, it can be expected that including the next nearest hopping energy deviation does not change our results for the optical conductivity in general. ## 4 Summary and conclusions In summary, we investigated the effects of a uniform Y-shaped Kekulé distortion of C-C bonds on the electronic band structure and optical conductivity of graphene. First we introduce our tight-binding model in which the effects of the Y-shaped Kekulé distortion was taken into account by including both on-site and hopping energy deviations in the minimal tight- binding Hamiltonian of graphene. Then by projecting the high energy bands onto the subspace defined by the low energy bands we derived a low-energy effective Hamiltonian which was found to be in excellent agreement with one calculated from a diagonalization of the full tight-binding Hamiltonian. It was shown that in addition the coupling of energy bands in different valleys caused by both on-site and nearest-neighbor hopping energy deviations, the other main effect of the on-site energy deviation on the low-energy band structure is that a set of bands gains an effective mass and a shift in energy, thus lifting the degeneracy of the conduction bands at the Dirac point. Then, using Kubo formula, we obtained an analytical expression for the real part of the optical conductivity. In the next section we presented our results for the optical conductivity of as a function of the photon energy. We found that in the zero limit of the chemical potential and the simultaneous presence of on- site and nearest-neighbor hopping energy deviations, the optical conductivity displays a dip-peak structure located at the photon energy corresponding to 2 times the effective hopping energy whose occurrence was explained by considering the allowed optical transitions. This is in contrast to the case of zero hopping energy deviation which similar to pristine graphene leads to a flat interband response at all frequencies. Furthermore it was shown that at finite chemical potential the interband absorbtion edge, which for pristine graphene is at $2\mu$, is splitted into two edges whose splitting depends on the amounts of the on-site and nearest-neighbor hopping energy deviations and also the chemical potential. As a notable result it was found that the Y-shaped Kekulé-patterned graphene at finite chemical potential also displays a large optical response called band nesting resonance. We showed that this effect is robust with respect to increasing temperature except when the chemical potential approaches the intersection of $E_{1\mathbf{k}}$ and $E_{3\mathbf{k}}$ energy bands. Therefore the occurrence of band nesting resonance is a significant optical signature to detect the Y-shaped Kekulé distortion in single layer graphene. We also discussed the effect of the next- nearest hopping energy deviations which is found to only lead to renormalization of the effective on-site energy deviation preserving our general results for the optical conductivity. ## 5 acknowledgment This work has been supported by Farhangian University. ## References * [1] C. Gutiérrez, C.-J. Kim, L. Brown, T. Schiros, D. Nordlund, E. B. Lochocki, K. M. Shen, J. Park and A. N. Pasupathy, Nat. Phys. 12 (2016) 950. * [2] O. V. Gamayun,V. P. Ostroukh, N. V. Gnezdilov, İ. Adagideli and C. W. J. Beenakker, New J. Phys. 20 (2018) 023016. * [3] Q.-P. Wu, L.-L. Chang, Y.-Z. Li, Z.-F. Liu, and X.-B. Xiao, Nanoscale Research Letters 15 (2020) 1. * [4] E. Andrade, R. Carrillo-Bastos, and G. G. Naumis, Phys. Rev. B 99 (2019) 035411. * [5] E. Andrade, R. Carrillo-Bastos, P. A. Pantaleón, and F. Mireles, J. Appl. Phys. 127 (2020) 054304. * [6] J. J. Wang, S. Liu, J. Wang, and J.-F. Liu, Phys. Rev. B 98 (2018) 195436. * [7] S. A. Herrera and G. G. Naumis, Phys. Rev. B 102 (2020) 205429. * [8] Y. Mohammadi, ECS J. Solid State Sci. Technol. 10(6) (2021) 061011. * [9] Y. Mohammadi and S. Bahrami, Chin. Phys. B 31 (1) (2022) 017305. * [10] M. A. Mojarro, V. G. Ibarra-Sierra, J. C. Sandoval-Santana, R. Carrillo-Bastos, and Gerardo G. Naumis, Phys. Rev. B 102 (2020) 165301. * [11] S. A. Herrera and Gerardo G. Naumis, Phys. Rev. B 104 (2021) 115424. * [12] W. A. Harrison, Solid State Theory, Dover Publications, 1970. * [13] Y. Ren, X. Deng, Z. Qiao, C. Li, J. Jung, C. Zeng, Z. Zhang, and Q. Niu, Phys. Rev. B 91 (2015) 245415. * [14] E. Andrade, G. G. Naumis and R Carrillo-Bastos, J. Phys.: Condens. Matter 33 (2021) 225301 * [15] T. Stauber, N. M. R. Peres, and A. K. Geim, Phys. Rev. B 78 (2008) 085432. * [16] G. D. Mahan, Many-Particle Physics, Plenum, New York (1990). * [17] E. J. Nicol and J. P. Carbotte, Phys. Rev. B 77 (2008) 155409. * [18] Y Mohammadi, R Moradian, F Shirzadi Tabar, Solid State Commun. 193 (2014) 1. * [19] Y. Mohammadi, Superlattices Microstruct. 156 (2021) 106955. * [20] H. Bruus and K. Flensberg, Many-Body Quantum Theory in Condensed Matter Physics: An Introduction, Oxford University Press, 2002. * [21] N. M. R. Peres, T. Stauber, Int. J. Mod. Phys. B 22 (2008) 2529. * [22] K. S. Novoselov, A. K. Geim, S. V. Morozov, D. Jiang, M. I. Katsnelson, I. V. Grigorieva, S. V. Dubonos, and A. A. Firsov, Nature 438, 197 (2005); Y. Zhang, Y.-W. Tan, H. L. Stormer, and P. Kim, Nature 438, (2005) 201. * [23] N. M. R. Peres, F. Guinea, and A. H. Castro Neto, Phys. Rev. B 73 (2006) 125411. * [24] S. A. Herrera and G. G. Naumis, Phys. Rev. B 101 (2020) 205413. * [25] A. Carvalho, R. M. Ribeiro, and A. H. Castro Neto, Phys. Rev. B 88 (2013) 115205. * [26] Z. Rashidian, Y. V. Bludov, R. M. Ribeiro, N. M. R. Peres and M. I Vasilevskiy, J. Phys.: Condens. Matter 26 (2014) 395301. * [27] L. Mennel, V. Smejkal, L. Linhart, J. Burgdörfer, F. Libisch and T. Mueller, Nano Lett. 20(6) (2020) 4242. * [28] C. J. Tabert and E. J. Nicol, Phys. Rev. B 86 (2012) 075439. * [29] V. P. Gusynin, S. G. Sharapov, and J. P. Carbotte, Phys. Rev. Lett. 96 (2006) 256802.
[orcid=0000-0003-3204-238X<EMAIL_ADDRESS>] [orcid=0000-0001-8116-9446<EMAIL_ADDRESS>] # Ontomathedu Ontology Enrichment Method Konstantin S. Nikolaev Federal State Institution of the Federal Research Center NIISI RAS Kazan Federal University Olga A. Nevzorova ###### Abstract Nowadays, distance learning technologies have become very popular. The recent pandemic has had a particularly strong impact on the development of distance education technologies. Kazan Federal University has a distance learning system based on LMS Moodle. This article describes the structure of the OntoMathEdu ecosystem aimed at improving the process of teaching school mathematics courses, and also provides a method for improving the OntoMathEdu ontology structure based on identifying new connections between contextually related concepts. ###### keywords: OntoMathEdu ontologysoftware serviceeducational applicationontology enrichment ## 1 Introduction The ontological approach is often used to organize training in a more structured format, as well as to help the student form a full picture of the current discipline, to more clearly understand the existing patterns and relationships between concepts in the problem area under consideration. The basis of such an educational infrastructure can serve as a qualitative ontology. Currently, there are many training systems that use various ontologies to organize subject data. For example, Lutoshkina et al. in the work [1], the ontology of the subject area is used to systematize the multimedia content of electronic courses (the connections between the concepts of ontology are used to construct an integral structure of the lesson). In addition, Stancin et al. [2] provide a detailed overview of the use of ontologies in education, among which the most interesting for us are ontologies describing the subject under study and e-learning services. Here are some examples of such solutions: 1\. LiReWiapproach (Conde et al. [3]). Educational ontology is used to describe the topics to be mastered by students and the pedagogical relations between the topics. 2\. MOOC-KG (Dang et al. [4]). To build a knowledge graph, an ontology is used that models knowledge about online learning resources. 3\. MKMSE (Martinez-Ramirez et al. [5]). Ontology is used to store mathematical knowledge. 4\. Ontology is used to represent the semantic description of document resources and the relationships between document resources and other types of resources (Hai [6]). 5\. CALMS (Erazo-Garzon et al. [7]). Ontologies are used to define concepts and semantic relations between academic information and conceptual aspects. 6\. Ontology is used as an expression of knowledge to support the ends of the content of education. (Kubekov et al. [8]) ## 2 Structure of the OntoMathEdu ecosystem The article discusses the structure and main tasks of the infrastructure of software components built around the ontology of school mathematics OntoMathEdu [9, 10]. Ontology OntoMathEdu is a knowledge system that describes a set of concepts, relationships between them and statements within the framework of a school course in planimetry. Figure 1 shows the composition of the OntoMathEdu digital ecosystem. The ecosystem consists of the following components. Figure 1: Components of the OntoMathEdu ecosystem 1. 1. Intelligent digital educational platform for school mathematics. This component occupies a central place in the OntoMathEdu ecosystem. Its main purpose is the application of semantic and ontological technologies in teaching school mathematics. 2. 2. Collection of questions. This collection is used as an input data set for various components of the ecosystem, in particular, for the module "Ontology Enrichment Services", as well as in the central component "Intelligent Digital Educational Platform for School Mathematics". The collection is constantly updated. 3. 3. Formula storage. This component contains formulas extracted from school geometry textbooks. Formulas are presented in various formats (plain text, LaTeX, OpenMath). We plan to use this repository as an auxiliary data set for building test tasks and other information components. 4. 4. Digital educational resources. This component combines all auxiliary data sources located on the Internet (for example, open educational databases, resources for the construction of geometric shapes, etc.) 5. 5. OntoMathEdu ontology. The OntoMathEdu ontology is a reflection of the level of knowledge corresponding to the level of school mathematics. It serves as the main repository of concepts and their relationships that are used by other services. 6. 6. Ontology enrichment service. This component includes a set of methods that are used to clarify the relationships between the concepts of ontologies and improve the horizontal connectivity of ontologies. 7. 7. Test generator. This component is used to automatically create new test tasks based on the analysis of the structure and concepts of existing tasks. 8. 8. Search by semantic formulas. A search module that performs semantic search of mathematical texts present in the ecosystem. As a result of the search, it returns a set of texts containing the requested concept and offers a list of related concepts for further navigation through the text database. 9. 9. Recommendation system. This module helps the user to study various sections of the training course, providing an individual learning trajectory. ## 3 Ontology enrichment service The ontology OntoMathEdu is currently in the active stage, and contains more than 900 concepts and more than 20 relationships. The task of introducing new relations between concepts is urgent. We have developed a method for introducing new relations between concepts based on the use of a collection of questions. To find new connections between concepts, we use a collection of test questions on school geometry. Our main idea is as follows. Each question contains some basic concepts of ontology, about which the question is asked. We assume that the answer to this question can be found with the help of ontology, in which there is a chain of concepts of limited length connecting the original concepts. In addition, the search for such connections allows us to evaluate the structural properties of the ontology, namely the horizontal connectivity of the resource. The source of the texts is a collection of school test tasks on geometry with the markup of nominal groups. These nominal groups will act as candidates to search for the corresponding concept in the OntoMathEdu ontology. A noun group of the first type (Noun phrase, NP) is understood as a syntactic construction in which there is a vertex (noun) and dependent words. An example of NP is a "finite set of points", where the real "set" is the vertex of a given phrase with two dependent words. A nominal group of the second type (Prepositional phrase, PP) is a phrase in which the main word is a preposition. An example of PP is the phrase "up to two characters after the dot", where the preposition "to" is the vertex of the nominal group. Then, for each pair of concepts, all existing relationships are searched, and the search is performed twice: in the first case, only hierarchical relationships (rdfs:subClassOf and ome:hasChild) are used as relationships between concepts, and in the second case, all possible relationships are used. We call the first type of relationship as a "hierarchical relationship", and the second as a "complete relationship". The presence of an optimal relationship is an event in which the shortest complete link is shorter than the hierarchical one. This connection is optimal, because any pair of concepts is somehow connected by a hierarchical connection (we can always climb from the first concept to the node shared with the second concept and descend to the second concept), but this connection will not be the shortest if there are direct connections between the concepts under consideration. The resulting set of shortest links with an indication of optimality is sent for analysis to a group of experts for manual updating of the ontology. It can also be used as auxiliary information when checking test tasks. Figure 2: Block diagram of the ontology enrichment method Figure 2 shows a block diagram of the task of enriching ontology through method names. The following sections will describe the main stages of the algorithm for finding optimal connections between concepts in the texts of questions with the name of the method. Here are some comments on Figure 2. 1. 1. Extraction of NP and PP. At this stage, we need to select a set of NP and PP in the texts of questions and text variants of answers (this means that we do not consider answer options for computational problems and for problems with the choice of an answer presented as a number or a set of characters that are not a dictionary unit). Thanks to the already preliminary markup of the text by the nominal groups NP (TERM1 tag) and PP (TERM2 tag), this stage is reduced to highlighting the contents of these tags using the standard xml library for Python 3. 2. 2. Tokenization and normalization of phrases from nominal groups (NP and PP): TokenizeAndLemmatizeTexts() function. Tokenization of phrases received at the last stage is performed using the nltk library. Normalization of the resulting set of tokens is carried out using the pymorphy2 library, which has the necessary functionality to bring words to a normal form. For example, for the word "triangles", the normal form is "triangle". 3. 3. Tokenization and normalization of ontology concepts: TokenizeAndLemmatizeOntology() function. Tokenization of phrases received at the last stage is performed using the nltk library. Tokens are normalized using the pymorphy2 library. 4. 4. Match search: Match() function. Next, we need to find similar sequences of tokens in the question text and names of ontological concepts, using a modified Jacquard measure to work with a sequence of strings. The sequence element in our case is a lemma from the phrases NP or PP. However, since the words in the compared sequences may not completely match, we also apply the Jacquard measure when searching for common words between sequences. In this case, the elements of the sequence will be individual letters. When performing this method, we will introduce two thresholds: a lower accuracy threshold at which two words are considered the same, and a lower accuracy threshold at which two sequences of words are considered similar. Their values were obtained experimentally on a specific set of questions. 5. 5. Search for connections through hierarchical relations: the SearchHierarchicalConnections() function. In this method, we look for connections between pairs of concepts located in the same question, using only hierarchical relationships (rdfs:subClassOf and ome:hasChild) as relationships between neighboring concepts. This type of connection is guaranteed to be present between any two concepts in the ontology, but it is not necessarily optimal. We select the shortest connection from the resulting set of this method. 6. 6. Search for connections through any relations: Search All Connections(). In this method, we look for connections between pairs of concepts located in the same question, using all types of relations as relations between neighboring concepts. We select the shortest connection from the resulting set of this method. 7. 7. Comparison of shortest distances with Previous methods: CompareDistances() function. In this method, we check a link using all types of links, which is shorter than the corresponding link using only hierarchical links. If such a connection is found, then the optimal connection between a pair of concepts is confirmed. Table 3 shows examples of optimal connections found in ontology, indicating the lengths of these connections. ## 4 Conclusion This article examines the structure of the OntoMathEdu ecosystem, designed to support the process of personalized teaching of school mathematics, and also provides an algorithm for enriching ontology, based on the contextual proximity of concepts in geometry tasks, which contributes to improving the horizontal connectivity of ontology. Future work is related to the implementation of the personalization component in all implemented test generators, as well as the overall development of these components. In addition, it is planned to develop a component for the automatic analysis of a detailed response by drawing up a framework (scheme) of the solution, and filtering tasks by types of these frameworks. Table 1: Examples of optimal connections found Source theorem | First concept | Second concept | Link length ---|---|---|--- If the diagonals of a given quadrilateral are perpendicular, then this quadrilateral has: 1) equal middle lines, 2)<perpendicular> <middle lines> 3) equal opposite angles | Perpendicular | Triangle middle line | 3 If a circle can be described around a quadrilateral, then 1) the center of the circle is equidistant from the sides of the quadrilateral 2) the bisectors of the angles of the quadrilateral intersect at one point 3) the sum of <opposite angles> is equal to <right angle> | Opposite corners of a quadrilateral | Right angle | 4 <Segments> connecting the midpoints of the sides of any quadrilateral and the segment connecting the midpoints of <diagonals> | The middle of the segment | Diagonal of the polygon | 3 The sum of two axial symmetries with intersecting axes can be replaced by <rotation> around the <point> at which the ax-es intersect at an angle equal to the angle between the axes | Point | Rotation angle | 4 ###### Acknowledgements. The article was prepared within the framework of the government task of the Federal State Institution of the Federal Research Center NIISI RAS for 2022-2024 (topic No. 0580-2022-0014 (FNEF-2022-0014)). ## References * Lutoshkina and Vysotin [2011] N. Lutoshkina, A. Vysotin, Use of ontologies in interactive courses, Info-Strategiya 2011: Obshchestvo. Gosudarstvo. Obrazovanie: Sbornik materialov konferencii (2011) 56–59. * Stancin et al. [2020] K. Stancin, P. Poscic, D. Jaksic, Ontologies in education – state of the art, Education and Information Technologies 25 (2020) 5301–5320. doi:10.1007/s10639-020-10226-z. * Conde et al. [2019] A. Conde, M. Larranaga, A. Arruarte, J. A. Elorriaga, A Combined Approach for Eliciting Relationships for Educational Ontologies Using General-Purpose Knowledge Bases, IEEE Access 7 (2019) 48339–48355. doi:10.1109/ACCESS.2019.2910079. * Dang et al. [2019] F. Dang, J. Tang, S. Li, MOOC-KG: A MOOC knowledge graph for cross-platform online learning resources, ICEIEC 2019 - Proceedings of 2019 IEEE 9th International Conference on Electronics Information and Emergency Communication (2019) 596–603. URL: https://ieeexplore.ieee.org/abstract/document/8784572?casa_token=Hn9FqPbXI1AAAAAA:X0MhRkIRwXQjleGUG0HWTmaPOj0D2jsvR9ATrUG1_wWOWHSlSrVkv400fbFbsYX_iKBDYP7vYyuv. doi:10.1109/ICEIEC.2019.8784572. * Martinez-Ramirez et al. [2018] Y. Martinez-Ramirez, A. Ramirez-Noriega, M. Zayas-Esquer, S. Miranda-Mondaca, J. Armenta-Bojorquez, M. Quintero-Fonseca, M. Gonzialez-Videgaray, C. Cortes-Velazquez, Architecture of mathematical knowledge management system in education: Ontology-based and case-based, in: 2018 International Symposium on Computers in Education, Proceedings, 2018, pp. 1–5. doi:10.1109/SIIE.2018.8586771. * Hai [2019] J. Hai, An efficient semantic retrieval method for network education information resources, Proceedings - 2019 11th International Conference on Measuring Technology and Mechatronics Automation, ICMTMA 2019 (2019) 522–526. doi:10.1109/ICMTMA.2019.00121. * Erazo-Garzon et al. [2019] L. Erazo-Garzon, A. Patino, P. Cedillo, A. Bermeo, CALMS: A context-aware learning mobile system based on ontologies, in: 2019 6th International Conference on eDemocracy and eGovernment, ICEDEG 2019, 2019, pp. 84–91. doi:10.1109/ICEDEG.2019.8734423. * Kubekov et al. [2018] B. Kubekov, N. Zhaksybaeva, V. Naumenko, A. Utegenova, Methodology of formation of educational resources on the basis of ontology, IEEE 12th International Conference on Application of Information and Communication Technologies, AICT 2018 - Proceedings (2018). doi:10.1109/ICAICT.2018.8747069. * Kirillovich et al. [2020] A. Kirillovich, O. Nevzorova, M. Falileeva, E. Lipachev, L. Shakirova, OntomathEdu: A linguistically grounded educational mathematical ontology 12236 LNAI (2020) 157–172. doi:10.1007/978-3-030-53518-6_10. * Kirillovich et al. [2019] A. Kirillovich, L. Shakirova, M. Falileeva, E. Lipachev, Towards an Educational Mathematical Ontology, INTED2019 Proceedings 1 (2019) 6823–6829. doi:10.21125/inted.2019.1657.
# Deep Neural Network Techniques for Monaural Speech Enhancement and Separation: State of the art Analysis Peter Ochieng Department of Computer Science. University of Cambridge. <EMAIL_ADDRESS> ###### Abstract Deep neural networks (DNN) techniques have become pervasive in domains such as natural language processing and computer vision. They have achieved great success in tasks such as machine translation and image generation. Due to their success, these data driven techniques have been applied in audio domain. More specifically, DNN models have been applied in speech enhancement and separation to achieve denoising, dereverberation, speaker extraction and speaker separation in monaural speech intelligibility improvement. In this paper, we review some dominant DNN techniques being employed to achieve speech enhancement and separation. The review looks at the pipeline of speech enhancement and separation techniques from feature extraction, how DNN-based tools models both global and local features of speech and model training (supervised and unsupervised). The review also covers the use of domain adaptation techniques and pre-trained models to boost speech enhancement process. ## 1 Introduction Techniques for monaural speech intelligibility improvement can be categorised either as speech enhancement or separation. Speech enhancement involves isolating a target speech either from noise [1] or a mixed speech [2]. Speech enhancement involves tasks such as dereverberation, denoising and speaker extraction. Speaker separation on the other hand seeks to estimate independent speeches composed in a mixed speech [3]. Speech enhancement and separation have applications in multiple domains such as automatic speech recognition, mobile speech communication and designing of hearing aids [4]. Initial research on speech enhancement and separation exploited techniques such as non-negative matrix factorization [5] [6] [7] probabilistic models [8] and computational auditory scene analysis (CASA)[9]. However, these techniques are tailored for closed-set speakers (i.e., do not work well with mixtures with unknown speakers) which significantly restricts their applicability in real environments. Due to the recent success of deep learning models in different domains such natural language processing and computer vision, these data driven techniques have been introduced to process audio dataset. In particular, DNN models have become popular in speech enhancement and separation and have achieved great performance in terms of boosting speech intelligibility and their ability to enhance speech with unknown speakers [10] [11]. In order to be effective in speech enhancement and separation, DNN models must extract important features of speech, maintain order of audio frames, exploit both local and global contextual information to achieve coherent separation of speech data. This necessitates that DNN models should include techniques tailored to meet these requirements. Discussion of these techniques is the core subject of this review. Further, in computer vision and text domain, large pre-trained models are used to extract universal representations that are beneficial to downstream tasks. The review discusses the impact of pre-trained models to the speech enhancement and separation domain. It also discusses DNN techniques being adopted by speech enhancement and separation tools to reduce computation complexity to enable them work in low latency and resource constrained environments. The review therefore focuses on the whole pipeline of DNN application to speech enhancement and separation, i.e., from feature extraction, model implementation, training and evaluation. Our goal is to uncover the dominant techniques at each level of DNN implementation. In each section, we highlight key emerging features and challenges that exist. A recent review [12] only looked at supervised techniques of performing speech separation and in this review, we discuss both supervised and unsupervised methods. Moreover, with the fast-growing field of deep learning, new techniques have emerged that necessitates a new look into how these techniques have been implemented in speech enhancement and separation. The review is constrained to discussing how DNN techniques are being applied to monaural speech enhancement and so we do not focus on multi- channel speech separation (which has been covered in [13]). The paper first explains the types of speech enhancement and separation (section 2) by highlighting their key elements and the tools that focus on each type. It discusses the key speech features that are being used by speech enhancement and separation tools in section 3. This section looks at how the features are derived and how they are used to train the DNN models in supervised learning technique. Section 5 discusses the techniques the tools use to model long dependencies that exist in speech. The paper discusses model size compression techniques in section 6. In section 7, the paper discusses some of the popular objective functions used in speech enhancement and separation. Section 8 discusses how some tools are implementing unsupervised techniques to achieve speech enhancement and separation. Section 9 discusses how the speech separation and enhancement tools are being adapted to the target environment. In section 10 the paper looks at how pre-trained models are being utilized in the speech enhancement and separation pipeline. Finally, section 11 looks at future direction. Figure 1 gives an overall organization and topics covered by the paper. forked edges, for tree= edge+=thick, inner color=gray!5, outer color=gray!20, rounded corners=2pt, draw, thick, tier/.option=level, align=center, font=, blur shadow, , where level<=1 rotate=90, anchor=east, , [Deep Neural Network Techniques for Monaural Speech Enhancement and Separation [, coordinate, [Types of SE and separation [Speech separation] [Dereverberation] [Denoising] [Speech extraction]]] [Features [Fourier spectrum [Log-power spectrum features][Mel-frequency spectrum features][DFT magnitude features][Complex DFT features][Complementary features][Supervised Training [Spectral Mapping Techniques][Spectral Masking Techniques][Generative Modelling]] [Phase handling]] [Time domain [Supervised training[Adaptive front-end method][Waveform mapping][Generative modelling]]] ] [Long term dependency modelling [RNN] [TCN] [Use of transformers]] [Model size reduction techniques, calign with current [Use of dilated convolution] [Parameter quantization] [Use of depthwise separable convolution] [Knowledge distillation] [Parameter pruning] [Weight sharing:] ] [Objective functions [SSTOI] [SI-SDR, calign with current] [PMSQR] ] [Unsupervised Techniques] [Domain Adaptation Techniques] [Use of Pre-trained Models] [Future directions]] ] [expatriate, inner color=green!10, outer color=green!25, child anchor=west, edge path’=(!u.parent anchor) -| (.child anchor), before drawing tree=y’+=7.5pt ] ] Figure 1: Overall structure of the topics covered by this review. ## 2 Types of speech separation and enhancement ### 2.1 Speech separation Scenarios arise where more than one target speech signals are composed in a given speech mixture and the goal is to isolate each independent speech composed in a mixture. This problem is known as speech separation. For a mixture that is composed of $C$ independent speech signals $x_{c}(n)$ with $c=1,\cdots,C$, a recording $y(n)$ composed of the $C$ speech signals can be represented as: $y(n)=\sum_{c=1}^{C}x_{c}(n)$ (1) Here, $n$ indexes time. The goal of speech separation is to estimate each independent $x_{c}$ speech signal composed in $y(n)$. Separating speech from another speech is a daunting task by the virtue that all speakers belong to the same class and share similar characteristics[14]. Some models such as [15] and [16] lessen this by performing speech separation on a mixed speech signal based on gender voices present. They exploit the fact that there is large discrepancy between male and female voices in terms of vocal track, fundamental frequency contour, timing, rhythm, dynamic range etc. This results in a large spectral distance between male and female speakers in most cases to facilitate a good gender segregation. For speech separation that the mixture involves speakers of the same gender, the separation task is much difficult since the pitch of the voice is in the same range [14]. Most speech separation tools that solve this task such as [17] [18] [19] [20] [14] and [10] cast the problem as a multi-class regression. In that case, training a DNN model involves comparing its output to a source speaker. DNN models always output a dimension for each target class and when multiple sources of the same type exist, the system needs to select arbitrarily which output dimension to map to each output and this raises a permutation problem (permutation ambiguity) [14]. Taking a case of a two speaker separation, if the model estimates $\hat{a_{1}}$ and $\hat{a_{2}}$ as the magnitude of the reference speech magnitudes $a_{1}$ and $a_{2}$ respectively, it is unclear the order in which the model will output the estimates i.e. the order of output can either be $\\{\hat{a}_{1},\hat{a}_{2}\\}$ or $\\{\hat{a}_{2},\hat{a}_{1}\\}$. A naive approach shown in figure 2 [21] is to present the reference speech magnitudes in a fixed order and hope that it is the same order in which the system will output its estimation. Figure 2: Naive approach of solving label matching problem for a two-talker speech separation model In case of a mismatch, the loss computation will be based on the wrong comparison resulting in low quality of separated speeches. Therefore, systems that perform speaker separation have an extra burden of designing mechanisms that are geared towards handling the permutation problem. There are several strategies that are being implemented by speech separation tools to tackle permutation problem. In [19], a number of DNN techniques are implemented that estimates two clean speeches contained in a two-talker mixed speech. They employ supervised training to train DNN models to discriminate the two speeches based on average energy, pitch and instantaneous energy of a frame. Work in [22] and [21] introduce permutation invariant training (PIT) technique of computing permutation loss such that permutations of reference labels are presented as a set to be compared with the output of the system. The permutation with the lowest loss is adopted as the correct order. For a a two- speaker separation system introduced earlier, the reference sources permutation will be $\\{a_{1},a_{2}\\}$ and $\\{a_{2},a_{1}\\}$ such that the possible permutation losses are computed as: $loss_{1}=D([a_{1},a_{2}],[\hat{a}_{1},\hat{a}_{2}])=D[a_{1},\hat{a}_{1}]+D[a_{2},\hat{a}_{2}]$ $loss_{2}=D([a_{1},a_{2}],[\hat{a}_{2},\hat{a}_{1}])=D[a_{1},\hat{a}_{2}]+D[a_{2},\hat{a}_{1}]$ The one that returns the lowest loss between the two is selected as the permutation loss to be minimized (see figure 3). For an $S$ speaker separation system a total of $S!$ permutations are generated. Figure 3: Permutation invariant training implementation of label matching for a two-talker speech separation model For a system that performs $S$ speaker separation and $S$ is high (e.g. 10), implementation of PIT which has a computation complexity of $O(S!)$ is computationally expensive [23] [24]. Due to this, [24] casts the permutation problem as a linear sum problem where Hungarian algorithm is exploited to find the permutation which minimizes the loss at computation complexity of $O(S^{3})$. Work in [23] proposes SinkPIT loss which is based on Sinkhorn’s matrix balancing algorithm. They utilize the loss to reduce the complexity of PIT loss from $O(C!)$ to $O(kC^{2})$. Work in [17] employs minimum loss permutation computation at each time step $t$. The best permutation (argmin) at each time-step is exploited to re-order the embedding vectors to be consistent with the training labels. To evade the permutation problem, they train two separate DNN models for each of the two speakers to be identified. Another prominent technique of handling permutation problem is to employ a DNN clustering technique [25] [26] [20] [27] [28] to identify the multiple speakers present in a mixed speech signal. The DNN $f_{\theta}$ accepts as its input the whole spectrogram $X$ and generates a $D$ dimension embedding vector V i.e., $V=f_{\theta}(X)\in R^{N\times D}$. Here, the embedding $V$ learns the features of the spectrogram $X$ and is considered a permutation- and cardinality-independent encoding of the network’s estimate of the signal partition. For the network $f_{\theta}$ to be learn how to generate an embedding vector $V$ given the input $X$, it is trained to minimize the cost function. $C_{Y}(V)=||VV^{T}-YY^{T}||_{F}^{2}=\sum_{ij}(<v_{i},v_{j}>-<y_{i},y_{j}>)^{2}$ (2) Here, $Y=\\{y_{i,c}\\}$ represents the target partition that maps the spectrogram $S_{i}$ to each of the $C$ clusters such that $y_{i,c=1}$ if element $i$ is in cluster $c$ . $YY^{T}$ is taken here as a binary affinity matrix that represents the cluster assignment in a partition-independent way. The goal in equation 2 is to minimise the distance between the network estimated affinity matrix $VV^{T}$ and the true affinity matrix $YY^{T}$. The minimization is done over the training examples. $||A||_{F}^{2}$ is the squared Frobenius norm. Once $V$ has been established, its rows are clustered into partitions that will represent the binary masks. To cluster the rows $v_{i}$ of $V$, K-means clustering algorithm is used. The resulting clusters of $V$ are then used as binary masks to separate the sources by applying the masks on mixed spectrogram $X$. The separated sources are then reconstructed separated by using inverse STFT. Even though PIT is popular in speech separation models, it is unable to handle the output dimension mismatch problem where there is a mismatch on the number of speakers between training and inference [29]. For example, training a speech separation model on $n$ speaker mixtures but testing it on $t\neq n$ speaker mixtures. The PIT-based methods cannot directly deal with this problem due to their fixed output dimension. Most speech separation models such as [30] [21] [31] [32] deal with the problem by setting a maximum number of sources $C$ that the model should output from any given mixture. If an inference mixture has $K$ sources, where $C>K$, $C-K$ outputs are invalid, and the model needs to have techniques to handle the invalid sources. In case of invalid sources, some models such as [31], [30], [21] design the model to output silences for invalid sources while [32] outputs the mixture itself which are then discarded by comparing the energy level of the outputs relative to the mixture. The challenge with models that output silences for invalid sources is that they rely on a pre-defined energy threshold, which may be problematic if the mixture also has a very low energy [32]. Some models handle the output dimension mismatch problem by generating a single speech in each iteration and subtracting it from the mixture until no speech is left [33], [34], [35][36], [37]. The iterative technique despite being trained with a mixture with low number of sources can generalize to mixtures with a higher number of sources [35]. It however faces criticism that setting iteration termination criteria is difficult and the separation performance decreases in later iterations due degradations introduced in prior iterations [35]. Other speech separation models include [38] [39][40] [31] [41] [19] [20] [42]. ### 2.2 Speaker extraction Some speech enhancement DNN models have been developed where in a mixed speech such as an equation 1, they design methods to extract a single target speech. These models focus only on a single target speech $x_{t}$ and treat all other speeches as interfering signals, therefore they modify equation 1 as shown in 3. $y(n)=\sum_{c=1}^{C}x_{c}(n)=x_{t}(n)+\sum_{c\not=t}^{C}x_{c}(n)$ (3) where $x_{t}(n)$ is the target speech. By focusing on only a single target speech, the permutation ambiguity problem is avoided. They formulate the speech extraction task into a binary classification problem, where the positive class is the target speech, and the negative class is formed by the combination of all other speakers. A popular technique of speaker extraction is to give as input to the DNN models additional speaker dependent information that can be used to isolate a target speaker [43]. Speaker dependent information can be injected into the DNN models by either concatenating speaker dependent auxiliary clues with the input features or adapting part of the DNN model parameters for each speaker [44]. This addition information about a speaker injects a bias that is necessary to differentiate the target speaker from the rest in the mixture [2]. Several auxiliary clues have been exploited by DNN models which include pre-recorded enrolment utterances of the target speaker [44] [2] [45] [46] [47] [48], electroglottographs (EGGs) of the target speaker [49] and i-vectors extracted at speaker level [50] [51]. Tool in [52] adapt parameters for each speaker by allocating a speaker dependent module to a selected intermediate layer of DNN. Speech extraction tool in [53] does not use auxiliary clues of the target speaker but design attractor points that are compared with the mixed speech embeddings to generate the mask used to extract the target speech. ### 2.3 Dereverberation This is a speech enhancement technique that seeks to eliminate the effect of reverberation contained in speech. When speech is captured in an enclosed space by a microphone that is at distance $d$ from the talker, the observed signal consists of a superposition of many delayed and attenuated copies of the speech resulting from reflections of the enclosed space walls and existing objects within the space (see figure 4) [54]. The signal received by the microphone consists of direct sound, reflections that arrive shortly after direct sound ( within approximately 50ms) i.e., early reverberation and reflections that arrive after early reverberation i.e., late reverberation [55]. Normally, early reverberation does not affect speech intelligibility much [56] and much of perceptual degradation of speech is attributed to late reverberation. Speech degradation due to reverberation can be attributed to two types of masking [57], overlap masking- where the energy of a preceding phoneme overlaps with the one following or self-masking-where internal temporal which refers to the time and frequency alterations of an individual phoneme. Reverberation therefore can be viewed as the convolution of the direct sound and the room impulse response (RIR). A reverberant speech can be formally represented according to equation 4: Figure 4: How Reverberation happens. $y(t)=h(t)*s(t)$ (4) Here, $*$ represents convolution, $s(t)$ is the clean anechoic speech. $h(t)$ represents room impulse response i.e., direct speech $h_{d}(t)$, early reverberation $h_{e}(t)$and late reverberation $h_{l}(t)$. Hence $h(t)$ can be represented as $h(t)=h_{d}(t)+h_{e}(t)+h_{l}(t)$ (5) Using the distributive property of convolution [58] equation 4 becomes: $y(t)=h_{d}(t)*s(t)+h_{e}(t)*s(t)+h_{l}(t)*s(t)$ (6) The goal of dereverberation is therefore to establish $s(t)$ from $y(t)$. Hence it can be viewed as a deconvolution between the speech signal and RIR [59]. Dereverberation is considered a more challenging task than denoising for a number of reasons. First, it is difficult to pinpoint direct speech from its copies especially when the reverberation is strong. Secondly, the key underlying assumption of sparsity and orthogonality of speech representations in the feature domain that is commonly used in monaural mask-based speech separation does not hold for speech under reverberation [60]. Due to these unique features of reverberation, most tools designed for denoising, or speaker separation are ill poised to perform dereverberation [60]. The DNN tools for speaker separation and denoising mostly make assumption that they are working on reverberation free speech hence do not make special consideration for eliminating reverberation (with exception of a few such as [61] [62]). For instance, in [60] they demonstrate that SepFormer [11] performance can significantly improve by making adjustments to include techniques that handle reverberation. Several deep learning models have been designed with a goal to estimate clean speech from a reverberant one. Similar to speech denoising and speech separation, DNN models performing dereverberation exploit these models to fit a nonlinear function to map features of a reverberant speech to features of clean anechoic speech either directly [63] [64] [65] [66] [67] [68] or by use of mask [18] [55] [69] [70] [65] [55] [70]. Therefore, one way of categorising the existing dereverberation DNN tools is based on the type of target (spectrogram or ratio mask) they employ. Another way in which dereverberation tools can be categorised is based on whether a tool performs general dereverberation ( i.e suppress $h(t)$ see equation 4) or focus only on eliminating late reverberation ($h_{l}*s(t)$ see equation 6). Tools such as [71] [59] [72] [73] [74] [75] explore elimination of late reverberation. This is because early reverberation does not affect speech intelligibility much. Finally, the DNN dereverberation tools can be categorised based on the type of training technique used (supervised or unsupervised). Tools such as [64] and [18] perform speech dereverberation by implementing supervised training where the DNN model is trained to directly estimate features clean speech when given features from a reverberant speech. $D(k,f)=M(k,f)\times Y(k,f)$ (7) Here $D(k,f)$, $M(k,f)$and $Y(k,f)$ are the STFTs of the clean speech, the ideal ratio mask, and the reverberant speech at time frame k and frequency channel f respectively. Work in [76] exploits conditional GAN to perform unsupervised dereverberation of a reverberant speech. Dereverberation in DTF magnitude domain: When dereverberation is to be performed in DFT magnitude domain (see section 3.1), a DFT has to be applied to equation 4 such that, $DFT(y(t))=Y(t,f)=H(t,f)\times S(t,f)$ (8) the assumption in equation 8 is that the convolution of the clean signal $s(t)$ with RIR $h(t)$ corresponds to the multiplications of their Fourier transform in the T-F domain. However, this is only true if the extent of $H(t,f)$ is smaller than the analysis window [60]. Therefore, when performing dereverberation in the TF domain the selection of the window is crucial on the performance of the DNN model [60]. Target selection in dereverberation:In dereverberation training, most tools use direct speech as the target. This therefore means that the estimated speech will have to be compared with the direct path speech via a selected loss function. This has the potential of resulting in large prediction errors which can cause speech distortion [59]. Due to this, recent work [75] proposes the use of a target that has early reverberation. By doing this, they suggest it will improve the quality of enhanced speech. In fact, experiments in [75] demonstrate that allowing early reverberation in the target speech improves the quality of enhanced speech. ### 2.4 Speech denoising This is a speech enhancement technique of separating background noise from the target speech. Formally, the noisy speech is represented as: $y_{t}=s_{t}+n_{t}$ (9) where $y_{t}$ is the noisy speech, $s_{t}$ is the target speech and $n_{t}$ is the noise. Speech denoising seeks to isolate a single target speech from noise. Hence data driven DNN models are optimized to predict $s_{t}$ from $y_{t}$. Since speech denoising has only a single target speech it does not suffer from global permutation ambiguity problem. Some DNN tools that perform speech denoising include [77] [78] [1] [79] [80] [81] [82] [83] [84] [85]. ## 3 Speech separation and enhancement features Speech enhancement and separation tools’ input features can be categorised into two: 1. 1. Fourier spectrum features. 2. 2. Time domain features. ### 3.1 Fourier spectrum features Speech enhancement and separation tools that use these features do not work directly on the raw signal (i.e., signal in the time domain) rather they incorporate the discrete Fourier transform (DFT) in their signal processing pipeline mostly as the first step to transform a time domain signal into frequency domain. These models recognise that speech signals are highly non- stationary, and their features vary in both time and frequency. Therefore, extracting their time-frequency features using DFT will better capture the representation of speech signal [86]. To demonstrate the DFT process we exploit a noisy speech signal shown in equation 10. The same process can be applied in speech separation. A noisy raw waveform signal of speech, $y(t)$, can be represented as in equation 10. $y(t)=x(t)+n(t)$ (10) where $x(t)$ and $n(t)$ represent discrete clean speech and noise respectively. Since speech is assumed to be statistically static for a short period of time, it is analysed frame-wise using DFT as shown in equation 11 [86] [87] [88]. $Y[t,k]=\sum_{m=\infty}^{-\infty}y(m)w(t-m)\exp^{-j2\pi km/L}$ (11) Here, $k$ represents the index ( frequency bin) of the discrete frequency, $L$ is the length of the frequency analysis and $w(n)$ is the analysis window. In speech analysis, the Hamming window is mostly used as $w(n)$ [89]. Once the DFT has been applied to the signal $y(t)$, it is transformed into time- frequency domain represented as: $Y[t,k]=X[t,k]+N[t,k]$ (12) $Y[t,k],X[t,k]$ and $N[t,k]$ are the DFT representations of the noisy speech, clean speech and noise respectively. Each term in equation 12 can be expressed in terms of DFT magnitude and phase spectrum. For example, the polar form (including magnitude and phase) of the noisy signal $Y[t,k]$ is: $Y[t,k]=|Y[t,k]|\exp^{j\angle{Y[t,k]}}$ (13) $|Y[t,k]|$ and $\angle{Y[t,k]}$ are the magnitude and phase spectra of $Y[t,k]$ respectively. Equation 13 can be written in Cartesian coordinates as shown in equation 14. $Y[t,k]=|Y[t,k]|(\cos{\theta}+i\sin{\theta})=|Y[t,k]|\cos{\theta}+i|Y[t,k]|\cos{\theta}$ (14) Both the phase and the magnitude are computed from the real and the imaginary part of $Y[t,k]$ i.e. $|Y[t,k]|=\sqrt{\mathbb{R}(Y[t,k])^{2}+\Im{(Y[t,k]^{2})}}$ (15) $\angle{Y[t,k]}=\tan^{-1}\frac{\Im{(Y[t,k])}}{\mathbb{R}(Y[t,k])}$ (16) All models that work with the Fourier spectrum features either use the DFT representations directly as the input of the model or further modify the DFT features. The features based on Fourier spectrum include: 1. 1. Log-power spectrum features. 2. 2. Mel-frequency spectrum features. 3. 3. DFT magnitude features. 4. 4. Complex DFT features. 5. 5. Complementary features. DFT magnitude features: These are features where the mixed raw waveform $y(t)$ is first converted into time-frequency (TF) representation (spectrogram) using DFT ( specifically, short-time Fourier transform (STFT) (equation 11) [90]. The magnitude of the time-frequency representation (equation 13) acts as the input to a deep neural network (DNN) model for speech separation. The DNN model is then trained to learn how to separate the TF-bins such that those that comprise each source speech are grouped together. DNN speech enhancements and separation models that exploit DFT features include systems such as [91] [83] [92] [93] [94] [95]. The use of DFT magnitude as features work with high frequency resolution hence necessitating the use of larger time window which is typically more that 32ms [20] [21] for speech and more than 90ms for music separation [96]. Due to this, these models must handle increased computational complexity [97]. This has motivated other speech separation models to work with lower dimensional features as compared to those of DFT magnitude. DFT complex features: Unlike the DFT magnitude features that only use the magnitude of T-F representations, tools that use DFT complex features include both the magnitude and the phase of the noisy (mixed) speech signal in the estimation of the enhanced or separated speech. Therefore, each T-F unit of a complex features is a complex number with a real and imaginary component (see equation 13). The magnitude and phase of a signal is computed according to equation 15 and 16 respectively. Tools that use DFT complex features include [98] [55] [99] [100]. Mel-frequency cepstral coefficients (MFCC) features: Given the mixed speech signal such as in equation 10, to extract Mel frequency cepstral features, the following steps are executed: 1. 1. Perform DFT of the input noisy signal $DFT(y(t))=Y[t,k]=X[t,k]+N[t,k]$ 2. 2. Given the DFT features $Y[n,k]$ of the input signal, a filterbank with M filters i.e. a $1\leq m\leq M$ is defined where $m$ is a triangular filter given by: $\displaystyle H_{m}[k]=\begin{cases}0&k<f[m-1]\\\ \frac{(2(k-f[k-m])}{f[m+1]-f[m-1])(f[m]-f[m-1])}&f[m-1]\leq k\leq f[m]\\\ \frac{2(f[m+1]-1)}{(f[m+1]-f[m-1])(f[m+1]-f[m])}&f[m]\leq k\leq f[m+1]\\\ 0&k>f[m+1]\end{cases}$ (17) The filters are used to compute the average spectrum around centre frequencies with increasing bandwidths as shown in figure 1. Here, $f[m]$ are uniformly spaced boundary points in the Mel-scale which is computed according to equation 18. The Mel-scale B is given by equation 19 and $B^{-1}$ which is its inverse is computed as shown in equation 20. $f[m]=\frac{N}{F_{s}}B^{-1}(B(f)+m\frac{B(f_{h})-B(f_{l})}{M+1})$ (18) $F_{s}$ is the sampling frequency, $f_{l}$ and $f_{h}$ represent the lowest and the highest frequencies of the filter bank in Hz. N is the size of DFT and M is the number of filters. $B(f)=1125\ln{(1+\frac{f}{700})}$ (19) $B^{-1}(b)=700(\exp{(\frac{b}{1125})-1})$ (20) Figure 5: Triangular filters used in the computation of the Mel-cepstrum using equation 18 3. 3. Scale the magnitude spectrum $|Y[t,k]|$ of the noisy signal in both frequency and magnitude using mel-filter bank $H(k,m)$ and then take the logarithm of the scaled frequency. $X^{\prime}(m)=\log(\sum_{k=0}^{N-1}|Y[t,k]|^{2}H_{m}[k]$ (21) for $m=0,\cdots,M$where $M$ is the number of filter banks. 4. 4. Compute the Mel frequency by computing the discrete cosine transform of the $m$ filter outputs as shown in equation 22. $c[n]=\sum_{m=0}^{m-1}X^{\prime}(m)\cos{(\pi n(m+1/2)/M)}$ (22) where $0\leq n<M$ The motivation for working with MFCC is that it results in reduced resolution space as compared to DFT features. Fewer parameters are easier to learn and may generalise better to unseen speakers and noise [97]. The challenge however with working on a reduced resolution such as MFCC is that the DNN estimated features must be extrapolated to the DFT feature space. Due to working on a reduced resolution, MFCC degree-of-freedom will be restricted by the dimensionality of the reduced resolution feature space which is much less than that of the DFT space. The low-rank approximation generates a sub-optimal Wiener filter which cannot account for all the added noise content and yields reduced SDR [97]. MFCC features have been exploited in tools such as [101] [68] [102] [83][103] [104]. Log-power spectra features: To compute these features, a short-time Fourier analysis is applied to the raw signal computing the DFT of each overlapping waveform (see equation 11). The log-power spectra are then computed from the output of the DFT. Consider a noisy speech signal in the time-frequency domain i.e., where DFT has been applied to the signal (see equation 12). From equation 14, the power spectrum of the noisy signal can be represented as in equation 23. $|Y[k]|^{2}=|X[k]|^{2}+N[k]|^{2}=|X[k]|^{2}+|N[k]|^{2}+2|X[k]||N[k]|\cos{\theta}$ (23) Here, $\theta$ represents the angle between the two complex variables $|X[k]|$ and $|N[k]|$. Most models that exploit log-power spectra features ignore the last term ( assume the value to be zero) and employ equation 24. $Y^{l}=log(|Y[k]|^{2})=log(|X[K]|^{2})+log(|N[k]|^{2})$ (24) Figure 6 [105] summarises the process of log-power feature extraction. Figure 6: Demonstrating feature extraction. Here, $Y^{t}$ represents the noisy signal in time domain, $Y^{f}$ represents the transformed signal in the frequency domain. $Y^{l}$ is the log power features of the input signal Examples of models that use Log-power spectra features include [98] [105] [106] [107]. Complementary features: Since different features strongly capture different acoustic features which characterise different properties of the speech signal, some DNN models exploit a combination of the features to perform speech separation. This is based on works such as [108] and [109] which demonstrated that complementary features significantly improve performance in speech recognition. The complementary features used in [109] [110] [111]include perceptual linear prediction, amplitude modulation spectrogram (AMS), relative spectral transform and perceptual linear product (RASTA-PLP), Gammatone frequency cepstral coefficient, MFCC, pitch-based features. The complementary features are combined by concatenation. Research in [111] reports that the use of complementary features registered better results as compared to those of DFT magnitude. The challenge with using complementary features is how to effectively combine the different features, such that those complementing each other are retained while redundant ones are eliminated [110]. #### 3.1.1 Supervised speech enhancement and separation training with Fourier spectrum features DNN models that are trained via supervised learning using Fourier spectrum features employ several strategies to learn how to generate estimated clean signal from a noisy (mixed) signal. These strategies can be classified into three categories based on the target of the model. 1. 1. Spectral mapping techniques. 2. 2. Spectral masking techniques. 3. 3. Generative modelling. Spectral mapping techniques These models fit a nonlinear function to learn a mapping from a mixed signal feature to an estimated clean signal feature (see figure 7). Figure 7: Supervised training of speech enhancement model using spectrogram as input and spectrogram as output. The training dataset of these models consist of a noisy speech signal (source) and clean speech (target) features. The process of training these models can be generalised in the following steps: 1. 1. Given $N$ raw waveforms of mixed (noisy) speech, convert the $N$ raw waveform of noisy speech to the desired representation (such as spectrogram). 2. 2. Convert the respective $N$ clean speech waveform in time domain to the same representation as that of the noisy speech. 3. 3. Create an annotated dataset consisting of a pair of noisy speech features and that of clean speech i.e., $<noisy\\_speech\\_features_{i},clean\\_speech\\_features_{i}>$ with $1\leq i\leq N$ 4. 4. Train a deep learning model $g_{\theta}$ to learn how to estimate clean features $clean\\_speech\\_features_{i}$ given a noisy speech feature as input $noisy\\_speech\\_features_{i}$ by minimizing an objective function. 5. 5. Given new a noisy speech features $x_{j}$ the trained model $g_{\theta}$ should estimate a clean speech feature $y_{j}$. 6. 6. Using the estimated clean speech features $y_{j}$, reconstruct its raw waveform by performing the inverse of the feature generation process (such as using the inverse short-time Fourier transform if the features are in time- frequency domain). The above generalisation has been exploited in [83] [92] [95] [106] [112] [113] [84] [85] to achieve speech enhancement and in [94]and [103] to perform speech separation and enhancement. Figure 8 gives a summary of the steps when time-frequency(spectrogram) is exploited as the input of the speech enhancement model. Figure 8: Showing steps involves to train speech enhancement model to fit a regression function from noisy spectrogram to an estimated clean speech spectrogram Spectral masking techniques Here, the task of estimating clean speech features from a noisy (mixed) speech input features is formulated as that of predicting real-valued or complex- valued masks [55]. The mask function is usually constrained to be in range the [0,1] even though different types of soft masks have been proposed (see [21] [114] [115]). Source separation based on masks is predicated on the assumption of sparsity and orthogonality of the sources in the domain in which the masks are computed [60]. Due to the sparsity assumption, the dominant signal at a given range (such as time-frequency bin) is taken to be the only signal at that range (i.e. all other signals are ignored except the dominant signal). In that case, the role of DNN estimated mask is to estimate the dominant source at a given range. To do this, the mask is applied on the input features such that it eliminates portion of the signal( where the mask has a value of 0) while allowing others (mask value of 1)[116] [117]. The masks are always established by computing the signal-to-noise (SNR) within each TF bin against a threshold or a local criterion [116]. It has been demonstrated experimentally that the use of masks significantly improves speech intelligibility when an original speech is composed of noise or a masker speech signal [118] [119]. For deep learning models working on the time- frequency domain, a model $g_{\theta}$ is designed such that given a noisy or mixed speech spectrogram $Y[t,n]$ at time frame $t$, it estimates the mask $m_{t}$ at that time frame. The established mask $m_{t}$ is then applied to the input spectrogram to estimate target or denoised spectrogram i.e., $\hat{S}_{t}=m_{t}\otimes Y[t,n]$ (see figure 9). Here, $\hat{S}_{t}$ is the spectrogram estimate of the clean speech at time frame $t$ and $\otimes$ denotes element wise multiplication. To train the model $g_{\theta}$, there are two key objective variants. The first type minimizes an objective function $D$ such as mean squared error (MSE) between the model estimated mask $\hat{m}_{m}$ and the target mask ($tm$). $\mathcal{L}=\sum_{u,t,f}D|tm_{u,t,f},\hat{m}_{u,t,f}|$ This approach however cannot effectively handle silences where $|Y[t,n]|=0$ and $|X[t,n]|=0,$ because the target masks $tm$ will be undefined at the silence bins. Note that target masks such ideal amplitude mask (IRM) that is defined as $IRM(t,f)=\frac{|X_{s}(t,f)|}{\sum_{i=1}^{s}|Y_{s}(t,f)|}$ involves division of $|X[t,n]|$ by $|Y[t,n]|$ hence silence regions will make the target mask undefined [21]. This cost function also focuses on minimizing the disparity between the masks instead of the features of estimated signal and the target clean signal [21]. The second type of cost function seeks to minimize the features of estimated signal $\hat{S}_{t}=m_{t}\otimes Y[t,n]$ and those of target clean signal $S$ directly as shown equation 25. $\mathcal{L}=\sum_{u,t,f}D|\hat{m}_{u,t}Y_{u,t,f},S_{u,t,f}|$ (25) The sum is over all the speech $u$ and time-frequency bin $(t,f)$. Here, $Y$ and $S$ represents noisy (mixed) and clean (target) speech respectively. So, for DNN tools using indirect estimation of clean signal features, instead of them estimating the clean features directly from the noisy features input, the models first estimate binary masks. The binary masks are then applied to the noisy features to separate the sources (see figure 5, here, the features are the TF spectrogram). This technique has been applied in [3] [20] [103] [84] [114] [120] [121] [25] [122] [123] [124] [125] [18] [123] [31]. Figure 9: DNN model for mask estimation from a noisy spectrogram. Generative modelling Given an observed sample $x$, the goal of a generative model is to model its true distribution $p(x)$. The established model can then be used to generate new samples that are similar to the observed samples $x$. In speech separation and enhancement, these models have been exploited almost exclusively to perform speech denoising. Several generative models have been employed in a supervised manner to generate clean speech from a noisy one. Generative adversarial network (GAN) [126] is an unsupervised model that constitutes two key parts: the generator $\mathcal{G}$ and the discriminator $\mathcal{D}$, where $\mathcal{G}$ generates samples which are then judged by the $\mathcal{D}$. The generator $\mathcal{G}$ generates the synthetic data by sampling from a simple prior $z\sim p(z)$ and the outputs a final sample $g_{\theta}(z)$ where $g_{\theta}$ is non-linear function more specifically a DNN. The discriminator $\mathcal{D}$ on the other hand must be able to catch synthetic data as fake and real data from $p(x)$ as real. The training objective is shown in equation 26. $\min_{\mathcal{G}}\max_{\mathcal{D}}=\mathbb{E}_{x\sim p(x)}[\log D(x)]-\mathbb{E}_{z\sim p(z)}[\log(1-D(G(z)))]$ (26) The objective function in equation 26 is maximised w.r.t to $D(.)$ and minimised w.r.t $G(.)$. In speech enhancement, GAN was first introduced by SEGAN( see section 3.2.3). SEGAN which works in time-domain, uses a conditioned version of the objective function of equation 26. In conditioned GAN, both the generator and the discriminator are given extra information. This allows GAN to perform classification and mapping. SEGAN uses the least- squares GAN loss as opposed to sigmoid cross-entropy loss used for training. CGAN [104] just like SEGAN uses conditioned GAN but works in T-F domain to generate a denoised speech. Since most automatic speech recognition (ASR) tools work in T-F domain, CGAN hypothesises that the generative model working in T-F domain will be more robust for ASR as compared to those working in raw waveform. Therefore, CGAN can be seen as version of SEGAN that accepts input in T-F domain. To address the problem of mismatch between the training objective used in CGAN and the evaluation metrics, MetricGAN [93] proposes to integrate evaluation metric in the discriminator. By doing this, instead of the generator giving a false (0) or true (1) discrete values, it will generate continuous values based on the evaluation metric. MetricGAN can therefore be trained to generate data according to the the selected metric score. Through this modification, MetricGAN produces more robust enhanced speech. Another common generative group of generative models is variational auto-encoder (VAE) technique [127]. Like GAN, VAE is mainly used for denoising i.e. where the mixture is modelled as: $x_{fn}=\sqrt{g_{n}}s_{fn}+b_{fn}$ (27) Here, $x_{fn}$ denotes the mixture at the frequency index $f$ and the time- frame index $n$, $g_{n}\in R_{+}$ is a frequency independent but frame dependent gain while $s_{fn}$ and $b_{fn}$ represent the clean speech and the noise respectively at the frequency index $f$ and the time-frame index $n$. We first give a brief overview of VAE before we discuss how it is adapted for speech enhancement. Mathematically, given an observable sample $s$, the goal of a generative VAE model is to model true data distribution $p(s)$. To do this, VAE assumes that the observed sample $s$ are generated by associated latent variable $z$ and their joint distribution is $p(s,z)$. The model therefore seeks to learn how to maximize the likelihood $p(s)$ over all observed data. $p(s)=\int p(s,z)dz$ Integrating out all the latent variables $z$ in the above equation is intractable. However, using Evidence lower bound (ELBO) which quantifies the log-likelihood of observed data $p(s)$ can be estimated. ELBO is given in equation 28 (refer to [81] to see derivation of relationship between $p(s)$ and ELBO). $\log p(s)\geq\mathbb{E}_{q_{\phi}(z|s)}[\log\frac{p(s,z)}{q_{\phi}(z\mid s)}]$ (28) Here, $q_{\phi}(z\mid s)$ is a flexible variational distribution with parameters $\phi$ that the model seeks to maximize. Equation 28 can be written as equation 29 using Bayes theorem. $\log p(s)\geq\mathbb{E}_{q_{\phi}(z|s)}[\log\frac{p_{\theta}(s\mid z)p(z)}{q_{\phi}(z\mid s)}]$ (29) Equation 29 can be expanded as: $\log p(s)\geq\mathbb{E}_{q_{\phi}(z\mid s)}[\log p_{\theta}(s\mid z)]+\mathbb{E}_{q_{\phi}(z\mid s)}[\frac{\log p(z)}{q_{\phi}(z\mid s)}]$ (30) Equation 30 can be expanded as: $\log p(s)\geq\mathbb{E}_{q_{\phi}(z\mid s)}[\log p_{\theta}(s\mid z)]+D_{KL}(q_{\phi}(z\mid s)\mid\mid p(z))$ (31) The second term on the right of equation 31 seeks to learn the prior $p(z)$ via $q_{\phi}(z\mid s)$ while the first term reconstructs data based on the learned latent variable $z$. $q_{\phi}(z\mid s)$ is always modelled by a DNN and referred to as encoder and the reconstruction term is another DNN referred to as decoder. Both the encoder and decoder are trained simultaneously. The encoder is normally chosen to model a multivariate Gaussian with diagonal covariance and the prior is often selected to be a standard multivariate Gaussian: $q_{\phi}(z\mid x)=\mathcal{N}(z;\mu_{\theta}(x),\delta^{2}_{\theta}(x)I)$ $p(z)=\mathcal{N}(z;0,I)$ To estimate clean speech based on variational-autoencoder pre-training, the tools execute several techniques that can be generalised into the following steps: 1. 1. Train a model such that it can maximise the likelihood $p_{\theta}(s\mid z)$. Here, $s$ denotes the clean speech dataset that is composed of F-dimensional samples i.e $s_{t}\in R^{F},1\leq t\leq T$. The variational autoencoder assumes a D-dimensional latent variable $z_{t}\in R^{D}$. The latent variable $z_{t}$ and the clean speech $s_{t}$ have the following distribution: $z_{t}\sim\mathcal{N}(0,I_{D})$ $s_{t}\sim p(s_{t}|z_{t})$ Here, $\mathcal{N}(\mu,\delta)$ denotes a Gaussian distribution with mean $\mu$ and variance $\delta$. Basically, a decoder $p_{\theta}(s_{t}\mid z_{t})$ is trained to generate clean speech $s_{t}$ when given the latent variable $z_{t}$, the decoder is parameterized by $\theta$. The decoder $p_{\theta}(s_{t}\mid z_{t})$ is learned by deep learning model during training. The encoder is trained to estimate the posterior $q_{\phi}(z_{t}|s_{t})$ using a DNN. The overall objective of the variational auto-encoder training is to maximise equation 32. $p(s)=\operatorname*{argmin}_{\theta,\phi}\sum_{i=1}^{L}\log p_{\theta}(s\mid z^{i})+D_{KL}(q_{\phi}(z\mid s)\mid\mid p(z))$ (32) The posterior estimator $q_{\phi}(z\mid s)$ is a Gaussian distribution with parameters $\mu_{d}$ and $\delta_{d}$. These parameters are to be established by the encoder deep neural network such that $\mu_{d}:R^{F}\rightarrow R$ and $\delta_{d}:R^{F}\rightarrow R_{+}$. 2. 2. Set up a noise model using unsupervised techniques such as NMF [128]. For example, in case of NMF the noise $b_{fn}$ in equation 27 can be modelled as $b_{bf};w_{b,f},h_{b,n}\sim\mathcal{N}(0,(W_{b},H_{b})_{f,n})$ (33) where $\mathcal{N}(0,\delta)$ is a Gaussian distribution with zero mean and variance of $\delta$. 3. 3. Set up a mixture model such that $p(x\mid z,\theta_{s},\theta_{u})$ is maximised. Here $x$ is the noisy speech signal, $\theta_{s}$ are parameters from the pre-trained model in step 1 i.e $\phi$ and $\theta$. $\theta_{u}={g_{n},(W_{b},H_{b})_{f,n}}$ represents the parameters to be optimised. The parameters are $\theta_{u}$ are optimised by appropriate Bayesian inference technique. 4. 4. Reconstruct the clean speech $\hat{s}$ such that $p(\hat{s}|\theta_{u},\theta_{s},x)$ is maximised based on the parameters $\theta_{u},\theta_{s}$ from step 1 and 3 respectively and the observed mixed speech $x$. Works that exploit different versions of variational auto-encoder technique include [77] [78] [1] [79] [129]. Another generative modelling technique that has been used in speech enhancement is the variational diffusion model (VDM)[130]. VDM is composed of two processes i.e., diffusion and reverse process. The diffusion process perturbs data to noise and the reverse process seeks to recover data from noise. The goal of diffusion therefore is to transform a given data distribution into a simple prior distribution mostly standard Gaussian while the reverse process recovers data by learning a decoder parameterised by DNN. Formally, representing true data samples and latent variables as $x_{t}$ where $t=0$ represents true data and $1\leq t\leq T$ represents a sequence of latent variables, the VDM posterior is represented as: $q(x_{1:T}\mid x_{0})=\prod_{t=1}^{T}q(x_{t}\mid x_{t-1)}$ (34) The VDM encoder $q(x_{t}\mid x_{t-1})$ unlike that of VAE, is not learned rather it is a predefined linear Gaussian model. The Gaussian encoder is parameterized with mean $u_{t}(x_{t})=\sqrt{\alpha_{t}}x_{t-1}$ and variance $\varepsilon_{t}=(1-\alpha_{t})I$. Therefore, the encoder $q(x_{t}\mid x_{t-1})$ can mathematically be represented as $q(x_{t}\mid x_{t-1})=\mathcal{N}(x_{t};\sqrt{\alpha_{t}}x_{t-1},(1-\alpha_{t})I).$ (35) $\alpha_{t}$ evolves over time such that the final distribution of the latent $p(x_{T})$ is a standard Gaussian. The reverse process seeks to train a decoder that starts from the standard Gaussian distribution $p(x_{T})$. Formally the reverse process can be represented as: $p(x_{0:T})=p(x_{T})\prod_{i=1}^{T}p_{\theta}(x_{t-1}\mid x_{t})$ (36) Here $p(x_{T})=\mathcal{N}(x_{T};0,I)$. The reverse process seeks to set up a decoder $p_{\theta}(x_{t-1}\mid x_{t})$ that optimizes the parameter $\theta$ such that: the conditionals $p_{\theta}(x_{t-1}\mid x_{t})$ are established. Once the VDM is optimized, a sample from the Gaussian noise $p(x_{T})$ can iteratively be denoised through transitions $p_{\theta}(x_{t-1}\mid x_{t})$ for T steps to generate a simulated $x_{0}$. Using reparameterization trick, $x_{t}$ in equation 34 can be rewritten as: $x_{t}=\sqrt{\alpha_{t}}x_{t-1}+\sqrt{1-\alpha_{t}}\epsilon$ (37) where $\epsilon\sim\mathcal{N}(\epsilon,O,I)$ Similarly, $x_{t-1}=\sqrt{\alpha_{t-1}}x_{t-2}+\sqrt{1-\alpha_{t-1}}\epsilon$ (38) Based on this and through iterative derivation of equation 34, it can be shown that: $x_{t}=\sqrt{\bar{\alpha_{t}}}x_{0}+\sqrt{1-\bar{\alpha_{t}}}\epsilon_{0}$ (39) In the reverse process in equation 36, the transition probability $p_{\theta}(x_{t-1}\mid x_{t})$ can be represented by two parameters $\mu_{\theta}$ and $\delta_{\theta}$ as $\mathcal{N}(x_{t-1};\mu_{\theta}(x_{t},t),\delta_{\theta}(x_{t},t)^{2}I)$ with $\theta$ being the learnable parameters. It has been shown in [131] that $\mu_{\theta}(x_{t},t)$ can be established as: $\mu_{\theta}(x_{t},t)=\frac{1}{\sqrt{\alpha_{t}}}x_{t}-\frac{1-\alpha_{t}}{\sqrt{1-\bar{\alpha_{t}}}\sqrt{\alpha_{t}}}\epsilon_{\theta}(x_{t},t)$ (40) Based on equation 40, to estimate $\mu_{\theta}(x_{t},t)$ the DNN $\epsilon_{\theta}(x_{t},t)$ needs to estimate the Gaussian noise $\epsilon$ in $x_{t}$ which was injected during the diffusion process. Like VAE, VDM uses ELBO objective for optimization. Please see [131] for a thorough discussion on VDM. In speech denoising, work in [81] uses conditional diffusion process to model the encoder $q(x_{t}\mid x_{t-1})$. In conditional encoder, instead of $q(x_{t}\mid x_{t-1})$, they define it as $q(x_{t}\mid x_{0},y)$ i.e., $q(x_{t}\mid x_{0},y)=\mathcal{N}(x_{t};(1-m_{t})\sqrt{\bar{\alpha}}x_{0}+m_{t}\sqrt{\bar{\alpha}}y,\delta_{t}I)$. Here $x_{0}$,$y$ represents the clean speech and noisy speech respectively. The encoder is modeled as a linear interpolation between clean speech $x_{0}$ and the noise speech $y$ with interpolation ratio $m_{t}$. The reverse process $p_{\theta}(x_{t-1}\mid x_{t})$ is also modified to $p_{\theta}(x_{t-1}\mid x_{t},y)=\mathcal{N}(x_{t-1};\mu_{\theta}(x_{t},y,t),\delta I)$. Here, $\mu_{\theta}(x_{t},y,t)$ is the mean of the conditional reverse process. similar to equation 40, $\mu_{\theta}(x_{t},y,t)$ is estimated as $\mu_{\theta}(x_{t},y,t)=c_{xt}x_{t}+c_{yt}y-c_{\epsilon t}\epsilon_{\theta}(x_{t},y,t)$ (41) where $\epsilon_{\theta}(x_{t},y,t)$ is a DNN model to estimate the combination of Gaussian and non-Gaussian noise. The coefficients $c_{xt}$, $c_{yt}$ and $c_{\epsilon t}$ are established via the ELBO optimization. Other generative modelling techniques for speech enhancement(denoising). #### 3.1.2 Highlights on Fourier spectrum features 1. 1. When performing a DFT on the input signal, an optimum window length must be selected. The choice of the window has a direct impact on the frequency resolution and the latency of the system. To achieve good performance, most systems use 32ms. This may limit the use of the DFT based models in environments which require short latency [132]. 2. 2. DFT is a generic method for signal transformation that may not be optimised for waveform transformation in speech separation. It is therefore important to know to what extent does it place an upper bound on the performance level of speech enhancement techniques. 3. 3. Accurate reconstruction of estimated clean speech from the estimated features is not easy and the erroneous reconstruction of clean speech places an upper bound on the accuracy of the reconstructed audio. 4. 4. Perhaps the biggest challenge when working in the frequency domain is how to handle the phase. Most DNN models only use the magnitude spectrum of the noisy signal to train the DNN then factor in the phase of the noisy signal during reconstruction. Recent works such as [89] have shown that this technique does not generate optimum results. 5. 5. While working in the frequency domain, experimental research has demonstrated that spectral masking generates better results in terms of enhanced speech quality as compared to the spectral mapping method [119]. #### 3.1.3 Handling of phase in frequency domain The assumption made by most DNN models that use Fourier spectrum features is that phase information is not crucial for human auditory. Therefore, they exploit only the magnitude or power of the input speech to train the DNN models to learn the magnitude spectrum of the clean signal and factor in the phase during the reconstruction of the signal( see figure 10)[113] [133] [105] [134] [135]. The use of the phase from the noisy signal to estimate the clean signal is based on works such as [136] that demonstrated that the optimal estimator of the clean signal is the phase of the noisy signal. Further, most speech separation models work on frames that are of size between 20-40 ms and believe that the short-time phase contain low information [137] [138] [139] [140] and therefore not crucial when estimating clean speech. However, recent research [89] have demonstrated through experiments that further improvements in quality of estimated clean speech can be attained by processing both the short-time phase and magnitude spectra. Further, the factoring in of the noisy input phase during reconstruction has been noted to be a problem since the phase errors in the input interact with the amplitude of the estimated clean signal hence causing the amplitude of the estimated clean signal to differ with the amplitude of the actual clean signal being estimated [115], [64]. Figure 10: Showing how DNN models exploit the phase of the noisy signal during reconstruction of the estimated signal Based on the realisation of the importance of phase, some studies have avoided factoring in the phase of the noisy signal but rather exploit a modified phase to estimate the clean signal. The existing techniques of modifying the phase by the DNN models working on Fourier spectrum features can be categorised into two: Phase learning: These models make the phase part of the objective function i.e., they learn the phase of the estimated clean signal during training. To integrate the phase in the learning process works such as [115] [21] use a phase sensitive objective by replacing equation 42 with 43. It essentially exploits a phase sensitive spectrum approximation objective by minimising the distance between the raw waveform of the estimated speech and that of the target clean speech. $\mathcal{L}=\sum_{u,t,f}D|\hat{m}_{u,t}Y_{u,t,f},S_{u,t,f}|$ (42) $\mathcal{L}=\sum_{u,t,f}D|\hat{m}_{u,t}Y_{u,t,f},S_{u,t,f}(\cos{\theta}_{y}-\cos{\theta}_{s})|$ (43) Here $D$, is a selected objective function such as MSE, $\theta_{y}$ and $\theta_{s}$ represent the phase of the noisy and clean (target) speech respectively. The sum is over all the speech $u$ and time-frequency bin $(t,f)$. Experiments conducted based on the objective function in equation 43 show superior results in terms of signal-to-distortion ratio (SDR)[115]. Work in [111] trains a DNN model to generate masks that are composed of both the real and imaginary part (see equation 44). The complex mask will then be applied to a complex representation of the noisy signal to generate the estimated clean signal. By learning a mask that has both the real and imaginary part, they integrate the phase as part of the learning. $L=\frac{1}{2N}\sum_{t}\sum_{f}[(O_{r}(t,f)-M_{r}(t,f))^{2}+(O_{i}(t,f)-M_{i}(t,f))^{2}]$ (44) $O_{r}$ is the real part of the mask estimated by the DNN model while $O_{i}$ is the imaginary part. $M_{r}$ is the real part of the target mask while $M_{i}$ is the imaginary part. $N$ is the number of frames and (t,f) is a given TF bin. The complex mask implementation has been exploited in [55] [115][141] where the targets are formulated in the complex coordinate system i.e. the magnitude and phase are composed as part of the learning process. Work in [42] proposes a model that learns the phase during training via input spectrogram inversion (MISI) algorithm [142]. Work in [143] proposes a generative adversarial network (GAN) [126] based technique of learning the phase during training. Other works that learn the phase during training include [63] and [144]. Techniques that include phase as part of the training face the difficulty of processing a phase spectrogram which is randomly distributed and highly unstructured [145]. To mitigate this problem and derive a highly structured phase-aware target masks, [145] employs instantaneous frequency (IF)[146] to extract structured patterns from phase spectrograms. Post-processing phase update: The models that use this technique, train the DNN models using only the magnitude spectrum. Once the model has been trained to estimate the magnitude spectrum of the clean signal, they iteratively update the phase of the noisy signal to be as close as possible to that of the target clean signal. The algorithm being exploited by the models performing post-processing phase update is based on the Griffin-Lim algorithm proposed in [147]. For example, in [64], they exploit the magnitude $X^{0}$ of the target clean signal to iteratively obtain an optimal phase $\phi$ from the phase of the noisy signal ( see algorithm 1). The obtained phase is then used in the reconstruction of the estimated clean signal together with the magnitude $\hat{X}$ estimated by the DNN. The technique is also used in [148]. Techniques that implement Griffin-Lim algorithm such as in algorithm 1 perform iterative phase reconstruction of each source independently and may not be effective for multiple source separation where the sources must sum up to the mixture [42]. Work in [42] proposes to jointly reconstruct the phase of all sources in a given mixture by exploiting their estimated magnitudes and the noisy phase using the multiple input spectrogram inversion (MISI) algorithm [142]. They ensure that the sum of the reconstructed time-domain signals after each iteration must sum to the mixture signal. Work [149] and [62] also uses post-processing to update the phase of the noisy signal. Algorithm 1 Iteratively updating the phase of a noisy signal Target clean magnitude $X^{0}$, noisy phase $\phi^{0}$, iteration N. $X\leftarrow X^{0},\phi\leftarrow\phi^{0},n\leftarrow 1$ while $n\leq N$ do $s^{n}\leftarrow iSTFT(X,\phi)$ $(X^{n},\phi^{n})\leftarrow STFT(s^{n})$ $X\leftarrow X^{0}$ $\phi\leftarrow\phi^{n}$ $n\leftarrow n+1$ end while $s\leftarrow s^{n}$ ### 3.2 Time-domain features Due to the challenges highlighted in section 3.1.2 of working in the time- frequency domain, different models such as [132] [150] [10] [151] [152] [11] [153] [154][155] [61] [156] [157] explore the idea of designing a deep learning model for speech separation that accepts speech signal in the time- domain. The fundamental concept for these models is to replace the DFT based input with a data-driven representation that is jointly learned during model training. The models therefore accept as their input the mixed raw waveform sound and then generates either the estimated clean sources or masks that are applied on the noisy waveform to generate clean sources. By working on the raw waveform, these models address two key limitations of DFT based models. First, the models are designed to fully learn the magnitude and phase information of the input signal during training [150]. Secondly, they avoid reconstruction challenges faced when working with DFT features. The time domain methods can broadly be classified into two categories [150]. #### 3.2.1 Adaptive front-end based method The models in this category can roughly be discussed as composed of three key modules i.e., the encoder, separation and decoder modules ( see figure 11). 1. 1. Encoder: The encoder can be regarded as an adaptive front-end which seeks to replace STFT with a differentiable transform that is jointly trained (learned) with the separation model. It accepts as its input a time-domain mixture signal then learns STFT-like representation [11] [155]. By working directly with the time-domain signal, these models avoid the decoupling of the magnitude and phase of the input signal [10]. Most systems employ 1-dimensional convolution as the encoder to learn the features of the input signal. The transform generated by the encoder is then passed to the separation module. Work in [158] demonstrates that learned bases from raw data produce better results for speech/non-speech separation. 2. 2. Separation module: This module is fed by the output of the encoder. It implements techniques to identify the different sources present in the input signal. 3. 3. Decoder: It accepts input from the separation module and sometimes from the encoder(for residual implementation ). It is mostly implemented as an inverse of the encoder in order to reconstruct the separated signals [132] [10] [11]. Figure 11: Showing adaptive front-end implementation. #### 3.2.2 Waveform Mapping The second category of systems implement end-to-end systems where they utilise deep learning models to fit a regression function that maps an input mixed signal to its constituent estimated clean signal without an explicit front-end encoder (see figure 12). The models are trained using a pair of mixed(noisy) and clean speech. The model is fed with features of mixed signal for it to estimate clean speech. The training involves minimising an objective function such as minimum mean square error(MMSE) between the features of the clean signal and the estimated clean signal generated by the model. This approach has been implemented in [159] [160] [161]. Figure 12: Direct approach training of DNN models using raw waveform. #### 3.2.3 Generative modelling SEGAN [162] is GAN based model for speech denoising that conditions both $G$ and $D$ of equation 26 on extra information $z$ representing latent representation of the input. To solve the problem of vanishing gradient associated with optimizing objective in equation 26, they replace the cross- entropy loss by a least square function in equation 45. $\min_{\mathcal{G}}=\mathbb{E}_{z\sim p(z),\bar{x}\sim p(\bar{x})}[(D(G(z,\bar{x}),\bar{x})-1)^{2})]+\lambda||G(z,\bar{x})-x)||_{1}]$ (45) Here, $\bar{x}$ is the noisy speech, $x$ is the clean speech, $z$ is the extra input latent representation and $||.||_{1}$ is the $l_{1}$ norm distance between the clean sample x and the generated sample $|G(z,\bar{x})$ to encourage the generator G to generate more realistic audio. Work in [163] improves SEGAN to handle a more generalised speech signal distortion case which involves distortions such as chunk removal, band reduction, clipping and whispered speech. Work [164] improves SEGAN by implementing multiple generators as opposed to one and demonstrates that by doing so the speech quality of the enhanced speech is better than when a single generator is used. Work in [165] proposes a variation of SEGAN that is more tailored towards speech synthesis and not ASR. They replace the original loss function used in SEGAN with Wasserstein distance with gradient penalty(WGAN) [166]. They also exploit gated linear unit as activation function which has been shown in [167] to be more robust in generating realistic speech. Other GAN based models for speech enhancement in the time domain include [168]. Other tools that implement supervised conditional GAN include [104] [169] [170] [93]. In [171], Bayesian network is exploited to generate estimated clean speech from a noisy one. ### 3.3 Challenges of working with time-domain features 1. 1. Time domain features lack direct frequency representation; this hinders the features from capturing speech phonetics that are present in the frequency domain. Due to this, artefacts are always introduced in the reconstructed speech in the time domain [172]. 2. 2. The time domain waveform has a large input space. Based on this, models working with raw waveforms are often deep and complex in order to effectively model the dependencies in the waveform. This is computationally expensive [72] [162] [11] [173]. . ## 4 Which feature produces superior quality of enhanced speech? We performed analysis of 500 papers that exploit DNN to perform speech enhancement(i.e., multi-talker speech separation or denoising or dereverberation). We selected papers published from 2018 to 2022. We were interested to answer the question, which features are more popular with these tools? The summary is presented in figure 13. Based on the analysis, time- domain features popularity has grown rapidly from 2018 to 2022. The use of DFT features has slightly dropped, however remains popular over the five years. The popularity of MFCC and LPS has diminished. The popularity of features that are computationally expensive such as time-domain and DFT features may be attributed to the improved computation power of computers and efficient sequence modelling techniques such as transformers and temporal convolutional networks (see section 5 for discussion). Features such as MFCC are becoming less popular due to their reduced resolution, which must be extrapolated during reconstruction hence placing an upper bound on the quality of enhanced speech. Figure 13: Feature popularity between 2018 and 2022 We also investigated whether DFT or time-domain features produced the highest quality enhanced speech. Several works have conducted experiments with the goal to answer this question. Notable works include[174] and [175]. For example, [174] investigates Conv-TasNet’s [10] performance under different input types in the encoder and decoder. Conv-TasNet uses a frame length of 4ms, stride of 2ms and overlap of 2ms. Sample results presented in [174] are presented in table 1 where evaluation parameters include scale-invariant signal-to-distortion (si_SDR), signal-to-distortion (SDR), word error rate (WER). Table 1: Comparison of different encoder and decoder combination using Lw = 4 ms and Ls = 2ms on the test set of the WSJ0-2mix database loss | Encoder | Decoder | si_SDR dB | SDB dB | WER % ---|---|---|---|---|--- $L^{SI\\_SDR}$ | Learned | Learned | 14.4 | 14.7 | 21.71 $L^{SI\\_SDR}$ | STFT | Learned | 13.9 | 14.3 | 21.92 $L^{SI\\_SDR}$ | Learned | iSTFT | 14.1 | 14.5 | 21.87 $L^{SI\\_SDR}$ | STFT | iSTFT | 12.4 | 12.8 | 24.69 The results in table 1 show that the Conv-TasNet model gives marginally better results in terms of $si\\_SDR$, $SDBdB$ and $WER$ when the input is in time domain where the signal representation is learned by the encoder and output is learned by the decoder. The results are significantly reduced in all the three parameters if STFT is used as the input and its inverse used in the decoder. For instance, Conv-TasNet model achieves a SDR of 14.7 when time-domain features are used. This drops to 12.8 when DFT features are used. This shows that working in the time domain may be better for this setting as compared to the frequency domain. Work in [175] also shows the same trend where working in time-domain provides better results as compared to frequency domain. However, for mixed speech with reverberation, the use of a time domain signal does not improve the same results as compared to the frequency domain and further investigation on behaviour of both time and frequency features in the presence of reverberation is needed[175]. ## 5 Long term dependencies modelling To effectively perform speech separation, the speech separation tools need to model both long and short sequences within the audio signal. To do this, existing tools have employed several techniques: ### 5.1 Use of RNN The initial speech separation models such as [85] [106] [176] relied on a feedforward DNN to estimate clean speech from a noisy one. However, feedforward DNN models are ill poised for speech data since they are unable to effectively model long dependencies across time that are present in the speech data. Due to this, researchers progressively introduced recurrent neural networks (RNN) which have a feedback structure such that the representations at given time step $t$ is a function of the data at time $t$, the hidden state and memory at time $t-1$. One such RNN that has been exploited in speech separation is long-short-term memory (LSTM) [177]. LSTM has memory blocks that are composed of a memory cell to remember the temporary state and several gates to control the information and gradient flow. LSTM structures can be used to model sequential prediction networks which can exploit long-term contextual information [177]. Works in [103] [120][178] exploit LSTM to perform speech separation while [115] uses bidirectional long short-term memory (BLSTM) networks to make use of contextual information from both sides in the sequence. Due to their inherently sequential nature, RNN models are unable to support parallelization of computation. This limits their use when working with large datasets with long sequences due to slow training [11]. Moreover, in speech separation, a typical frame(input features) is usually 25ms which corresponds to 400 samples at a 16kHz sampling rate, for LSTM to work directly on the raw waveform, it would require unrolling the LSTM for an unrealistic large number of time steps to cover an audio of modest length [179]. Other models that use different versions of RNN include [180]. Models such as [181] use the gated recurrent unit (GRU)[182] to perform speech denoising. ### 5.2 Use of temporal convolution network Conventional convolution neural networks(CNN)have been used to design speech separation models[94] [183]. However, CNNs are limited in their ability to model long-range dependencies due to limited receptive fields [184]. They are therefore mainly tailored to learn local features. They exploit local window which maintain translation equivariance to learn a shared position-based kernel[185]. For CNN to capture long range dependencies ( i.e., to enlarge the receptive field), there is a need to stack many layers. This increases computation cost due to the large number of parameters. These shortcomings of the CNN and RNN, have motivated the use of dilated temporal convolution network (TCN) in speech separation to encode long-range dependencies using hierarchical convolutional layers [186] [187] [188] [17] [189]. TCN is composed of two key distinguishing characteristics: the convolution in the model must be causal i.e., a given activation of a certain layer $l$ at time $t$ is only influenced by activations of the previous layer $l-1$ from time steps that are less that $t$, 2) the model takes the sequence of any length and maps it into an output sequence of the same length. To achieve the second characteristic, TCN models are implemented using a 1-dimensional convolutional network such that each hidden layer is the same length as the input layer. To ensure same length, a zero padding of length $filtersize-1$ is added to keep subsequent layers the same length as previous ones [190] (see figure 14).The first property is achieved through the use of causal convolutions i.e. where an output at time $t$ is convolved only with elements from time t and earlier in the previous layer. To increase the receptive fields, models implement dilated TCN. Dilated convolution is where the filter is applied to a region larger than its size [191]. This is achieved by skipping input with certain specified steps (see figure 10). More formally, for 1D sequence such as speech signal, the input $x\in R^{n}$ and the kernel $f:\\{0,\cdots,k-1\\}\rightarrow R$, the dilated convolution operation $F$ on an element $s$ of a given sequence is defined according to equation 20 [190]. $F(s)=(x\ast_{d}f)(s)=\sum_{i=0}^{k-1}f(i)x_{s-di}$ (46) where $x$ is the $1D$ input signal, $k$ is the kernel and $d$ is the dilation factor. The effect of this is to expand the receptive field without loss of resolution and drastically increase the number of parameters. Stacked dilated convolution expands the receptive field with only a few layers. The expanded receptive field allows the network to capture temporal dependence of various resolutions with the input sequences [152]. In effect, TCN introduces the idea of time-hierarchy where the upper layers of the network model longer input sequences on larger timescales while local information are modelled by lower layers and are mainly maintained in the network through residuals and skip connections [152]. TCN also uses causal convolution where a given output at layer $l$ in time step $t$ is computed only based on time steps up to time step $t-1$ in the previous layer. The dilated TCN is exploited by [10] to model sequences that exist within the input speech signal. They implement TCN such that each layer is composed of 1-D dilated convolution blocks. The layers have 1-D CNN blocks with increasing dilation factors. This is to uncover long range dependencies that exist in the audio input. The dilation factors increase exponentially over the layers in order to cover a large temporal context window to exploit the long-range dependencies that exist within a speech signal. $y(m,n)=\mathop{\sum_{i=1}^{M}\sum_{j=1}^{N}}x(m+r\times i,n+r\times j)w(i,j)$ (47) Here, $y(m,n)$ is the output of a given layer of dilated convolution, $x(m,n)$ is the input and $w(i,j)$ is the filter with the length and the width of $M$ and $N$ respectively. The parameter $r$ is the dilation rate. Note that if $r=1$, the dilated convolution becomes the normal convolution convolution. Figure 14: Dilated TCN with four layers. . ### 5.3 Use of transformers A transformer[192] is an attention-based deep learning technique that has been successful in modelling sequences and allows uncovering of dependencies that exist within an input without regard to the distance between any two values of the input. Transformers consist only of feed-forward layers which allows them to exploit the parallel processing capabilities of GPUs leading to fast training [192]. In speech separation, [11] introduces a speech separation system that fully relies on transformers to model the dependencies that exist in the mixed audio signal. This is used to extract a mask for each of the speakers in the audio mixture. The transformer is used to uncover both the short-term dependencies (within a frame) and long-term dependencies (between frames). Work in [184] also exploits transformers in the encoder to model the dependencies that exist in the mixed audio while [67] uses transformers to perform speech dereverberation. Despite their ability to model long-range dependencies and ability to work well with parallelization, the attention mechanism of transformers, has $O(N^{2})$ complexity that brings a major memory bottleneck [193]. For a sequence of length $N$, the transformer needs to compare $N^{2}$ elements which results in a computational bottleneck especially for long signals such as speech. Transformers also use many parameters aggravating the memory problem further. Several versions of transformers such as Longformer[194], LinFormer[195] and Reformer[196] have been proposed with a goal to reduce the computation complexity of the transformers. Work in [197] investigates the performance of the three versions of transformers in speech separation and concludes that they are suitable for speech separation applications since they achieve a highly favourable trade- off between performance and computational requirements. Work in [198] proposes a technique of parameter sharing to reduce the computation complexity of the transformer while [193] reduces complexity by avoiding frame overlap. In [199], a teacher-student speech separation model based on transformer is proposed. The student model which is much smaller than the teacher model is used to reduce computation complexity. Other transformer-based speech enhancement tools include [173] [200]. Another key limitation of a transformer is that while it can model long-range global context, they do not extract fine-grained local features patterns well. Based on this, transformer-based speech separation tools apply attention within a frame (chunk) to capture local features and between frames(chunks) to capture global features [11] [201]. ## 6 Model size reduction techniques To achieve high performance i.e., generate speech with high intelligibility, DNN models for speech enhancements are becoming large by exploiting millions of parameters [202]. High number of parameters increase the memory requirements, computation complexity and latency. To reduce these parameters significantly without compromising quality and make speech enhancement tools to work in resource constrained platform, several techniques are being exploited. The techniques include: Use of dilated convolution: To increase the receptive field of 1D CNN and subsequently increase the temporal window and model long range dependencies within a speech, speech separation such as [132] and [186] implement dilated CNN. Dilated convolution initially introduced by [167] involves a convolution where a kernel is applied to an area that is larger that it. This is achieved by skipping input values by a defined step. It is like implementing a sparse kernel ( i.e., dilating the kernel with zeros). When dilated convolution is applied in a stacked network, it enables the network to increase its receptive field with few layers hence minimizing parameters and reducing computation [188]( see figure 14). This ensures that the models can capture long range dependencies while keeping the number of parameters at minimum. The dilating factors are made to increase exponentially per layer( see figure 10). Parameter quantization: To reduce computation, inference complexity of DNN models and to scale down the number of parameters, models such as [203] [204] [205] [206] [207] use parameter quantization. In quantization, the objective is to reduce the precision of model parameters and activation values to a low precision with minimal effects on the generalization capability of the DNN model. To achieve this, a quantization operator $Q$ is defined that maps a floating value to a quantized one [208]. Use of depthwise separable convolution: This type of convolution, decouples the convolution process into two i.e. depthwise convolution where a single filter is applied to each input channel and pointwise convolution which is applied to the output of depthwise convolution to achieve a linear convolution of the depthwise layer. Depthwise separable convolution has been shown to reduce the number of parameters as compared to the convectional one [209] [210]. Speech enhancement tools that exploit depthwise separable convolution include [10] [26] [67]. Knowledge distillation: Knowledge distillation involves training a large teacher model which can easily extract the structure of data then the knowledge learned by the teacher is distilled down to a smaller model called the student. Here, the student is trained under the supervision of the teacher[211] [212] [213]. The student model must mimic the teacher and by doing so achieve superior or similar performance but at reduced computation cost due to reduced parameters. Knowledge distillation technique has been exploited to reduce latency in speech enhancement tool [199] [214]. Parameter pruning: In order to reduce the number of parameters and hence speed up computation, some speech enhancement tools use parameter pruning [206] [215]. Pruning involves converting a dense DNN model into a sparse one by significantly scaling down the number of parameters without compromising model’s output’s quality. In [216] they train a speech enhancement DNN model to obtain an initial parameter set $\Theta$, they then prune the parameters by dropping the weights whose absolute values are below a set pruning threshold. The sparse network is again re-trained to obtain final parameters. Work in [203] estimates the sparsity $S(k)$ of a given channel $F_{jk}$. If the sparsity $S(k)>\theta$ where $\theta$ is a predefined threshold, the weights within $F_{jk}$ is set to zero and the model is retrained. After several iterations, the channel $F_{jk}$ is dropped. Weight sharing: This involves identifying clusters of weights that have a common value. The clusters are normally identified using K-means algorithm. So instead of storing each weight value, only the indexes of the shared values are stored. Through this memory requirements of the model is reduced [217]. Speech enhancement tools that use weight sharing include [204], [218]. ## 7 Objective functions for speech enhancement and separation Most DNN monaural speech enhancement and separation models especially those working on features in the frequency domain exploit mean-square-error (MSE) as the training objective [21] [38] [136] [151]. The DNN models that have the mask as the target use the training objective to minimise the MSE between the estimated mask and the ideal mask target. For models that predict estimated features (such as T-F spectrogram) of the clean source speech, MSE is used to minimise the difference between target features and the estimated features by the model. Despite the dominance of MSE as an objective function in the speech enhancement tools, it has been criticised since it is not closely related to human auditory perception [93]. Its major weakness is that it treats estimation elements independently and equally. For instance, it treats each time-frequency unit separately rather than whole spectral [106]. This leads to muffled sound quality and compromises intelligibility [106]. MSE also treats every estimation element with equal importance which is not the case [219]. It also does not discriminate between the positive or negative differences between the clean and estimated spectra. A positive difference between the clean and estimated spectra represents attenuation distortion, while a negative spectral difference represents amplification distortion. MSE treats the effects of these two distortions on speech intelligibility as equivalent which is problematic [220] [221]. Moreover, the MSE is usually defined in the linear frequency scale while the human auditory perception is on the Mel- frequency scale. To avoid the problem of treating every estimation element with equal importance [222], [223] propose a weighted MSE. Due to the shortcomings of MSE, objective functions that are closely related to the human auditory perception have been introduced to train the DNN [219] [224] [225] [226] [120]. Some of the human auditory perception training objectives being used by speech enhancement tools are also used as metrics for perceptual evaluation. They include: 1. 1. Short-time objective intelligibility (STOI)[227]. 2. 2. Scale invariant signal-to-distortion ratio(SI-SDR)[228]. 3. 3. Perceptual metric for speech quality evaluation(PMSQE). Scale invariant signal-to-distortion ratio Work in [228] proposes an intelligibility measure such that given the target signal $s$ and the model estimated signal $\hat{s}$, they re-scale either $s$ or $\hat{s}$ such that the residual $(s-\beta\hat{s})$ after scaling $\hat{s}$ or $(\alpha s-\hat{s})$ after scaling the target $s$ is orthogonal to the target as: $(s-\beta\hat{s}).s=0$ or $(\hat{s}-\alpha s).s=0$ based on this,$\alpha$ can be computed as: $(\hat{s}-\alpha s).s=0$ $\hat{s}.s-\alpha s.s=0\\\ $ $\alpha=\frac{\hat{s}^{T}s}{s.s}$ based on scaling of the target $s$, the signal-to-noise ratio(SNR) equation $SDR=10\log_{10}\frac{||s||^{2}}{||s-\hat{s}||^{2}}$ (48) is transformed to: $SDR=10\log_{10}\frac{||\alpha s||^{2}}{||s-\hat{\alpha s}||^{2}}$ (49) replacing the $\alpha$ we get the SI-SDR: $SI- SDR=10\log_{10}\frac{||\frac{\hat{s}^{T}s}{||s^{2}||}s||^{2}}{||\frac{\hat{s}^{T}s}{||s^{2}||}s-\hat{s}||^{2}}$ (50) This objective function has been used in [11] [10] [175] [30][193] [229], [230] [26] [28] [202]. Short-time objective intelligibility: This objective has been used in [225] [226] [219] [93]. STOI [227][231] is a speech intelligibility measure that is achieved by executing the following steps: 1. 1. Given discrete time signals of clean speech signal $x(n)$ and enhanced speech $y(n)$, perform a DFT on both $x(n)$ and $y(n)$ i.e $X(n,k)=DFT(y(n))$ and $Y(n,k)=DFT(y(n))$. Here, $k$ refers to the index of the discrete frequency. 2. 2. Remove silences in both the clean signal and the enhanced signals. Silences are removed by first identifying the frame with maximum energy($max_{energy}$) in the clean signal. All frames with energy of 40 dB less than $max_{energy}$ are dropped. 3. 3. Reconstruct both the clean and enhanced speech signals. 4. 4. Perform a one-third band octave analysis on both clean and enhanced speech by grouping DFT bins i.e the complex-valued STFT coefficients, $X(n,k)$, are combined into J third-octave bands by computing the TF units. $X_{j}(m)=\sqrt{\sum_{k=k_{1}(j)}^{k_{2}(j)-1}|X(n,k)|^{2}}j=1,\cdots,J$ (51) Here, $k_{1}$ and $k_{2}$ represent the one-third octave band edges. The same octave analysis is performed on the enhanced speech. The one-third octave of the enhanced speech is defined in a similar manner. 5. 5. Define a short temporal envelope of both enhanced and clean speech as: $Y_{j,m}=[Y_{j}(m-N+1),Y_{j}(m-N+2),\cdots,X_{j}(m)]^{T}$ and $X_{j,m}=[X_{j}(m-N+1),X_{j}(m-N+2),\cdots,X_{j}(m)]^{T}$ respectively. STOI exploits correlation coefficients to compare the temporal envelopes of clean and enhanced speech for a short time region. Note that N=30. 6. 6. Normalise the short temporal envelopes of the enhanced speech. Let $y_{j,m}(n)$ denote the $n^{th}$ envelope of enhanced speech. The normalised enhanced speech $y^{\prime}_{j,m}(n)$ of $Y_{j,m}(n)$ is given by $y^{\prime}_{j,m}(n)=\frac{|X_{j,m}|}{||Y_{j,m}||}Y_{j,m}(n)$. $||.||$ is the $l_{2}$ norm. The intuition behind normalisation of the enhanced speech is to reduce global level differences between clean and enhanced speech. These global level differences should not have a strong effect on speech intelligibility. 7. 7. Clip the normalised enhanced speech as $\bar{y}_{j,m}(n)=\min(y^{\prime}_{j,m},(1+10^{\frac{\beta}{20}})x_{j,m}(n))$. Clipping is done to ensure the effects of severely degraded frames of the enhanced speech on the model is upper bounded. Here, $\beta=-15dB$ is the lower signal-to-distortion(SDR) bound. 8. 8. Compute intermediate intelligibility measure as $d_{j,m}=\frac{(x_{j,m}-\mu_{x_{j,m}})^{T}(y_{j,m}-\mu_{y_{j,m}})}{||x_{j,m}-\mu_{x_{j,m}}||y_{j,m}-\mu_{y_{j,m}}||}$ (52) Here,$\mu_{(.)}$ refers to the sample mean of the corresponding vector. 9. 9. Compute the average intermediate intelligibility of all frames as $d=\frac{1}{JM}\sum_{j,m}d_{j,m}$ (53) where $M$ represents the total number of frames and $J$ the number of one- third octave band. Short-time spectral amplitude mean square error. Let $X[n,k]$ with $1\leq n\leq N$ and $1\leq k\leq K$ be an $N$ point DFT of $x$ and $K$ is the number of frames.Let $A[n,k]=X[n,k]$ with $k,\cdots,\frac{N}{2}+1$ and $k=1,\cdots,K$ denote the single sided amplitude spectra of $X[n,k]$. Let $\hat{A}[n,k]$ be an estimate of $A[n,k]$, the short-time spectral amplitude mean square error(STSA-MSE) is given by $\mathcal{L}_{STSA- MSE}=\frac{1}{(N/2+1)K}\sum_{n=1}^{N=K/2+1}\sum_{k=1}^{k=K}(\hat{A}[n,k]-A[n,k])^{2}$ (54) Equation 36 represents the mean square error between single-sided amplitude spectra of the clean speech $x$ and the DNN estimated speech. Equation 36 is not sensitive to the phase spectrum of the two signals $\hat{x}$ Perceptual metric for speech quality evaluation(PMSQE). This is an objective function that is based on the adaptation of perceptual evaluation of speech quality (PESQ) algorithm [232]. Given the MSE loss in the log-power spectrum with mean and variance normalisation i.e. $MSE_{t}=\frac{1}{k}\sum_{n=1}^{K}(\frac{\log|x[n,k]|^{2}-\mu_{k}}{\delta_{k}}-\frac{\log|\hat{X}[n,k]|^{2}-\mu_{k}}{\delta_{k}})^{2}=\frac{1}{K}\sum_{k=1}^{k=K}\frac{1}{\delta_{k}^{2}}(\log\frac{|X[n,k]|^{2}}{|\hat{X}[n,k]|^{2}})^{2}$ (55) Here, $X[n,k]|^{2}$ and $\hat{X}[n,k]|^{2}$ represent the power spectra of the clean and enhanced speech respectively. $\mu_{k}$ is the mean log-power spectrum and $\delta_{k}$ is its standard deviation. The indices n and k represent the frame and frequency, while $K$ is the number of frequency bins. From equation 55, the MSE is entirely dependent on power spectra across frequency bands hence not factoring in the perceptual factors such as loudness difference, masking and threshold effects[233]. To factor in the perceptual factors in the MSE, PMSQE modifies the MSE loss by incorporating two disturbance terms (symmetrical disturbance and asymmetrical disturbance) which are based on the PESQ algorithm both computed in a frame-by-frame basis[233]. $MSE_{t}=\sum_{t}{}MSE_{t}+\alpha D_{t}^{s}+\beta D_{t}^{a}$ (56) Here $D_{t}^{s}$ and $D_{t}^{a}$ represent symmetrical and asymmetrical disturbances respectively. The parameters $\alpha$ and $\beta$ are weighting factors which are determined experimentally. Work in [233] describes how to arrive at the values of $D_{t}^{s}$ and $D_{t}^{a}$. Since PESQ is non- differentiable, the PMSQE objective function provides a way of estimating it. PMSQE objective function is designed to be inversely proportional to PESQ, such that a low PMSQE value corresponds to a high PESQ value and vice versa. The key question here is: Which objective function is superior? work in [234] tries to answer this question where they evaluate six objective functions. Their conclusion is that the evaluation metric should be a major factor in deciding on the objective function to use in the speech enhancement model. In case a given model targets to improve a specific evaluation metric, then selection of an objective function related to that metric is advantageous. ## 8 Unsupervised techniques for speech enhancement Although supervised techniques of speech enhancement and separation have achieved great success towards improving speech intelligibility, the inherent problems associated with supervised learning still prohibits their applications in all scenarios. First, collecting parallel data of clean and noisy (mixed) data remains costly and time consuming. This limits the amount of data that can be used to train these models. Consequently, the models are not exposed to enough variations of the recording during training hence affecting their generalizability to noise types and acoustic conditions that were not seen during training [235] [236]. Collecting clean audio is always difficult and requires a well-controlled studio exacerbating the already high cost of data annotation [236]. Unsupervised learning offers an alternative to solving these problems. The existing unsupervised techniques for speech enhancement and separation can roughly be categorised into three: MixIT based techniques, generative modelling technique and teacher-student based techniques. Few novel techniques have also been proposed that fall outside these three dominant categories. Work in [237] proposes mixture invariant training (MixIT) to perform unsupervised speech separation. Given a set of $X$ that is composed of mixed speech i.e. $X=\\{x_{1},x_{2},\cdots,x_{n}\\}$ where each mixture $x_{i}$ is composed of up to $N$ sources, mixtures are drawn at random from the set $X$ without replacement and a mixture of mixture (MoM) created by adding the drawn mixtures, for example if two mixtures $x_{1}$ and $x_{2}$ are drawn from the set $X$, $MoM$ $\bar{x}$ is created by adding $x_{1}$ and $x_{2}$ i.e $\bar{x}=x_{1}+x_{2}$. The MoM $\bar{x}$ is the input to a DNN model which is trained to estimate sources $\hat{s}$ composed in $x_{1}$ and $x_{2}$. The DNN model is trained to minimize the loss function in equation 57. $L_{MixIT}=\min_{A}\sum_{i=1}^{2}L(x_{i},[A\hat{s}]_{i})$ (57) For a case where MoM is composed of only two mixtures, $A\in B^{2\times M}$ is a set of binary matrices where each column sums to 1. The loss function is trained to minimize the loss between mixtures $x_{i}$ and the remixed separated sources $A\hat{s}$. MixIT has been criticised for over-separation where it outputs estimated sources greater than the actual number of underlying sources in the mixtures $x_{i}$ [238]. Further, MixIT does not work well for speech enhancement ( i.e., denoising) [239]. MixIT teacher-student unsupervised model has been proposed in [238] to tackle the problem of over- separation in MixIT. It trains a student model such that its output matches the number of sources in the mixed speech $x$. Another MixIT based technique for solving over-separation problem is discussed in MixCycle [240]. Work in [241] proposes to improve MixIT to make it more tailored for denoising by exploiting an ASR pre-trained model to modify MixIT’s loss function. Work in [239] also seeks to improve MixIT for denoising by improving loss function and noise augmentation scheme. RemixIT [242] is an unsupervised speech denoising tool that exploits teacher-student DNN model. Given a batch of noisy speeches of size $b$, the teacher estimates the clean speech sources $\hat{s}_{i}$ and noises $\hat{n}_{i}$ where $1\leq i\leq b$. The teacher estimated noises $\hat{n}_{i}$ are mixed at random to generate $n^{p}$. The mixed noise $n^{p}$ together with the teacher estimated sources are used to generate new mixtures $\hat{m}_{i}=\hat{s}_{i}+n^{p}$. The new mixtures $\hat{m}_{i}$ are used as input to the student. The student is optimised to generate $\hat{s}_{i}$ and noise $n^{p}$ i.e., $\hat{s}_{i}$ and $n^{p}$ are the targets. Through this the teacher-student model is trained to denoise the speech. In RemixIT, a pre- trained speech enhancement model is used as the teacher model. Motivated by RemixIT, [243] also proposes a speech denoising unsupervised tool using teacher-student DNN model. They propose various techniques of student training. MetricGAN-U [76] is an unsupervised GAN based speech enhancement tool that trains a conditioned GAN discriminator without a reference clean speech, MetricGAN-U employs objective in equation 58 to train the speech enhancement model. $L=\mathbb{E}_{x}[(D(G(x))-Q\prime(G(x)))^{2}+(D(x)-Q\prime(G(x))^{2}]$ (58) In equation 58, $Q\prime$ is a non-intrusive metric (i.e does not require reference clean speech) that is used to score the enhanced speech from the generator. The scores obtained by $Q\prime$ are used to optimize the model. In MetricGAN-U, DNSMOS [244] is used as $Q\prime$. Another GAN based technique for unsupervised learning is [245] which exploits CycleGAN [246] multi- objective learning to perform parallel-data-free speech enhancement. Tools in [235] and [247] propose unsupervised speech denoising technique based on variations of VAE. Work in [236] proposes a speech denoising technique that uses only the noisy speech. It exploits the idea that was first proposed in [248] where they demonstrated that it is possible to recover signals under corruptions without observing clean signals. Predicating their work on these findings, given a noisy speech signal $x$, and noise $n$, [236] creates a more noisy speech $y=x+n$. They then train a DNN model to predict an enhanced speech $\hat{s}$ by having the more noisy input $y$ as the input and noisy speech $x$ as the target. Consequently, the DNN is trained by minimizing the loss in equation 59. $L=\frac{1}{M}\sum_{i=1}^{M}D(\hat{s}_{m},x_{m})$ (59) Here, $D$ is the objective function and $M$ is the sample size. This technique works on the basis that DNN cannot predict random noise hence the noise component in the training data is mapped to their expected values. Therefore, by assuming the noise as zero mean random variable, the objective function eliminates the noise [236]. Work in [16] proposes unsupervised techniques to perform speech separation based on gender. They exploit i-vectors to model large discrepancy in vocal tract, fundamental frequency contour, timing, rhythm, dynamic range, etc between speakers of different genders. In this case DNN model can be viewed as gender separator. ## 9 Domain adaptive techniques for speech enhancement and Separation Training data used to train speech enhancement and separation tools mostly have acoustic features that are significantly different from the acoustic features of the speech signals where the tools are deployed. This mismatch between the training data and target data leads to degradation in the tool’s performance in their deployed environment [249].The target environment dataset’s acoustic features may vary from the training data in noise type, speaker and signal-noise-ratio [247]. One potential way of tackling this problem is to collect massive training data that covers different variation of deployment environment. However, this is mostly not possible due to prohibitive cost. Due to this, some tools are proposing DNN based techniques for domain adaptation. Domain adaptation seeks to exploit either labelled or unlabelled target domain data to transfer a given tool from training data domain to the target data domain. Basically, domain adaptation seeks to reduce the covariance shift between the source and target domains data. The domain adaptation techniques in literature for speech separation and enhancement tools can be categorised into two: Unsupervised domain adaptation techniques such as [250] [251] which use unlabelled target domain dataset to adapt a DNN model and supervised domain adaptation techniques such as [252] [247] [253] which exploit limited labelled target domain dataset to perform domain adaptation of a DNN model for speech enhancement or separation. To make speech enhancement tools portable to new a new language, [252] proposes to use transfer learning. Transfer learning entails tailoring trained DNN models to apply knowledge acquired during training to a new domain where there is some commonality in type of task. The tool fine-tunes the top layers of a trained DNN model for speech enhancement by using labelled data of a new language while freezing the lower layer which are composed of parameters acquired during training of the original language. Work in [253] also uses transfer learning to show that pre-trained SEGAN can achieve high performance in new languages with unseen speakers and noise with just short training time. To make it more adaptable to different types of noise, tool in [247] proposes to employ multiple encoders where each encoder is trained in a supervised manner to focus only on given acoustic feature. The features are categorized into two i.e., utterance-level features such as gender, age, ascent of the speaker, signal-to-noise ratio and noise type and the signal-level features such high and low frequency of the speech parts. Feature focused encoders are trained to learn how to extract a given feature representation such as gender representation composed in the speech. Through the feature focused encoders, the experimental results show that the tool can adapt more to unseen noise types as compared to using a single global encoder. To adapt the DNN speech enhancement model to unseen noise type, work in [250] utilizes domain adversarial training (DAT)[254] to train an encoder to extract noise-invariant features. To do this, it utilizes the labelled source data and unlabelled target data. Through the feedback from the discriminator which gives the probability distribution over multiple noise types, the encoder is trained to produce noise-invariant features, hence reducing the mismatch problem. Work in [251] also exploits unsupervised DAT for speaker mismatch resolution. Work in [247] exploits importance-weighting (IW) using the classifiers of the networks to classify the source domain samples from the outlier weights and hence reduces the shift between the source and target domain. ## 10 Use of pre-trained models in speech separation and enhancement Pre-trained models have become popular especially in Natural language processing(NLP) and Computer vision. In NLP, for example, large corpus of text can be used to learn universal language representations which are beneficial for downstream NLP tasks. Due to their success in domains such as NLP and computer vision, pre-trained models based on unsupervised learning have been introduced in audio data [255] [256] [257] [258][259][260]. Such pre-trained models are beneficial in several ways: 1. 1. Pre-trained models are trained in large speech dataset hence can learn universal speech representations which can be beneficial to speech separation by boosting the quality of enhanced speech generated by these models. 2. 2. Pre-trained models provide models with better initialization which can result in better generalization and speed up convergence during training of speech enhancement models. 3. 3. Pre-trained speech models can act as regularizers to help speech enhancements models to avoid over fitting. Work in [261] seeks to establish if pre-trained speech models will help generate more robust features for downstream speech denoising and separation task as compared to features established without pre-trained models. To do this they use 13 speech pre-trained models to generate features of a noisy speech which are then passed through a three-layer BLSTM network which generates speech denoising or separation mask. They compare the performance of these features with those of baseline STFT and mel filerbank (FBANK) features. Their experiments establish that the 13 pre-trained models used do not significantly improve feature representations as compared to those of baselines. Hence the quality of enhanced and separated speech generated by features of pre-trained models are only slightly better or worse in some cases as compared to those generated based on the baseline features. They attribute this to domain mismatch and information loss. Since most of the pre-trained models were trained with clean speech, they are not portable to a noisy speech domain. Pre-trained models are usually trained to model global features and long-term dependencies hence some local features of the noisy or mixed speech signal may be lost due to this during feature extraction. Using HuBERT Large model [262], they demonstrate that the last layer of the model does not produce the best feature representation for speech enhancement and separation. In fact, for speech separation, the higher layers features are of low quality as compared to lower layers. They show that the weighted-sum representations of the representations from the different layers of pre-trained models where lower layers are given more weight generate better speech the enhancement and separation results as compared to isolated layers representations. They hypothesise that this could be due loss of some local signal information necessary for speech reconstruction tasks in deeper layers. To address the problem of information loss in pre-trained model, [263] proposes two solutions, first they utilize cross-domain features as model inputs to compensate lost information and secondly, they fine-tune a pre-trained model by using a speech enhancement model such that the extracted features are more tailored towards speech enhancement. Research in [264] seeks to synthesise clean speech directly from a noisy one using pre-trained model and HiFiGAN [265] speech synthesis model. It exploits the pre-trained model to extract features of the noisy model. The features are then used as input of HiFiGAN which generates estimated clean speech from these features. Based on the results reported in [261] that demonstrated that the final layer of pre- trained model does not give optimized representation for speech enhancement, they exploit weighted average of the representations of all the layers of the pre-trained model to generate representations of the noisy speech. The novelty of this work is that they do not use a model dedicated for speech denoising rather show that given features of a noisy speech, speech synthesis model can perform denoising. In [266], a pre-trained model has been exploited to design the loss function. Given a pre-trained model $\Phi$ with $m$ layers, the tool uses weighted $L^{1}$ loss to compute the difference between the feature activations of clean and noisy speech generated by different layers of the pre-trained model according to equation 60. $L=\sum_{i=1}^{m}\lambda_{w}||\Phi_{m}(s)-\Phi_{m}(g_{\theta}(x))||$ (60) Here, $s$ and $x$ are the clean and noisy speech respectively, $g_{\theta}$ is the denoising DNN model and $\lambda_{m}$ are the weights of contribution of each pre-trained model layer features to the loss function. Work in [267] proposes a two-stage speech enhancement framework where they first pre-train a model using unpaired noisy and clean data and utilize the pre-train model to perform speech enhancement. Unlike the previous works that use general pre- trained models for audio, the pre-trained model in [267] is trained using speech enhancement dataset. They report state of the art results in speech enhancement and ability of the tool to generalize to unseen noise. ## 11 Future direction for research in speech enhancement Unsupervised techniques for speech separation: Majority of speech enhancements tools use supervised learning technique. For those that use unsupervised learning discussed in section 8, they almost entirely focus on speech denoising and not speaker separation and dereverberation. There is therefore a gap of extending the unsupervised DNN techniques to perform multi-speaker speech separation and dereverberation. Dimension mismatch problem: Most speech separation tools set a fixed number $C$ of speakers therefore cannot deal with an inference mixture with $K$ sources, where $C\neq K$. Currents tools deal with the dimension mismatch problem by either outputting silences if $C>K$ or performing speech separation through iteration. However, both techniques have been found to be inefficient (see discussion in section 2.1). Therefore, there need to explore on developing DNN techniques for speech separation which are dynamic to the number of speakers present in inference mixture and adapt appropriately. Focus on data: Most model compression techniques for speech enhancement speed up the model performance by reducing or optimizing the model parameters. They don’t focus on the data-side impact on the model performance. For instance, what is the ideal sequence length and chunk overlap when working in time- domain that can speed up the speech enhancement process without compromising the quality of enhancement. More focus needs to turn towards exploring the data modifications that can speech up the speech enhancement process. Dataset modification: In dereverberation, tools are beginning to explore the use of speech with early reverberation [75] as the target as opposed to using anechoic target. Experiments in [75] demonstrate that allowing early reverberation in the target speech improves the quality of enhanced speech. There is need to develop a standardized dataset where the target is composed of early reverberation to allow for standardized evaluation of the tools on this dataset. Pre-trained Model: The pre-trained models that have been utilized for speech enhancement or separation have been trained on clean dataset hence failing potability test when used to generate features of a noisy speech signal [261]. There is need for development of a pre-trained model tailored for speech separation and enhancement. ## 12 Conclusion This review gives a discussion on how DNN techniques are being exploited by monaural speech enhancement tools. The objective was to uncover the key trends and dominant techniques being used by DNN tools at each stage of speech enhancement process. The review therefore discusses the type of features being exploited by these tools, the modelling of speech contextual information, how the models are trained (both supervised and unsupervised), key objective functions, how pre-trained speech models are being utilized and dominant dataset for evaluating the performance of the speech enhancement tools. In each section we highlight the standout challenges and how tools are dealing with these challenges. Our target is to give an entry into the speech enhancement domain by getting a thorough overview of the concepts and the trends of the domain. The hope is that the review gives a snapshot of the current research on DNN application to speech enhancement. ## References * Bando et al. [2018] Y. Bando, M. Mimura, K. Itoyama, K. Yoshii, and T. Kawahara, “Statistical Speech Enhancement Based on Probabilistic Integration of Variational Autoencoder and Non-Negative Matrix Factorization,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2018-April, no. Mcmc, pp. 716–720, 2018. * Xiao et al. [2019] X. Xiao, Z. Chen, T. Yoshioka, H. Erdogan, C. Liu, D. Dimitriadis, J. Droppo, and Y. Gong, “Single-channel speech extraction using speaker inventory and attention network,” in _ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2019, pp. 86–90. * Wang and Wang [2013] Y. Wang and D. L. Wang, “Towards scaling up classification-based speech separation,” _IEEE Transactions on Audio, Speech and Language Processing_ , vol. 21, no. 7, pp. 1381–1390, 2013. * Wang et al. [2014] Y. Wang, A. Narayanan, and D. L. Wang, “On training targets for supervised speech separation,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 22, no. 12, pp. 1849–1858, 2014. * Schmidt and Olsson [2006] M. N. Schmidt and R. K. Olsson, “Single-channel speech separation using sparse non-negative matrix factorization,” _INTERSPEECH 2006 and 9th International Conference on Spoken Language Processing, INTERSPEECH 2006 - ICSLP_ , vol. 5, pp. 2614–2617, 2006. * Wang and Sha [2014] Z. Wang and F. Sha, “Discriminative Non-Negative Matrix Factorization For Single-Channel Speech Separation,” _2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP)_ , pp. 3777–3781, 2014. [Online]. Available: https://pdfs.semanticscholar.org/854a/454106bd42a8bca158426d8b12b48ba0cae8.pdf * Virtanen and Cemgil [2009] T. Virtanen and A. T. Cemgil, “Mixtures of gamma priors for non-negative matrix factorization based speech separation,” _Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)_ , vol. 5441, no. 3, pp. 646–653, 2009. * Virtanen [2006] T. Virtanen, “Speech recognition using factorial hidden Markov models for separation in the feature space,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 1, pp. 89–92, 2006. * Shao and Wang [2006] Y. Shao and D. Wang, “Model-based sequential organization in cochannel speech,” _IEEE Transactions on Audio, Speech and Language Processing_ , vol. 14, no. 1, pp. 289–298, 2006. * Luo and Mesgarani [2019] Y. Luo and N. Mesgarani, “Conv-TasNet: Surpassing Ideal Time-Frequency Magnitude Masking for Speech Separation,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 27, no. 8, pp. 1256–1266, 2019. * Subakan et al. [2021] C. Subakan, M. Ravanelli, S. Cornell, M. Bronzi, and J. Zhong, “Attention is all you need in speech separation,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2021-June, pp. 21–25, 2021. * Wang and Chen [2018] D. Wang and J. Chen, “Supervised speech separation based on deep learning: An overview,” _IEEE/ACM Transactions on Audio, Speech, and Language Processing_ , vol. 26, no. 10, pp. 1702–1726, 2018. * Gannot et al. [2017] S. Gannot, E. Vincent, S. Markovich-Golan, and A. Ozerov, “A Consolidated Perspective on Multimicrophone Speech Enhancement and Source Separation,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 25, no. 4, pp. 692–730, 2017. * Hershey et al. [2016a] J. R. Hershey, Z. Chen, J. Le Roux, and S. Watanabe, “Deep clustering: Discriminative embeddings for segmentation and separation,” in _2016 IEEE international conference on acoustics, speech and signal processing (ICASSP)_. IEEE, 2016, pp. 31–35. * Wang et al. [2017] Y. Wang, J. Du, L. R. Dai, and C. H. Lee, “A gender mixture detection approach to unsupervised single-channel speech separation based on deep neural networks,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 25, no. 7, pp. 1535–1546, 2017. * Wang et al. [2016] Y. Wang, J. Du, L.-R. Dai, and C.-H. Lee, “Unsupervised single-channel speech separation via deep neural network for different gender mixtures,” in _2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA)_. IEEE, 2016, pp. 1–4. * Zeghidour and Grangier [2021] N. Zeghidour and D. Grangier, “Wavesplit: End-to-End Speech Separation by Speaker Clustering,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 29, no. iv, pp. 2840–2849, 2021. * Huang et al. [2011] P.-S. Huang, M. Kim, M. Hasegawa-Johnson, and P. Smaragdis, “Deep Learning For Monaural Speech Separation,” _Acta Physica Polonica B_ , vol. 42, no. 1, pp. 33–44, 2011. * Weng et al. [2015] C. Weng, D. Yu, M. L. Seltzer, and J. Droppo, “Deep neural networks for single-channel multi-talker speech recognition,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 23, no. 10, pp. 1670–1679, 2015. * Isik et al. [2016] Y. Isik, J. Le Roux, Z. Chen, S. Watanabe, and J. R. Hershey, “Single-channel multi-speaker separation using deep clustering,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 08-12-Sept, pp. 545–549, 2016. * Kolbæk et al. [2017a] M. Kolbæk, D. Yu, Z.-H. Tan, and J. Jensen, “Multitalker speech separation with utterance-level permutation invariant training of deep recurrent neural networks,” _IEEE/ACM Transactions on Audio, Speech, and Language Processing_ , vol. 25, no. 10, pp. 1901–1913, 2017. * Yu et al. [2017] D. Yu, M. Kolbaek, Z. H. Tan, and J. Jensen, “Permutation invariant training of deep models for speaker-independent multi-talker speech separation,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , pp. 241–245, 2017. * Tachibana [2021] H. Tachibana, “Towards listening to 10 people simultaneously: An efficient permutation invariant training of audio source separation using Sinkhorn’s algorithm,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2021-June, pp. 491–495, 2021\. * Dovrat et al. [2021] S. Dovrat, E. Nachmani, and L. Wolf, “Many-speakers single channel speech separation with optimal permutation training,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 4, pp. 2408–2412, 2021. * Hershey et al. [2016b] J. R. Hershey, Z. Chen, J. L. Roux, and S. Watanabe, “Deep Clustering : Discriminative Embeddings For Segmentation And Separation,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , pp. 31–35, 2016. * Byun and Shin [2021] J. Byun and J. W. Shin, “Monaural speech separation using speaker embedding from preliminary separation,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 29, pp. 2753–2763, 2021. * Qin et al. [2020] S. Qin, T. Jiang, S. Wu, N. Wang, and X. Zhao, “Graph convolution-based deep clustering for speech separation,” _IEEE Access_ , vol. 8, pp. 82 571–82 580, 2020. * Lee et al. [2022] J. H. Lee, J. H. Chang, J. M. Yang, and H. G. Moon, “NAS-TasNet: Neural Architecture Search for Time-Domain Speech Separation,” _IEEE Access_ , vol. 10, pp. 56 031–56 043, 2022. * Jiang and Duan [2020] F. Jiang and Z. Duan, “Speaker attractor network: Generalizing speech separation to unseen numbers of sources,” _IEEE Signal Processing Letters_ , vol. 27, pp. 1859–1863, 2020. * Nachmani et al. [2020] E. Nachmani, Y. Adi, and L. Wolf, “Voice separation with an unknown number of multiple speakers,” _37th International Conference on Machine Learning, ICML 2020_ , vol. PartF16814, pp. 7121–7132, 2020. * Liu and Wang [2019] Y. Liu and D. Wang, “Divide and Conquer: A Deep CASA Approach to Talker-Independent Monaural Speaker Separation,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 27, no. 12, pp. 2092–2102, 2019. * Luo and Mesgarani [2020] Y. Luo and N. Mesgarani, “Separating varying numbers of sources with auxiliary autoencoding loss,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2020-October, pp. 2622–2626, 2020. * Shi et al. [2018] J. Shi, J. Xu, G. Liu, and B. Xu, “Listen, think and listen again: Capturing top-down auditory attention for speaker-independent speech separation,” _IJCAI International Joint Conference on Artificial Intelligence_ , vol. 2018-July, pp. 4353–4360, 2018. * Kinoshita et al. [2018] K. Kinoshita, L. Drude, M. Delcroix, and T. Nakatani, “Listening to each speaker one by one with recurrent selective hearing networks,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2018-April, pp. 5064–5068, 2018. * Takahashi et al. [2019] N. Takahashi, S. Parthasaarathy, N. Goswami, and Y. Mitsufuji, “Recursive speech separation for unknown number of speakers,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2019-September, pp. 1348–1352, 2019. * Neumann et al. [2019] T. V. Neumann, K. Kinoshita, M. Delcroix, S. Araki, T. Nakatani, and R. Haeb-Umbach, “All-neural Online Source Separation, Counting, and Diarization for Meeting Analysis,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2019-May, pp. 91–95, 2019. * von Neumann et al. [2020] T. von Neumann, C. Boeddeker, L. Drude, K. Kinoshita, M. Delcroix, T. Nakatani, and R. Haeb-Umbach, “Multi-talker ASR for an unknown number of sources: Joint training of source counting, separation and ASR,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2020-October, pp. 3097–3101, 2020. * Luo et al. [2018] Y. Luo, Z. Chen, and N. Mesgarani, “Speaker-Independent Speech Separation with Deep Attractor Network,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 26, no. 4, pp. 787–796, 2018. * Yul et al. [2017] D. Yul, M. Kalbcek, Z.-h. Tan, and J. Jensen, “SPEAKER-INDEPENDENT MULTI-TALKER SPEECH SEPARATION,” in _IEEE International Conference on Acoustics, Speech and Signal Processing_ , 2017, pp. 241–245. * Chang et al. [2020a] X. Chang, W. Zhang, Y. Qian, J. Le Roux, and S. Watanabe, “End-to-end multi-speaker speech recognition with transformer,” in _ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2020, pp. 6134–6138. * Chang et al. [2020b] X. Chang, W. Zhang, Y. Qian, J. L. Roux, and S. Watanabe, “End-To-End Multi-Speaker Speech Recognition with Transformer,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2020-May, pp. 6134–6138, 2020. * Wang et al. [2018a] Z. Q. Wang, J. L. Roux, and J. R. Hershey, “Alternative objective functions for deep clustering,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2018-April, pp. 686–690, 2018. * Veselỳ et al. [2016] K. Veselỳ, S. Watanabe, K. Žmolíková, M. Karafiát, L. Burget, and J. H. Černockỳ, “Sequence summarizing neural network for speaker adaptation,” in _2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2016, pp. 5315–5319. * Wang et al. [2018b] J. Wang, J. Chen, D. Su, L. Chen, M. Yu, Y. Qian, and D. Yu, “Deep extractor network for target speaker recovery from single channel speech mixtures,” _arXiv preprint arXiv:1807.08974_ , 2018. * Wang et al. [2018c] Q. Wang, H. Muckenhirn, K. Wilson, P. Sridhar, Z. Wu, J. Hershey, R. A. Saurous, R. J. Weiss, Y. Jia, and I. L. Moreno, “Voicefilter: Targeted voice separation by speaker-conditioned spectrogram masking,” _arXiv preprint arXiv:1810.04826_ , 2018. * Ji et al. [2020] X. Ji, M. Yu, C. Zhang, D. Su, T. Yu, X. Liu, and D. Yu, “Speaker-aware target speaker enhancement by jointly learning with speaker embedding extraction,” in _ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2020, pp. 7294–7298. * Zhang et al. [2021a] C. Zhang, M. Yu, C. Weng, and D. Yu, “Towards robust speaker verification with target speaker enhancement,” in _ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2021, pp. 6693–6697. * Delcroix et al. [2018] M. Delcroix, K. Zmolikova, K. Kinoshita, A. Ogawa, and T. Nakatani, “Single channel target speaker extraction and recognition with speaker beam,” in _2018 IEEE international conference on acoustics, speech and signal processing (ICASSP)_. IEEE, 2018, pp. 5554–5558. * Chen et al. [2023a] L. Chen, Z. Mo, J. Ren, C. Cui, and Q. Zhao, “An electroglottograph auxiliary neural network for target speaker extraction,” _Applied Sciences_ , vol. 13, no. 1, p. 469, 2023. * Miao et al. [2015] Y. Miao, H. Zhang, and F. Metze, “Speaker adaptive training of deep neural network acoustic models using i-vectors,” _IEEE/ACM Transactions on Audio, Speech, and Language Processing_ , vol. 23, no. 11, pp. 1938–1949, 2015\. * Senior and Lopez-Moreno [2014] A. Senior and I. Lopez-Moreno, “Improving dnn speaker independence with i-vector inputs,” in _2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2014, pp. 225–229. * Ochiai et al. [2014] T. Ochiai, S. Matsuda, X. Lu, C. Hori, and S. Katagiri, “Speaker adaptive training using deep neural networks,” in _2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2014, pp. 6349–6353. * Chen et al. [2017] Z. Chen, Y. Luo, and N. Mesgarani, “Deep attractor network for single-microphone speaker separation,” in _2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2017, pp. 246–250. * Patrick A. Naylor [2010] N. D. G. Patrick A. Naylor, _Speech Dereverberation_ , N. D. G. Patrick A. Naylor, Ed. London: Springer-Verlag London, 2010, vol. 53, no. 1. * Williamson and Wang [2017a] D. S. Williamson and D. Wang, “Time-frequency masking in the complex domain for speech dereverberation and denoising,” _IEEE/ACM transactions on audio, speech, and language processing_ , vol. 25, no. 7, pp. 1492–1501, 2017\. * Arweiler and Buchholz [2011] I. Arweiler and J. M. Buchholz, “The influence of spectral characteristics of early reflections on speech intelligibility,” _The Journal of the Acoustical Society of America_ , vol. 130, no. 2, pp. 996–1005, 2011. * Nábělek et al. [1989] A. K. Nábělek, T. R. Letowski, and F. M. Tucker, “Reverberant overlap- and self-masking in consonant identification,” _Journal of the Acoustical Society of America_ , vol. 86, no. 4, pp. 1259–1265, 1989. * Oppenheim [1999] A. V. Oppenheim, _Discrete-Time Signal Processing_ , 2nd ed. upper Saddle River, NJ,USA:: Prentice-Hall, 1999\. * Zhou et al. [2022] R. Zhou, W. Zhu, and X. Li, “Single-Channel Speech Dereverberation using Subband Network with A Reverberation Time Shortening Target,” _arXiv preprint arXiv:2210.11089_ , 2022. [Online]. Available: http://arxiv.org/abs/2204.08765 * Cord-Landwehr et al. [2021] T. Cord-Landwehr, C. Boeddeker, T. von Neumann, C. Zorila, R. Doddipatla, and R. Haeb-Umbach, “Monaural source separation: From anechoic to reverberant environments,” in _In 2022 International Workshop on Acoustic Signal Enhancement (IWAENC)_ , 2021, pp. 1–5. [Online]. Available: http://arxiv.org/abs/2111.07578 * Su et al. [2020] J. Su, Z. Jin, and A. Finkelstein, “HiFi-GAN: High-fidelity denoising and dereverberation based on speech deep features in adversarial networks,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2020-October, no. 3, pp. 4506–4510, 2020. * Choi et al. [2020] H.-S. Choi, H. Heo, J. H. Lee, and K. Lee, “Phase-aware Single-stage Speech Denoising and Dereverberation with U-Net,” _arXiv preprint arXiv:2006.00687_ , 2020. [Online]. Available: http://arxiv.org/abs/2006.00687 * Wang et al. [2019] Z. Q. Wang, K. Tan, and D. Wang, “Deep Learning Based Phase Reconstruction for Speaker Separation: A Trigonometric Perspective,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2019-May, pp. 71–75, 2019. * Han et al. [2015] K. Han, Y. Wang, D. L. Wang, W. S. Woods, I. Merks, and T. Zhang, “Learning Spectral Mapping for Speech Dereverberation and Denoising,” _IEEE Transactions on Audio, Speech and Language Processing_ , vol. 23, no. 6, pp. 982–992, 2015. * Jiang et al. [2014] Y. Jiang, D. L. Wang, R. S. Liu, and Z. M. Feng, “Binaural classification for reverberant speech segregation using deep neural networks,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 22, no. 12, pp. 2112–2121, 2014. * Gamper and Tashev [2018] H. Gamper and I. J. Tashev, “Blind reverberation time estimation using a convolutional neural network,” _16th International Workshop on Acoustic Signal Enhancement, IWAENC 2018 - Proceedings_ , pp. 136–140, 2018. * Zhao et al. [2020] Y. Zhao, D. Wang, B. Xu, and T. Zhang, “Monaural Speech Dereverberation Using Temporal Convolutional Networks with Self Attention,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 28, pp. 1598–1607, 2020. * Ueda et al. [2016] Y. Ueda, L. Wang, A. Kai, X. Xiao, E. S. Chng, and H. Li, “Single-channel Dereverberation for Distant-Talking Speech Recognition by Combining Denoising Autoencoder and Temporal Structure Normalization,” _Journal of Signal Processing Systems_ , vol. 82, no. 2, pp. 151–161, 2016. * Williamson and Wang [2017b] D. S. Williamson and D. Wang, “Speech Dereverberation And Denoising Using Complex Ratio Masks ,” _IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP) 2017_ , pp. 5590–5594, 2017. * Jin and Wang [2009] Z. Jin and D. Wang, “A supervised learning approach to monaural segregation of reverberant speech,” _IEEE Transactions on Audio, Speech and Language Processing_ , vol. 17, no. 4, pp. 625–638, 2009. * León and Tobar [2021] D. León and F. Tobar, “Late reverberation suppression using U-nets,” _arXiv preprint arXiv:2110.02144._ , no. 1, 2021. [Online]. Available: http://arxiv.org/abs/2110.02144 * Défossez et al. [2020] A. Défossez, G. Synnaeve, and Y. Adi, “Real time speech enhancement in the waveform domain,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2020-October, pp. 3291–3295, 2020. * Isik et al. [2020] U. Isik, R. Giri, N. Phansalkar, J. M. Valin, K. Helwani, and A. Krishnaswamy, “PoCoNet: Better speech enhancement with frequency-positional embeddings, semi-supervised conversational data, and biased loss,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2020-October, pp. 2487–2491, 2020. * Li et al. [2021a] A. Li, W. Liu, X. Luo, G. Yu, C. Zheng, and X. Li, “A simultaneous denoising and dereverberation framework with target decoupling,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2, pp. 796–800, 2021. * Valin et al. [2022] J.-M. Valin, R. Giri, S. Venkataramani, U. Isik, and A. Krishnaswamy, “To Dereverb Or Not to Dereverb? Perceptual Studies On Real-Time Dereverberation Targets,” _arXiv:2206.07917_ , 2022. [Online]. Available: http://arxiv.org/abs/2206.07917 * Fu et al. [2022] S.-W. Fu, C. Yu, K.-H. Hung, M. Ravanelli, and Y. Tsao, “Metricgan-u: Unsupervised speech enhancement/dereverberation based only on noisy/reverberated speech,” in _ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2022, pp. 7412–7416. * Leglaive et al. [2019] S. Leglaive, U. Simsekli, A. Liutkus, L. Girin, and R. Horaud, “Speech Enhancement with Variational Autoencoders and Alpha-stable Distributions,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2019-May, pp. 541–545, 2019. * Leglaive et al. [2018] S. Leglaive, L. Girin, and R. Horaud, “A variance modeling framework based on variational autoencoders for speech enhancement,” _IEEE International Workshop on Machine Learning for Signal Processing, MLSP_ , vol. 2018-September, 2018. * Leglaive et al. [2020a] S. Leglaive, X. Alameda-Pineda, L. Girin, and R. Horaud, “A Recurrent Variational Autoencoder for Speech Enhancement,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2020-May, pp. 371–375, 2020. * Kolbæk et al. [2017b] M. Kolbæk, Z. H. Tan, and J. Jensen, “Speech Intelligibility Potential of General and Specialized Deep Neural Network Based Speech Enhancement Systems,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 25, no. 1, pp. 149–163, 2017. * Lu et al. [2022] Y.-J. Lu, Z.-Q. Wang, S. Watanabe, A. Richard, C. Yu, and Y. Tsao, “Conditional diffusion probabilistic model for speech enhancement,” in _ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2022, pp. 7402–7406. * Lu et al. [2021] Y.-J. Lu, Y. Tsao, and S. Watanabe, “A study on speech enhancement based on diffusion probabilistic model,” in _2021 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC)_. IEEE, 2021, pp. 659–666. * Fu et al. [2018a] S. W. Fu, T. W. Wang, Y. Tsao, X. Lu, H. Kawai, D. Stoller, S. Ewert, S. Dixon, X. Lu, Y. Tsao, S. Matsuda, C. Hori, Y. Xu, J. Du, L. R. Dai, C. H. Lee, T. Gao, J. Du, L. R. Dai, C. H. Lee, S. W. Fu, Y. Tsao, X. Lu, F. Weninger, J. R. Hershey, J. Le Roux, B. Schuller, Y. Xu, J. Du, L. R. Dai, C. H. Lee, F. Lluís, J. Pons, and X. Serra, “Speech enhancement based on deep denoising autoencoder,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 08-12-Sept, no. 1, pp. 7–19, 2018. * Fu et al. [2016] S. W. Fu, Y. Tsao, and X. Lu, “SNR-aware convolutional neural network modeling for speech enhancement,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 08-12-Sept, pp. 3768–3772, 2016. * Gao et al. [2016] T. Gao, J. Du, L. R. Dai, and C. H. Lee, “SNR-based progressive learning of deep neural network for speech enhancement,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 08-12-Sept, pp. 3713–3717, 2016. * Portnoff [1980] M. R. Portnoff, “Time-Frequency Representation of . Digital Signals,” _IEEE Transactions on Acoustics, Speech and Signal Processing ASSP_ , vol. 28, no. 1, pp. 55–69, 1980. * Allen [1982] J. B. Allen, “Applications of the short time Fourier transform to speech processing and spectral analysis,” in _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 1982-May, 1982, pp. 1012–1015. * Allen and Rabiner [1977] J. B. Allen and L. R. Rabiner, “A Unified Approach to Short-Time Fourier Analysis and Synthesis,” _Proceedings of the IEEE_ , vol. 65, no. 11, pp. 1558–1564, 1977. * Paliwal et al. [2011] K. Paliwal, K. Wójcicki, and B. Shannon, “The importance of phase in speech enhancement,” _Speech Communication_ , vol. 53, no. 4, pp. 465–494, 2011. * Natsiou and O’Leary [2021] A. Natsiou and S. O’Leary, “Audio representations for deep learning in sound synthesis: A review,” _Proceedings of IEEE/ACS International Conference on Computer Systems and Applications, AICCSA_ , vol. 2021-Decem, 2021\. * Nossier et al. [2020a] S. A. Nossier, J. Wall, M. Moniri, C. Glackin, and N. Cannings, “A Comparative Study of Time and Frequency Domain Approaches to Deep Learning based Speech Enhancement,” in _Proceedings of the International Joint Conference on Neural Networks_ , 2020. * Grais and Plumbley [2018] E. M. Grais and M. D. Plumbley, “Single channel audio source separation using convolutional denoising autoencoders,” _2017 IEEE Global Conference on Signal and Information Processing, GlobalSIP 2017 - Proceedings_ , vol. 2018-Janua, pp. 1265–1269, 2018. * Fu et al. [2019] S. W. Fu, C. F. Liao, Y. Tsao, and S. D. Lin, “MetricGAN: Generative adversarial networks based black-box metric scores optimization for speech enhancement,” _36th International Conference on Machine Learning, ICML 2019_ , vol. 2019-June, pp. 3566–3576, 2019. * Jansson et al. [2017] A. Jansson, E. Humphrey, N. Montecchio, R. Bittner, A. Kumar, and T. Weyde, “Singing voice separation with deep U-Net convolutional networks,” _Proceedings of the 18th International Society for Music Information Retrieval Conference, ISMIR 2017_ , pp. 745–751, 2017. * Kim and Smaragdis [2015] M. Kim and P. Smaragdis, “Adaptive denoising autoencoders: A fine-tuning scheme to learn from test mixtures,” _Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)_ , vol. 9237, pp. 100–107, 2015. * Luo et al. [2017] Y. Luo, Z. Chen, J. R. Hershey, J. Le Roux, and N. Mesgarani, “Deep clustering and conventional networks for music separation: Stronger together,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , pp. 61–65, 2017. * Baby et al. [2014] D. Baby, T. Virtanen, T. Barker, and H. Van Hamme, “Coupled dictionary training for exemplar-based speech enhancement,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , pp. 2883–2887, 2014. * Fu et al. [2017] S. W. Fu, T. Y. Hu, Y. Tsao, and X. Lu, “Complex spectrogram enhancement by convolutional neural network with multi-metrics learning,” _IEEE International Workshop on Machine Learning for Signal Processing, MLSP_ , vol. 2017-September, pp. 1–6, 2017. * Kothapally and Hansen [2022a] V. Kothapally and J. H. Hansen, “Complex-valued time-frequency self-attention for speech dereverberation,” _arXiv preprint arXiv:2211.12632_ , 2022. * Kothapally and Hansen [2022b] ——, “Skipconvgan: Monaural speech dereverberation using generative adversarial networks via complex time-frequency masking,” _IEEE/ACM Transactions on Audio, Speech, and Language Processing_ , vol. 30, pp. 1600–1613, 2022. * Liu et al. [2022] H. Liu, X. Liu, Q. Kong, Q. Tian, Y. Zhao, D. Wang, C. Huang, and Y. Wang, “VoiceFixer: A Unified Framework for High-Fidelity Speech Restoration,” _arXiv preprint arXiv:2204.05841._ , no. September, pp. 4232–4236, 2022. * Du et al. [2020] Z. Du, X. Zhang, and J. Han, “A Joint Framework of Denoising Autoencoder and Generative Vocoder for Monaural Speech Enhancement,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 28, pp. 1493–1505, 2020. * Weninger et al. [2014] F. Weninger, J. R. Hershey, J. Le Roux, and B. Schuller, “Discriminatively trained recurrent neural networks for single-channel speech separation,” _2014 IEEE Global Conference on Signal and Information Processing, GlobalSIP 2014_ , pp. 577–581, 2014. * Donahue et al. [2018] C. Donahue, B. Li, and R. Prabhavalkar, “Exploring speech enhancement with generative adversarial networks for robust speech recognition,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2018-April, no. Figure 1, pp. 5024–5028, 2018\. * Du and Huo [2008] J. Du and Q. Huo, “A speech enhancement approach using piecewise linear approximation of an explicit model of environmental distortions,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , pp. 569–572, 2008. * Xu et al. [2015] Y. Xu, J. Du, L. R. Dai, and C. H. Lee, “A regression approach to speech enhancement based on deep neural networks,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 23, no. 1, pp. 7–19, 2015. * Du et al. [2014] J. Du, Y. Tu, Y. Xu, L. Dai, and C. H. Lee, “Speech separation of a target speaker based on deep neural networks,” _International Conference on Signal Processing Proceedings, ICSP_ , vol. 2015-January, no. October, pp. 473–477, 2014. * Garau and Renals [2008] G. Garau and S. Renals, “Combining spectral representations for large-vocabulary continuous speech recognition,” _IEEE Transactions on Audio, Speech and Language Processing_ , vol. 16, no. 3, pp. 508–518, 2008. * Zolnay et al. [2007] A. Zolnay, D. Kocharov, R. Schlüter, and H. Ney, “Using multiple acoustic feature sets for speech recognition,” _Speech Communication_ , vol. 49, no. 6, pp. 514–525, 2007. * Wang et al. [2013] Y. Wang, K. Han, and D. Wang, “Exploring monaural features for classification-based speech segregation,” _IEEE Transactions on Audio, Speech and Language Processing_ , vol. 21, no. 2, pp. 270–279, 2013. * Williamson et al. [2016] D. S. Williamson, Y. Wang, and D. L. Wang, “Complex ratio masking for monaural speech separation,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 24, no. 3, pp. 483–492, 2016. * Lu et al. [2013] X. Lu, Y. Tsao, S. Matsuda, and C. Hori, “Speech enhancement based on deep denoising autoencoder,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , no. August, pp. 436–440, 2013. * Xu et al. [2014a] Y. Xu, J. Du, L. R. Dai, and C. H. Lee, “An experimental study on speech enhancement based on deep neural networks,” _IEEE Signal Processing Letters_ , vol. 21, no. 1, pp. 65–68, 2014. * Narayanan and Wang [2013] A. Narayanan and D. Wang, “Ideal Ratio Mask Estimation Using Deep Neural Networks For Robust Speech Recognition,” _2013 IEEE International Conference on Acoustics, Speech and Signal Processing_ , pp. 7092–7096, 2013. * Erdogan et al. [2015] H. Erdogan, J. R. Hershey, S. Watanabe, and J. Le Roux, “Phase-sensitive and recognition-boosted speech separation using deep recurrent neural networks,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2015-Augus, pp. 708–712, 2015. * Kjems et al. [2009] U. Kjems, J. B. Boldt, M. S. Pedersen, T. Lunner, and D. Wang, “Role of mask pattern in intelligibility of ideal binary-masked noisy speech,” _The Journal of the Acoustical Society of America_ , vol. 126, no. 3, pp. 1415–1426, 2009. * Wang [2008] D. Wang, “Time – Frequency Masking for Speech Hearing Aid Design,” _Trends In Amplification_ , vol. 12, pp. 332–353, 2008. [Online]. Available: http://www.ncbi.nlm.nih.gov/pubmed/18974204 * Brungart et al. [2006] D. S. Brungart, P. S. Chang, B. D. Simpson, and D. Wang, “Isolating the energetic component of speech-on-speech masking with ideal time-frequency segregation,” pp. 4007–4018, 2006. * Nossier et al. [2020b] S. A. Nossier, J. Wall, M. Moniri, C. Glackin, and N. Cannings, “A Comparative Study of Time and Frequency Domain Approaches to Deep Learning based Speech Enhancement,” _Proceedings of the International Joint Conference on Neural Networks_ , 2020. * Chen et al. [2015] Z. Chen, S. Watanabe, H. Erdogan, and J. R. Hershey, “Speech enhancement and recognition using multi-task learning of long short-term memory recurrent neural networks,” in _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2015-Janua, 2015, pp. 3274–3278. * Huang et al. [2015] P. S. Huang, M. Kim, M. Hasegawa-Johnson, and P. Smaragdis, “Joint Optimization of Masks and Deep Recurrent Neural Networks for Monaural Source Separation,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 23, no. 12, pp. 2136–2147, 2015. * Grais et al. [2014] E. M. Grais, M. U. Sen, and H. Erdogan, “Deep neural networks for single channel source separation,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , pp. 3734–3738, 2014. * Zhang and Wang [2016] X. L. Zhang and D. Wang, “A deep ensemble learning method for monaural speech separation,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 24, no. 5, pp. 967–977, 2016. * Narayanan and Wang [2015] A. Narayanan and D. Wang, “Improving robustness of deep neural network acoustic models via speech separation and joint adaptive training,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 23, no. 1, pp. 92–101, 2015. * Weninger et al. [2015] F. Weninger, H. Erdogan, S. Watanabe, E. Vincent, J. Le Roux, J. R. Hershey, and B. Schuller, “Speech enhancement with LSTM recurrent neural networks and its application to noise-robust ASR,” _Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)_ , vol. 9237, pp. 91–99, 2015. * Goodfellow [2016] I. Goodfellow, “NIPS 2016 Tutorial: Generative Adversarial Networks,” _arXiv preprint arXiv_ , 2016. [Online]. Available: http://arxiv.org/abs/1701.00160 * Kingma and Welling [2014] D. P. Kingma and M. Welling, “Auto-encoding variational bayes,” _2nd International Conference on Learning Representations, ICLR 2014 - Conference Track Proceedings_ , no. Ml, pp. 1–14, 2014. * Hien et al. [2015] T. D. Hien, D. V. Tuan, P. V. At, and L. H. Son, “Novel algorithm for non-negative matrix factorization,” _New Mathematics and Natural Computation_ , vol. 11, no. 2, pp. 121–133, 2015. * Leglaive et al. [2020b] S. Leglaive, X. Alameda-Pineda, L. Girin, and R. Horaud, “A recurrent variational autoencoder for speech enhancement,” in _ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2020, pp. 371–375. * Sohl-Dickstein et al. [2015] J. Sohl-Dickstein, E. Weiss, N. Maheswaranathan, and S. Ganguli, “Deep unsupervised learning using nonequilibrium thermodynamics,” in _International Conference on Machine Learning_. PMLR, 2015, pp. 2256–2265. * Luo [2022] C. Luo, “Understanding diffusion models: A unified perspective,” _arXiv preprint arXiv:2208.11970_ , 2022. * Luo and Mesgarani [2018] Y. Luo and N. Mesgarani, “TaSNet: Time-Domain Audio Separation Network for Real-Time, Single-Channel Speech Separation,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2018-April, pp. 696–700, 2018. * Kumar and Florencio [2016] A. Kumar and D. Florencio, “Speech enhancement in multiple-noise conditions using deep neural networks,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 08-12-September-2016, pp. 3738–3742, 2016. * Tu et al. [2014] Y. Tu, J. Du, Y. Xu, L. Dai, and C. H. Lee, “Deep neural network based speech separation for robust speech recognition,” _International Conference on Signal Processing Proceedings, ICSP_ , vol. 2015-January, no. October, pp. 532–536, 2014. * Li et al. [2017] H. Li, X. Zhang, H. Zhang, and G. Gao, “Integrated Speech Enhancement Method Based on Weighted Prediction Error and DNN for Dereverberation and Denoising,” _arXiv preprint arXiv:1708.08251_ , no. 2, 2017. [Online]. Available: http://arxiv.org/abs/1708.08251 * Ephraim and Malah [1984] Y. Ephraim and D. Malah, “Speech Enhancement Using a Minimum Mean-Square Error Short-Time Spectral Amplitude Estimator,” _IEEE Transactions on Acoustics, Speech, and Signal Processing_ , vol. 32, no. 6, pp. 1109–1121, 1984\. * Lim and Oppenheim [1979] J. S. Lim and A. V. Oppenheim, “Enhancement and Bandwidth Compression of Noisy Speech,” _Proceedings of the IEEE_ , vol. 67, no. 12, pp. 1586–1604, 1979. * Oppenheim and Lim [1981] A. V. Oppenheim and J. S. Lim, “The Importance of Phase in Signals,” _Proceedings of the IEEE_ , vol. 69, no. 5, pp. 529–541, 1981. * Vary and Eurasip [1985] P. Vary and M. Eurasip, “Noise suppression by spectral magnitude estimation -mechanism and theoretical limits-,” _Signal Processing_ , vol. 8, no. 4, pp. 387–400, 1985. * Wang and Lim [1982] D. L. Wang and J. S. Lim, “The Unimportance of Phase in Speech Enhancement,” _IEEE Transactions on Acoustics, Speech, and Signal Processing_ , vol. 30, no. 4, pp. 679–681, 1982. * Lee et al. [2017] Y.-S. Lee, C.-Y. Wang, S.-F. Wang, J.-C. Wang, and C.-H. Wu, “Fully complex deep neural network for phase-incorporating monaural source separation,” in _2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2017, pp. 281–285. * Gunawan and Sen [2010] D. Gunawan and D. Sen, “Iterative phase estimation for the synthesis of separated sources from single-channel mixtures,” _IEEE Signal Processing Letters_ , vol. 17, no. 5, pp. 421–424, 2010. * Ai et al. [2021] Y. Ai, H. Li, X. Wang, J. Yamagishi, and Z. Ling, “Denoising-and-Dereverberation Hierarchical Neural Vocoder for Robust Waveform Generation,” _2021 IEEE Spoken Language Technology Workshop, SLT 2021 - Proceedings_ , pp. 477–484, 2021. * Le Roux et al. [2019] J. Le Roux, G. Wichern, S. Watanabe, A. Sarroff, and J. R. Hershey, “Phasebook and friends: Leveraging discrete representations for source separation,” _IEEE Journal on Selected Topics in Signal Processing_ , vol. 13, no. 2, pp. 370–382, 2019. * Zheng and Zhang [2019] N. Zheng and X. L. Zhang, “Phase-aware speech enhancement based on deep neural networks,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 27, no. 1, pp. 63–76, 2019. * Friedman [1985] D. H. Friedman, “Instantaneous-Frequency Distribution Vs. Time: An Interpretation Of The Phase Structure Of Speech.” in _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , 1985, pp. 1121–1124. * Griffin and Lim [1984] D. W. Griffin and J. S. Lim, “Signal Estimation from Modified Short-Time Fourier Transform,” _IEEE Transactions on Acoustics, Speech, and Signal Processing_ , vol. 32, no. 2, pp. 236–243, 1984. * Zhao et al. [2019] Y. Zhao, Z. Q. Wang, and D. Wang, “Two-stage deep learning for noisy-reverberant speech enhancement,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 27, no. 1, pp. 53–62, 2019. * Li et al. [2016] K. Li, B. Wu, and C. H. Lee, “An iterative phase recovery framework with phase mask for spectral mapping with an application to speech enhancement,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 08-12-Sept, pp. 3773–3777, 2016\. * Luo et al. [2020] Y. Luo, Z. Chen, and T. Yoshioka, “Dual-Path RNN: Efficient Long Sequence Modeling for Time-Domain Single-Channel Speech Separation,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2020-May, pp. 46–50, 2020. * Venkataramani et al. [2018] S. Venkataramani, J. Casebeer, and P. Smaragdis, “End-To-End Source Separation With Adaptive Front-Ends,” _2018 52nd Asilomar Conference on Signals, Systems, and Computers_ , no. 1, pp. 684–688, 2018. * Zhang et al. [2020a] L. Zhang, Z. Shi, J. Han, A. Shi, and D. Ma, “FurcaNeXt: End-to-end monaural speech separation with dynamic gated dilated temporal convolutional networks,” _Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)_ , vol. 11961 LNCS, pp. 653–665, 2020. * Tzinis et al. [2020a] E. Tzinis, S. Venkataramani, Z. Wang, C. Subakan, and P. Smaragdis, “Two-Step Sound Source Separation: Training on Learned Latent Targets,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2020-May, pp. 31–35, 2020. * Tzinis et al. [2020b] E. Tzinis, Z. Wang, and P. Smaragdis, “Sudo RM -RF: Efficient networks for universal audio source separation,” _IEEE International Workshop on Machine Learning for Signal Processing, MLSP_ , vol. 2020-Septe, 2020. * Kong et al. [2022] Z. Kong, W. Ping, A. Dantrey, and B. Catanzaro, “Speech Denoising in the Waveform Domain With Self-Attention,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2022-May, pp. 7867–7871, 2022. * Lam et al. [2021a] M. W. Lam, J. Wang, D. Su, and D. Yuy, “Sandglasset: A Light Multi-Granularity Self-Attentive Network for Time-Domain Speech Separation,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2021-June, pp. 5759–5763, 2021\. * Lam et al. [2021b] M. W. Lam, J. Wang, D. Su, and D. Yu, “Effective Low-Cost Time-Domain Audio Separation Using Globally Attentive Locally Recurrent Networks,” _2021 IEEE Spoken Language Technology Workshop, SLT 2021 - Proceedings_ , pp. 801–808, 2021. * Kavalerov et al. [2019] I. Kavalerov, S. Wisdom, H. Erdogan, B. Patton, K. Wilson, J. Le Roux, and J. R. Hershey, “Universal sound separation,” _IEEE Workshop on Applications of Signal Processing to Audio and Acoustics_ , vol. 2019-October, pp. 175–179, 2019. * Stoller et al. [2018] D. Stoller, S. Ewert, and S. Dixon, “Wave-u-net: A multi-scale neural network for end-to-end audio source separation,” _arXiv preprint arXiv:1806.03185_ , 2018. * Fu et al. [2018b] S. W. Fu, T. W. Wang, Y. Tsao, X. Lu, and H. Kawai, “End-to-End Waveform Utterance Enhancement for Direct Evaluation Metrics Optimization by Fully Convolutional Neural Networks,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 26, no. 9, pp. 1570–1584, 2018. * Lluís et al. [2019] F. Lluís, J. Pons, and X. Serra, “End-to-end music source separation: Is it possible in the waveform domain?” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2019-Septe, pp. 4619–4623, 2019. * Pascual et al. [2017] S. Pascual, A. Bonafonte, and J. Serra, “SEGAN: Speech enhancement generative adversarial network,” in _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2017-Augus, no. D, 2017, pp. 3642–3646. * Pascual et al. [2019] S. Pascual, J. Serrà, and A. Bonafonte, “Towards generalized speech enhancement with generative adversarial networks,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2019-September, pp. 1791–1795, 2019. * Phan et al. [2020] H. Phan, I. V. McLoughlin, L. Pham, O. Y. Chen, P. Koch, M. De Vos, and A. Mertins, “Improving GANs for Speech Enhancement,” _IEEE Signal Processing Letters_ , vol. 27, pp. 1700–1704, 2020. * Adiga et al. [2019] N. Adiga, Y. Pantazis, V. Tsiaras, and Y. Stylianou, “Speech enhancement for noise-robust speech synthesis using wasserstein gan.” in _INTERSPEECH_ , 2019, pp. 1821–1825. * Gulrajani et al. [2017] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, and A. C. Courville, “Improved training of wasserstein gans,” _Advances in neural information processing systems_ , vol. 30, 2017. * van den Oord et al. [2016] A. van den Oord, S. Dieleman, H. Zen, K. Simonyan, O. Vinyals, A. Graves, N. Kalchbrenner, A. Senior, and K. Kavukcuoglu, “Wavenet: A generative model for raw audio,” _arXiv preprint arXiv:1609.03499_ , pp. 1–15, 2016. [Online]. Available: http://arxiv.org/abs/1609.03499 * Xiao et al. [2021] F. Xiao, J. Guan, Q. Kong, and W. Wang, “Time-domain speech enhancement with generative adversarial learning,” _arXiv preprint arXiv:2103.16149_ , 2021\. * Li et al. [2018a] Z. X. Li, L. R. Dai, Y. Song, and I. McLoughlin, “A Conditional Generative Model for Speech Enhancement,” _Circuits, Systems, and Signal Processing_ , vol. 37, no. 11, pp. 5005–5022, 2018. [Online]. Available: https://doi.org/10.1007/s00034-018-0798-4 * Qin and Jiang [2018] S. Qin and T. Jiang, “Improved Wasserstein conditional generative adversarial network speech enhancement,” _Eurasip Journal on Wireless Communications and Networking_ , vol. 2018, no. 1, 2018. * Qian et al. [2017] K. Qian, Y. Zhang, S. Chang, X. Yang, D. Florêncio, and M. Hasegawa-Johnson, “Speech enhancement using bayesian wavenet.” in _Interspeech_ , 2017, pp. 2013–2017. * Cao et al. [2022] R. Cao, S. Abdulatif, and B. Yang, “CMGAN: Conformer-based Metric GAN for Speech Enhancement,” _arXiv preprint arXiv:2209.11112_ , pp. 936–940, 2022\. * Wang et al. [2021] K. Wang, B. He, and W. P. Zhu, “Tstnn: Two-Stage Transformer Based Neural Network for Speech Enhancement in the Time Domain,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2021-June, pp. 7098–7102, 2021. * Heitkaemper et al. [2020] J. Heitkaemper, D. Jakobeit, C. Boeddeker, L. Drude, and R. Haeb-Umbach, “Demystifying TasNet: A Dissecting Approach,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2020-May, pp. 6359–6363, 2020. * Bahmaninezhad et al. [2019] F. Bahmaninezhad, J. Wu, R. Gu, S. X. Zhang, Y. Xu, M. Yu, and D. Yu, “A comprehensive study of speech separation: Spectrogram vs waveform separation,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2019-September, pp. 4574–4578, 2019. * Xu et al. [2014b] Y. Xu, J. Du, L.-R. Dai, and C.-H. Lee, “Global variance equalization for improving deep neural network based speech enhancement,” in _2014 IEEE China Summit & International Conference on Signal and Information Processing (ChinaSIP)_. IEEE, 2014, pp. 71–75. * Hochreiter and Schmidhuber [1997] S. Hochreiter and J. Schmidhuber, “Long short-term memory,” _Neural computation_ , vol. 9, no. 8, pp. 1735–1780, 1997. * Han et al. [2019] C. Han, J. O’Sullivan, Y. Luo, J. Herrero, A. D. Mehta, and N. Mesgarani, “Speaker-independent auditory attention decoding without access to clean speech sources,” _Science Advances_ , vol. 5, no. 5, pp. 1–12, 2019. * Sainath et al. [2015] T. N. Sainath, R. J. Weiss, A. Senior, K. W. Wilson, and O. Vinyals, “Learning the speech front-end with raw waveform CLDNNs,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2015-Janua, pp. 1–5, 2015. * Parveen and Green [2004] S. Parveen and P. Green, “Speech enhancement with missing data techniques using recurrent neural networks,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 1, no. Figure 1, pp. 13–16, 2004. * Wichern and Lukin [2017] G. Wichern and A. Lukin, “Low-Latency approximation of bidirectional recurrent networks for speech denoising,” _IEEE Workshop on Applications of Signal Processing to Audio and Acoustics_ , vol. 2017-October, pp. 66–70, 2017. * Cho et al. [2014] K. Cho, B. Van Merriënboer, C. Gulcehre, D. Bahdanau, F. Bougares, H. Schwenk, and Y. Bengio, “Learning phrase representations using RNN encoder-decoder for statistical machine translation,” _EMNLP 2014 - 2014 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference_ , pp. 1724–1734, 2014. * Chandna et al. [2017] P. Chandna, M. Miron, J. Janer, and E. Gómez, “Monoaural audio source separation using deep convolutional neural networks,” _Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)_ , vol. 10169 LNCS, pp. 258–266, 2017. * Chen et al. [2020] J. Chen, Q. Mao, and D. Liu, “Dual-path transformer network: Direct context-aware modeling for end-to-end monaural speech separation,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2020-Octob, pp. 2642–2646, 2020\. * Gulati et al. [2020] A. Gulati, J. Qin, C. C. Chiu, N. Parmar, Y. Zhang, J. Yu, W. Han, S. Wang, Z. Zhang, Y. Wu, and R. Pang, “Conformer: Convolution-augmented transformer for speech recognition,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2020-October, pp. 5036–5040, 2020. * Rethage et al. [2018] D. Rethage, J. Pons, and X. Serra, “A wavenet for speech denoising,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2018-April, pp. 5069–5073, 2018. * Li et al. [2018b] Y. Li, X. Zhang, and D. Chen, “CSRNet: Dilated Convolutional Neural Networks for Understanding the Highly Congested Scenes,” _Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition_ , pp. 1091–1100, 2018. * Lea et al. [2017] C. Lea, M. D. Flynn, R. Vidal, A. Reiter, and G. D. Hager, “Temporal convolutional networks for action segmentation and detection,” in _proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 2017, pp. 156–165. * Zhang et al. [2020b] Q. Zhang, A. Nicolson, M. Wang, K. K. Paliwal, and C. Wang, “DeepMMSE: A Deep Learning Approach to MMSE-Based Noise Power Spectral Density Estimation,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 28, pp. 1404–1415, 2020. * Bai et al. [2018] S. Bai, J. Z. Kolter, and V. Koltun, “An empirical evaluation of generic convolutional and recurrent networks for sequence modeling,” _arXiv preprint arXiv:1803.01271_ , 2018. * He and Zhao [2019] Y. He and J. Zhao, “Temporal Convolutional Networks for Anomaly Detection in Time Series,” _Journal of Physics: Conference Series_ , vol. 1213, no. 4, 2019. * Vaswani et al. [2017] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” _Advances in Neural Information Processing Systems_ , vol. 2017-Decem, no. Nips, pp. 5999–6009, 2017. * Subakan et al. [2022a] C. Subakan, M. Ravanelli, S. Cornell, F. Lepoutre, and F. Grondin, “Resource-Efficient Separation Transformer,” _arXiv preprint arXiv:2206.09507_ , pp. 1–5, 2022. [Online]. Available: http://arxiv.org/abs/2206.09507 * Beltagy et al. [2020] I. Beltagy, M. E. Peters, and A. Cohan, “Longformer: The Long-Document Transformer,” _arXiv preprint arXiv:_ , 2020. [Online]. Available: http://arxiv.org/abs/2004.05150 * Wang et al. [2020] S. Wang, B. Z. Li, M. Khabsa, H. Fang, and H. Ma, “Linformer: Self-Attention with Linear Complexity,” _arXiv preprint arXiv_ , vol. 2048, no. 2019, 2020\. [Online]. Available: http://arxiv.org/abs/2006.04768 * Kitaev et al. [2020] N. Kitaev, Ł. Kaiser, and A. Levskaya, “Reformer: The Efficient Transformer,” in _International Conference on Learning Representations_ , 2020, pp. 1–12. [Online]. Available: http://arxiv.org/abs/2001.04451 * Subakan et al. [2022b] C. Subakan, M. Ravanelli, S. Cornell, F. Grondin, and M. Bronzi, “On Using Transformers for Speech-Separation,” in _International Workshop on Acoustic Signal Enhancement_ , vol. 14, no. 8, 2022, pp. 1–10. [Online]. Available: http://arxiv.org/abs/2202.02884 * Luo et al. [2022] J. Luo, J. Wang, N. Cheng, E. Xiao, X. Zhang, and J. Xiao, “Tiny-Sepformer: A Tiny Time-Domain Transformer Network For Speech Separation,” _arXiv preprint arXiv:2206.13689_ , no. 1, pp. 5313–5317, 2022. * Chen et al. [2021] S. Chen, Y. Wu, Z. Chen, J. Wu, T. Yoshioka, S. Liu, J. Li, and X. Yu, “Ultra fast speech separation model with teacher student learning,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 3, pp. 2298–2302, 2021. * de Oliveira et al. [2022] D. de Oliveira, T. Peer, and T. Gerkmann, “Efficient Transformer-based Speech Enhancement Using Long Frames and STFT Magnitudes,” _arXiv preprint arXiv:2206.11703._ , no. 1, pp. 2948–2952, 2022. * Qiu and Hu [2022] W. Qiu and Y. Hu, “Dual-Path Hybrid Attention Network for Monaural Speech Separation,” _IEEE Access_ , vol. 10, pp. 78 754–78 763, 2022. * Lutati et al. [2022] S. Lutati, E. Nachmani, and L. Wolf, “SepIt: Approaching a Single Channel Speech Separation Bound,” in _arXiv preprint arXiv:2205.11801_ , 2022, pp. 5323–5327. * Wu et al. [2019] J. Y. Wu, C. Yu, S. W. Fu, C. T. Liu, S. Y. Chien, and Y. Tsao, “Increasing Compactness of Deep Learning Based Speech Enhancement Models with Parameter Pruning and Quantization Techniques,” _IEEE Signal Processing Letters_ , vol. 26, no. 12, pp. 1887–1891, 2019. * Sun and Li [2017] H. Sun and S. Li, “An optimization method for speech enhancement based on deep neural network,” _IOP Conference Series: Earth and Environmental Science_ , vol. 69, no. 1, 2017. * Lin et al. [2019] Y. C. Lin, Y. T. Hsu, S. W. Fu, Y. Tsao, and T. W. Kuo, “IA-Net: Acceleration and Compression of Speech Enhancement using Integer-adder Deep Neural Network,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2019-September, pp. 1801–1805, 2019. * Fedorov et al. [2020] I. Fedorov, M. Stamenovic, C. Jensen, L. C. Yang, A. Mandell, Y. Gan, M. Mattina, and P. N. Whatmough, “TinyLSTMs: Efficient neural speech enhancement for hearing aids,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2020-October, pp. 4054–4058, 2020. * Hsu et al. [2019] Y. T. Hsu, Y. C. Lin, S. W. Fu, Y. Tsao, and T. W. Kuo, “A Study on Speech Enhancement Using Exponent-Only Floating Point Quantized Neural Network (EOFP-QNN),” _2018 IEEE Spoken Language Technology Workshop, SLT 2018 \- Proceedings_ , pp. 566–573, 2019. * Gholami et al. [2022] A. Gholami, S. Kim, Z. Dong, Z. Yao, M. W. Mahoney, and K. Keutzer, “A Survey of Quantization Methods for Efficient Neural Network Inference,” _Low-Power Computer Vision_ , pp. 291–326, 2022. * Avery et al. [2014] K. R. Avery, J. Pan, C. C. Engler-Pinto, Z. Wei, F. Yang, S. Lin, L. Luo, and D. Konson, “Fatigue Behavior of Stainless Steel Sheet Specimens at Extremely High Temperatures,” _SAE International Journal of Materials and Manufacturing_ , vol. 7, no. 3, pp. 560–566, 2014. * Howard et al. [2017] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, and H. Adam, “MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications,” _arXiv preprint arXiv:1704.04861._ , 2017. [Online]. Available: http://arxiv.org/abs/1704.04861 * Hinton et al. [2015] G. Hinton, O. Vinyals, and J. Dean, “Distilling the knowledge in a neural network,” _arXiv preprint arXiv:1503.02531 2.7_ , pp. 1–9, 2015. [Online]. Available: http://arxiv.org/abs/1503.02531 * Gou et al. [2021] J. Gou, B. Yu, S. J. Maybank, and D. Tao, “Knowledge distillation: A survey,” _International Journal of Computer Vision_ , vol. 129, pp. 1789–1819, 2021\. * Wang and Yoon [2022] L. Wang and K. J. Yoon, “Knowledge Distillation and Student-Teacher Learning for Visual Intelligence: A Review and New Outlooks,” _IEEE Transactions on Pattern Analysis and Machine Intelligence_ , vol. 44, no. 6, pp. 3048–3068, 2022. * Aihara et al. [2019] R. Aihara, T. Hanazawa, Y. Okato, G. Wichern, and J. L. Roux, “Teacher-student Deep Clustering for Low-delay Single Channel Speech Separation,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2019-May, pp. 690–694, 2019\. * Tan and Wang [2021] K. Tan and D. Wang, “Towards Model Compression for Deep Learning Based Speech Enhancement,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 29, pp. 1785–1794, 2021. * Ye et al. [2019] F. Ye, Y. Tsao, and F. Chen, “Subjective feedback-based neural network pruning for speech enhancement,” _2019 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA ASC 2019_ , no. November, pp. 673–677, 2019. * Dupuis et al. [2020] E. Dupuis, D. Novo, I. O’Connor, and A. Bosio, “Sensitivity Analysis and Compression Opportunities in DNNs Using Weight Sharing,” in _Proceedings - 2020 23rd International Symposium on Design and Diagnostics of Electronic Circuits and Systems, DDECS 2020_ , 2020. * Hu et al. [2021] X. Hu, K. Li, W. Zhang, Y. Luo, J. M. Lemercier, and T. Gerkmann, “Speech Separation Using an Asynchronous Fully Recurrent Convolutional Neural Network,” _Advances in Neural Information Processing Systems_ , vol. 27, no. NeurIPS, pp. 22 509–22 522, 2021. * Zhang et al. [2018] H. Zhang, X. Zhang, and G. Gao, “Training Supervised Speech Separation System To Improve Stoi And Pesq Directly,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , pp. 5374–5378, 2018. * Loizou and Kim [2011] P. C. Loizou and G. Kim, “Reasons why current speech-enhancement algorithms do not improve speech intelligibility and suggested solutions,” _IEEE Transactions on Audio, Speech and Language Processing_ , vol. 19, no. 1, pp. 47–56, 2011. * P. C. Loizou [2013] P. C. Loizou, _Speech enhancement: theory and practice_. CRC press, 2013. * Xia and Bao [2014] B. Xia and C. Bao, “Wiener filtering based speech enhancement with Weighted Denoising Auto-encoder and noise classification,” pp. 13–29, 2014. * Shivakumar and Georgiou [2016] P. G. Shivakumar and P. Georgiou, “Perception optimized deep denoising autoencoders for speech enhancement,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 08-12-September-2016, pp. 3743–3747, 2016. * Koizumi et al. [2017] Y. Koizumi, K. Niwa, Y. Hioka, K. Kobayashi, and Y. Haneda, “DNN-based source enhancement self-optimized by reinforcement learning using sound quality measurements,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , pp. 81–85, 2017. * Kolbcek et al. [2018] M. Kolbcek, Z. H. Tan, and J. Jensen, “Monaural Speech Enhancement Using Deep Neural Networks by Maximizing a Short-Time Objective Intelligibility Measure,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2018-April, pp. 5059–5063, 2018. * Yan et al. [2018] Z. Yan, X. Buye, G. Ritwik, and Z. Tao, “Perceptually Guided Speech Enhancement Using Deep Neural Networks,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , pp. 5074–5078, 2018. [Online]. Available: https://cliffzhao.github.io/Publications/ZXGZ.icassp18.pdf%0Ahttp://arxiv.org/abs/1312.6114%0Ahttps://arxiv.org/pdf/1312.6114.pdf * Taal et al. [2011] C. H. Taal, R. C. Hendriks, R. Heusdens, and J. Jensen, “An algorithm for intelligibility prediction of time-frequency weighted noisy speech,” _IEEE Transactions on Audio, Speech and Language Processing_ , vol. 19, no. 7, pp. 2125–2136, 2011. * Roux et al. [2019] J. L. Roux, S. Wisdom, H. Erdogan, and J. R. Hershey, “SDR - Half-baked or Well Done?” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2019-May, pp. 626–630, 2019\. * Li et al. [2022] H. Li, K. Chen, L. Wang, J. Liu, B. Wan, and B. Zhou, “Sound source separation mechanisms of different deep networks explained from the perspective of auditory perception,” _Applied Sciences_ , vol. 12, no. 2, p. 832, 2022. * Fan et al. [2020] C. Fan, J. Tao, B. Liu, J. Yi, Z. Wen, and X. Liu, “End-to-End Post-Filter for Speech Separation with Deep Attention Fusion Features,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 28, pp. 1303–1314, 2020. * Taal et al. [2010] C. H. Taal, R. C. Hendriks, R. Heusdens, and J. Jensen, “A Short-Time Objective Intelligibility Measure For Time-Frequency Weighted Noisy Speech,” _IEEE International Conference on Acoustics, Speech, and Signal Processing_ , pp. 4214–4217, 2010. [Online]. Available: http://cas.et.tudelft.nl/pubs/Taal2010.pdf * Rix et al. [2001] A. W. Rix, J. G. Beerends, M. P. Hollier, and A. P. Hekstra, “Perceptual evaluation of speech quality (PESQ) - A new method for speech quality assessment of telephone networks and codecs,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2, pp. 749–752, 2001. * Martin-Donas et al. [2018] J. M. Martin-Donas, A. M. Gomez, J. A. Gonzalez, and A. M. Peinado, “A deep learning loss function based on the perceptual evaluation of the speech quality,” _IEEE Signal Processing Letters_ , vol. 25, no. 11, pp. 1680–1684, 2018. * Kolbaek et al. [2020] M. Kolbaek, Z. H. Tan, S. H. Jensen, and J. Jensen, “On Loss Functions for Supervised Monaural Time-Domain Speech Enhancement,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 28, pp. 825–838, 2020\. * Bie et al. [2022] X. Bie, S. Leglaive, X. Alameda-Pineda, and L. Girin, “Unsupervised speech enhancement using dynamical variational autoencoders,” _IEEE/ACM Transactions on Audio, Speech, and Language Processing_ , vol. 30, pp. 2993–3007, 2022. * Fujimura et al. [2021] T. Fujimura, Y. Koizumi, K. Yatabe, and R. Miyazaki, “Noisy-target training: A training strategy for dnn-based speech enhancement without clean speech,” in _2021 29th European Signal Processing Conference (EUSIPCO)_. IEEE, 2021, pp. 436–440. * Wisdom et al. [2020] S. Wisdom, E. Tzinis, H. Erdogan, R. J. Weiss, K. Wilson, and J. R. Hershey, “Unsupervised sound separation using mixture invariant training,” in _Advances in Neural Information Processing Systems_ , vol. 2020-Decem, jun 2020. [Online]. Available: http://arxiv.org/abs/2006.12701 * Zhang et al. [2021b] J. Zhang, C. Zorila, R. Doddipatla, and J. Barker, “Teacher-student mixit for unsupervised and semi-supervised speech separation,” _arXiv preprint arXiv:2106.07843_ , 2021. * Saito et al. [2021] K. Saito, S. Uhlich, G. Fabbro, and Y. Mitsufuji, “Training speech enhancement systems with noisy speech datasets,” _arXiv preprint arXiv:2105.12315_ , 2021\. * Karamatlı and Kırbız [2022] E. Karamatlı and S. Kırbız, “Mixcycle: Unsupervised speech separation via cyclic mixture permutation invariant training,” _IEEE Signal Processing Letters_ , 2022. * Trinh and Braun [2022] V. A. Trinh and S. Braun, “Unsupervised speech enhancement with speech recognition embedding and disentanglement losses,” in _ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2022, pp. 391–395. * Tzinis et al. [2022] E. Tzinis, Y. Adi, V. K. Ithapu, B. Xu, P. Smaragdis, and A. Kumar, “Remixit: Continual self-training of speech enhancement models via bootstrapped remixing,” _IEEE Journal of Selected Topics in Signal Processing_ , vol. 16, no. 6, pp. 1329–1341, 2022. * Chen et al. [2023b] L.-W. Chen, Y.-F. Cheng, H.-S. Lee, Y. Tsao, and H.-M. Wang, “A training and inference strategy using noisy and enhanced speech as target for speech enhancement without clean speech,” in _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , 2023, pp. 5315–5319. * Reddy et al. [2021] C. K. Reddy, V. Gopal, and R. Cutler, “Dnsmos: A non-intrusive perceptual objective speech quality metric to evaluate noise suppressors,” in _ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2021, pp. 6493–6497. * Xiang and Bao [2020] Y. Xiang and C. Bao, “A parallel-data-free speech enhancement method using multi-objective learning cycle-consistent generative adversarial network,” _IEEE/ACM Transactions on Audio, Speech, and Language Processing_ , vol. 28, pp. 1826–1838, 2020. * Zhu et al. [2017] J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” in _Proceedings of the IEEE international conference on computer vision_ , 2017, pp. 2223–2232. * Li et al. [2021b] Y. Li, Y. Sun, K. Horoshenkov, and S. M. Naqvi, “Domain adaptation and autoencoder-based unsupervised speech enhancement,” _IEEE Transactions on Artificial Intelligence_ , vol. 3, no. 1, pp. 43–52, 2021. * Lehtinen et al. [2018] J. Lehtinen, J. Munkberg, J. Hasselgren, S. Laine, T. Karras, M. Aittala, and T. Aila, “Noise2noise: Learning image restoration without clean data,” _arXiv preprint arXiv:1803.04189_ , 2018. * Pan et al. [2010] S. J. Pan, I. W. Tsang, J. T. Kwok, and Q. Yang, “Domain adaptation via transfer component analysis,” _IEEE transactions on neural networks_ , vol. 22, no. 2, pp. 199–210, 2010. * Liao et al. [2018] C.-F. Liao, Y. Tsao, H.-Y. Lee, and H.-M. Wang, “Noise adaptive speech enhancement using domain adversarial training,” _arXiv preprint arXiv:1807.07501_ , 2018. * Wang et al. [2018d] Q. Wang, W. Rao, S. Sun, L. Xie, E. S. Chng, and H. Li, “Unsupervised domain adaptation via domain adversarial training for speaker recognition,” in _2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2018, pp. 4889–4893. * Xu et al. [2014c] Y. Xu, J. Du, L.-R. Dai, and C.-H. Lee, “Cross-language transfer learning for deep neural network based speech enhancement,” in _The 9th International Symposium on Chinese Spoken Language Processing_. IEEE, 2014, pp. 336–340. * Pascual et al. [2018] S. Pascual, M. Park, J. Serrà, A. Bonafonte, and K.-H. Ahn, “Language and noise transfer in speech enhancement generative adversarial network,” in _2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2018, pp. 5019–5023. * Ganin et al. [2016] Y. Ganin, E. Ustinova, H. Ajakan, P. Germain, H. Larochelle, F. Laviolette, M. Marchand, and V. Lempitsky, “Domain-adversarial training of neural networks,” _The journal of machine learning research_ , vol. 17, no. 1, pp. 2096–2030, 2016. * Chung et al. [2019] Y. A. Chung, W. N. Hsu, H. Tang, and J. Glass, “An unsupervised autoregressive model for speech representation learning,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2019-September, pp. 146–150, 2019. * Chung et al. [2020] Y. A. Chung, H. Tang, and J. Glass, “Vector-quantized autoregressive predictive coding,” _Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH_ , vol. 2020-October, no. 1, pp. 3760–3764, 2020. * Liu et al. [2020] A. T. Liu, S. W. Yang, P. H. Chi, P. C. Hsu, and H. Y. Lee, “Mockingjay: Unsupervised speech representation learning with deep bidirectional transformer encoders,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2020-May, pp. 6419–6423, 2020. * Liu et al. [2021] A. T. Liu, S. W. Li, and H. Y. Lee, “TERA: Self-Supervised Learning of Transformer Encoder Representation for Speech,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 29, pp. 2351–2366, 2021. * Baevski et al. [2020] A. Baevski, H. Zhou, A. Mohamed, and M. Auli, “wav2vec 2.0: A framework for self-supervised learning of speech representations,” _Advances in Neural Information Processing Systems_ , vol. 2020-Decem, no. Figure 1, pp. 1–19, 2020. * Hsu et al. [2021a] W. N. Hsu, B. Bolte, Y. H. H. Tsai, K. Lakhotia, R. Salakhutdinov, and A. Mohamed, “HuBERT: Self-Supervised Speech Representation Learning by Masked Prediction of Hidden Units,” _IEEE/ACM Transactions on Audio Speech and Language Processing_ , vol. 29, no. Cv, pp. 3451–3460, 2021. * Huang et al. [2022] Z. Huang, S. Watanabe, S. W. Yang, P. García, and S. Khudanpur, “Investigating Self-Supervised Learning for Speech Enhancement and Separation,” _ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings_ , vol. 2022-May, pp. 6837–6841, 2022\. * Hsu et al. [2021b] W.-N. Hsu, B. Bolte, Y.-H. H. Tsai, K. Lakhotia, R. Salakhutdinov, and A. Mohamed, “Hubert: Self-supervised speech representation learning by masked prediction of hidden units,” _IEEE/ACM Transactions on Audio, Speech, and Language Processing_ , vol. 29, pp. 3451–3460, 2021. * Hung et al. [2022] K.-H. Hung, S.-w. Fu, H.-H. Tseng, H.-T. Chiang, Y. Tsao, and C.-W. Lin, “Boosting self-supervised embeddings for speech enhancement,” _arXiv preprint arXiv:2204.03339_ , 2022. * Irvin et al. [2023] B. Irvin, M. Stamenovic, M. Kegler, and L.-C. Yang, “Self-supervised learning for speech enhancement through synthesis,” in _ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)_. IEEE, 2023, pp. 1–5. * Kong et al. [2020] J. Kong, J. Kim, and J. Bae, “Hifi-gan: Generative adversarial networks for efficient and high fidelity speech synthesis,” _Advances in Neural Information Processing Systems_ , vol. 33, pp. 17 022–17 033, 2020. * Germain et al. [2018] F. G. Germain, Q. Chen, and V. Koltun, “Speech denoising with deep feature losses,” _arXiv preprint arXiv:1806.10522_ , 2018. * Hao et al. [2023] X. Hao, C. Xu, and L. Xie, “Neural speech enhancement with unsupervised pre-training and mixture training,” _Neural Networks_ , vol. 158, pp. 216–227, 2023.
Dedicated to the 70th anniversary of Peter Olver # Natural differential invariants and equivalence of third order nonlinear differential operators Valentin Lychagin Institute of Control Sciences of RAS, Moscow, Russia <EMAIL_ADDRESS>and Valeriy Yumaguzhin Program Systems Institute of RAS, Pereslavl’-Zalesskiy, Russia<EMAIL_ADDRESS> ###### Abstract. We give a description of the field of rational natural differential invariants for a class of nonlinear differential operators of the third order on a two dimensional manifold and show their application to the equivalence problem of such operators. ###### Key words and phrases: 3rd order linear partial differential operator, jet bundle, differential invariant, equivalence problem ###### 2010 Mathematics Subject Classification: Primary: 58J70,53C05,35A3; Secondary: 35G05,53A55 V. Yumaguzhin is the corresponding author ## 1\. Introduction In this paper, we continue to study rational differential invariants of differential operators on two-dimensional manifolds. Here we consider a class of nonlinear operators that in local coordinates $x_{1},x_{2}$ have the following form: $A_{w}\colon f\longmapsto\sum\limits_{i,j,k}a_{ijk}(x,f)\frac{\partial^{3}f}{\partial x_{i}\partial x_{j}\partial x_{k}}+\sum\limits_{i,j}a_{ij}\left(x,f\right)\frac{\partial^{2}f}{\partial x_{i}\partial x_{j}}\\\ +\sum\limits_{i,j}a_{i}\left(x,f\right)\frac{\partial f}{\partial x_{i}}+a_{0}\left(x,f\right)f,$ (1) where $x=(x_{1},x_{2})$, $f=f(x)$. We call such operators as weakly nonlinear operators. In paper [5], we found rational differential invariants of the 2nd order weakly nonlinear operators and used them to solve the local equivalence problem. Here we use the methods of [5] to find rational differential invariants for the 3rd order weakly nonlinear operators and apply them to solve the local as well as the global equivalence problem. Additionally we will assume that all coefficients $a(x,y)$ of these operators are smooth in $x$ and rational in $y$. It is easy to see that the pseudogroup of local diffeomorphisms (in variables $x$) naturally acts in this class of operators. The main step in studies of such operators is a representation of operator $A_{w}$ as a pair $(A,f)$ (we call it as a (related pair), where $A=\sum\limits_{i,j,k}a_{ijk}\left(x,y\right)\frac{\partial^{3}}{\partial x_{i}\partial x_{j}\partial x_{k}}+\sum\limits_{i,j}a_{ij}\left(x,y\right)\frac{\partial^{2}}{\partial x_{i}\partial x_{j}}\\\ +\sum\limits_{i}a_{i}\left(x,y\right)\frac{\partial}{\partial x_{i}}+a_{0}(x,y)$ (2) is a linear third order linear differential operator on the extended space $(x,y)$. The procedure of descent allows us to get differential invariants of weakly nonlinear operators from invariants of related pairs and linear operators of the form $A_{f}=\sum\limits_{i,j,k}a_{ijk}\left(x,f\right)\frac{\partial^{3}}{\partial x_{i}\partial x_{j}\partial x_{k}}+\sum\limits_{i,j}a_{ij}\left(x,f\right)\frac{\partial^{2}}{\partial x_{i}\partial x_{j}}\\\ +\sum\limits_{i}a_{i}\left(x,f\right)\frac{\partial}{\partial x_{i}}+a_{0}\left(x,f\right).$ (3) Using this procedure and a machinery of reduced Gröbner bases, we propose a method to find invariants of the 3rd order weakly nonlinear operators. ## 2\. Notations In this paper, we use the same notations as in [3, 4]. Let $M$ be $n$-dimensional manifold and let $\tau\colon TM\to M$ and $\tau^{*}\colon T^{*}M\to M$ be respectively tangent and cotangent bundles over $M$. We will denote by $\mathit{Diff}_{k}(M)$ the module of linear $k$-th order operators, acting in $C^{\infty}(M)$. The corresponding vector bundle of such operators we denote by $\chi_{k}\colon\mathbf{Diff}_{k}\to M$, thus $\mathit{Diff}_{k}(M)$ is also the module of smooth sections of $\chi_{k}$, $\mathit{Diff}_{k}(M)=C^{\infty}(\chi_{k})$. The group of diffeomorphisms of $M$ will be denoted by $\mathcal{G}(M)$ and the multiplicative group of nowhere vanishing functions on $M$ will be denoted as $\mathcal{F}(M)$. We denote by $\Sigma^{k}(M)$ and $\Sigma_{k}(M)$ the modules of symmetric $k$-forms and $k$-vectors respectively. Also, we denote by $\Omega^{k}$ the modules of exterior $k$-forms. ## 3\. Linear differential operators on $M$ In this section, we recall the necessary facts from [3] on scalar linear third order differential operators on two-dimensional manifolds. Let $M$ be 2-dimensional smooth oriented connected manifold and let $\pi\colon M\times\mathbb{R}\longrightarrow M$ be trivial line bundle. Denote by $\pi_{k}\colon\mathbf{J}^{k}(\pi)\longrightarrow M$ the bundles of $k$-jets of smooth sections of $\pi$ (i.e. $k$-jets of smooth functions on $M$). Let $\mathcal{G}(M)$ be the group of diffeomorphisms of the manifold $M$. Then, together with the action of $\mathcal{G}\left(M\right)$ on $M$, we have also the actions $\mathcal{G}\left(M\right)$ in the bundles $\pi_{k}$ by prolongations of diffeomorphisms. The prolongations of diffeomorphisms $\phi\in\mathcal{G}(M)$ in the bundles $\pi_{k}$ will be denoted by $\phi^{(k)}$. ### 3.1. Let $\chi_{3}\colon\mathbf{Diff}_{3}\left(M\right)\longrightarrow M$ be the bundle of the third order linear differential operators on $M$ and $\chi_{3,k}\colon\mathbf{J}^{k}(\chi_{3})\longrightarrow M$ be bundles of their $k$-jets of sections of $\chi_{3}$. The group $\mathcal{G}\left(M\right)$ acts on operators $A\in\mathit{Diff}_{3}(M)$ in natural way: $\phi_{\ast}\colon A\rightarrow\phi_{\ast}(A)=\phi_{*}\circ A\circ\phi_{*}^{-1},$ where the morphism $\phi_{\ast}\colon C^{\infty}(M)\longrightarrow C^{\infty}(M)$ is defined by the formula $\phi_{\ast}=(\phi^{-1})^{\ast}\colon h\mapsto h\circ\phi^{-1}.$ ### 3.2. Symbols By the symbol $\sigma_{3,A}$ of the operator $A$, we means the equivalence class $\sigma_{3,A}=A\\!\\!\\!\mod\\!\mathit{Diff}_{2}(M).$ Let an operator $A\in\mathit{Diff}_{3}(M)$ be represented in local coordinates $x_{1},x_{2}$ of $M$ in the form $A=a_{1}\partial^{3}_{1}+3a_{2}\partial^{2}_{1}\partial_{2}+3a_{3}\partial_{1}\partial^{2}_{2}+a_{4}\partial^{3}_{2}\\\ +b_{1}\partial^{2}_{1}+2b_{2}\partial_{1}\partial_{2}+b_{3}\partial^{2}_{2}+c_{1}\partial_{1}+c_{2}\partial_{2}+a_{0},$ (4) where the all coefficients are smooth functions of $x=(x_{1},x_{2})$, $\partial_{1}$ and $\partial_{2}$ are $\partial_{x_{1}}$ and $\partial_{x_{2}}$ respectively. Then the symbol $\sigma_{3,A}$ is identified with the symmetric 3-vector $\sigma_{3,A}=a_{1}\partial^{3}_{1}+3a_{2}\partial^{2}_{1}\partial_{2}+3a_{3}\partial_{1}\partial^{2}_{2}+a_{4}\partial^{3}_{2}\in\Sigma_{3}(M),$ (5) where dots and degrees are symmetric products of the vector fields $\partial_{1},\;\partial_{2}$. The symbol $\sigma_{3,A}$ is regular (at point or in domain) if it as a homogeneous cubic polynomial on the cotangent bundle $T^{*}(M)$ has distinct roots. Denote by $\Delta(\sigma_{3,A})$ the discriminant of $\sigma_{3,A}$, then $\Delta(\sigma_{3,A})=6a_{1}a_{2}a_{3}a_{4}-4(a_{1}a_{3}^{3}+a_{4}a_{2}^{3})+3a_{2}^{2}a_{3}^{2}-a_{1}^{2}a_{4}^{2}.$ Recall that the symbol $\sigma_{3,A}$ has three distinct real roots if $\Delta(\sigma_{3,A})>0$ and one real and two complex roots if $\Delta(\sigma_{3,A})<0$. Thus, we say that the operator $A$ is regular if its symbol $\sigma_{3,A}$ is regular or $\Delta(\sigma_{3,A})\neq 0$ More over, we say that an operator $A$ is hyperbolic if $\Delta(\sigma_{3,A})>0$ and ultrahyperbolic if $\Delta(\sigma_{3,A})<0$. Locally, the symbol of a hyperbolic operator can be presented as a symmetric product of pair wise linear independent vector fields $\sigma_{3,A}=X_{1}\cdot X_{2}\cdot X_{3}$, and therefore there are local coordinates $x_{1},x_{2}$ such that $\sigma_{3,A}=(a\partial_{1}+b\partial_{2})\cdot\partial_{1}\cdot\partial_{2},$ (6) where $a$ and $b$ are smooth functions and $ab\neq 0$. For the case of ultrahyperbolic operators we have $\sigma_{3,A}=X\cdot q$, where $X$ is a nonzero vector field and $q\in\Sigma_{2}$ is a positive symmetric 2-vector. Therefore, there are local coordinates $x_{1},x_{2}$ such that $\sigma_{3,A}=(a\partial_{1}+b\partial_{2})\cdot(\partial^{2}_{1}+\partial^{2}_{2}),$ (7) where $a$ and $b$ are smooth functions and $a^{2}+b^{2}>0$. ### 3.3. Wagner connections The following result due to Wagner [7, 3]. ###### Theorem 1. 1. (1) Let $A\in{\it Diff}_{3}(M)$ be a regular differential operator. Then there exist a unique linear connection $\nabla$ such that the symbol $\sigma_{3,A}$ is parallel with respect to $\nabla$. 2. (2) The curvature tensor of $\nabla$ is zero. We call this connection Wagner’s connection. In the case of hyperbolic symbol, we choose local coordinates in $M$ such that $\sigma$ has form (6). Then the non zero Christoffel coefficients $\Gamma^{i}_{jk}$ of the Wagner connection are the following: $\displaystyle\Gamma^{1}_{11}=\frac{1}{3}\big{(}\ln\frac{b}{a^{2}}\big{)}_{x_{1}},\quad\Gamma^{2}_{22}=\frac{1}{3}\big{(}\ln\frac{a}{b^{2}}\big{)}_{x_{2}},$ $\displaystyle\Gamma^{1}_{12}=\frac{1}{3}\big{(}\ln\frac{b}{a^{2}}\big{)}_{x_{2}},\quad\Gamma^{2}_{21}=\frac{1}{3}\big{(}\ln\frac{a}{b^{2}}\big{)}_{x_{1}}.$ In the ultrahyperbolic case (7) we have the following nonzero Chritoffel coefficients: $\displaystyle\Gamma^{1}_{12}=\Gamma^{2}_{22}=-\frac{1}{6}\big{(}\ln(a^{2}+b^{2})\big{)}_{x_{2}},$ $\displaystyle\Gamma^{1}_{21}=-\Gamma^{2}_{11}=\frac{a_{x_{1}}b-ab_{x_{1}}}{a^{2}+b^{2}},\quad\Gamma^{1}_{22}=-\Gamma^{2}_{12}=\frac{ab_{x_{2}}-a_{x_{2}}b}{a^{2}+b^{2}}.$ ###### Corollary 2. 1 The torsion form $\theta$ of the Wagner connection is $\theta=\frac{1}{3}\big{(}\ln\frac{b^{2}}{a}\big{)}_{x_{1}}dx_{1}+\frac{1}{3}\big{(}\ln\frac{a^{2}}{b}\big{)}_{x_{2}}dx_{2}$ for the hyperbolic case and $\theta=\frac{ab_{x_{2}}-a_{x_{2}}b}{a^{2}+b^{2}}dx_{1}+\frac{ab_{x_{1}}-a_{x_{1}}b}{a^{2}+b^{2}}dx_{2}-\frac{1}{6}d\big{(}\ln(a^{2}+b^{2})\big{)}$ for the ultrahyperbolic case. ### 3.4. Symols and quantization Let $\Sigma^{\cdot}=\oplus_{k\geq 0}\Sigma^{k}(M)$ be the graded algebra of symmetric differential forms and let $\nabla$ be the Wagner connection associated with a regular symbol from $\Sigma_{3}(M)$. Then the covariant differential $d_{\nabla}:\Omega^{1}(M)\longrightarrow\Omega^{1}(M)\otimes\Omega^{1}(M)$ define derivation $d_{\nabla}^{s}\colon\Sigma^{\cdot}\longrightarrow\Sigma^{\cdot+1}$ of degree one in graded symmetric algebra $\Sigma^{\cdot}\oplus_{k\geq 0}\Sigma^{k}(M)$. Namely, this derivation is defined by its action on generators, and we have $\displaystyle d^{s}_{\nabla}=d:C^{\infty}(M)\longrightarrow\Omega^{1}(M)=\Sigma^{1},$ $\displaystyle d^{s}_{\nabla}:\Omega^{1}(M)=\Sigma^{1}\stackrel{{\scriptstyle d_{\nabla}}}{{\longrightarrow}}\Omega^{1}(M)\otimes\Omega^{1}(M)\stackrel{{\scriptstyle\mathrm{Sym}}}{{\longrightarrow}}\Sigma^{2}.$ Let now $\alpha_{k}\in\Sigma_{k}(M)$. We define a differential operator $\mathcal{Q}(\alpha_{k})\in\mathit{Diff}_{k}(M)$ as follows: $\mathcal{Q}(\alpha_{k})(h)\stackrel{{\scriptstyle\mathrm{def}}}{{=}}\frac{1}{k!}\left\langle\,\alpha_{k},\,\big{(}d^{s}_{\nabla}\big{)}^{k}(h)\,\right\rangle,$ where $h\in C^{\infty}(M)$, $\big{(}d^{s}_{\nabla}\big{)}^{k}(h)\in\Sigma^{k}(M)$, and $\langle\cdot\,,\cdot\rangle$ is the standard convolution $\Sigma_{k}(M)\otimes\Sigma^{k}(M)\longrightarrow C^{\infty}(M).$ Remark, that the value of the symbol of the derivation $d^{s}_{\nabla}$ at a covector $\theta$ equals to the symmetric product by $\theta$ into the module $\Sigma^{\cdot}$. Therefore, the symbol of operator $\mathcal{Q}(\alpha_{k})$ equals $\alpha_{k}$ as the symbol of a composition of operators equals the composition of symbols. We call differential operator $\mathcal{Q}(\alpha_{k})$ a quantization of symbol $\alpha_{k}$. Let now $A\in\mathit{Diff}_{3}(M)$ and $\sigma_{3,A}$ be its symbol. Then operator $A-\mathcal{Q}\big{(}\sigma_{3,A}\big{)}$ has order $2$, and let $\sigma_{2,A}$ be its symbol. Then operator $A-\mathcal{Q}\big{(}\sigma_{3,A}\big{)}-\mathcal{Q}\big{(}\sigma_{2,A}\big{)}$ has order $1$ and let $\sigma_{1,A}$ be its symbol. Thus we get subsymbols $\sigma_{i,A}\in\Sigma_{i}(M)$, $0\leq i\leq 2$, such that $A=\mathcal{Q}\big{(}\sigma_{(3)}(A)\big{)},$ where $\sigma_{(3)}(A)=\sigma_{3,A}+\sigma_{2,A}+\sigma_{1,A}+\sigma_{0,A}$ (8) is the total symbol and $\mathcal{Q}\big{(}\sigma_{(3)}(A)\big{)}=\mathcal{Q}\big{(}\sigma_{3,A}\big{)}+\mathcal{Q}\big{(}\sigma_{2,A}\big{)}+\ldots+\mathcal{Q}\big{(}\sigma_{0,A}\big{)}.$ #### 3.4.1. Coordinates Let $x_{1},x_{2}$ be local coordinates in a neighborhood $\mathcal{O}\subset M$, where the symbol $\sigma_{3,A}$ is regular. Denote by $x_{1},x_{2},w_{1},w_{2}$ induced standard coordinates in the tangent bundle over $\mathcal{O}$. Then $d_{\nabla}(dx_{k})=-\sum\Gamma^{k}_{ij}dx_{i}\otimes dx_{j}$, where $\Gamma^{k}_{ij}$ are the Christoffel symbols of the Wagner connection $\nabla$. Thus, in coordinates $x,w$ we have $d^{s}_{\nabla}(w_{k})=-\sum\Gamma^{k}_{ij}w_{i}w_{j}$ and the derivation $d^{s}_{\nabla}$ has the form: $d^{s}_{\nabla}=\sum w_{i}\partial_{x_{i}}-\sum\Gamma^{k}_{ij}w_{i}w_{j}\partial_{w_{k}}.$ ### 3.5. Universal differential operator We define a total operator of third order $\square_{3}\colon C^{\infty}(J^{k}\chi_{3})\longrightarrow C^{\infty}(J^{k+3}\chi_{3})$ as it was done in [3]. In local coordinates this operator has the form $\square_{3}=6\sum_{\alpha,0\leq|\alpha|\leq 3}\frac{u^{\alpha}}{\alpha!}\Big{(}\frac{d}{dx}\Big{)}^{\alpha},$ (9) where $\alpha=(\alpha_{1},\alpha_{2})$, $|\alpha|=\alpha_{1}+\alpha_{2}$, $(d/dx)^{\alpha}=(d/dx_{1})^{\alpha_{1}}(d/dx_{2})^{\alpha_{2}}$, and $d/dx_{1},d/dx_{2}$ are total derivatives. ###### Theorem 3. 1 The operator $\square_{3}$ commutes with the action of the group $\mathcal{G}(M)$ on the jet bundles. ### 3.6. Natural differential invariants of regular operators By natural differential invariants of order k we mean a function on $J^{k}(\chi_{3})$ which are $\mathcal{G}(M)$-invariant and rational along fibers of the projection $\chi_{3,k}$. ###### Theorem 4. If $I$ is a natural differential invariant of order $\leq k$ for differential operators of the third order, then $\square_{3}(I)$ is a natural differential invariant of the order $\leq(k+3)$ for these operators. We say that two natural differential invariants $I_{1},I_{2}$ are in general position if $\hat{d}I_{1}\wedge\hat{d}I_{2}\neq 0,$ (10) where $\hat{d}$ is the total differential, and denote by $\mathcal{O}(I_{1},I_{2})\subset J^{\infty}(\chi_{3})$ the open domain, where condition (10) holds. ###### Theorem 5. Let natural differential invariants $I_{1},I_{2}$ are in general position and let $J^{\alpha}=\square_{3}(I^{\alpha}),$ where $\alpha=(\alpha_{1},\alpha_{2})$ and $0\leq|\alpha|\leq 3$. Then the field of natural differential invariants for differential operators of the third order in the domain $\mathcal{O}(I_{1},I_{2})$ is generated by invariants $I_{1},I_{2},J^{\alpha}$ and all their Tresse derivatives $\frac{d^{l}J^{\alpha}}{dI_{1}^{l_{1}}dI_{2}^{l_{2}}},$ where $l=l_{1}+l_{2}$. Thus, to obtain the field of natural differential invariants of linear regular differential operators of order 3, it is enough to have two natural differential invariants $I_{1}$, $I_{2}$ in general position. The invariants can be obtained by various methods. For example, as the invariant $I_{1}$ one can take the free term $u_{0}$ of the universal operator $\square_{3}$, or $I_{1}$ is natural invariant such that $I_{1}(A)$ is the natural convolution of the torsion form $\theta$ of the Wagner connection and the subsymbol $\sigma_{1,A}$. As the natural invariant $I_{2}$ one can take, for example, the invariant $\square_{3}(I_{1})$. #### 3.6.1. Let $\mathbf{F}_{k}^{(3)}$ be the field of all natural differential invariants of order $\leq k$ of linear scalar differential operators of order $\leq 3$ on $M$. The $\mathcal{G}(M)$-action on $M$ is transitive and, therefore, the set of all such invariants forms an $\mathbb{R}$-field $\mathbf{F}_{k}^{(3)}$. The natural projections $\chi_{k,l}\colon\mathbf{J}^{k}(\chi)\rightarrow\mathbf{J}^{l}(\chi),\;k>l$, define the embeddings $\mathbf{F}_{l}^{(3)}\hookrightarrow\mathbf{F}_{k}^{(3)}$ and their inductive limit $\mathbf{F}_{*}^{(3)}$ is the field of natural differential invariants of linear differential operators on the manifold $M$. Remark that the $\mathcal{G}(M)$-action on differential operators satisfies the conditions of the Lie-Tresse theorem (see [2]) and, therefore, the field $\mathbf{F}_{*}^{(3)}$ separates regular $\mathcal{G}(M)$-orbits. ## 4\. Linear differential operators on $M\times\mathbb{R}$ Consider now linear differential operators $A$ of the third order on the manifold $\mathbf{J}^{0}\left(\pi\right)=M\times\mathbb{R}$, that satisfy the following condition: $\big{(}A-A(1)\big{)}(y)=0,$ where $y$ is the fibrewise coordinate on $\mathbf{J}^{0}(\pi)$. In local coordinates $(x_{1},x_{2})$ on $M,$ we have representation (2) of operators of such type: $A=\sum\limits_{i,j,k}a_{ijk}\left(x,y\right)\frac{\partial^{3}}{\partial x_{i}\partial x_{j}\partial x_{k}}+\sum\limits_{i,j}a_{ij}\left(x,y\right)\frac{\partial^{2}}{\partial x_{i}\partial x_{j}}\\\ +\sum\limits_{i}a_{i}\left(x,y\right)\frac{\partial}{\partial x_{i}}+a_{0}\left(x,y\right).$ The module of these operators we will denote by $\mathit{Diff}_{3}\left(M,\mathbb{R}\right)$ and by $\mathbb{\zeta}\colon\mathbf{Diff}_{3}\left(M,\mathbb{R}\right)\longrightarrow M\times\mathbb{R}$ we will denote the corresponding bundle of these operators. As before, elements of the module of $\mathit{Diff}_{3}(M,\mathbb{R})$ are just sections of the bundle $\mathbb{\zeta}$ with the correspondence $\mathit{Diff}_{3}\left(M,\mathbb{R}\right)\equiv C^{\infty}(\mathbb{\zeta}),\quad A\mapsto s_{A},$ where $s_{A}\left(a,y\right)=A_{\left(a,y\right)}$ and $\left(a,y\right)\in M\times\mathbb{R}$. The diffeomorphism group $\mathcal{G}(M)$ acts by prolongation $\phi\rightarrow\phi^{(0)}$ on the manifold $\mathbf{J}^{0}(\pi)=M\times\mathbb{R}$, preserves the function $y$, and therefore acts in the bundle $\mathbb{\zeta}$ as well as in the $k$-jet bundles $\mathbb{\zeta}_{k}\colon\mathbf{J}^{k}\left(\mathbb{\zeta}\right)\rightarrow M\times\mathbb{R}$. ### 4.1. Linear differential operators $\mathbf{A_{f}}$ Given an operator $A\in\mathit{Diff}_{3}(M,\mathbb{R})$ and a function $f\in C^{\infty}(M)$ we define the operator $A_{f}=s_{f}^{\ast}\circ A\circ\pi^{\ast}\in\mathit{Diff}_{3}(M)$ as the operator that corresponds to the restriction of the section $s_{A}$ to the graph of the function $f$. Here $s_{f}\colon M\rightarrow M\times\mathbb{R}$ is the section of the bundle $\pi$ that corresponds to the function $f$. In local coordinates, we get the above representation (3) for this type of operators: $A_{f}=\sum\limits_{i,j,k}a_{ijk}\left(x,f\right)\frac{\partial^{3}}{\partial x_{i}\partial x_{j}\partial x_{k}}+\sum\limits_{i,j}a_{ij}\left(x,f\right)\frac{\partial^{2}}{\partial x_{i}\partial x_{j}}\\\ +\sum\limits_{i}a_{i}\left(x,f\right)\frac{\partial}{\partial x_{i}}+a_{0}\left(x,f\right).$ ### 4.2. Weakly nonlinear differential operators Define now thespace of weakly nonlinear operators of the third order $\mathit{Diff}_{3}^{w}(M)$ as the space of differential operators $A_{w}$ on $C^{\infty}\left(M\right)$ of the form $A_{w}\left(f\right)=A_{f}\left(f\right),$ (11) where $A\in\mathit{Diff}_{3}\left(M,\mathbb{R}\right)$ and $f\in C^{\infty}(M)$. ###### Proposition 6. The mappings $\Theta\colon\mathit{Diff}_{3}(M,\mathbb{R})\times C^{\infty}(M)\longrightarrow\mathit{Diff}_{3}(M),\quad\Theta\colon(A,f)\mapsto A_{f},$ and $\Theta_{w}\colon\mathit{Diff}_{3}(M,\mathbb{R})\longrightarrow\mathit{Diff}_{3}^{w}(M),\quad\Theta_{w}\colon A\mapsto A_{w},$ are natural in the following sense : $\displaystyle\phi_{\ast}\left(A,f\right)$ $\displaystyle=$ $\displaystyle\left(\phi_{\ast}^{\left(0\right)}\left(A\right),\phi_{\ast}\left(f\right)\right),$ $\displaystyle\phi_{\ast}\left(A_{f}\right)$ $\displaystyle=$ $\displaystyle\big{(}\phi_{\ast}^{(0)}(A)\big{)}_{\phi_{\ast}(f)},$ and $\phi_{\ast}(A_{w})=\big{(}\phi_{\ast}^{(0)}(A)\big{)}_{w},$ for all diffeomorphisms $\phi\in\mathcal{G}\left(M\right)$. ###### Proof. The proof is almost verbatim repetition of the proof of Proposition 1 in [5]. ∎ ## 5\. Related pairs and their invariants The above proposition has the following consequences: * • By differential $\mathcal{G}(M)$-invariants of weakly nonlinear operators we will mean differential $\mathcal{G}(M)$-invariants of operators in$\mathit{Diff}_{3}(M,\mathbb{R})$, that are $\mathcal{G}(M)$-invariant functions on jet spaces $\mathbf{J}^{k}(\mathbb{\zeta})$. * • In what follows, we will require that operators $A\in\mathit{Diff}_{3}(M,\mathbb{R})$ under consideration are rational in $y$ (as sections of the bundle $\mathbb{\zeta}$). Also, by differential invariants of rational weakly nonlinear operators we mean $\mathcal{G}(M)$-invariant functions on jet spaces $\mathbf{J}^{k}(\mathbb{\zeta})$ that are rational along fibres of the projections $\pi\circ\mathbb{\zeta}_{k}\colon\mathbf{J}^{k}(\mathbb{\zeta})\rightarrow M$. * • The group $\mathcal{G}\left(M\right)$ acts transitively on the manifold $M$ and therefore rational differential invariants of order $\leq k$ for rational weakly nonlinear operators form a field $\mathbf{F}_{k}^{w}$. We have the embedding $\mathbb{\zeta}_{k,l}^{\ast}:\mathbf{F}_{l}^{w}\hookrightarrow\mathbf{F}_{k}^{w},$ if $k\geq l,$ where $\mathbb{\zeta}_{k,l}:\mathbf{J}^{k}\left(\mathbb{\zeta}\right)\rightarrow\mathbf{J}^{l}\left(\mathbb{\zeta}\right)$ are the natural projections, and $\mathbf{F}^{w}=\bigcup\limits_{k\geq 0}\mathbf{F}_{k}^{w}$ is the field of all rational differential invariants of rational weakly nonlinear operators. We consider the following vector bundles over manifold $\mathbf{J}^{0}\left(\pi\right)=M\times\mathbb{R}$: 1. (1) jet bundles of functions on manifold: $\pi_{l,0}\colon\mathbf{J}^{l}(\pi)\longrightarrow\mathbf{J}^{0}(\pi),$ 2. (2) jet bundles of operators: $\mathbb{\zeta}_{k}\colon\mathbf{J}^{k}(\mathbb{\zeta})\longrightarrow M\times\mathbb{R},$ 3. (3) the Whitney sum of vector bundles $\mathbb{\zeta}_{k}$ and $\pi_{l,0}$ $r_{k,l}\colon\mathbf{RP}^{k,l}=\mathbb{\zeta}_{k}\oplus_{M\times\mathbb{R}}\pi_{l,0}\longrightarrow M\times\mathbb{R}.$ We call the last bundle as the bundle of related pairs. Elements of the total space of this bundle are related pairs $\left([A]_{\left(x,y\right)}^{k},[f]_{x}^{l}\right)$ consisting of $k$-jet $[A]_{\left(x,y\right)}^{k}$ of operator $A\in\mathit{Diff}_{3}\left(M,\mathbb{R}\right)$ at the point $\left(x,y\right)\in M\times\mathbb{R}$ and $l$ -jet $[f]_{x}^{l}$ of function $f\in C^{\infty}(M)$ at the point $x\in M$ under condition that $f\left(x\right)=y$. The group $\mathcal{G}(M)$ acts by prolongations in the bundles $r_{k,l}$ and by invariants of this action (or invariants of related pairs) we mean functions on the total space $\mathbf{CP}^{k,l}=\mathbf{J}^{k}(\zeta)\oplus_{M\times\mathbb{R}}\mathbf{J}^{l}(\pi)$. that are $\mathcal{G}\left(M\right)$-invariant and rational along fibers of the projection $\pi\circ r_{k,l}:\mathbf{RP}^{k,l}\rightarrow M$. Because of transitivity of $\mathcal{G}(M)$-action on $M,$ all such functions are completely determined by their values on fibre $\left(\pi\circ r_{k,l}\right)^{-1}(a)$ at a base point $a\in M$. Therefore, $\mathcal{G}\left(M\right)$-invariants of related pairs form an $\mathbb{R}$-field $\mathbf{F}_{k,l},$ that is a subfield of the filed $\mathbf{Q}_{k,l}$ of all rational functions on the fibre $\left(\pi\circ r_{k,l}\right)^{-1}\left(a\right)$. The natural projections $\mathbf{CP}^{k^{\prime},l^{\prime}}\rightarrow\mathbf{CP}^{k,l},$ where $k\leq k^{\prime},\,l\leq l^{\prime}$, give us embeddings of fields $\mathbf{F}_{k,l}\subset\mathbf{F}_{k^{\prime},l^{\prime}}$, $\mathbf{Q}_{k,l}\subset\mathbf{Q}_{k^{\prime},l^{\prime}}$ and we define the fields $\mathbf{F}_{l},\;\mathbf{Q}_{l}$ by the inductive limits: $\mathbf{F}_{l}=\bigcup\limits_{k\geq 0}\mathbf{F}_{k,l},\quad\mathbf{Q}_{l}=\bigcup\limits_{k\geq 0}\mathbf{Q}_{k,l}.$ Remark that $\mathbf{F}_{0}=\mathbf{F}^{w}$ is just the field of rational differential invariants of rational weakly nonlinear differential operators. Below we will discuss various methods of finding invariants, but first of all we remark that the vector field $\partial_{y}$ on $\mathbf{J}^{0}(\pi)=M\times\mathbb{R}$ is an invariant of the $\mathcal{G}(M)$-action. Therefore its $l$-th prolongation $\partial_{y}^{(l)}$ on $\mathbf{J}^{l}(\pi)$ is also $\mathcal{G}(M)$-invariant. The same is valid for the total derivation $\displaystyle\frac{d}{dy}$ that acts in $\mathbf{J}^{\infty}(\zeta)$. All together they define $\mathcal{G}\left(M\right)$-invariant derivation $\nabla$ in the fields $\mathbf{Q}_{l}$, as well as in $\mathbf{F}_{l}$, where $\nabla(\alpha\beta)=\frac{d\alpha}{dy}\beta+\alpha\partial_{y}^{(l)}(\beta),$ $\alpha\in\mathbf{Q}_{0}$ and $\beta$ is a function on $\mathbf{J}^{l}(\pi)$. ## 6\. Construction of invariants At first we consider $y$ as a parameter and identify operators $A\in\mathit{Diff}_{3}(M,\mathbb{R})$ with 1-parametric family $A_{y}$ of operators in $\mathrm{Diff}_{3}(M)$. Remark that $y$ is a $\mathcal{G}(M)$-invariant. Therefore, for any invariant $I\in\mathbf{F}_{k}^{(3)}$ of linear differential operators on manifold $M$, the function $\widehat{I}\colon J^{k}(\zeta)\rightarrow\mathbb{R}$, where $\widehat{I}\big{(}[A]^{k}_{(x,y)}\big{)}=I\big{(}[A_{y}]^{k}_{x}\big{)}$ is a $\mathcal{G}(M)$-invariant too. Thus we get a mapping $\mathbf{F}^{(3)}_{*}\longrightarrow\mathbf{F}_{0}=\mathbf{F}^{w},\quad I\mapsto\widehat{I},$ that immediately gives us invariants of weakly nonlinear operators. Moreover, application of the invariant derivation $\nabla$ essentially increases their amount. As we have seen, Proposition 6, the mapping $\Theta_{l}\colon\mathbf{RP}^{l,l}\longrightarrow\mathbf{J}^{l}(\chi),\quad\Theta_{l}\colon([A]_{(x,y)}^{l},[f]_{x}^{l})\mapsto[A_{f}]_{x}^{l},$ commutes with the $\mathcal{G}(M)$-action. Therefore, $\Theta_{l}^{*}(I)\in\mathbf{F}_{l},$ i.e., it is an invariant of related pairs for any invariant $I\in\mathbf{F}_{l}^{(3)}$. #### 6.0.1. Descent procedure 1 To get invariants of weakly nonlinear operators from invariants of related pairs we consider the following descent procedure for invariants $\mathbf{F}_{l}\longrightarrow\mathbf{F}^{w}.$ Let $I_{0}\in\mathbf{F}_{l}$ be an invariant of related pairs and let $I_{1}=\nabla(I_{0})\in\mathbf{F}_{l},\ldots,I_{i+1}=\nabla(I_{i})\in\mathbf{F}_{l}$ be its invariant derivatives. Remark that the transcendence degree of the field $\mathbf{Q}_{l}$ over $\mathbf{Q}_{0}$ equals $N=\dim\pi_{l,0}$. Therefore, ([8]), there are polynomial relations between rational functions ${I_{0},...,I_{N}}$. Denote by $J\left(I\right)\subset\mathbf{Q}_{0}[X_{0},...,X_{N}]$ the ideal of these relations. ###### Theorem 7. Let $b_{1},..,b_{r}\in\mathbf{Q}_{0}[X_{0},...,X_{N}]$ be the reduced Gröbner basis in the ideal $J(I)$ with respect to the standard lexicographic order. Then the coefficients of polynomials $b_{i}$ are natural invariants of weakly nonlinear operators. ###### Proof. The action of the diffeomorphism group preserves the ideal $J\left(I\right)$ as well as the lexicographic order. The reduced Gröbner basis in an ideal with respect to the lexicographic order is unique ([1]), and therefore, the action preserves elements of the basis and their coefficients. ∎ ## 7\. Example $n=1,k=3$ Let $M=\mathbb{R}$. Then in coordinate $x$ on M, an operator $A\in\mathrm{Diff}_{3}(M)$ has the form $A=a_{3}\partial^{3}+a_{2}\partial^{2}+a_{1}\partial+a_{0},$ where the all coefficients are smooth functions on $x$ and $\partial=\partial_{x}$. The symbol $a_{3}\partial^{3}$ of $A$ defines an invariant connection on the line $\mathbb{R}$ with the Christoffel coefficient $\Gamma$: $\Gamma=-\frac{a^{\prime}_{3}}{3a_{3}},$ the total symbol $\sigma_{(3)}$ of $A$ with respect to the Wagner connection $\nabla$ is the following: $\sigma_{(3)}=\sigma_{3}+\sigma_{2}+\sigma_{1}+\sigma_{0},$ where $\displaystyle\sigma_{3}=$ $\displaystyle a_{3}\partial^{3},$ (12) $\displaystyle\sigma_{2}=$ $\displaystyle(a_{2}-a^{\prime}_{3})\partial^{2},$ $\displaystyle\sigma_{1}=$ $\displaystyle\big{(}a_{1}-\frac{2(a^{\prime}_{3})^{2}a_{3}+3a^{\prime}_{3}-a^{\prime\prime}_{3}a_{3}}{9a_{3}}-(a_{2}-a^{\prime}_{3})\frac{a^{\prime}_{3}}{3a_{3}}\big{)}\partial,$ $\displaystyle\sigma_{0}=$ $\displaystyle a_{0}$ and the natural splitting of the operator $A$ is the following: $A=\widehat{\sigma_{(3)}}=\widehat{\sigma_{3}}+\widehat{\sigma_{2}}+\widehat{\sigma_{1}}+\widehat{\sigma_{0}},$ (13) where the differential operators $\widehat{\sigma_{k}}$ are the quantizations of the symbols $\sigma_{k}$, see [3]: $\displaystyle\widehat{\sigma_{3}}$ $\displaystyle=a_{3}\partial^{3}+a^{\prime}_{3}\partial^{2}+\frac{2(a^{\prime}_{3})^{2}a_{3}+3a^{\prime}_{3}-a^{\prime\prime}_{3}a_{3}}{9a_{3}}\,\partial,$ $\displaystyle\widehat{\sigma_{2}}$ $\displaystyle=(a_{2}-a^{\prime}_{3})\big{(}\partial^{2}+\frac{a^{\prime}_{3}}{3a_{3}}\,\partial\big{)},$ $\displaystyle\widehat{\sigma_{1}}$ $\displaystyle=\big{(}\,a_{1}-\frac{2(a^{\prime}_{3})^{2}a_{3}+3a^{\prime}_{3}-a^{\prime\prime}_{3}a_{3}}{9a_{3}}-(a_{2}-a^{\prime}_{3})\frac{a^{\prime}_{3}}{3a_{3}}\,\big{)}\partial,$ $\displaystyle\widehat{\sigma_{0}}$ $\displaystyle=a_{0}.$ Therefore, we get from (12) the following $\mathcal{G}(M)$-invariants of operators $A$: $\displaystyle I_{0}$ $\displaystyle=a_{0},$ $\displaystyle I_{1}$ $\displaystyle=\big{\langle}\,\sigma_{1},\,da_{0}\,\big{\rangle}=\big{(}\,a_{1}-\frac{2(a^{\prime}_{3})^{2}a_{3}+3a^{\prime}_{3}-a^{\prime\prime}_{3}a_{3}}{9a_{3}}-(a_{2}-a^{\prime}_{3})\frac{a^{\prime}_{3}}{3a_{3}}\,\big{)}a^{\prime}_{0},$ $\displaystyle I_{2}$ $\displaystyle=\big{\langle}\,\sigma_{2},\,da_{0}^{2}\,\big{\rangle}=(a_{2}-a^{\prime}_{3})(a_{0}^{\prime})^{2},$ $\displaystyle I_{3}$ $\displaystyle=\big{\langle}\,\sigma_{3},\,da_{0}^{3}\,\big{\rangle}=a_{3}(a_{0}^{\prime})^{3}.$ Now take an operator $A\in\mathrm{Diff}_{3}(M,\mathbb{R})$. $\displaystyle A$ $\displaystyle=a_{3}(x,y)\partial_{x}^{3}+a_{2}(x,y)\partial_{x}^{2}+a_{1}(x,y)\partial_{x}+a_{0}(x,y),$ $\displaystyle A_{f}$ $\displaystyle=a_{3}(x,f)\partial_{x}^{3}+a_{2}(x,f)\partial_{x}^{2}+a_{1}(x,f)\partial_{x}+a_{0}(x,f),\quad f\in C^{\infty}(M).$ Then, the corresponding invariants of relates pairs are the following: $\displaystyle I_{0}=$ $\displaystyle a^{0},$ $\displaystyle I_{1}=$ $\displaystyle\Big{(}\,a_{1}-\frac{2(a_{3,x}+a{3,y}f^{\prime})^{2}a_{3}+3(a_{3,x}+a_{3,y}f^{\prime})}{9a_{3}}$ $\displaystyle-\frac{\big{(}a_{3,xx}+2a_{3,xy}f^{\prime}+a_{3,yy}(f^{\prime})^{2}+(a_{3,x}+a_{3,y})f^{\prime\prime}\big{)}a_{3}}{9a_{3}}$ $\displaystyle-\big{(}a_{2}-(a_{3,x}+a_{3,y}f^{\prime})\big{)}\frac{(a_{3,x}+a_{3,y}f^{\prime})}{3a_{3}}\,\Big{)}(a_{0,x}+a_{0,y}f^{\prime}),$ $\displaystyle I_{2}=$ $\displaystyle\big{(}a_{2}-a_{3,x}-a_{3,y}f^{\prime}\big{)}(a_{0,x}+a_{0,y}f^{\prime})^{2},$ $\displaystyle I_{3}=$ $\displaystyle a_{3}(a_{0,x}+a_{0,y}f^{\prime})^{3}.$ From the expression of $I_{3}$, we get $f^{\prime}=\big{(}(\,I_{3}/a_{3}\,)^{1/3}-a_{0,x}\,\big{)}/a_{0,y}.$ Substituting it into $I_{2}$, we get a relation of the form $KI_{2}+K_{1}I_{3}^{1/3}+K_{2}I_{3}^{2/3}+K_{3}I_{3}+K_{0}=0,$ where $\frac{K_{1}}{K},\quad\frac{K_{2}}{K},\quad\frac{K_{3}}{K},\quad\frac{K_{0}}{K}$ are invariants of operators $A\in\mathrm{Diff}_{3}(M,\mathbb{\mathbb{R}})$, i.e. invariants of weakly nonlinear operators. ## 8\. Equivalence of weakly nonlinear operators Let $z$ be natural differential invariant of weakly nonlinear operators and $A\in\mathit{Diff}_{3}(M,\mathbb{R})$. Then the values $z(A,y_{0})=s_{A}^{\ast}(z)$ is function rational in $y$ with coefficients in $C^{\infty}(M)$. Values $z(A,y_{0})$ of this function for a given value $y_{0}$ is a smooth function on $M$. We say that the operator $A$ is in general position if for any point $a\in M$ there are: natural invariants $z_{1},z_{2}$, a value $y_{0}$ of $y$, and a neighborhood $U\subset M$, $a\in U$, such that the mapping $Z_{A,y_{0}}\colon U\mathcal{\longrightarrow}\,\mathbf{D}\subset\mathbb{R}^{2},\quad Z_{A,y_{0}}\colon x\mapsto\big{(}z_{1}(A,y_{0})(x),z_{2}(A,y_{0})(x)\big{)}$ is a local diffeomorphism. We say, that the mapping $Z_{A,y_{0}}$ is a natural chart of the operator $A$ and the functions $z_{1}(A,y_{0}),z_{2}(A,y_{0})$ are natural coordinates on $U$. We call the atlas of these charts $\\{U_{\alpha},\phi_{\alpha}\colon U_{\alpha}\rightarrow\mathbf{D}_{\alpha}\subset\mathbb{R}^{2}\\}$ natural if coordinates $\phi_{\alpha}=\big{(}z_{1}^{\alpha}(A,y_{0}),z_{2}^{\alpha}(A,y_{0})\big{)}$ are given by distinct invariants $\big{(}z_{1}^{\alpha}(A,y_{0}),z_{2}^{\alpha}(A,y_{0})\big{)}\neq\big{(}z_{1}^{\beta}(A,y_{0}),z_{2}^{\beta}(A,y_{0})\big{)}$, when $\alpha\neq\beta$. We denote by $\mathbf{D}_{\alpha\beta}=\phi_{\alpha}(U_{\alpha}\cap U_{\beta})$ and we assume that domains $\mathbf{D_{\alpha}}$ and $\mathbf{D_{\alpha\beta}}$ are connected and simply connected. Let $A_{\alpha}=\phi_{\alpha*}(A|_{U_{\alpha}})$, $A_{\alpha\beta}=\phi_{\alpha*}(A|_{U_{\alpha}\cap U_{\beta}})$ be the images of the operator $A$ in these coordinates. Then $\phi_{\alpha\beta*}(A_{\alpha\beta})=A_{\beta\alpha}$, where $\phi_{\alpha\beta}=\phi_{\beta}\circ\phi_{\alpha}^{-1}\colon\mathbf{D_{\alpha\beta}}\rightarrow\mathbf{D_{\beta\alpha}}$ are the transition mapping. Take now two operators $A,A^{\prime}\in\mathit{Diff}_{3}(M,\mathbb{R})$ and consider their natural charts $Z_{A,y_{0}}$ and $Z_{A^{\prime},y_{0}^{\prime}}$. The $\mathcal{G}(M)$-equivalence of these operators means that the local diffeomorphism $Z_{A^{\prime},y_{0}^{\prime}}^{-1}\circ Z_{A,y_{0}}$ does not depend on choice of $y_{0}$ and $y_{0}^{\prime}$. In this case we will say that these charts are coordinated. Summarizing we get the following result. ###### Theorem 8. Let operators $A,A^{\prime}\in\mathit{Diff}_{3}(M,\mathbb{R})$ be in general position, then they are $\mathcal{G}(M)$-equivalent if and only if the following conditions hold: 1. (1) the mappings $\\{\phi^{\prime}_{\alpha}\colon U^{\prime}_{\alpha}\rightarrow\mathbf{D}_{\alpha}\\}$, where $\phi^{\prime}_{\alpha}=\big{(}z_{1}^{\alpha}(A^{\prime},y^{\prime}_{0}),z_{2}^{\alpha}(A^{\prime},y^{\prime}_{0})\big{)}$ and $U^{\prime}_{\alpha}=(\phi^{\prime}_{\alpha})^{-1}(\mathbf{D}_{\alpha})$, constitute a natural atlas for the operator $A^{\prime}$, 2. (2) the charts $\phi_{\alpha}\in\\{\phi_{\alpha}\colon U_{\alpha}\rightarrow\mathbf{D}_{\alpha}\\}$ and $\phi^{\prime}_{\alpha}\in\\{\phi^{\prime}_{\alpha}\colon U^{\prime}_{\alpha}\rightarrow\mathbf{D}_{\alpha}\\}$ are coordinated, 3. (3) $\phi^{\prime}_{\alpha\beta}=\phi_{\alpha\beta}$, 4. (4) $A_{\alpha}=\phi^{\prime}_{\alpha*}\big{(}A^{\prime}\big{|}_{U^{\prime}_{\alpha}}\big{)}$ and $A_{\alpha\beta}=\phi^{\prime}_{\alpha*}\big{(}A^{\prime}\big{|}_{U^{\prime}_{\alpha\cap\beta}}\big{)}$. ###### Proof. Any diffeomorphism $\psi\colon M\rightarrow M$ such that $\psi_{*}(A)=A^{\prime}$ transform natural atlas to the natural one and because of $\psi^{*-1}\big{(}z(A)\big{)}=z\big{(}\psi_{*}(A)\big{)}$, for any natural invariant $z$, this diffeomorphism has the form of the identity map in the natural coordinates. ∎ ###### Corollary 9. Let operators $A$ and $A^{\prime}\in\mathit{Diff}_{3}(M,\mathbb{R})$ be in general position. Then the weakly nonlinear differential operators $A_{w}$ and $A^{\prime}_{w}$ are $\mathcal{G}(M)$\- equivalent if and only if the operators $A$ and $A^{\prime}$ are $\mathcal{G}(M)$-equivalent. ## References * [1] Cox, D.; Little, J.; O’Shea, D.: Ideals, Varieties, and Algorithms: An Introduction to Computational Algebraic Geometry and Commutative Algebra, Springer, (1997), ISBN 0-387-94680-2. * [2] Kruglikov, Boris, Lychagin, Valentin, Global Lie-Tresse theorem, Selecta Math. (N.S.) 22 (2016), no. 3, 1357–1411. * [3] Lychagin, V.V., Yumaguzhin, V.A., On equivalence of third order linear differential operators on two-dimensional manifolds, Journal of geometry and physics, Vol. 146, December 2019, 103507. * [4] Valentin Lychagin, Valeriy Yumaguzhin, On structure of linear differential operators, acting in line bundles, Journal of geometry and physics, Volume 148, February 2020, 103549. * [5] Valentin Lychagin, Valeriy Yumaguzhin, Natural differential invariants and equivalence of nonlinear second order differential operators, Journal of geometry and physics, 178(2022)104549. * [6] Peter J. Olver, Applications of Lie groups to differential equations, Springer-Verlag * [7] Wagner,V.V., Two dimensional space with cubic metric, Scientific notes of Saratov State University, Vol. 1(XIV), Ser. FMI, No. 1, 1938. (in Russian). * [8] Zariski O., Samuel P.:Commutative Algebra, vol. 1, 2, Van Nostrand, 1960.
# Universal scaling regimes in rotating fluid turbulence Abhik Basu abhik.basu<EMAIL_ADDRESS>Theory Division, Saha Institute of Nuclear Physics, Calcutta 700064, West Bengal, India Jayanta K Bhattacharjee<EMAIL_ADDRESS>Department of Theoretical Physics, Indian Association for the Cultivation of Science, 2A and 2B Raja S C Mullick Road, Calcutta 700032, West Bengal, India ###### Abstract We analyse the scaling properties of the energy spectra in fully developed incompressible turbulence in forced, rotating fluids in three dimensions (3D), which are believed to be characterised by universal scaling exponents in the inertial range. To elucidate the scaling regimes, we set up a scaling analysis of the 3D Navier-Stokes equation for a rotating fluid that is driven by large- scale external forces. We use scaling arguments to extract the scaling exponents, which characterise the different scaling regimes of the energy spectra. We speculate on the intriguing possibility of two-dimensionalisation of 3D rotating turbulence within our scaling theory. Our results can be tested in large scale simulations and relevant laboratory-based experiments. ## I Introduction Nonequilibrium systems are described by the appropriate equations of motion for the relevant dynamical variables and exhibit much richer universal behavior than usually observed in equilibrium critical dynamics hal . Hydrodynamic turbulence in fluids, described by the Navier-Stokes equation land ; frish for the evolution of the velocity field $\bf v$, is a prime example of an out of equilibrium system, due to the external drive acting on the fluids. Interestingly, fully developed fluid turbulence in three- (3D) and two- (2D) dimensions show markedly different behavior: In 3D, the energy spectra follow the well-known K41 result for homogeneous and isotropic 3D hydrodynamics turbulence where the one-dimensional energy spectra $E(k)\sim k^{-5/3}$ (hereafter K41) in the inertial range, where $k$ is a wavevector k41 . This K41 result is quite robust and universal, and found in wide-ranging natural systems, e.g., shear flows shear , viscoelastic fluids visco and jet flows jet . In contrast, 2D turbulence is characterised by an inverse cascade of energy at very large scales with $E(k)\sim k^{-5/3}$, and forward cascade of enstrophy with $E(k)\sim k^{-3}$ at intermediate scales 2d1 ; 2d2 ; 2d3 ; 2d4 ; 2d5 . Rotating turbulence, i.e., turbulence in a rotating fluid, is a naturally occurring phenomenon in many astrophysical and geophysical flows, as well as in laboratory-based engineering fluid flows. The presence of the Coriolis forces is the distinctive feature of rotating turbulence, which should affect the large-scale scaling properties of rotating turbulence. In spite of extensive studies, there is still no good agreement on the scaling of the energy spectra in rotating turbulence, in particular how the Coriolis forces affect the scaling of the spectra at very large scales. Varieties of analytical, numerical, or experimental investigations of either the forced or the decaying rotating turbulent fluid suggest that the kinetic energy spectra in the rotation-dominated small-$k$ regions should scale as $E(k)\sim k^{-m},\,m\in(2,3)$ 2ds1 ; 2ds2 ; 2ds3 ; 2ds4 ; 2ds5 ; 2ds6 ; 2ds7 ; 2ds8 ; 2ds9 ; 2ds10 ; 2ds11 ; 2ds12 ; 2ds13 ; 2ds14 ; 2ds15 . Recent perturbative studies indicate that the one-dimensional kinetic energy spectra made out of the velocity component parallel to the rotation axis scales as $k^{-5/3}$, indistinguishable from the K41 prediction. In contrast, the one-dimensional kinetic energy spectra made out of the velocity components lying in a plane normal to the rotation axis scales as $k^{-3}$, different from the K41 scaling ab-jkb-rot . The precise forms of the scaling of the energy spectra in rotating turbulence however is still not well-settled. There is a degree of formal similarity between the (linearised) equations of motion of rotating turbulence, which is nothing but the Navier-Stokes equation in a rotating frame (see below) and the equations for magnetohydrodynamic turbulence (MHD) in the presence of a mean magnetic field $B_{0}$ jackson ; arnab . In the former case, the Coriolis forces lead to oscillatory modes, whereas in the MHD case, a non-zero $B_{0}$ gives rise to propagating Alfv́en waves. Strong Alfv́en waves are known to make the energy spectra in MHD anisotropic, and change the scaling as well jkb-mhd . In the same vein, strong Coriolis forces should make the scaling of energy spectra in rotating turbulence anisotropic and also different from its isotropic counterpart (i.e., the K41 scaling). In this work, we revisit the universal scaling of energy spectra in forced, statistically steady rotating turbulence in its inertial range. To this end, we have set up a scaling theory to study the scaling of the energy spectra in the inertial range. We cover both the weak and strong rotation limits. In the former case, unsurprisingly, the K41 result is found. With stronger rotation, anisotropic scaling with different exponents ensues. In particular, in the wavevector region $k_{\perp}\gg k_{\parallel}$, our scaling theory gives the scaling of the 2D spectra $E(k_{\perp},k_{\parallel})$, where ${\bf k}_{\perp}$ and $k_{\parallel}$ are the components of the wavevector $\bf k$ in the plane perpendicular to the rotation axis (here $\hat{z}$-axis) and along the rotation axis, respectively. We find $E(k_{\perp},k_{\parallel})\sim k_{\perp}^{-5/2}k_{\parallel}^{-1/2}$ for $k_{\perp}\gg k_{\parallel}$, which agrees with the Kuznetsov–Zakharov–Kolmogorov spectra predicted by the weak inertial-wave turbulence theory for the rotating fluids 2ds7 . We show that this result is unaffected by nonlinear fluctuation corrections at the one-loop order. We further demonstrate that this result could be obtained by demanding that the cascade of the kinetic energy flux is hindered by a non-zero helicity, which is naturally present in a rotating fluid. In the opposite limit of $k_{\perp}\ll k_{\parallel}$, we get $E(k_{\perp},k_{\parallel})\sim k_{\perp}^{-1}k_{\parallel}^{-2}$. We also show perturbatively in the rotation $\Omega$ that the kinetic energy flux is indeed reduced by it. The remainder of the article is organised as follows. In Sec. II, we set up the forced Navier-Stokes equation in a rotating fluid. Then in Sec. III we set up the scaling arguments. Then next in Sec. III.1, we revisit the K41 scaling scaling in an isotropic, nonrotating fluid turbulence, and show how are scaling theory reproduces it. Next, in Sec. III.2, we show that for weak rotation, the energy spectra again show the K41 scaling. Then in Sec. III.3 we study the 2D anisotropic energy spectra in the opposite limit of large $\Omega$. In Sec. IV we discuss and summarise our results. We provide some technical results, including a perturbative demostration of the reduction of the kinetic energy flux by helicity, in the Appendix for interested readers. ## II Turbulence in a rotating fluid The Navier-Stokes equation for the velocity field ${\bf v}({\bf r},t)$ in a rotating frame with rotation ${\boldsymbol{\Omega}}=\omega\hat{z}$ is given by $\frac{\partial\bf v}{\partial t}+2({\boldsymbol{\Omega}}\times{\bf v})+\lambda({\bf v}\cdot{\boldsymbol{\nabla}}){\bf v}=-\frac{{\boldsymbol{\nabla}}p^{*}}{\rho}+\nu\nabla^{2}{\bf v}+{\bf f},$ (1) where $p^{*}=p+\frac{1}{2}|{\boldsymbol{\Omega}}\times{\bf v}|^{2}$ is the effective pressure. We assume ${\boldsymbol{\Omega}}=\Omega\hat{z}$, i.e., the rotation is about the $z$-axis; see Fig. 1. Figure 1: Geometry of the rotating fluid. We assume the rotation to be about the $z$-axis. In this coordinate system, ${\bf k}_{\perp}=(k_{x},k_{y})$ and $k_{\parallel}=k_{z}$. In this case (1) may be written in terms of components as $\displaystyle\frac{\partial v_{z}}{\partial t}+\lambda({\bf v}\cdot{\boldsymbol{\nabla}})v_{z}=-\frac{\partial_{z}p^{*}}{\rho}+\nu\nabla^{2}v_{z}+f_{z},$ (2) $\displaystyle\frac{\partial v_{x}}{\partial t}-2\Omega v_{y}+\lambda({\bf v}\cdot{\boldsymbol{\nabla}})v_{x}=-\frac{\partial_{x}p^{*}}{\rho}+\nu\nabla^{2}v_{x}+f_{x},$ (3) $\displaystyle\frac{\partial v_{y}}{\partial t}+2\Omega v_{x}+\lambda({\bf v}\cdot{\boldsymbol{\nabla}})v_{y}=-\frac{\partial_{y}p^{*}}{\rho}+\nu\nabla^{2}v_{y}+f_{y}.$ (4) Here, $\lambda=1$. These equations in 3D admit two conserved quantities in the inviscid limit: (i) kinetic energy $E=\int d^{3}x\rho\,v^{2}/2$ and (ii) helicity $H=\int d^{3}x\,{\bf v}\cdot{\boldsymbol{\nabla}}\times{\bf v}={\bf v}\cdot{\boldsymbol{\omega}}$, where $\boldsymbol{\omega}\equiv{\boldsymbol{\nabla}}\times{\bf v}$ is the local vorticity. In the viscous steady states, in a Kolmogorov-like picture neglecting intermittency, $E$ and $H$ should have constant (i.e., scale independent) fluxes. Clearly, $E/H$ has the dimension of a length, which allows us to define a length-scale $l^{*}=E/H$. We assume the external forces to be non-helical, i.e., no helicity injection by the forces. Thus helicity is generated in the bulk only by the global rotation. We consider the incompressible limit, i.e., the mass density $\rho=const.$, or equivalently ${\boldsymbol{\nabla}}\cdot{\bf v}=0$. At this stage, it is useful to set up the notations. Below we use $\tilde{\omega}$ and $\tilde{\Omega}$ to denote Fourier frequencies, while $\Omega$ and $\boldsymbol{\omega}$ represent the global rotation frequency and vorticity, respectively. ## III Scaling analysis To classify the scaling regimes, we first define the following dimensionless numbers (i) Rossby number $R_{o}=U/(2\Omega L)$, (ii) Reynolds number $R_{e}=\frac{LU}{\nu}$ and (iii) Ekman number $Ek=R_{o}/R_{e}=\nu/(2\Omega L^{2})$, where $L$ is the linear system size, and $U$ is a typical velocity. We expect to find two distinct scaling regimes as characterised by $R_{o}$ (or $\Omega$): (i) Weak rotation $\Omega\rightarrow 0$, or $R_{o}\rightarrow\infty$ (iii) Large rotation $\Omega\rightarrow\infty$, or $R_{o}\rightarrow 0$. Since the rotation picks up a direction (the axis of rotation, here the $z$-axis), system is generally anisotropic. We therefore construct an anisotropic scaling theory of the system: we assume scaling under the transformations ${\bf r}_{\perp}\rightarrow l_{\perp}{\bf r}_{\perp},\,z\rightarrow l_{\perp}^{\xi}z,\,{\bf v}_{\perp}\rightarrow l_{\perp}^{a_{\perp}}{\bf v}_{\perp},\,v_{z}\rightarrow l_{\perp}^{a_{z}}v_{z}.$ (5) Here, ${\bf r}_{\perp}\equiv(x,\,y)$, ${\bf v}_{\perp}\equiv(v_{x},\,v_{y})$. In a general anisotropic situation, $\xi\neq 1$. We also allow for the possibility $a_{\perp}\neq a_{z}$, i.e., ${\bf v}_{\perp}$ and $v_{z}$ may not scale in the same way under spatial rescaling. We further define time-scale $t$ to scale as $t\sim l_{\perp}^{\tilde{z}},$ (6) where $\tilde{z}$ is a dynamic exponent. Furthermore, we define a phenomenological dimensionless constant $\tilde{\Omega}$ by $\tilde{\Omega}=\frac{[{\boldsymbol{\Omega}}]}{[{\boldsymbol{\omega}}]},$ (7) Here, […] implies “in a dimensional sense” jkb-mhd . Clearly, the two limiting cases $\tilde{\Omega}\rightarrow 0$ and $\tilde{\Omega}\rightarrow\infty$ phenomenologically correspond to $R_{o}\rightarrow\infty$ and $R_{o}\rightarrow 0$. We note that by balancing the Coriolis force terms against the advective nonlinear terms, we can extract a length-scale $L_{o}$. In a scaling sense, we set $\Omega v\sim\frac{v^{2}}{L_{0}},$ (8) giving $v\sim\Omega L_{0}$. Dimensionally speaking, energy dissipation $\epsilon\sim\frac{v^{3}}{L_{0}}\sim\Omega^{3}L_{0}^{2}.$ (9) This gives $L_{0}\sim\sqrt{\frac{\epsilon}{\Omega^{3}}}.$ (10) The corresponding wavevector $k_{0}\equiv 2\pi/L_{0}$ is the Zeeman wavevector. For length-scales $L\gg L_{0}$ (or equivalently, for wavevector $k\ll k_{0}$), we expect Coriolis force terms to be important, effects of the rotation should be strong, and hence non-K41 spectra should follow. In the opposite limit of $L\ll L_{0}$ (or $k\gg k_{0}$), Coriolis forces should be irrelevant, and hence K41 scaling should follow. Thus a dual scaling is believed to exist 2ds12 ; dual1 ; dual2 ; dual3 . In terms of the dimensionless numbers, we are interested in $R_{e}\rightarrow\infty$ for fully developed turbulence. Together with $R_{e}\rightarrow\infty$ (implying fully developed turbulence) and $Ek\rightarrow 0$ (implying Coriolis forces dominating over the viscous damping at large scales), we can have two situations: (i) $R_{o}\rightarrow\infty$ for weak rotation, and (ii) $R_{o}\rightarrow 0$ for strong rotation. Lastly, one has the dissipation scale $\eta_{d}$, such that for length scales smaller than $\eta_{d}$, the dissipation range ensues. Then in terms of the length scales defined above, we can have the following scenarios. In a sufficiently large system, there should be adequate scale separations, such that $\eta_{d}\ll L_{0}$, i.e., $L_{0}$ should belong to the inertial range. This should allow for both the scaling regimes, viz. K41 and non-Kolmogorov scaling regimes to be observed. ### III.1 Nonrotating isotropic case For a nonrotating, isotropic fluid, $\Omega=0$ in (3)-(4) gives the usual 3D isotropic Navier-Stokes equation. Let us briefly revisit the extraction of the K41 scaling by applying the scaling arguments on the usual 3D Navier-Stokes equation first. Due to the isotropy of the system, we expect $\chi=1$ strictly, and make no distinction between $l_{\perp}$ and $l_{\parallel}$, the rescaling factors of ${\bf r}_{\perp}$ and $z$ respectively: $l_{\perp}\sim l_{\parallel}\sim l$. Demanding scale invariance jkb-mhd , we find $\frac{\partial{\bf v}}{\partial t}\sim{\bf v}\cdot{\boldsymbol{\nabla}}{\bf v}\implies l^{a-\tilde{z}}=l^{2a-1}\implies a=1-\tilde{z}.$ (11) Next, in a mean-field like approach, we assume the kinetic energy flux or the kinetic energy dissipation per unit mass is scale invariant in the inertial range. This gives $\frac{\partial v^{2}}{\partial t}\sim l^{0}\implies 2a=\tilde{z}.$ (12) Combining then, we get $a=1/3,\,\tilde{z}=2/3$. This corresponds to a 1D kinetic energy spectra $E(k)\sim k^{-5/3}$, the expected K41 result. ### III.2 Weak rotation effects: $k\gg 2\pi/L_{0}$ In order to study the effects of weak rotation on the scaling of the energy spectra, we consider the limit $R_{o}\rightarrow\infty$, or $\Omega\rightarrow 0$. Equivalently, we consider length scales $L\ll L_{0}$, with the understanding that $L\gg\eta_{r}$, the dissipation scale. In this case, the Coriolis force is unimportant. Hence $\langle H\rangle\approx 0$, where $<...>$ implies averages over the statistical steady states. Thus, the flux of $E$ is the relevant (in the Kolmogorov sense) flux. Then, proceeding as in Ref. jkb-mhd , we unsurprisingly recover the K41 scaling: $a_{\perp}=a_{z}=1/3,\,\tilde{z}=2/3,\,\xi=1.$ (13) The last of the above naturally means isotropic scaling (although geometry remains anisotropic). Furthermore, if we let $R_{o}\sim l^{\eta}$ and demand scale invariance of all the terms (including the Coriolis force terms), we find $\eta=\tilde{z}=2/3.$ (14) Thus $R_{o}\rightarrow\infty$ as $l\rightarrow\infty$, which corresponds to nonrotating and isotropic fully developed turbulence. We have assumed that the nonlinear coupling constant does not scale under spatial rescaling, which is consistent with the nonrenormalisation of $\lambda$ due to the Galilean invariance of the Navier-Stokes equation. The scaling of the viscosity is controlled by the dynamic exponent $\tilde{z}$. ### III.3 Strong rotation effects: $k\ll 2\pi/L_{0}$ We next consider the large rotation case, i.e., when $R_{o}\geq{\cal O}(1)$, or $\Omega\geq{\cal O}(1)$. It is now expected that anisotropy is significant. Below we analyse the scaling in several equivalent ways. Balancing different terms of (3) and (4) we find $\tilde{z}=0,\,a_{\perp}=1,\,a_{z}=\xi.$ (15) To proceed further, we allow for the possibility that not only the spatial scaling may be anisotropic, there may be different dynamic exponents for ${\bf v}_{\perp}$ and $v_{z}$, with $\tilde{z}$ being identified as the dynamic exponent $\tilde{z}_{\perp}$ of ${\bf v}_{\perp}$. To study this, we separately consider the contribution to the kinetic energy from the in-plane velocity ${\bf v}_{\perp}$ and normal component of the velocity $v_{z}$. Interestingly, the exponents in (15) mean that the flux of the “in-plane kinetic energy” $E_{\perp}\equiv\int d^{3}x\,v_{\perp}^{2}$ cannot be scale- independent! Let us now consider the kinetic energy $E_{z}\equiv\int d^{3}x\,v_{z}^{2}$ flux of $v_{z}$. If we assume $\tilde{z}=0$ is the dynamic exponent of $v_{z}$ also, then $a_{z}=0=\xi$ can actually keep the flux of $E_{z}$ scale-independent. However, $a_{z}=0$ is unexpected, as it means $v_{z}$ does not scale with $l$ at all. Assume the dynamics $z_{\parallel}$ of $v_{z}$ be non-zero: $z_{\parallel}>0$. Now consider Eq. (2) and balance $\frac{\partial v_{z}}{\partial t}\sim v_{z}\partial_{z}v_{z}\implies a_{z}=1-z_{\parallel}.$ (16) Next, demanding scale-invariance of the flux of $E_{z}$ gives $2a_{z}=\tilde{z}_{\parallel}\implies a_{z}=\frac{1}{3},\,\tilde{z}_{\parallel}=\frac{2}{3}.$ (17) This further means $\xi=a_{z}=1/3$. Notice that with $a_{\perp}=1,\,\xi=1/3$, $({\bf v}_{\perp}\cdot{\boldsymbol{\nabla}}_{\perp})v_{z}$ and $v_{z}\partial_{z}v_{z}$ scale the same way. Since we get $a_{\perp}>a_{z}$, we should have $\langle v_{\perp}^{2}\rangle\gg\langle v_{z}^{2}\rangle$ in the long wavelength limit, suggesting concentration of the kinetic energy in a plane normal to the rotation axis gode . This implies an effective two- dimensionalisation. On the other hand, $\tilde{z}<\tilde{z}_{\parallel}$ implies ${\bf v}_{\perp}\ll v_{z}$ in the long time limit, a conclusion contradictory to our above inference. In fact, this alternative scenario implies a type of dimensional reduction, where most of the energy is confined to the $z$-direction. We are unable to conclusively predict which of these two scenarios actually holds. Numerical studies should be useful in this regard. Is the ensuing flow field in the limit $R_{o}\rightarrow\infty$ truly 2D? In our opinion, the answer is no. First of all, the flow remains overall 3D incompressible. This means the effective 2D flow field might be 2D compressible, which is an interesting possibility. Secondly, it is not whether the direction of the kinetic energy cascade becomes backward, a hallmark of pure 2D turbulence. Thirdly, enstrophy is a conserved quantity in the inviscid limit of pure 2D turbulence, whereas it is not expected to be so in the 3D rotating case even in the limit of high rotation. Therefore, notwithstanding the dominance of $v_{\perp}$ over $v_{z}$, the resulting flow field should be fundamentally different from pure 2D nonrotating turbulence. Lastly, since $z_{\perp}\neq z_{\parallel}$, we find weak dynamic scaling jkb-mhd ; dib . We now calculate the scaling of the two-dimensional kinetic energy spectra $E_{\perp}(k_{\perp},k_{\parallel})$ and $E_{z}(k_{\perp},k_{z})$, such that total kinetic energy $E_{\text{tot}}=\int dk_{\perp}dk_{z}[E_{\perp}(k_{\perp},k_{z})+E_{z}(k_{\perp},k_{z})]$. The scaling of $E_{\perp}(k_{\perp},k_{z})$ and $E_{z}(k_{\perp},k_{z})$ can be obtained as follows. We use the general definition to write in a dimensional/scaling sense $\displaystyle v_{m}({\bf k},\tilde{\omega})\sim\int v_{m}({\bf x},t)\exp(i{\bf k_{\perp}}\cdot{\bf x_{\perp}})$ (18) $\displaystyle\times\exp(ik_{\parallel}z)\exp(i\tilde{\omega}t)d^{2}x_{\perp}dzdt$ $\displaystyle\sim$ $\displaystyle l_{\perp}^{a_{m}}l_{\perp}^{2}l_{\parallel}\sim\frac{1}{k_{\perp}^{a_{m}+2}}\frac{1}{k_{\parallel}}.$ Here, $m=\perp,\,\parallel$. Furthermore, $\displaystyle\langle{\bf v}_{\perp}({\bf k}_{1})\cdot{\bf v}_{\perp}({\bf k_{2}})\rangle=F_{\perp}(k_{1})\delta({\bf k}_{1}+{\bf k}_{2}),$ (19) $\displaystyle\langle{v}_{z}({\bf k}_{1}){v}_{z}({\bf k_{2}})\rangle=F_{\parallel}(k_{1})\delta({\bf k}_{1}+{\bf k}_{2}).$ (20) Dimensionally then, $F_{\perp}\sim\frac{1}{k_{\perp}^{a_{\perp}+2}k_{\parallel}},\,F_{\parallel}\sim\frac{1}{k_{\perp}^{a_{\parallel}+2}k_{\parallel}}.$ (21) This gives for the two-dimensional energy spectra $E_{\perp}(k_{\perp},k_{\parallel})\sim k_{\perp}^{-3},\,E_{z}(k_{\perp},k_{z})\sim k_{\perp}^{-5/3}.$ (22) If we ignore anisotropy, we can define two corresponding one dimensional energy spectra $E_{\perp}(k)$ and $E_{z}(k)$ from (22) by $E_{\text{tot}}=\int dk[E_{\perp}(k)+E_{z}(k)]$. Notice that, neglecting anisotropy, the one- dimensional spectra corresponding to $E_{\perp}(k)\sim k_{\perp}^{-3}$ should scale as $k^{-2}$ as argued above, in agreement with Refs. dual1 ; dual2 ; dual3 . Nonetheless, in spite of this agreement, we notice that our results (22) appear to suggest that $E_{\perp}(k_{\perp},k_{z})$ and $E_{z}(k_{\perp},k_{z})$ have no $k_{z}$-dependence, which should be unphysical. We try to rectify this below. First of all, for large $\Omega$, the scaling should be dominated by the Coriolis forces. The vorticity ${\boldsymbol{\omega}}$ satisfies 2ds7 $\partial_{t}{\boldsymbol{\omega}}({\bf k})=-2\Omega\frac{k_{\parallel}\hat{e}_{\bf k}\times{\boldsymbol{\omega}}_{\bf k}}{k},$ (23) giving time-scale $\tau\sim k/k_{\parallel}\sim k_{\perp}/k_{\parallel}$ for $k_{\perp}\gg k_{\parallel}$. It is thus reasonable to assume that $\tau$ as defined above is the relevant time-scale when $R_{o}\rightarrow 0$. In what follows below, we do not make any distinction between $v_{\perp}$ and $v_{z}$. We now impose the scale- independence of the kinetic energy flux $\Pi$. The energy flux may be calculated from the Navier-Stokes equations (1). We find $\displaystyle\Pi$ $\displaystyle=$ $\displaystyle-2\lambda^{2}\int_{{\bf k,q},\tilde{\omega},\tilde{\Omega}}\bigg{[}M_{imn}({\bf k})M_{ijp}(-{\bf k})\langle v_{m}({-k+q},-\tilde{\omega}+\tilde{\Omega})v_{p}({\bf k-q},\tilde{\omega}-\tilde{\Omega})\rangle$ (24) $\displaystyle+$ $\displaystyle M_{jmn}({\bf k})M_{ijp}({\bf q})\langle v_{i}({-k},-\tilde{\omega})v_{n}({\bf k},\tilde{\omega})\rangle\langle v_{m}({\bf-k+q},-\tilde{\omega}+\tilde{\Omega})v_{p}({\bf k-q},\tilde{\omega}-\tilde{\Omega})\rangle$ (25) $\displaystyle+$ $\displaystyle M_{jmn}({\bf k-q})M_{ijp}({\bf k})\langle v_{i}({-\bf k},-\tilde{\omega})v_{n}({\bf k},\tilde{\omega})\rangle\langle v_{j}({\bf q},\tilde{\Omega})v_{m}({\bf-q},-\tilde{\Omega})\rangle\bigg{]}.$ (26) Here, $M_{ijp}({\bf k})=P_{ij}({\bf k})k_{p}+P_{ip}({\bf k})k_{j}$. Since we are interested in the scaling, it suffices to consider the scaling of $\Pi$, suppressing indices and wavevector labels. At this one-loop order, suppressing indices and wavevector labels, and assuming $k_{\perp}\gg k_{\parallel}$ $\Pi\sim\int d^{2}k_{\perp}d^{2}p_{\perp}dk_{\parallel}dp_{\parallel}\int dt\int d^{2}k_{\perp}\,k_{\perp}^{2}C^{2}\sim const.,$ (27) where $C\equiv\langle v\,v\rangle$ is the correlation function, again suppressing indices and wavevector labels. This gives $C\sim k_{\perp}^{-7/2}k_{\parallel}^{-1/2},$ (28) where we have used $\int dt\sim\tau$. This implies for the two-dimensional kinetic energy spectra $E_{\perp}(k_{\perp},k_{\parallel})\sim Ck_{\perp}\sim k_{\perp}^{-5/2}k_{\parallel}^{-1/2},$ (29) as in the wave turbulence theory. It now behooves us to show that the scaling of $\tau$ with wavevector does not get renormalised at the one-loop order. We restrict ourselves here to a scaling-level demonstration. As shown in Appendix, the one-loop self-energy $\Sigma_{ik}({\bf k},\tilde{\omega})$ has the form $\displaystyle\Sigma_{ij}({\bf k},\tilde{\omega})\sim\int d^{3}qd\tilde{\Omega}M_{imn}({\bf k})\langle v_{m}({\bf q},\tilde{\Omega})v_{r}({-\bf q},-\tilde{\Omega})\rangle$ (30) $\displaystyle\times$ $\displaystyle G_{ns}({\bf k-q},\tilde{\omega}-\tilde{\Omega})M_{srj}({\bf k-q}).$ Here, $G_{ns}({\bf k},\tilde{\omega})$ is the propagator defined via $G_{ns}({\bf k},\tilde{\omega})\equiv\bigg{\langle}\frac{\delta v_{n}({\bf k},\tilde{\omega})}{\delta f_{s}({\bf k},\tilde{\omega})}\bigg{\rangle}.$ (31) Considering the one-loop self-energy $\Sigma$ and suppressing indices and wavevector labels [see also Appendix], we obtain $\Sigma\sim k_{\perp}^{2}\int d^{2}q_{\perp}dq_{z}\int dtG\,C,$ (32) where $G$ is a propagator. Assuming the dominant time-scale in the above time- integral is given by $\tau$, we get (in a scaling sense) $k_{\perp}^{2}\int d^{2}q_{\perp}dq_{z}\frac{q_{\perp}}{q_{z}}q_{\perp}^{-7/2}k_{z}^{-1/2}\sim[k_{\perp}]^{3/2}[k_{\parallel}]^{1/2},$ (33) which is less singular than the the bare form of $\tau$. Thus our scaling results on the 2D kinetic energy spectra remain unaffected by the advective nonlinearities in the asymptotic long wavelength limit. Interestingly, we can also derive the above results by using phenomenological arguments of suppression of the kinetic energy flux by the helicity generated by the rotation, which are similar to the arguments set up in Ref. jkb-mhd for scaling of the energy spectra in the presence of a strong mean magnetic field in magnetohydrodynamic turbulence. It is known that the predominant role of a (large) non-zero helicity flux is to hinder the cascade of the kinetic energy flux kraich . In fact, it is easy to see from (3) and (4) that a large $\Omega$ should suppress the nonlinear effects, relative to the Coriolis force terms. Since the nonlinear terms are responsible for the cascade phenomena, we expect the flux of $E_{\perp}$ to be suppressed by a large $\Omega$. This is similar to the suppression of the energy flux by a strong mean magnetic field in fully developed magnetohydrodynamic turbulence jkb-mhd ; see also similar treatment for turbulence in a stably stratified fluid ab-jkb-bolgiano . We write $\bigg{[}\frac{\partial v_{\perp}^{2}}{\partial t}\bigg{]}\bigg{[}\frac{\omega^{2}}{\Omega^{2}}\bigg{]}\sim l^{0}$ (34) as the condition of the flux being scale independent. Since dimensionally, $[\omega^{2}]\sim[v_{\perp}^{2}/l^{2}]$, we get $v_{\perp}\sim l_{\perp}^{3/4}l_{\parallel}^{-1/4}.$ (35) This gives $E_{\perp}(k_{\perp},k_{\parallel})\sim k_{\perp}^{-5/2}k_{\parallel}^{-1/2},$ (36) in agreement with the conclusion from the wave turbulence theory approaches 2ds7 . It is easy to get the spectra in the opposite limit $k_{\parallel}\gg k_{\perp}$. In this limit, $\tau\sim k_{\perp}^{0}k_{\parallel}^{0}$. At this one-loop order, suppressing indices and wavevector labels, and assuming $k_{\perp}\ll k_{z}$. $\Pi\sim\int d^{2}k_{\perp}d^{2}p_{\perp}dk_{z}dp_{\parallel}\int d^{2}k_{\perp}\,k_{z}^{2}C^{2}\sim const.,$ (37) Proceeding as before, we find $E_{\perp}(k_{\perp},k_{\parallel})\sim k_{\perp}^{-1}k_{\parallel}^{-2}.$ (38) In each of these cases, the corresponding one-dimensional spectra, without making any distinction between $k_{\perp}$ and $k_{\parallel}$ scale as $k^{-2}$, consistent with the recent shell-model studies on rotating turbulence new-shell . A pictorial summary of the scaling regimes are shown in Fig. 2. Figure 2: Schematic diagram illustrating the scaling regimes at different length scales. Dissipation range for small scales, K41 scaling regime at the intermediate scales and rotation dominated scaling regimes at the largest scales are shown. In the above, we have implicitly assumed that the kinetic energy flux $E$ is the relevant flux (in the Kolmogorov sense). As we have discussed above this holds for length scales $\gg l^{*}$. An interesting case may arise if $l^{*}\gg L_{0}$, in which case in the window between $L_{0}$ and $l^{*}$, helicity flux $H$ dominates over $E$. If this indeed holds, the scaling of the energy spectra might change within this window. We do not discuss this further here. ## IV Summary and outlook In this work we have developed a scaling theory for fully developed incompressible hydrodynamic turbulence in a rotating fluid in the inertial range. We have studied the scaling of the energy spectra in the inertial range for weak and strong rotations, i.e., for small and large Rossby number $R_{o}$. We argue that for wavevectors smaller than the Zeeman wavevector rotation is important, whereas in the opposite limit, rotation is unimportant. It is therefore expected that in the former regime the scaling may be different from the K41 scaling, but in the other regime K41 scaling should ensue. The scaling theory that we developed here bears this out. Our scaling theory reveals that in the rotation dominated regime, not only the scaling itself is anisotropic (i.e., different dependence on $k_{\perp}$ and $k_{\parallel}$), the scaling of $v_{\perp}$ and $v_{z}$ are different. Even the dynamic exponents of $v_{\perp}$ and $v_{z}$ are different, indicating weak dynamic scaling. This suggests that in the limit of a large rotation, the flow fields are dominated by only some of the velocity components. However, our theory cannot conclusively predict whether $v_{\perp}$ or $v_{z}$ will be the dominant part. Numerical simulations should be useful to make further conclusions. We have throughout assumed that the kinetic energy flux is the relevant flux (in the Kolmogorov sense), neglecting the helicity flux. However, for sufficiently large rotation, there may be a window of length scales where the helicity flux is the dominant flux. In this regime, the scaling of the energy spectra should be different. This will be discussed elsewhere. ## V Acknowledgement AB thanks the SERB, DST (India) for partial financial support through the MATRICS scheme [file no.: MTR/2020/000406]. ## Appendix A Reduction of the kinetic energy flux in rotating turbulence: perturbation theory We now calculate the kinetic energy flux $\Pi$ in the perturbation theory. We derive (26) and calculate $\Pi$ to lowest non-trivial order in $\Omega$. We start by eliminating pressure from the Navier-Stokes equation (1) in a rotating frame. We obtain $\displaystyle(-i\tilde{\omega}+\nu k^{2})v_{i}+2P_{im}({\bf k})\Omega\epsilon_{mzp}v_{p}({\bf k},\tilde{\omega})$ $\displaystyle+i\frac{\lambda}{2}M_{ijp}({\bf k})\sum_{{\bf q},\tilde{\omega}}v_{j}({\bf q},\tilde{\omega})v_{p}({\bf k-q},\,\tilde{\omega}-\tilde{\Omega})=f_{i}.$ (39) To the lowest order in $\Omega$, we expect $\Pi$ to depend on $\Omega$ quadratically, since the energy cascade should be independent of the sense of rotation around the $z$-axis, i.e., should be the same for clockwise and anticlockwise rotations. To this order in $\Omega$, it suffices to expand (39) to ${\cal O}(\Omega)$ and construct an effective equation: $\displaystyle(-i\tilde{\omega}+\nu k^{2})v_{i}+\frac{i\lambda}{2}M_{ijp}({\bf k})\sum_{{\bf q},\tilde{\omega}}v_{j}({\bf q},\tilde{\Omega})v_{p}({\bf k-q},\tilde{\omega}-\tilde{\Omega})$ (40) $\displaystyle=$ $\displaystyle f_{i}-2P_{im}({\bf k})\epsilon_{mzp}\Omega\frac{f_{p}}{-i\tilde{\omega}+\nu k^{2}}.$ Clearly, the last term on the rhs of (40), an effective noise, is the dominant noise in the hydrodynamic limit. We use (40) to calculate the kinetic energy flux $\Pi$ to ${\cal O}(\Omega^{2})$. The flux $\Pi$ follows $\displaystyle\frac{d}{dt}\langle v^{2}\rangle$ $\displaystyle=$ $\displaystyle-\frac{i\lambda}{2}\int_{{\bf k},\tilde{\omega}}\langle\bigg{[}M_{ijp}({\bf k})\sum_{{\bf q},\tilde{\omega}}v_{i}({\bf-k},-\tilde{\omega})v_{j}({\bf q},\tilde{\Omega})v_{p}({\bf k-q},\tilde{\omega}-\tilde{\Omega})$ (41) $\displaystyle+$ $\displaystyle M_{ijp}({\bf-k})\sum_{{\bf q},\tilde{\omega}}v_{i}({\bf k},\tilde{\omega})v_{j}({\bf q},\tilde{\Omega})v_{p}({\bf- k-q},-\tilde{\omega}-\tilde{\Omega})\bigg{]}\rangle.$ Next we iterate and expand the rhs of (41) up to the one-loop order, which gives (26) above. Now, to the linear order in $\Omega$, $\displaystyle\langle v_{j}({\bf q},\tilde{\Omega})v_{n}({\bf-q},-\tilde{\Omega})\rangle$ $\displaystyle=$ $\displaystyle\frac{\langle f_{j}({\bf q},\tilde{\Omega})f_{n}({-\bf q},-\tilde{\Omega})\rangle}{\tilde{\Omega}^{2}+\nu^{2}q^{4}}-2P_{js}({\bf q})\epsilon_{szp}G_{0}^{2}(q,\tilde{\Omega})\Omega\frac{\langle f_{p}({\bf q},\tilde{\Omega})f_{n}({\bf-q},-\tilde{\Omega})\rangle}{i\tilde{\Omega}+\nu q^{2}}$ (42) $\displaystyle-$ $\displaystyle 2P_{ns}({\bf q})\epsilon_{szp}G_{0}^{2}(-q,-\tilde{\Omega})\tilde{\Omega}\frac{\langle f_{p}({\bf-q},-\tilde{\Omega})f_{j}({\bf q},\tilde{\Omega})\rangle}{-i\tilde{\Omega}+\nu q^{2}}.$ Substituting in (26) and evaluating, we find $\Pi(\Omega)<\Pi(\Omega=0)$, indicating reduction of the kinetic energy flux by rotation. Since the helicity generated by the rotation scales with $\Omega$, we conclude that with a rising helicity, the kinetic energy flux is suppressed. ## References * (1) P. C. Hohenberg and B. I. Halperin, Rev. Mod. Phys. 49, 435 (1977). * (2) L. D. Landau and E. M. Lifshitz, Fluid Mechanics, Butterworth and Heinemann, Oxford (1998). * (3) U Frisch, Turbulence: the legacy of A.N. Kolmogorov, Cambridge University Press, Cambridge, 1995. * (4) A. N. Kolmogorov, “The local structure of turbulence in incompressible viscous fluid for very large Reynolds numbers,” Dokl. Akad. Nauk SSSR 30, 301–305 (1941); A. N. Kolmogorov, “Dissipation of energy in locally isotropic turbulence,” Dokl. Akad. Nauk SSSR 32, 16–18 (1941). * (5) G. E. Elsinga and I. Marusic, “The anisotropic structure of turbulence and its energy spectrum,” Phys. Fluids 28, 011701 (2016). * (6) P. C. Valente, C. B. da Silva, and F. T. Pinho, “Energy spectra in elasto-inertial turbulence,” Phys. Fluids 28, 075108 (2016). * (7) P. Buchhave and C. M. Velte, “Measurement of turbulent spatial structure and kinetic energy spectrum by exact temporal-to-spatial mapping,” Phys. Fluids 29, 085109 (2017). * (8) R. H. Kraichnan, “Inertial ranges in two-dimensional turbulence,” Phys. Fluids 10, 1417 (1967). * (9) J. Paret and P. Tabeling, “Experimental observation of the two-dimensional inverse energy cascade,” Phys. Rev. Lett. 79, 4162 (1997). * (10) S. Chen, R. E. Ecke, G. L. Eyink, M. Rivera, M. Wan, and Z. Xiao, “Physical mechanism of the two-dimensional inverse energy cascade,” Phys. Rev. Lett. 96, 084502 (2006). * (11) A. Vallgren and E. Lindborg, “The enstrophy cascade in forced two- dimensional turbulence,” J. Fluid Mech. 671, 168–183 (2011). * (12) G. Boffetta and R. E. Ecke, “Two-dimensional turbulence,” Annu. Rev. Fluid. Mech. 44, 427–451 (2012). * (13) J. Bardina, J. H. Ferziger, and R. S. Rogallo, “Effect of rotation on isotropic turbulence: Computation and modelling,” J. Fluid Mech. 154, 321–336 (1985). * (14) Y. Zhou, “A phenomenological treatment of rotating turbulence,” Phys. Fluids 7, 2092 (1995). * (15) V. M. Canuto and M. S. Dubovikov, “Physical regimes and dimensional structure of rotating turbulence,” Phys. Rev. Lett. 78, 666–669 (1997). * (16) V. M. Canuto and M. S. Dubovikov, “A dynamical model for turbulence. V. The effect of rotation,” Phys. Fluids 9, 2132–2140 (1997). * (17) L. M. Smith and F. Waleffe, “Transfer of energy to two-dimensional large scales in forced, rotating three-dimensional turbulence,” Phys. Fluids 11, 1608 (1999). * (18) C. N. Baroud, B. B. Plapp, Z.-S. She, and H. L. Swinney, “Anomalous self- similarity in a turbulent rapidly rotating fluid,” Phys. Rev. Lett. 88, 114501 (2002). * (19) S. Galtier, “Weak inertial-wave turbulence theory,” Phys. Rev. E 68, 015301 (2003). * (20) W.-C. Müller and M. Thiele, “Scaling and energy transfer in rotating turbulence,” Europhys. Lett. 77, 34003 (2007). * (21) S. Chakraborty and J. K. Bhattacharjee, “Third-order structure function for rotating three-dimensional homogeneous turbulent flow,” Phys. Rev. E 76, 036304 (2007). * (22) S. Chakraborty, “Signatures of two-dimensionalisation of 3D turbulence in the presence of rotation,” Europhys. Lett. 79, 14002 (2007). * (23) D. Mininni, A. Alexakis, and A. Pouquet, “Scale interactions and scaling laws in rotating flows at moderate Rossby numbers and large Reynolds numbers,” Phys. Fluids 21, 015108 (2009). * (24) S. Chakraborty, M. H. Jensen, and A. Sarkar, “On two-dimensionalization of three-dimensional turbulence in shell models,” Eur. Phys. J. B 73, 447–453 (2010). * (25) M. K. Sharma, M. K. Verma and S. Chakraborty, On the energy spectrum of rapidly rotating forced turbulence, Physics of Fluids 30, 115102 (2018). * (26) M. K. Sharma, M. K. Verma and S. Chakraborty, Anisotropic energy transfers in rapidly rotating turbulence, Phys. Fluids 31, 085117 (2019). * (27) S. K. Rathor, S. Chakraborty and S. S. Ray, Dynamic scaling in rotating turbulence: A shell model study, Phys. Rev. E 105, L063102 (2022). * (28) A. Basu and J. K. Bhattacharjee, Fluctuating hydrodynamics and turbulence in a rotating fluid: Universal properties, Phys. Rev. E 85, 026311 (2012). * (29) J. D. Jackson, Classical Electrodynamics, 2nd ed. (Wiley Eastern, New Delhi, 1975). * (30) A. Raichaudhuri, The Physics of Fluids and Plasmas (Cambridge University Press, Cambridge, 1998). * (31) O. Zeman, A note on the spectra and decay of rotating homogeneous turbulence, Phys. Fluids 6, 3221 (1994). * (32) Y. Zhou, A phenomenological treatment of rotating turbulence, Phys. Fluids 7, 2092 (1995). * (33) S. Chakraborty, M. H. Jensen, and A. Sarkar, On two-dimensionalization of three-dimensional turbulence in shell models, Eur. Phys. J. B 73, 447 (2010). * (34) A. Basu and J. K. Bhattacharjee, Varieties of scaling regimes in hydromagnetic turbulence, Phys. Rev. E 98, 062143 (2018). * (35) C. Cambon, N. N. Mansour, and F. S. Godefard, Energy transfer in rotating turbulence, J. Fluid Mech. 337, 303 (1997). * (36) D. Das, A. Basu, M. Barma, and S. Ramaswamy, Phys. Rev. E 64, 021402 (2001). * (37) R. H. Kraichnan, Helical turbulence and absolute equilibrium, J. Fluid Mech. 59, 745 (1973). * (38) A. Basu and J. K. Bhattacharjee, Kolmogorov or Bolgiano-Obukhov scaling: Universal energy spectra in stably stratified turbulent fluids, Phys. Rev. E 100, 033117 (2019). * (39) S. K. Rathor, M. K. Sharma, S. S. Ray and S. Chakraborty, Bridging inertial and dissipation range statistics in rotating turbulence, Phys. Fluids 32, 095104 (2020).
# A Noise-tolerant Differentiable Learning Approach for Single Occurrence Regular Expression with Interleaving Rongzhen Ye, 1 Tianqu Zhuang, 1 Hai Wan, 1* Jianfeng Du, 2 Weilin Luo, 1 Pingjia Liang, 1 Both Hai Wan and Jianfeng Du are corresponding authors. ###### Abstract We study the problem of learning a _single occurrence regular expression with interleaving_ (SOIRE) from a set of text strings possibly with _noise_. SOIRE fully supports interleaving and covers a large portion of regular expressions used in practice. Learning SOIREs is challenging because it requires heavy computation and text strings usually contain noise in practice. Most of the previous studies only learn restricted SOIREs and are not robust on noisy data. To tackle these issues, we propose a noise-tolerant differentiable learning approach SOIREDL for SOIRE. We design a neural network to simulate SOIRE matching and theoretically prove that certain assignments of the set of parameters learnt by the neural network, called _faithful encodings_ , are one-to-one corresponding to SOIREs for a bounded size. Based on this correspondence, we interpret the target SOIRE from an assignment of the set of parameters of the neural network by exploring the nearest faithful encodings. Experimental results show that SOIREDL outperforms the state-of-the-art approaches, especially on noisy data. ## Introduction Learning regular expressions (REs) is a fundamental task in Machine Learning. For example, REs are the target in EXtensible Markup Language (XML) schema inference, for covering a set of text strings about an XML element. The regular expression with interleaving, denoted as RE(&), is an extension of regular expressions, where the operator interleaving (&) is added to interleave two strings. RE(&) has been widely used in various areas, ranging from XML database system (Makoto and Clark 2003; Colazzo et al. 2013; Martens et al. 2017) to system verification (Gischer 1981; Bojanczyk et al. 2006) and natural language processing (Kuhlmann and Satta 2009; Nivre 2009), etc. We focus on a subclass of RE(&), _single occurrence regular expression with interleaving_ (SOIRE), where each symbol occurs at most once in a SOIRE. Learning SOIREs is still meaningful, as SOIREs have the second highest coverage rate of REs on the schema database Relax NG among all well-known subclasses (Li et al. 2019b). Although a subclass is generally easier to learn than the full class, it is still challenging to learn SOIREs. On one hand, learning SOIREs is a search problem requiring heavy computation. On the other hand, real-life text strings usually contain noise (Kearns and Li 1988; Galassi and Giordana 2005; Bex et al. 2006). For example, in XML schema inference the XML data may contain incorrect symbols (Bex et al. 2006). The presence of noise makes the problem of learning SOIREs more challenging. There have been a number of proposals for learning either the full class or its subclasses of SOIRE, from a set of text strings (Freydenberger and Kötzing 2015; Zhang et al. 2018; Li et al. 2019a, 2020b). However, they are hard to guarantee that the learnt REs reach the full declared expressive power. For example, Li et al. (2019a) claimed to learn SOIREs, but Wang and Zhang (2021) showed that they just learn special cases of SOIRE. Besides, existing approaches are not robust on noisy data because any modification to given strings will alter the patterns in learnt REs. As far as we know, there is no approach to learning SOIREs that works well with both noise-free data and noisy data. In this paper, we propose a noise-tolerant differentiable learning approach SOIREDL for SOIREs. Specifically, we design a neural network to simulate SOIRE matching for text strings. Since existing work for SOIRE matching (Wang 2021b, 2022) is not suitable for differentiable learning, we propose a new SOIRE matching algorithm SOIRETM based on the syntax tree of SOIRE, and accordingly, an algorithm for converting SOIRETM to a neural network. We theoretically prove that certain assignments of the set of parameters learnt by the neural network, called _faithful encodings_ , one-to-one correspond to SOIREs for a bounded size. This correspondence allows us to interpret the target SOIRE from an assignment of the set of learnt parameters of the neural network by exploring the nearest faithful encodings. To evaluate the performance of SOIREDL on noisy data, we extract $30$ SOIREs from the RE database built by Li et al. (2018) to make a group of datasets covering five domains with different noise levels. Experimental results show that at all noise levels, the average accuracy of SOIREDL is higher than state-of-the-art (SOTA) approaches and the faithfulness of SOIREDL is always beyond $80\%$. In particular, the average accuracy of SOIREDL only decreases slightly with increasing noise levels, which suggests that SOIREDL is robust on noisy data. ## Related Work Matching algorithms for regular expressions. It has been shown (Mayer and Stockmeyer 1994) that the matching problem of RE(&) is NP-hard. For SOIREs, Wang (2021b) proposed a finite automata with interleaving, written FA(&), to solve the matching problem. Wang (2022) proposed a single occurrence finite automata, written SFA(&, #), for matching single occurrence regular expressions with interleaving and counting. All above studies do not consider converting the matching algorithm into a neural network. In contrast, we present not only a matching algorithm for SOIREs but also the way to convert it to a neural network. Learning approaches for regular expressions. There also exists work for learning different subclasses of REs from a set of text strings, such as the deterministic regular expressions with counting (Wang and Chen 2018), the deterministic regular expressions with unorder (Wang and Chen 2020), and the deterministic regular expression with counting and unorder (Wang 2021a). Some subclasses of the extension RE(&) have also been explored, such as chain regular expression with interleaving (ICHARE) (Zhang et al. 2018), restricted SOIRE (RSOIRE) (Li et al. 2019a), and k-occurrence regular expression with interleaving (kOIRE) (Li et al. 2020a). All of them are learnt from positive strings only. Li et al. (2020b) learnt a subclass of ICHARE, called SIRE, from both positive and negative strings based on a genetic algorithm. The relation of expressive powers supported by these classes is SIRE $\subset$ ICHARE $\subset$ RSOIRE $\subset$ SOIRE $\subset$ kOIRE. Li et al. (2021) proposed a natural language processing based RE synthesizer to learn REs from natural language descriptions together with positive and negative strings. This problem setting is different from ours. Differentiable learning. Differentiable learning has attracted much research interest recently. Most studies focus on learning logical rules from knowledge bases (Yang, Yang, and Cohen 2017; Sadeghian et al. 2019; Cohen, Yang, and Mazaitis 2020; Huang et al. 2021) or on neural logic programming (Yang and Song 2020; Gao et al. 2022; Wang et al. 2020). In particular, some studies (Rocktäschel and Riedel 2017; Minervini et al. 2020b, a) focus on neural theorem proving. They convert the symbolic operations into differentiable modules to enhance the reasoning ability of the neural network. Mensch and Blondel (2018) utilized a strongly convex regularizer to smooth the max operator and convert a broad class of dynamic programming (DP) algorithms into differentiable operators. Wang et al. (2019) proposed a differentiable MaxSAT solver integrated into the deep learning networks to solve the problems like visual Sudoku, which has implicit satisfiability constraints. For REs, Jiang et al. (2020) injected a weighted finite-state automaton (FSA) of REs into the recurrent neural network (RNN) to improve the performance of text classification. Further, Jiang, Jin, and Tu (2021) injected a finite-state transducer of REs into RNN for slot filling. Both of the above methods fine- tune initial regular expressions given from the expert knowledge to obtain better results. Different from all above studies, we further study the one-to- one correspondence between parameters of a neural network and SOIREs. ## Preliminaries A regular expression with interleaving, written RE(&), over an alphabet $\Sigma$ is defined recursively as follows (Mayer and Stockmeyer 1994): $r:=\epsilon\big{|}a\big{|}r_{1}^{*}\big{|}r_{1}\cdot r_{2}\big{|}r_{1}\&r_{2}\big{|}r_{1}|r_{2}$ where $\epsilon$ is empty string, the symbol $a\in\Sigma$, and $r_{1},r_{2}$ are RE(&). The operator $*$ denotes Kleene-Star, $\cdot$ denotes concatenation (it can be omitted if there is no ambiguity), $\&$ denotes interleaving, $|$ denotes disjunction. The operators $?$ and $+$ are commonly used for repetition. They are defined as $r^{?}:=r\big{|}\epsilon$ and $r^{+}:=r\cdot r^{*}$, respectively. The operator $\&$ for two strings $s_{1},s_{2}$ is defined as follows: $s_{1}\&s_{2}:=\left\\{\begin{array}[]{cc}s_{2}&\text{if }s_{1}=\epsilon\\\ s_{1}&\text{if }s_{2}=\epsilon\\\ a(s_{1}^{\prime}\&s_{2})|b(s_{1}\&s_{2}^{\prime})&otherwise\\\ \end{array}\right.$ where $s_{1}=as_{1}^{\prime}$, $s_{2}=bs_{2}^{\prime}$, $a,b\in\Sigma$. ###### Definition 1 (Single occurrence regular expression with interleaving (SOIRE) (Li et al. 2019a)). SOIRE is a RE(&) where each symbol occurs at most once in the expression. ###### Example 1. $(a\&b)c^{*}$ is a SOIRE. $(a\&b)a^{*}$ is a RE(&) but not a SOIRE, because the symbol $a$ occurs twice. Each SOIRE can be expressed by its prefix notation where operators are written in front of operands rather than written in the middle as the infix notation. By $\texttt{PreForm}(r)$ we denote the prefix notation of a SOIRE $r$. For the SOIRE $r=(a\&b)c^{*}$ given in Example 1, $\texttt{PreForm}(r)$ is $\cdot\&ab*c$. Syntax tree. Each SOIRE can also be represented as a binary tree, called _syntax tree_ , where each inner vertex in the tree represents an operator and each leaf represents a symbol. By $\texttt{RE2Tree}(r)$ we denote the syntax tree of a SOIRE $r$. In a syntax tree, each leaf represents a symbol in $\Sigma$ and each symbol occurs at most once, while each inner vertex represents an operator and the number of children of it is equal to the number of its operands. Figure 1 shows the syntax tree of $(a\&b)c^{*}$ in Example 1. The size of a SOIRE $r$, denoted by $|r|$, is defined as the number of vertices in the syntax tree of $r$. For example, the size of $(a\&b)c^{*}$ is $6$. Obviously, the preorder traversal sequence of the syntax tree of $r$ is $\texttt{PreForm}(r)$ and each subtree represents a subexpression. In this preorder traversal sequence, vertex $t+1$ is the left child for any inner vertex $t$. We use $r^{t}(1\leq t\leq|r|)$ to denote the corresponding SOIRE of the subtree of $\texttt{RE2Tree}(r)$ whose root is vertex $t$. Further, if vertex $t$ represents a binary operator, we use $\eta^{t}$ to denote the sequential number of its right child. For $r=(a\&b)c^{*}$ in Example 1, $r^{2}=a\&b$ and $\eta^{2}=4$. From prefix notation to syntax tree. We show a way to realize $\texttt{RE2Tree}(r)$. It scans $\texttt{PreForm}(r)$ from back to front and maintains a stack of syntax trees. Take Figure 1 for instance, $\texttt{PreForm}(r)=\cdot\&ab*c$. $\texttt{RE2Tree}(r)$ scans $\texttt{PreForm}(r)$ from $c$ to $\cdot$. When scanning $c$, push $c$ ($v_{6}$) into the stack. When scanning $*$, pop $c$ ($v_{6}$) from the stack and make $c$ the left child of $*$, then push $*c$ ($v_{5}$) into the stack. When scanning $b$ and $a$, push $b$ ($v_{4}$) and $a$ ($v_{3}$) into the stack. When scanning $\&$, pop $a$ ($v_{3}$) and $b$ ($v_{4}$) from the stack and make $a$ the left child of $\&$ and $b$ the right child, then push $\&ab$ ($v_{2}$) into the stack. In this way, the syntax tree in Figure 1 is built. We use $\texttt{Tree2RE}(\xi)$ to denote the inverse function of $\texttt{RE2Tree}(r)$, which returns $\texttt{PreForm}(r)$ from the syntax tree $\xi=\texttt{RE2Tree}(r)$ by preorder traversal. Figure 1: The syntax tree of SOIRE $(a\&b)c^{*}$. $v_{1}$, …, $v_{6}$ represent $\cdot$, $\&$, $a$, $b$, $*$ and $c$, respectively. SOIRE matching. By $r\models s$ we denote that a SOIRE $r$ matches a string $s$. Then the problem of SOIRE matching is to check whether $r\models s$, which can be decided in the following way: $\displaystyle r\models s:=$ $\displaystyle\left\\{\begin{array}[]{ll}s=a&\text{if }r=a\\\ s=\epsilon\vee r_{1}\models s&\text{if }r=r_{1}^{?}\\\ s=\epsilon\vee\exists s_{1}\exists s_{2},(s=s_{1}s_{2}\wedge s_{2}\neq\epsilon\\\ \wedge r_{1}^{*}\models s_{1}\wedge r_{1}\models s_{2})&\text{if }r=r_{1}^{*}\\\ \exists s_{1}\exists s_{2},(s=s_{1}s_{2}\wedge r_{1}^{*}\models s_{1}\wedge r_{1}\models s_{2})&\text{if }r=r_{1}^{+}\\\ \exists s_{1}\exists s_{2},(s=s_{1}s_{2}\wedge r_{1}\models s_{1}\wedge r_{2}\models s_{2})&\text{if }r=r_{1}\cdot r_{2}\\\ \exists s_{1}\exists s_{2},(s=s_{1}\&s_{2}\wedge r_{1}\models s_{1}\wedge r_{2}\models s_{2})&\text{if }r=r_{1}\&r_{2}\\\ r_{1}\models s\vee r_{2}\models s&\text{if }r=r_{1}|r_{2}\\\ \end{array}\right.$ (9) where $r_{1},r_{2}$ are SOIREs and $a\in\\{\epsilon\\}\cup\Sigma$. SOIRE learning. Given a set of strings $\Pi=\Pi^{+}\cup\Pi^{-}$, where $\Pi^{+}$ (resp. $\Pi^{-}$) denotes the set of strings with the positive (resp. negative) label, the problem of SOIRE learning is to find a SOIRE $r$ that maximizes $\mathrm{acc}(r)$ defined below. $\displaystyle\mathrm{acc}(r)=\frac{|\\{s|r\models s,s\in\Pi^{+}\\}|+|\\{s|r\not\models s,s\in\Pi^{-}\\}|}{|\Pi|}$ (10) ## The Proposed SOIREDL Approach In this section, we first describe a new matching algorithm SOIRETM, then show how to convert SOIRETM to a neural network. Afterwards, we show correspondence between parameters of the neural network and SOIREs. Finally, we show how to interpret the target SOIRE from the parameters. First of all, we introduce a variant problem of SOIRE matching, called _filter matching_. Filter matching for a SOIRE $r$ and a string $s$ is to check if $r$ matches $filter(s,\alpha(r))$, where $\alpha(r)$ denotes the set of symbol in a SOIRE $r$, and the filter function $filter(s,V)$ returns a string that only retains symbols in $V$, where $V\subseteq\Sigma$. For Example 1 and $s=dbac$, the corresponding problem of filter matching is to check if $(a\&b)c^{*}$ matches $filter(dbac,\\{a,b,c\\})=bac$, as $\alpha((a\&b)c^{*})=\\{a,b,c\\}$. Given a SOIRE $r$ and a string $s$, we use $g_{i,j}^{t}\in\\{0,1\\}$ ($1\leq t\leq|r|$, $1\leq i,j\leq|s|$) to denote whether $r^{t}$ matches $filter(s_{i,j},\alpha(r^{t}))$, where $s_{i,j}$ denotes the substring of $s$ from $i$ to $j$, and where $s_{1,0}=\epsilon$ specially. If $r^{t}$ matches $filter(s_{i,j},\alpha(r^{t}))$, then $g_{i,j}^{t}=1$, otherwise $g_{i,j}^{t}=0$. In particular, $g_{1,0}^{t}$ denotes if $r^{t}$ matches $\epsilon$ since for all $t$ from $1$ to $|r|$, $filter(s_{1,0},\alpha(r^{t}))=\epsilon$. For Example 1 and $s=dbac$, $g_{1,2}^{2}$ denotes if $a\&b$ matches $ba$ and $g_{1,2}^{2}=1$. SOIRE | Semantics of $r\models filter(s,\alpha(r))$ ---|--- $r=a\in\Sigma$ | 1\. $filter(s,{a})=a$. $r=r_{1}^{?}$ | 1\. $filter(s,\alpha(r_{1}^{?}))=\epsilon$. $=\epsilon|r_{1}$ | 2\. $r_{1}\models filter(s,\alpha(r_{1}))$. $r=r_{1}^{*}$ | 1\. $filter(s,\alpha(r_{1}^{*}))=\epsilon$. $=\epsilon|r_{1}|r_{1}^{*}\cdot r_{1}$ | 2\. $r_{1}\models filter(s,\alpha(r_{1}))$. | 3\. $r_{1}^{*}\models filter(s_{1},\alpha(r_{1}^{*}))$ and $r_{1}\models filter(s_{2},\alpha(r_{1}))$, | where $s=s_{1}s_{2}(s_{1},s_{2}\neq\epsilon)$. $r=r_{1}^{+}$ | 1\. $r_{1}\models filter(s,\alpha(r_{1}))$. $=r_{1}|r_{1}^{+}\cdot r_{1}$ | 2\. $r_{1}^{+}\models filter(s_{1},\alpha(r_{1}^{+}))$ and $r_{1}\models filter(s_{2},\alpha(r_{1}))$, | where $s=s_{1}s_{2}(s_{1},s_{2}\neq\epsilon)$. $r=r_{1}\cdot r_{2}$ | 1\. $r_{1}\models filter(s,\alpha(r_{1}\cdot r_{2}))$ and $r_{2}\models\epsilon$. | 2\. $r_{2}\models filter(s,\alpha(r_{1}\cdot r_{2}))$ and $r_{1}\models\epsilon$. | 3\. $r_{1}\models filter(s_{1},\alpha(r_{1}\cdot r_{2}))$ and | $r_{2}\models filter(s_{2},\alpha(r_{1}\cdot r_{2}))$, where $s=s_{1}s_{2}(s_{1},s_{2}\neq\epsilon)$. $r=r_{1}\&r_{2}$ | $r_{1}\models filter(s,\alpha(r_{1}))$ and $r_{2}\models filter(s,\alpha(r_{2}))$. $r=r_{1}|r_{2}$ | 1\. $r_{1}\models filter(s,\alpha(r_{1}|r_{2}))$. | 2\. $r_{2}\models filter(s,\alpha(r_{1}|r_{2}))$. Table 1: The semantics of filter matching, where $r,r_{1},r_{2}$ are SOIREs and $s,s_{1},s_{2}$ are strings. $r\models filter(s,\alpha(r))$ if and only if at least one condition on the right is satisfied. ### SOIRE Matching by SOIRETM We propose a new matching algorithm for SOIRE, named SOIRETM, based on dynamic programming. Generally, SOIRETM divides the original matching problem into smaller ones to conquer. We observe that SOIRE matching can be simplified to filter matching, as shown in the following theorem. ###### Theorem 1. Given a SOIRE $r$ and a string $s$, $r\models s$ iff $filter(s,\alpha(r))=s$ and $r\models filter(s,\alpha(r))$.111All the proofs of theorems/lemmas/propositions are provided in appendix B. Theorem 1 presents a necessary and sufficient condition for SOIRE matching. The condition $filter(s,\alpha(r))=s$ guarantees that $\alpha(r)$ contains all symbols in $s$, which is easy to check. Therefore, the primary problem is to check if $r$ matches $filter(s,\alpha(r))$. We build the syntax tree of $r$ and compute the result of filter matching of $s$ and $r$, namely $g_{1,|s|}^{1}$. The proposed algorithm SOIRETM is detailed in Algorithm 1. Initially, line 1 checks if $filter(s,\alpha(r))=s$. Then we calculate $g_{i,j}^{t}$ from shorter substrings to longer ones and from bottom to top of the syntax tree. The statements in Line 7-20 conform to the semantics of each operator, given by Table 1 and Lemma 2. ###### Lemma 2. Given a SOIRE $r$ and a string $s$. If $r=r_{1}\cdot r_{2}$ or $r=r_{1}\&r_{2}$ or $r=r_{1}|r_{2}$, then for all $i\in\\{1,2\\}$, $r_{i}\models filter(s,\alpha(r))$ iff $filter(s,\alpha(r_{i}))=filter(s,\alpha(r))$ and $r_{i}\models filter(s,\alpha(r_{i}))$. The time complexity of Algorithm 1 is $O(|s|^{3}|r|)$. It is sound and complete according to the following theorem. ###### Theorem 3. Given a SOIRE $r$ and a string $s$, $r\models s$ iff $\texttt{SOIRETM}(r,s)=1$. Algorithm 1 Matching Algorithm for SOIRE, SOIRETM. Input: A SOIRE $r$ and a string $s$. Output: The answer of whether $r$ matches $s$. 1:if $filter(s,\alpha(r))\neq s$ then 2: Return $0$; 3:Build the syntax tree of $r$ by $\texttt{RE2Tree}(r)$; 4:Let $flag_{i,j}^{t,t^{\prime}}$ denote $1[filter(s_{i,j},\alpha(r^{t}))=filter(s_{i,j},\alpha(r^{t^{\prime}}))]$, where $1[\mu]=1$ iff $\mu$ is true; 5:for all substring $s_{i,j}$ from $\epsilon$ to $s$ (shorter to longer) do 6: for $t\leftarrow|r|$ downto $1$ do 7: if $r^{t}=a\in\alpha(r)$ then 8: $g_{i,j}^{t}\leftarrow 1[filter(s_{i,j},\\{a\\})=a]$; 9: else if $r^{t}=(r^{t+1})^{?}$ then 10: $g_{i,j}^{t}\leftarrow 1[filter(s_{i,j},\alpha(r^{t}))=\epsilon]\vee g_{i,j}^{t+1}$; 11: else if $r^{t}=(r^{t+1})^{*}$ then 12: $g_{i,j}^{t}\leftarrow 1[filter(s_{i,j},\alpha(r^{t}))=\epsilon]\vee g_{i,j}^{t+1}\vee\bigvee_{k=i}^{j-1}(g_{i,k}^{t}\wedge g_{k+1,j}^{t+1})$; 13: else if $r^{t}=(r^{t+1})^{+}$ then 14: $g_{i,j}^{t}\leftarrow g_{i,j}^{t+1}\vee\bigvee_{k=i}^{j-1}(g_{i,k}^{t}\wedge g_{k+1,j}^{t+1})$; 15: else if $r^{t}=r^{t+1}\cdot r^{\eta^{t}}$ then 16: $g_{i,j}^{t}\leftarrow(flag_{i,j}^{t,t+1}\wedge g_{i,j}^{t+1}\wedge g_{1,0}^{\eta^{t}})\vee(flag_{i,j}^{t,\eta^{t}}\wedge g_{i,j}^{\eta^{t}}\wedge g_{1,0}^{t+1})\vee\bigvee_{k=i}^{j-1}(flag_{i,k}^{t,t+1}\wedge g_{i,k}^{t+1}\wedge flag_{k+1,j}^{t,\eta^{t}}\wedge g_{k+1,j}^{\eta^{t}})$; 17: else if $r^{t}=r^{t+1}\&r^{\eta^{t}}$ then 18: $g_{i,j}^{t}\leftarrow g_{i,j}^{t+1}\wedge g_{i,j}^{\eta^{t}}$; 19: else if $r^{t}=r^{t+1}|r^{\eta^{t}}$ then 20: $g_{i,j}^{t}\leftarrow(flag_{i,j}^{t,t+1}\wedge g_{i,j}^{t+1})\vee(flag_{i,j}^{t,\eta^{t}}\wedge g_{i,j}^{\eta^{t}})$; 21:Return $g_{1,|s|}^{1}$; ### From SOIRETM to Neural Network We now detail how to convert SOIRETM to a trainable neural network to simulate SOIRE matching. SOIRETM uses the syntax tree for SOIRE matching, so the trainable parameters of an expected neural network can be defined by constructs of the syntax tree. There are two parts of parameters used in an expected neural network, $\theta=(w,u)$, where $w\in[0,1]^{T\times|\mathbb{B}|}$, $u\in[0,1]^{T\times T}$ and $\mathbb{B}=\Sigma\cup\\{?,*,+,\cdot,\&,|,\text{none}\\}$, and where $T$ is the bounded size of the target SOIRE. For $1\leq t\leq T,a\in\mathbb{B}$, $w^{t}_{a}$ denotes the probability of vertex $t$ representing a symbol in $\Sigma$ or an ordinary operator or the none operator. For $1\leq t\leq T$ and $t+2\leq t^{\prime}\leq T$, $u^{t}_{t^{\prime}}$ denotes the probability of vertex $t$ choosing vertex $t^{\prime}$ as its right child. The total number of the parameters to be learnt is $T|\mathbb{B}|+\frac{(T-1)(T-2)}{2}$. Example 2 shows the parameters of the neural network $\theta=(w,u)$ corresponding to the syntax tree in Figure 1. ###### Example 2. When $T=6$, $w^{1}_{\cdot}$, $w^{2}_{\&}$, $w^{3}_{a}$, $w^{4}_{b}$, $w^{5}_{*}$, $w^{6}_{c}$, $u^{1}_{5}$, $u^{2}_{4}$ are $1$s, whereas other parameters are $0$s. When $T=8$, $w^{7}_{\text{none}}$, $w^{8}_{\text{none}}$ are $1$s in addition to the above parameters. There are four parts that should be considered during the conversion of SOIRETM to neural network: $\alpha(r^{t})$, $flag_{i,j}^{t,t^{\prime}},g^{t}_{i,j}$, and the return value of Algorithm 1. Recall that $\alpha(r^{t})$ represents the set of symbols in $r^{t}$, which is also the set of symbols occurring in the subtree whose root is $t$. We use $\rho^{t}_{a}$ to denote the probability of symbol $a\in\Sigma$ that occurs in the subtree whose root is $t$. We calculate $\rho^{t}_{a}$ from bottom to top of the syntax tree by Equation 11, where $\sigma_{01}(x)=\min(\max(x,0),1)$. For all $t>T$ and $a\in\Sigma$, $\rho^{t}_{a}$ is set to $0$. Note that $\min(x)=-\max(-x)$, and the $\max$ function amounts to ReLU and can be approximated by a more differentiable LeakyReLU function. $\rho^{t}_{a}=\sigma_{01}(w^{t}_{a}+\sum_{\begin{subarray}{c}o\in\\{?,*,+,\\\ \cdot,\&,|\\}\end{subarray}}w^{t}_{o}\rho^{t+1}_{a}+\sum_{\begin{subarray}{c}o\in\\\ \\{\cdot,\&,|\\}\end{subarray}}w^{t}_{o}\sum_{t^{\prime}=t+2}^{T}u^{t}_{t^{\prime}}\rho^{t^{\prime}}_{a})$ (11) For converting $flag_{i,j}^{t,t^{\prime}}$, we treat it as the probability that there does not exist a symbol occurring in both $s_{i,j}$ and $\alpha(r^{t})$ but not occurring in $\alpha(r^{t^{\prime}})$, as defined in Equation 12. Note that $t^{\prime}$ can only be either $t+1$ or $\eta^{t}$. $flag_{i,j}^{t,t^{\prime}}=1-\sigma_{01}(\sum_{a\in\Sigma}\sigma_{01}(1[a\in s_{i,j}]+(\rho^{t}_{a}-\rho^{t^{\prime}}_{a})-1))$ (12) For converting $g_{i,j}^{t}$, we introduce $p_{i,j}^{t}(?)$ (resp. $p_{i,j}^{t}(*)$ or $p_{i,j}^{t}(+)$) to denote the probability that the right-hand side of Line 10 (resp. 12 or 14) in Algorithm 1 evaluates to $1$, as well as $p_{i,j}^{t}(\cdot,t^{\prime})$ (resp. $p_{i,j}^{t}(\&,t^{\prime})$ or $p_{i,j}^{t}(|,t^{\prime})$) the probability that the right-hand side of Line 16 (resp. 18 or 20) with $\eta^{t}$ substituted by $t^{\prime}$ evaluates to $1$. For example, $p^{8}_{1,3}(\&,10)$ denotes the probability that $g_{1,3}^{9}\wedge g_{1,3}^{10}$ evaluates to $1$. Since a right-hand side may contain logical connectives $\wedge$ or $\vee$, we apply the transformations given in Table 2 to estimate the ultimate probability that the right-hand side evaluates to $1$, where the special transformation is only used in the third term of Line 12 and Line 16 as well as the second term of Line 14 since these terms may have a large number of operands, while anywhere else the general transformations are used. With the probabilities that the right-hand sides evaluate to $1$, the probability that $g_{i,j}^{t}$ evaluates to $1$ can be defined recursively by Equation From SOIRETM to Neural Network, where we reuse $g_{i,j}^{t}$ to represent such a probability. $\displaystyle g_{i,j}^{t}=\sum_{a\in\Sigma}w^{t}_{a}\cdot 1[filter(s_{i,j},{a})=a]$ $\displaystyle+\sum_{o\in\\{?,*,+\\}}w^{t}_{o}p^{t}_{i,j}(o)+\sum_{o\in\\{\cdot,\&,|\\}}w^{t}_{o}\sum_{t^{\prime}=t+2}^{T}u^{t}_{t^{\prime}}p^{t}_{i,j}(o,t^{\prime})$ (13) Logical operation | General transformation | Special transformation ---|---|--- $A\wedge B$ | $\min(p_{A},p_{B})$ | - $A\vee B$ | $\sigma_{01}(p_{A}+p_{B})$ | $\max(p_{A},p_{B})$ Table 2: The transformation from logical operations to numerical computations, where $A$ and $B$ are formulae, and $p_{A}$ (resp. $p_{B}$) is the probability that $A$ (resp. $B$) evaluates to $1$. For converting the return value of Algorithm 1, we consider Line 1 and Line 21 in Algorithm 1 and compute the return value by Equation 14, where the second term in Equation 14 ensures that all symbols occurring in $s$ appear in the target SOIRE too. $\hat{y}=g^{1}_{1,|s|}-\max_{a\in\Sigma}\sigma_{01}(1[a\in s]-\rho^{1}_{a})$ (14) The converted neural network is trained to minimize the objective function $\frac{1}{2}(\hat{y}-y)^{2}$, where $y\in\\{0,1\\}$ is the ground-truth label for $r$ matching $s$. ### Faithful Encoding We simply refer to an assignment of the set of trainable parameters of the converted neural network as an _encoding_ of SOIREs. We find that encodings can one-to-one correspond to prefix notations of SOIREs for a bounded size, when it satisfies certain conditions given by Definition 2. ###### Definition 2 (Faithful encoding). An encoding $\theta=(w,u)$ of SOIREs with length $T$ is said to be _faithful_ if it satisfies all the following conditions: 1. 1. $\forall 1\leq t\leq T,w^{t}$ is a one-hot vector. 2. 2. $\forall 1\leq t\leq T,u^{t}$ is either a one-hot vector or an all-zero vector. 3. 3. $\forall 1\leq t\leq T,\sum_{t^{\prime}=t+2}^{T}u^{t}_{t^{\prime}}+\sum_{a\in\Sigma\cup\\{?,*,+,\text{none}\\}}w^{t}_{a}=1$. 4. 4. $\forall 1\leq t\leq T-1,w^{t+1}_{\text{none}}-w^{t}_{\text{none}}\geq 0$. 5. 5. $\forall 2\leq t\leq T,\sum_{a\in\\{?,+,*,\cdot,\&,|\\}}w^{t-1}_{a}+\sum_{t^{\prime}=1}^{t-2}u^{t^{\prime}}_{t}+w^{t}_{\text{none}}=1$. 6. 6. $\forall 3\leq t\leq T,\forall 1\leq p\leq t-2,(t-1-p)u^{p}_{t}+\sum_{p^{\prime}=p+1}^{t-1}\sum_{t^{\prime}=t+1}^{T}u^{p^{\prime}}_{t^{\prime}}\leq t-1-p$. 7. 7. $\forall a\in\Sigma,\sum_{t=1}^{T}w^{t}_{a}\leq 1$. All conditions in a faithful encoding are translated from the construction constraints of a syntax tree. Condition 1 guarantees that each vertex in the syntax tree represents a symbol, an ordinary operator or the none operator. Condition 2 guarantees that each vertex has at most one right child. Condition 3 guarantees that each vertex either has a right child, or represents a symbol, a unary operator or the none operator. Condition 4 guarantees that if vertex $t$ represents the none operator, then vertex $t+1$ is also the none operator. Condition 5 guarantees that if vertex $t$ represents the none operator, then vertex $t$ is not the child of any vertex; otherwise, vertex $t$ is the child of exactly one vertex. Condition 6 guarantees that the vertices are numbered in the order of preorder traversal. Condition 7 guarantees that each symbol in $\Sigma$ occurs at most once in the syntax tree. Example 2 shows two faithful encodings with lengths $6$ and $8$, respectively. By $\texttt{Enc2Pre}(\theta)$ we denote the prefix notation of the SOIRE interpreted from a faithful encoding $\theta$. $\texttt{Enc2Pre}(\theta)$ decodes $w^{t}$ and $u^{t^{\prime}}_{t}$ ($1\leq t^{\prime}\leq t-2$) into the syntax tree of $r$ from $t=1$ to $T$ progressively until $w^{t}_{\text{none}}=1$, and then translates the constructed syntax tree of $r$ to the prefix notation of $r$. Take Example 2 for instance, $\texttt{Enc2Pre}(\theta)=\cdot\&ab*c$ is decoded from the faithful encoding $\theta$ with length $T=8$. Proposition 4 shows that $\texttt{Enc2Pre}(\theta)$ always yields the prefix notation of a SOIRE. ###### Proposition 4. For any faithful encoding $\theta$, there exists a SOIRE $r$ such that $\texttt{Enc2Pre}(\theta)=\texttt{PreForm}(r)$. Proposition 5 and Proposition 6 show that $\texttt{Enc2Pre}(\theta)$ is surjective and injective, respectively, for a bounded size. ###### Proposition 5. Given a bounded size $T\in\mathbb{Z}^{+}$, for any SOIRE $r$ such that $|r|\leq T$, there exists a faithful encoding $\theta$ with length $T$ such that $\texttt{Enc2Pre}(\theta)=\texttt{PreForm}(r)$. ###### Proposition 6. Given a bounded size $T\in\mathbb{Z}^{+}$, for any two different faithful encodings $\theta_{1}$ and $\theta_{2}$ with length $T$, we have $\texttt{Enc2Pre}(\theta_{1})\neq\texttt{Enc2Pre}(\theta_{2})$. Since $\texttt{Enc2Pre}(\theta)$ is both injective and surjective, faithful encodings are one-to-one corresponding to the prefix notations of SOIREs for a bounded size. ###### Theorem 7. Given a bounded size $T\in\mathbb{Z}^{+}$, prefix notations of SOIREs $r$ with $|r|\leq T$ and faithful encodings $\theta$ with length $T$ have a one-to-one correspondence, i.e., $\texttt{Enc2Pre}(\theta)=\texttt{PreForm}(r)$. To make an encoding more faithful, we add one regularization for each condition to the objective function222Details of regularizations are provided in Appendix C.. Algorithm 2 SOIRE Interpretation. Input: A training set $(\Pi_{\text{train}}^{+},\Pi_{\text{train}}^{-})$ on the alphabet $\Sigma$, an encoding $\theta=(w,u)$ and the the beam width $\beta$. Output: The infix notation of the target SOIRE. 1:Let $T$ be the first dimension of $w$; 2:Let $C^{t}$ be the set of candidate solutions $(r,e)$ of the subtree with the root $t$, where $r$ is the infix notation of a SOIRE and $e$ is its score; 3:for $t$ from $T$ down to $1$ do 4: for all $a\in\Sigma$ do 5: Add $(a,w^{t}_{a})$ into $C^{t}$; 6: for all $(r_{i},e_{i})\in C^{t+1}$ do 7: Add $(r_{i}^{?},e_{i}w^{t}_{?})$ into $C^{t}$; 8: Add $(r_{i}^{*},e_{i}w^{t}_{*})$ into $C^{t}$; 9: Add $(r_{i}^{+},e_{i}w^{t}_{+})$ into $C^{t}$; 10: for $t^{\prime}$ from $t+2$ to $T$ do 11: for all $((r_{i},e_{i}),(r_{j},e_{j}))\in C^{t+1}\times C^{t^{\prime}}$ do 12: if $\alpha(r_{i})\cap\alpha(r_{j})=\emptyset$ then 13: Add $((r_{i})\cdot(r_{j}),e_{i}e_{j}w^{t}_{\cdot})$ into $C^{t}$; 14: Add $((r_{i})\&(r_{j}),e_{i}e_{j}w^{t}_{\&})$ into $C^{t}$; 15: Add $((r_{i})|(r_{j}),e_{i}e_{j}w^{t}_{|})$ into $C^{t}$; 16: Sort all $(r,e)$ in $C^{t}$ according to the descending order of ${e}^{\frac{1}{|r|}}$ and keep only the top-$\beta$ elements; 17:Get the accuracy of $r$ on $(\Pi_{\text{train}}^{+},\Pi_{\text{train}}^{-})$ for all $r$ in $C^{1}$; 18:Return $r$ in $C^{1}$ that has the highest accuracy; ### SOIRE Interpretation We apply beam search to find a faithful encoding nearby the learnt encoding and then interpret it to the target SOIRE. The algorithm is shown in Algorithm 2. The interpretation steps are conducted from bottom to top of the syntax tree. We keep $\beta$ candidate SOIREs for each subtree according to the score of each candidate SOIRE, which is defined as the geometric mean of the probabilities of all operators and symbols. At each step, we select different operators and candidate SOIREs from the left child and right child (if any) and merge them to generate new candidates. At last, we calculate the accuracy of each SOIRE in the last step on the training set and pick out the SOIRE with the highest accuracy. ## Evaluation We conduct experiments to evaluate the performance of SOIREDL on both noise- free data and noisy data. Datasets. We extract $30$ SOIREs from the RE database built by Li et al. (2018) and generate datasets with noise from them. The SOIREs are randomly chosen from different classes: SIRE, ICHARE, RSOIRE, SOIRE. We set $\Sigma$ as $10$ letters. For each SOIRE $r$, we generate a dataset $(\Pi^{+},\Pi^{-})$ randomly, making sure that for all $s\in\Pi^{-}$, there exists $s^{\prime}$ such that $r\models s^{\prime}$ and $s$ can be modified to $s^{\prime}$ by deleting a character, inserting a character at any position, replacing a character with another one, or moving a character to any other position. For example, string $abc$ can be modified to $ac$, $abac$, $acc$ or $bca$. We set the maximum length of strings to $20$ in the dataset. For training sets and test sets, we set $|\Pi^{+}|=|\Pi^{-}|=250$, whereas for validation sets, we set $|\Pi^{+}|=|\Pi^{-}|=50$. The noise levels are set to $\delta=\\{0,0.05,0.1,0.15,0.2\\}$, where for each $\delta$, we reverse the labels for $|\Pi^{+}|\delta$ positive strings and $|\Pi^{-}|\delta$ negative strings in the training and validation sets. Competitors. We choose approaches iSOIRE (Li et al. 2019a), GenICHARE (Zhang et al. 2018), iSIRE (Li et al. 2020b) and RE2RNN (Jiang et al. 2020) as competitors. iSOIRE learns RSOIREs and GenICHARE learns ICHAREs from positive strings only. iSIRE learns SIREs from both positive and negative strings. RE2RNN embeds a weighted FSA to improve the performance on text classification. It can also learn an automaton if we randomly initialize the parameters. Thus we also compare the performance between SOIREDL and RE2RNN. Settings. We implement iSOIRE, GenICHARE, iSIRE according to their papers, and reuse the source code of RE2RNN. The hyper-parameters of RE2RNN are set as default, except that the number of states is set to $100$ and the threshold in interpretation to $0.12$ for achieving the best accuracy. We train SOIREDL with the AdamW optimizer. The hyper-parameters of SOIREDL are set as follows: the bounded size $T$ is $4|\Sigma|-2$ according to Proposition 8, the batch size is $64$, the regularization coefficient $\lambda$ is $0$, and the beam width $\beta$ is $500$. The optimal $\lambda$ is selected from $\\{0,10^{-3},10^{-2},10^{-1},1,10\\}$ and $\beta$ from $\\{10,50,100,300,500,1000\\}$ for maximizing the accuracy on the validation set. iSOIRE, GenICHARE and iSIRE use the union of the training and validation sets for learning. All experiments were conducted on a Linux machine equipped with an Intel Xeon Gold 6248R processor with $126$ GB RAM and a single NVIDIA A$100$. We train SOIREDL with five learning rates $0.01,0.05,0.1,0.15,0.2$ and select the SOIRE achieving the best accuracy on the validation set. Therefore, the running time of SOIREDL is the sum of training time and interpretation time with five learning rates. The time limit for each approach is set to $5000$ seconds. ###### Proposition 8. For any SOIRE $r$ over $\Sigma$, there exists another SOIRE $r^{\prime}$ over $\Sigma$ and a faithful encoding $\theta$ with length $4|\Sigma|-2$ such that $\texttt{PreForm}(r^{\prime})=\texttt{Enc2Pre}(\theta)$ and $\\{s|r^{\prime}\models s\\}=\\{s|r\models s\\}$. Evaluate metrics. We use accuracy on the test set to evaluate the performance of all approaches. For SOIREDL and RE2RNN, We introduce faithfulness defined as $\frac{N_{=}}{\left|\Pi^{+}\right|+\left|\Pi^{-}\right|}$ to evaluate the consistency between the neural network and the interpreted SOIRE (SOIREDL) or automata (RE2RNN), where $N_{=}$ is the number of test strings that the neural network and the SOIRE or automata predicts the same label. | Positive and negative strings | Positive strings only ---|---|--- Data | iSI | RE2RNN | SOIREDL | iSO | GenIC | iSI | RE2RNN | SOIREDL set | RE | IRE | HARE | RE 1 | 86.0 | 52.4 (94.4) | 100.0 (100.0) | 89.4 | 89.4 | 83.8 | 48.0 (50.0) | 57.0 (57.0) 2 | 77.4 | 48.8 (91.4) | 100.0 (100.0) | 100.0 | 100.0 | 100.0 | 50.4 (50.0) | 100.0 (100.0) 3 | 90.8 | 50.6 (77.4) | 100.0 (100.0) | 100.0 | 97.2 | 95.6 | 50.6 (49.6) | 65.4 (65.4) 4 | 72.4 | 49.0 (80.2) | 99.6 (73.4) | 73.8 | 72.6 | 73.4 | 49.6 (49.8) | 61.4 (61.4) 5 | 90.8 | 52.6 (91.4) | 58.2 (87.2) | 100.0 | 100.0 | 86.2 | 49.6 (50.2) | 52.8 (52.8) 6 | 77.8 | 51.6 (62.2) | 93.4 (93.0) | 100.0 | 100.0 | 70.4 | 50.2 (50.0) | 52.6 (52.6) 7 | 81.2 | 52.2 (95.8) | 99.2 (99.2) | 100.0 | 96.4 | 89.0 | 49.2 (49.8) | 69.4 (69.4) 8 | 76.2 | 49.8 (88.0) | 100.0 (100.0) | 100.0 | 100.0 | 100.0 | 49.8 (50.0) | 100.0 (100.0) 9 | 93.8 | 44.2 (48.4) | 98.4 (98.4) | 99.8 | 98.8 | 98.0 | 47.2 (49.6) | 81.6 (81.6) 10 | 94.8 | 50.2 (89.8) | 99.8 (100.0) | 99.8 | 99.8 | 82.8 | 44.4 (50.2) | 61.2 (61.2) 11 | 91.0 | 52.4 (88.6) | 89.2 (91.0) | 100.0 | 100.0 | 91.4 | 47.6 (50.2) | 57.4 (57.4) 12 | 78.6 | 51.2 (98.4) | 100.0 (100.0) | 100.0 | 100.0 | 100.0 | 50.8 (49.4) | 100.0 (100.0) 13 | 96.6 | 52.8 (50.2) | 100.0 (100.0) | 100.0 | 100.0 | 83.6 | 51.2 (49.6) | 67.0 (67.0) 14 | 74.4 | 50.0 (79.0) | 84.8 (71.6) | 70.0 | 74.4 | 68.8 | 49.0 (50.0) | 54.4 (54.4) 15 | 95.2 | 54.0 (49.8) | 96.2 (100.0) | 100.0 | 100.0 | 100.0 | 47.8 (50.2) | 66.2 (66.2) 16 | 96.8 | 49.0 (71.4) | 94.8 (100.0) | 100.0 | 100.0 | 75.4 | 49.4 (50.0) | 75.4 (75.4) 17 | 91.0 | 46.2 (87.2) | 100.0 (100.0) | 100.0 | 100.0 | 100.0 | 50.6 (50.0) | 100.0 (100.0) 18 | 81.0 | 42.8 (78.8) | 87.4 (99.2) | 87.4 | 100.0 | 82.0 | 40.2 (50.2) | 53.0 (53.0) 19 | 88.2 | 50.0 (63.8) | 100.0 (93.4) | 100.0 | 100.0 | 88.0 | 50.4 (50.0) | 54.6 (54.6) 20 | 93.4 | 45.2 (54.4) | 83.4 (90.0) | 100.0 | 99.2 | 95.8 | 47.4 (50.0) | 56.0 (51.2) 21 | 69.8 | 48.8 (96.2) | 100.0 (100.0) | 71.2 | 71.2 | 71.2 | 49.8 (50.4) | 71.2 (71.2) 22 | 90.6 | 47.6 (50.6) | 100.0 (100.0) | 100.0 | 100.0 | 91.4 | 50.0 (50.2) | 58.2 (58.2) 23 | 85.6 | 32.8 (84.4) | 86.8 (65.2) | 90.0 | 90.0 | 88.0 | 50.0 (48.8) | 56.8 (56.8) 24 | 69.4 | 49.0 (57.2) | 74.6 (76.8) | 77.6 | 76.0 | 71.4 | 47.6 (50.0) | 54.2 (54.2) 25 | 67.8 | 54.2 (93.4) | 99.8 (74.6) | 70.0 | 70.0 | 70.0 | 50.2 (50.0) | 69.6 (69.6) 26 | 80.2 | 50.4 (50.4) | 63.8 (98.2) | 100.0 | 100.0 | 85.2 | 51.0 (49.6) | 63.8 (63.8) 27 | 92.0 | 52.0 (91.6) | 96.4 (96.4) | 100.0 | 100.0 | 97.4 | 52.2 (50.2) | 67.6 (67.6) 28 | 65.0 | 54.8 (95.2) | 100.0 (98.6) | 65.6 | 65.6 | 65.6 | 50.0 (50.0) | 65.6 (65.6) 29 | 93.4 | 56.2 (75.8) | 97.6 (97.2) | 100.0 | 99.8 | 81.2 | 49.2 (49.8) | 54.2 (54.2) 30 | 60.4 | 41.4 (61.6) | 78.8 (79.6) | 61.0 | 61.0 | 60.4 | 49.8 (49.8) | 54.8 (54.8) Avg. | 83.4 | 49.4 (76.6) | 92.7 (92.8) | 91.9 | 92.0 | 84.9 | 49.1 (49.9) | 66.7 (66.6) Table 3: Accuracy (%) on noise-free data with best results in bold. For X(Y), X denotes the accuracy of the learnt SOIRE or automaton, and Y the accuracy of the neural network. Comparison on noise-free data. The results of different approaches on noise- free data are shown in Table 3. For learning from both positive and negative strings, SOIREDL outperforms iSIRE and RE2RNN on almost all datasets, achieves comparable performance with iSOIRE and GenICHARE (both of which learn from positive strings only), and achieves the highest average accuracy among all approaches. Regarding the accuracy of the intermediate neural network, SOIREDL is also superior to RE2RNN. These results show that SOIREDL achieves the SOTA performance on noise-free data. Comparison on noisy data. The average accuracy of different approaches on noisy data is shown in Figure 2 (a). The performance of iSOIRE and GenICHARE drops sharply when noise is present, suggesting that they are not robust on noisy data. The average accuracy of RE2RNN also decreases quickly when the noise level increases, and so does the neural network of it. Both SOIREDL and iSIRE perform well on noisy data. The average accuracy of SOIREDL slightly decreases when the noise level increases, but it still keeps higher than others at all noise levels. This suggests that SOIREDL is the most robust on noisy data. Figure 2: Average accuracy (%) on test sets at different noise levels $\delta$. SOIREDL-ipt and RE2RNN-ipt represent the learnt SOIREs or automata, whereas SOIREDL-net and RE2RNN-net represent the neural networks. (a) Positive and negative strings. (b) Positive strings only. Comparison in terms of faithfulness. The average faithfulness of SOIREDL and RE2RNN is shown in Figure 3. Obviously, the faithfulness of SOIREDL is much higher than that of RE2RNN and keeps over $80\%$ at all noise levels, suggesting that the neural network of SOIREDL and its interpreted SOIRE are more consistent in performing SOIRE matching. Figure 3: Average faithfulness (%) of SOIREDL and RE2RNN on test sets at different noise levels $\delta$. Case study. We pick one RE from each subclass of SOIREs considered in our experiments to show their learnt results, reported in Table 4. Although the ground-truth REs and the SOIREs learnt by SOIREDL are not exactly the same in the subclasses ICHARE and RSOIRE, they still belong to the same subclass. These results show that SOIREDL is able to learn different subclasses of SOIREs. Subclass | Dataset | Ground-truth SOIREs | SOIREDL ---|---|---|--- SIRE | 13 | $a^{?}\&b^{*}\&c^{?}$ | $a^{?}\&c^{?}\&b^{*}$ ICHARE | 22 | $\left(a|b\right)^{*}c^{*}d^{*}$ | $\left(a^{*}\&b^{*}\right)c^{*}d^{*}$ RSOIRE | 3 | $a^{+}|\left(b|c\right)^{*}|d^{+}$ | $\left(b^{+}|c\right)^{*}|a^{*}|d^{*}$ SOIRE | 1 | $\left(\left(a|b\right)c^{*}\right)^{+}d$ | $\left(\left(a|b\right)c^{*}\right)^{+}d$ Table 4: The ground-truth SOIREs and the SOIREs learnt by SOIREDL on different subclasses of SOIREs. We also analyse the relation between the accuracy of the SOIRE learnt by SOIREDL and the size of the ground-truth SOIRE. Figure 4 shows that the accuracy decreases when the size of the ground-truth SOIRE increases. This may be due to the difficulty for a neural network to capture the long-distance dependency in SOIRE matching. Figure 4: The relation of the accuracy (%) of a SOIRE learnt by SOIREDL and the size $|r|$ of the ground-truth SOIRE $r$. The necessity for using negative strings. The results of learning from positive strings only are shown in Table 3 and Figure 2 (b). Neural networks do not perform well because they prefer to classify unseen strings as positive ones when training on positive strings only. Even worse, when noise is present, the accuracy of any competitor drops sharply to no more than $60\%$ as shown in Figure 2 (b). This suggests that negative strings are crucial in effectively learning with noisy data, possibly because they infer what kinds of strings that the target SOIRE cannot match. By learning from positive and negative strings, both SOIREDL and RE2RNN achieve significantly better performance; in particular, SOIREDL outperforms iSOIRE and GenICHARE in terms of average accuracy, as shown in Table 3 and Figure 2 (a). ## Conclusion and Future Work In this paper, we have proposed a noise-tolerant differentiable learning approach SOIREDL and a matching algorithm SOIRETM based on the syntax tree for SOIREs. The neural network introduced in SOIREDL simulates the matching process of SOIRETM. Theoretically, the faithful encodings learnt by SOIREDL one-to-one correspond to SOIREs for a bounded size. Experimental results have demonstrated higher performance compared with the SOTA approaches. In the future work, we will tackle the problem of long dependency in SOIRE matching and extend our approach to other subclasses of REs. ## Acknowledgments We thank Kunxun Qi for discussion on the paper and anonymous referees for helpful comments. This paper was supported by the National Natural Science Foundation of China (No. 62276284, 61976232, 61876204), the National Key Research and Development Program of China (No. 2021YFA1000504), Guangdong Basic and Applied Basic Research Foundation (No. 2022A1515011355), Guangzhou Science and Technology Project (No. 202201011699), Guizhou Science Support Project (No. 2022-259), as well as Humanities and Social Science Research Project of Ministry of Education (No. 18YJCZH006). ## References * Bex et al. (2006) Bex, G. J.; Neven, F.; Schwentick, T.; and Tuyls, K. 2006. Inference of Concise DTDs from XML Data. In _VLDB_ , 115–126. * Bojanczyk et al. (2006) Bojanczyk, M.; Muscholl, A.; Schwentick, T.; Segoufin, L.; and David, C. 2006. Two-Variable Logic on Words with Data. In _LICS_ , 7–16. * Cohen, Yang, and Mazaitis (2020) Cohen, W. W.; Yang, F.; and Mazaitis, K. 2020. TensorLog: A Probabilistic Database Implemented Using Deep-Learning Infrastructure. _J. Artif. Intell. Res._ , 67: 285–325. * Colazzo et al. (2013) Colazzo, D.; Ghelli, G.; Pardini, L.; and Sartiani, C. 2013. Efficient asymmetric inclusion of regular expressions with interleaving and counting for XML type-checking. _Theor. Comput. Sci._ , 492: 88–116. * Freydenberger and Kötzing (2015) Freydenberger, D. D.; and Kötzing, T. 2015. Fast Learning of Restricted Regular Expressions and DTDs. _Theory Comput. Syst._ , 57(4): 1114–1158. * Galassi and Giordana (2005) Galassi, U.; and Giordana, A. 2005. Learning Regular Expressions from Noisy Sequences. In _SARA_ , volume 3607, 92–106. * Gao et al. (2022) Gao, K.; Inoue, K.; Cao, Y.; and Wang, H. 2022. Learning First-Order Rules with Differentiable Logic Program Semantics. In _IJCAI_ , 3008–3014. * Gischer (1981) Gischer, J. L. 1981. Shuffle Languages, Petri Nets, and Context-Sensitive Grammars. _Commun. ACM_ , 24(9): 597–605. * Huang et al. (2021) Huang, J.; Li, Z.; Chen, B.; Samel, K.; Naik, M.; Song, L.; and Si, X. 2021. Scallop: From Probabilistic Deductive Databases to Scalable Differentiable Reasoning. In _NeurIPS_ , 25134–25145. * Jiang, Jin, and Tu (2021) Jiang, C.; Jin, Z.; and Tu, K. 2021. Neuralizing Regular Expressions for Slot Filling. In _EMNLP_ , 9481–9498. * Jiang et al. (2020) Jiang, C.; Zhao, Y.; Chu, S.; Shen, L.; and Tu, K. 2020. Cold-Start and Interpretability: Turning Regular Expressions into Trainable Recurrent Neural Networks. In _EMNLP_ , 3193–3207. * Kearns and Li (1988) Kearns, M. J.; and Li, M. 1988. Learning in the Presence of Malicious Errors (Extended Abstract). In _STOC_ , 267–280. * Kuhlmann and Satta (2009) Kuhlmann, M.; and Satta, G. 2009. Treebank Grammar Techniques for Non-Projective Dependency Parsing. In _EACL_ , 478–486. * Li et al. (2020a) Li, Y.; Cao, J.; Chen, H.; Ge, T.; Xu, Z.; and Peng, Q. 2020a. FlashSchema: Achieving High Quality XML Schemas with Powerful Inference Algorithms and Large-scale Schema Data. In _ICDE_ , 1962–1965. * Li et al. (2020b) Li, Y.; Chen, H.; Zhang, L.; Huang, B.; and Zhang, J. 2020b. Inferring Restricted Regular Expressions with Interleaving from Positive and Negative Samples. In _PAKDD_ , volume 12085, 769–781. * Li et al. (2019a) Li, Y.; Chen, H.; Zhang, X.; and Zhang, L. 2019a. An effective algorithm for learning single occurrence regular expressions with interleaving. In _IDEAS_ , 24:1–24:10. * Li et al. (2018) Li, Y.; Chu, X.; Mou, X.; Dong, C.; and Chen, H. 2018. Practical Study of Deterministic Regular Expressions from Large-scale XML and Schema Data. In _IDEAS_ , 45–53. * Li et al. (2021) Li, Y.; Li, S.; Xu, Z.; Cao, J.; Chen, Z.; Hu, Y.; Chen, H.; and Cheung, S. 2021\. TRANSREGEX: Multi-modal Regular Expression Synthesis by Generate-and-Repair. In _ICSE_ , 1210–1222. * Li et al. (2019b) Li, Y.; Zhang, X.; Cao, J.; Chen, H.; and Gao, C. 2019b. Learning k-Occurrence Regular Expressions with Interleaving. In _DASFAA_ , volume 11447, 70–85. * Makoto and Clark (2003) Makoto, M.; and Clark, J. 2003. RELAX NG. https://relaxng.org/. * Martens et al. (2017) Martens, W.; Neven, F.; Niewerth, M.; and Schwentick, T. 2017. BonXai: Combining the Simplicity of DTD with the Expressiveness of XML Schema. _ACM Trans. Database Syst._ , 42(3): 15:1–15:42. * Mayer and Stockmeyer (1994) Mayer, A. J.; and Stockmeyer, L. J. 1994. The Complexity of Word Problems - This Time with Interleaving. _Inf. Comput._ , 115(2): 293–311. * Mensch and Blondel (2018) Mensch, A.; and Blondel, M. 2018. Differentiable Dynamic Programming for Structured Prediction and Attention. In _ICML_ , volume 80, 3459–3468. * Minervini et al. (2020a) Minervini, P.; Bosnjak, M.; Rocktäschel, T.; Riedel, S.; and Grefenstette, E. 2020a. Differentiable Reasoning on Large Knowledge Bases and Natural Language. In _AAAI_ , 5182–5190. * Minervini et al. (2020b) Minervini, P.; Riedel, S.; Stenetorp, P.; Grefenstette, E.; and Rocktäschel, T. 2020b. Learning Reasoning Strategies in End-to-End Differentiable Proving. In _ICML_ , volume 119, 6938–6949. * Nivre (2009) Nivre, J. 2009. Non-Projective Dependency Parsing in Expected Linear Time. In _ACL_ , 351–359. * Rocktäschel and Riedel (2017) Rocktäschel, T.; and Riedel, S. 2017. End-to-end Differentiable Proving. In _NeurIPS_ , 3788–3800. * Sadeghian et al. (2019) Sadeghian, A.; Armandpour, M.; Ding, P.; and Wang, D. Z. 2019. DRUM: End-To-End Differentiable Rule Mining On Knowledge Graphs. In _NeurIPS_ , 15321–15331. * Wang et al. (2019) Wang, P.; Donti, P. L.; Wilder, B.; and Kolter, J. Z. 2019. SATNet: Bridging deep learning and logical reasoning using a differentiable satisfiability solver. In _ICML_ , volume 97, 6545–6554. * Wang et al. (2020) Wang, P.; Stepanova, D.; Domokos, C.; and Kolter, J. Z. 2020. Differentiable learning of numerical rules in knowledge graphs. In _ICLR_. * Wang (2021a) Wang, X. 2021a. Inferring Deterministic Regular Expression with Unorder and Counting. In _DASFAA_ , volume 12682, 235–252. * Wang (2021b) Wang, X. 2021b. Learning Finite Automata with Shuffle. In _PAKDD_ , volume 12713, 308–320. * Wang (2022) Wang, X. 2022. Membership Algorithm for Single-Occurrence Regular Expressions with Shuffle and Counting. In _DASFAA_ , volume 13245, 526–542. * Wang and Chen (2018) Wang, X.; and Chen, H. 2018. Inferring Deterministic Regular Expression with Counting. In _Conceptual Modeling - 37th International Conference, ER_ , volume 11157, 184–199. * Wang and Chen (2020) Wang, X.; and Chen, H. 2020. Inferring Deterministic Regular Expression with Unorder. In _SOFSEM_ , volume 12011, 325–337. * Wang and Zhang (2021) Wang, X.; and Zhang, X. 2021. Discovering an Algorithm Actually Learning Restricted Single Occurrence Regular Expression with Interleaving. _CoRR_ , abs/2103.10546. * Yang, Yang, and Cohen (2017) Yang, F.; Yang, Z.; and Cohen, W. W. 2017. Differentiable Learning of Logical Rules for Knowledge Base Reasoning. In _NeurIPS_ , 2319–2328. * Yang and Song (2020) Yang, Y.; and Song, L. 2020. Learn to Explain Efficiently via Neural Logic Inductive Learning. In _ICLR_. * Zhang et al. (2018) Zhang, X.; Li, Y.; Cui, F.; Dong, C.; and Chen, H. 2018. Inference of a Concise Regular Expression Considering Interleaving from XML Documents. In _PAKDD_ , volume 10938, 389–401. ## Appendix A ### From prefix notation to the infix notation of a SOIRE Algorithm 3 is a function to obtain the infix notaiton of a SOIRE from the prefix notaiton. Algorithm 3 From prefix notation to the infix notation of a SOIRE. Input: The prefix notation of the SOIRE $r_{p}$. Output: The infix notation of the SOIRE $r_{i}$. 1:Let $\Gamma$ be an empty stack; 2:for $i\leftarrow|r_{p}|$ downto $1$ do 3: Let $o^{i}$ be the $i$ th character of $r_{p}$; 4: if $o^{i}\in\Sigma$ then 5: Push $(o^{i})$ into stack $\Gamma$; 6: if $o^{i}\in\\{?,*,+\\}$ then 7: Pop an element from stack $\Gamma$, denoted as $r$; 8: Push $\text{concat}(`(^{\prime},r,o^{i},`)^{\prime})$ into stack $\Gamma$; 9: if $o^{i}\in\\{\cdot,\&,|\\}$ then 10: Pop an element from stack $\Gamma$, denoted as $r_{2}$; 11: Pop an element from stack $\Gamma$, denoted as $r_{1}$; 12: Push $\text{concat}(`(^{\prime},r_{1},o^{i},r_{2},`)^{\prime})$ into stack $\Gamma$; 13:Pop an element from stack $\Gamma$, denoted as $r_{i}$; 14:Return $r_{i}$; ### Faithful encoding interpreter Algorithm 4 Faithful encoding interpreter Enc2Pre. Input: The faithful encoding of an SOIRE, $\theta=(w,u)$. Output: The prefix notation of an SOIRE $r$. 1:Let $r$ be a empty string; 2:Let $T$ be bounded size of $r$; 3:for $t\leftarrow 1$ to $T$ do 4: if $w^{t}_{\text{none}}=1$ then 5: Break; 6: for all $a\in\mathbb{B}\setminus\\{\text{none}\\}$ do 7: if $w^{t}_{a}=1$ then 8: $r\leftarrow\text{concat}(a,r)$; 9:Return $r$; ## Appendix B ### Proof of Theorem 1 Proof: 1. 1. If $r\models s$, then $filter(s,\alpha(r))=s$ and $r\models filter(s,\alpha(r))$. Because $r\models s$, $\alpha(r)$ contains all the symbols in $s$ and $filter(s,\alpha(r))=s$ (Definition of SOIRE matching). Because $r\models s$ and $filter(s,\alpha(r))=s$, $r\models filter(s,\alpha(r))$. 2. 2. If $filter(s,\alpha(r))=s$ and $r\models filter(s,\alpha(r))$, than $r\models s$. Because $r\models filter(s,\alpha(r))$ and $filter(s,\alpha(r))=s$, $r\models s$. Theorem 1 has been proved. ### Proof of Lemma 2 Proof: We just prove that is correct for $r_{1}$ and the proof for $r_{2}$ is nearly the same. 1. 1. If $r_{1}\models filter(s,\alpha(r))$, then $filter(s,\alpha(r_{1}))=filter(s,\alpha(r))$ and $r_{1}\models filter(s,\alpha(r_{1}))$. Because $r_{1}\models filter(s,\alpha(r))$, $\alpha(r_{1})$ contains all the symbols in $filter(s,\alpha(r))$ (Definition of SOIRE matching). Because $\alpha(r_{1})\subseteq\alpha(r)$, $filter(s,\alpha(r_{1}))=filter(s,\alpha(r))$. Therefore, $r_{1}\models filter(s,\alpha(r_{1}))$. 2. 2. If $filter(s,\alpha(r_{1}))=filter(s,\alpha(r))$ and $r_{1}\models filter(s,\alpha(r_{1}))$, then $r_{1}\models filter(s,\alpha(r))$. Because $filter(s,\alpha(r_{1}))=filter(s,\alpha(r))$ and $r_{1}\models filter(s,\alpha(r_{1}))$, $r_{1}\models filter(s,\alpha(r))$. Lemma 2 has been proved. ### Proof of Theorem 3 Theorem 1 shows that $r\models s$ iff $filter(s,\alpha(r))=s$ and $r\models filter(s,\alpha(r))$. In Algorithm 1, Line 1 checks if $filter(s,\alpha(r))=s$. Therefore, we need to prove that $g_{1,|s|}^{1}=1$ iff $r\models filter(s,\alpha(r))$. Algorithm 1 calculate $g_{i,j}^{t}$ for $s_{i,j}$ from short to long and for $r^{t}$ from bottom to top of the syntax tree. We apply mathematical induction to the length of the substring $|s_{i,j}|$ and structural induction to $r^{t}$. Proof: Mathematical induction base: $|s_{i,j}|=l=0$, i.e. $s_{i,j}=\epsilon$ and $i=1,j=0$. Structural induction: Each case is corresponding to one statement in Line 8-20 in Algorithm 1. 1. 1. $r^{t}=a\in\alpha(r)$. $r^{t}=a$ does not match $filter(s_{1,0},\\{a\\})=\epsilon$, so $g_{1,0}^{t}=1[filter(s_{1,0},\\{a\\})=a]=0$ is correct. 2. 2. $r^{t}=(r^{t+1})^{?}$. $(r^{t+1})^{?}\models filter(s_{1,0},\alpha((r^{t+1})^{?}))=\epsilon$, so $g_{1,0}^{t}=g_{1,0}^{t+1}\ \vee\ 1[filter(s_{1,0},\alpha((r^{t+1})^{?}))=\epsilon]=1$ is correct. 3. 3. $r^{t}=(r^{t+1})^{*}$. $(r^{t+1})^{*}\models filter(s_{1,0},\alpha((r^{t+1})^{*}))=\epsilon$, so $g_{1,0}^{t}=1[filter(s_{1,0},\alpha(r^{t}))=\epsilon]\vee g_{1,0}^{t+1}\vee\vee_{k=i}^{j-1}(g_{i,k}^{t}\ \wedge\ g_{k+1,j}^{t+1})=1$ is correct. 4. 4. $r^{t}=(r^{t+1})^{+}$. Because $i=1>j-1=-1$, $\vee_{k=i}^{j-1}(g_{i,k}^{t}\ \wedge\ g_{k+1,j}^{t+1})=0$. Therefore $g_{1,0}^{t}=g_{1,0}^{t+1}\vee\vee_{k=i}^{j-1}(g_{i,k}^{t}\ \wedge\ g_{k+1,j}^{t+1})=g_{1,0}^{t+1}$. Because $g_{1,0}^{t}=1$ iff $(r^{t+1})^{+}\models\epsilon$, and $(r^{t+1})^{+}\models\epsilon$ iff $r^{t+1}\models\epsilon$, $g_{1,0}^{t}=1$ iff $r^{t+1}\models\epsilon$. Because $r^{t+1}\models\epsilon$ iff $g_{1,0}^{t+1}=1$ (Structural inductive hypothesis), $g_{1,0}^{t}=1$ iff $g_{1,0}^{t+1}=1$. Therefore, $g_{1,0}^{t}=g_{1,0}^{t+1}\vee\vee_{k=i}^{j-1}(g_{i,k}^{t}\ \wedge\ g_{k+1,j}^{t+1})$ is correct. 5. 5. $r^{t}=r^{t+1}\cdot r^{\eta^{t}}$. Because $i=1>j-1=-1$, $\vee_{k=i}^{j-1}(1[filter(s_{i,k},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{i,k},\alpha(r^{t+1}))]\ \wedge\ g_{i,k}^{t+1}\ \wedge\ 1[filter(s_{k+1,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{k+1,j},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{k+1,j}^{\eta^{t}})=0$. In addition, $filter(s_{1,0},\alpha(r^{t+1}))=\epsilon,filter(s_{1,0},\alpha(r^{\eta^{t}}))=\epsilon$, i.e. $filter(s_{1,0},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{1,0},\alpha(r^{t+1}))=filter(s_{1,0},\alpha(r^{\eta^{t}}))=\epsilon$. Therefore, $g_{1,0}^{t}=(1[filter(s_{1,0},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{1,0},\alpha(r^{t+1}))]\ \wedge\ g_{1,0}^{t+1}\ \wedge\ g_{1,0}^{\eta^{t}})\ \vee\ (1[filter(s_{1,0},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{1,0},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{1,0}^{\eta^{t}}\ \wedge\ g_{1,0}^{t+1})\vee\vee_{k=i}^{j-1}(1[filter(s_{i,k},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{i,k},\alpha(r^{t+1}))]\ \wedge\ g_{i,k}^{t+1}\ \wedge\ 1[filter(s_{k+1,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{k+1,j},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{k+1,j}^{\eta^{t}})=g_{1,0}^{t+1}\ \wedge\ g_{1,0}^{\eta^{t}}$. Because $g_{1,0}^{t}=1$ iff $r^{t+1}\cdot r^{\eta^{t}}\models\epsilon$, and $r^{t+1}\cdot r^{\eta^{t}}\models\epsilon$ iff both $r^{t+1},r^{\eta^{t}}\models\epsilon$, $g_{1,0}^{t}=1$ iff both $r^{t+1},r^{\eta^{t}}\models\epsilon$. Because $r^{t+1}$ (reps. $r^{\eta^{t}}$) matches $\epsilon$ iff $g_{1,0}^{t+1}=1$ (reps. $g_{1,0}^{\eta^{t}}=1$) (Structural inductive hypothesis), $g_{1,0}^{t}=1$ iff $(g_{1,0}^{t+1}\wedge g_{1,0}^{\eta^{t}})=1$. Therefore, $g_{1,0}^{t}=(1[filter(s_{1,0},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{1,0},\alpha(r^{t+1}))]\ \wedge\ g_{1,0}^{t+1}\ \wedge\ g_{1,0}^{\eta^{t}})\ \vee\ (1[filter(s_{1,0},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{1,0},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{1,0}^{\eta^{t}}\ \wedge\ g_{1,0}^{t+1})\vee\vee_{k=i}^{j-1}(1[filter(s_{i,k},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{i,k},\alpha(r^{t+1}))]\ \wedge\ g_{i,k}^{t+1}\ \wedge\ 1[filter(s_{k+1,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{k+1,j},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{k+1,j}^{\eta^{t}})$ is correct. 6. 6. $r^{t}=r^{t+1}\&r^{\eta^{t}}$. Because $g_{1,0}^{t}=1$ iff $r^{t+1}\&r^{\eta^{t}}\models\epsilon$, and $r^{t+1}\&r^{\eta^{t}}\models\epsilon$ iff both $r^{t+1},r^{\eta^{t}}\models\epsilon$, $g_{1,0}^{t}=1$ iff both $r^{t+1},r^{\eta^{t}}\models\epsilon$. Because $r^{t+1}$ (reps. $r^{\eta^{t}}$) matches $\epsilon$ iff $g_{1,0}^{t+1}=1$ (reps. $g_{1,0}^{\eta^{t}}=1$) (Structural inductive hypothesis), $g_{1,0}^{t}=1$ iff $(g_{1,0}^{t+1}\wedge g_{1,0}^{\eta^{t}})=1$. Therefore, $g_{1,0}^{t}=g_{1,0}^{t+1}\ \wedge\ g_{1,0}^{\eta^{t}}$ is correct. 7. 7. $r^{t}=r^{t+1}|r^{\eta^{t}}$. Because $filter(s_{1,0},\alpha(r^{t+1}))=\epsilon,filter(s_{1,0},\alpha(r^{\eta^{t}}))=\epsilon$, i.e. $filter(s_{1,0},\alpha(r^{t+1}|r^{\eta^{t}}))=filter(s_{1,0},\alpha(r^{t+1}))=filter(s_{1,0},\alpha(r^{\eta^{t}}))=\epsilon$, $g_{1,0}^{t}=(1[filter(s_{1,0},\alpha(r^{t+1}|r^{\eta^{t}}))=filter(s_{1,0},\alpha(r^{t+1}))]\ \wedge\ g_{1,0}^{t+1})\ \vee\ (1[filter(s_{1,0},\alpha(r^{t+1}|r^{\eta^{t}}))=filter(s_{1,0},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{1,0}^{\eta^{t}})=g_{1,0}^{t+1}\ \vee\ g_{1,0}^{\eta^{t}}$. Because $g_{1,0}^{t}=1$ iff $r^{t+1}|r^{\eta^{t}}\models\epsilon$, and $r^{t+1}|r^{\eta^{t}}\models\epsilon$ iff $r^{t+1}$ or $r^{\eta^{t}}\models\epsilon$, $g_{1,0}^{t}=1$ iff $r^{t+1}$ or $r^{\eta^{t}}\models\epsilon$. Because $r^{t+1}$ (reps. $r^{\eta^{t}}$) matches $\epsilon$ iff $g_{1,0}^{t+1}=1$ (reps. $g_{1,0}^{\eta^{t}}=1$) (Structural inductive hypothesis), $g_{1,0}^{t}=1$ iff $(g_{1,0}^{t+1}\vee g_{1,0}^{\eta^{t}})=1$. Therefore, $g_{1,0}^{t}=(1[filter(s_{1,0},\alpha(r^{t+1}|r^{\eta^{t}}))=filter(s_{1,0},\alpha(r^{t+1}))]\ \wedge\ g_{1,0}^{t+1})\ \vee\ (1[filter(s_{1,0},\alpha(r^{t+1}|r^{\eta^{t}}))=filter(s_{1,0},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{1,0}^{\eta^{t}})$ is correct. Mathematical induction step: $|s_{i,j}|=l>0$. Suppose that for all $s_{i^{\prime},j^{\prime}}$ ($|s_{i^{\prime},j^{\prime}}|=l^{\prime}<l$) and $1\leq t\leq|r|$, $g_{i^{\prime},j^{\prime}}^{t}=1$ iff $r^{t}\models s_{i^{\prime},j^{\prime}}$. Structural induction: Each case is corresponding to one statement in Line 8-20 in Algorithm 1. 1. 1. $r^{t}=a\in\alpha(r)$. Because $g_{i,j}^{t}=1$ iff $r^{t}=a\in\alpha(r)\models filter(s_{i,j},\\{a\\})$, i.e. $filter(s_{i,j},\\{a\\})=a$, so $g_{i,j}^{t}=1$ iff $filter(s_{i,j},\\{a\\})=a$. Therefore, $g_{i,j}^{t}=1[filter(s_{i,j},\\{a\\})=a]$ is correct. 2. 2. $r^{t}=(r^{t+1})^{?}$. $g_{i,j}^{t}=1$ iff $(r^{t+1})^{?}\models filter(s_{i,j},\alpha((r^{t+1})^{?}))$, which is also equivalent to that $r^{t+1}\models filter(s_{i,j},\alpha((r^{t+1})^{?}))$ or $filter(s_{i,j},\alpha((r^{t+1})^{?}))=\epsilon$. Because $\alpha((r^{t+1})^{?})=\alpha(r^{t+1})$, $filter(s_{i,j},\alpha((r^{t+1})^{?}))=filter(s_{i,j},\alpha(r^{t+1}))$. Therefore, $(r^{t+1})^{?}\models filter(s_{i,j},\alpha((r^{t+1})^{?}))$ iff $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$ or $filter(s_{i,j},\alpha((r^{t+1})^{?}))=\epsilon$. $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$ iff $g_{i,j}^{t+1}=1$ (Structural inductive hypothesis), so $g_{i,j}^{t}=1$ iff $g_{i,j}^{t+1}=1$ or $1[filter(s_{i,j},\alpha((r^{t+1})^{?}))=\epsilon]$. Therefore, $g_{i,j}^{t}=g_{i,j}^{t+1}\vee 1[filter(s_{i,j},\alpha((r^{t+1})^{?}))=\epsilon]$ is correct. 3. 3. $r^{t}=(r^{t+1})^{*}$. $g_{i,j}^{t}=1$ iff $(r^{t+1})^{*}\models filter(s_{i,j},\alpha((r^{t+1})^{*}))$, which is also equivalent to that $filter(s_{i,j},\alpha((r^{t+1})^{*}))=\epsilon$ or $r^{t+1}\models filter(s_{i,j},\alpha((r^{t+1})^{*}))$, or $\exists i\leq k<j,(r^{t+1})^{*}\models filter(s_{i,k},\alpha((r^{t+1})^{*}))$ and $r^{t+1}\models filter(s_{k+1,j},\alpha((r^{t+1})^{*}))$. Because $\alpha((r^{t+1})^{*})=\alpha(r^{t+1})$, $filter(s_{i,j},\alpha((r^{t+1})^{*}))=filter(s_{i,j},\alpha(r^{t+1}))$ and $filter(s_{k+1,j},\alpha((r^{t+1})^{*}))=filter(s_{k+1,j},\alpha(r^{t+1}))$. Therefore, $(r^{t+1})^{*}\models filter(s_{i,j},\alpha((r^{t+1})^{*}))$ iff $filter(s_{i,j},\alpha((r^{t+1})^{*}))=\epsilon$ or $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$, or $\exists i\leq k<j,(r^{t+1})^{*}\models filter(s_{i,k},\alpha((r^{t+1})^{*}))$ and $r^{t+1}\models filter(s_{k+1,j},\alpha(r^{t+1}))$. $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$ iff $g_{i,j}^{t+1}=1$ (Structural inductive hypothesis). Because $|s_{i,k}|=k-i+1<l$, $(r^{t+1})^{*}\models filter(s_{i,k},\alpha((r^{t+1})^{*}))$ iff $g_{i,k}^{t}=1$ (Mathematical inductive hypothesis). Because $|s_{k+1,j}|=j-(k+1)+1<l$, $r^{t+1}\models filter(s_{k+1,j},\alpha(r^{t+1}))$ iff $g_{k+1,j}^{t+1}=1$ (Mathematical inductive hypothesis). Therefore, $\exists i\leq k<j,(r^{t+1})^{*}\models filter(s_{i,k},\alpha((r^{t+1})^{*}))$ and $r^{t+1}\models filter(s_{k+1,j},\alpha(r^{t+1}))$ iff $\vee_{k=i}^{j-1}(g_{i,k}^{t}\ \wedge\ g_{k+1,j}^{t+1})=1$. Therefore, $g_{i,j}^{t}=1$ iff $1[filter(s_{i,j},\alpha(r^{t}))=\epsilon]\vee g_{i,j}^{t+1}=1\vee\vee_{k=i}^{j-1}(g_{i,k}^{t}\ \wedge\ g_{k+1,j}^{t+1})=1$. Therefore, $g_{i,j}^{t}=1[filter(s_{i,j},\alpha(r^{t}))=\epsilon]\vee g_{i,j}^{t+1}\vee\vee_{k=i}^{j-1}(g_{i,k}^{t}\ \wedge\ g_{k+1,j}^{t+1})$ is correct. 4. 4. $r^{t}=(r^{t+1})^{+}$. $g_{i,j}^{t}=1$ iff $(r^{t+1})^{+}\models filter(s_{i,j},\alpha((r^{t+1})^{+}))$, which is also equivalent to that $r^{t+1}\models filter(s_{i,j},\alpha((r^{t+1})^{+}))$, or $\exists i\leq k<j,(r^{t+1})^{+}\models filter(s_{i,k},\alpha((r^{t+1})^{+}))$ and $r^{t+1}\models filter(s_{k+1,j},\alpha((r^{t+1})^{+}))$. Because $\alpha((r^{t+1})^{+})=\alpha(r^{t+1})$, $filter(s_{i,j},\alpha((r^{t+1})^{+}))=filter(s_{i,j},\alpha(r^{t+1}))$ and $filter(s_{k+1,j},\alpha((r^{t+1})^{+}))=filter(s_{k+1,j},\alpha(r^{t+1}))$. Therefore, $(r^{t+1})^{+}\models filter(s_{i,j},\alpha((r^{t+1})^{+}))$ iff $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$, or $\exists i\leq k<j,(r^{t+1})^{+}\models filter(s_{i,k},\alpha((r^{t+1})^{+}))$ and $r^{t+1}\models filter(s_{k+1,j},\alpha(r^{t+1}))$. $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$ iff $g_{i,j}^{t+1}=1$ (Structural inductive hypothesis). Because $|s_{i,k}|=k-i+1<l$, $(r^{t+1})^{+}\models filter(s_{i,k},\alpha((r^{t+1})^{+}))$ iff $g_{i,k}^{t}=1$ (Mathematical inductive hypothesis). Because $|s_{k+1,j}|=j-(k+1)+1<l$, $r^{t+1}\models filter(s_{k+1,j},\alpha(r^{t+1}))$ iff $g_{k+1,j}^{t+1}=1$ (Mathematical inductive hypothesis). Therefore, $\exists i\leq k<j,(r^{t+1})^{+}\models filter(s_{i,k},\alpha((r^{t+1})^{+}))$ and $r^{t+1}\models filter(s_{k+1,j},\alpha(r^{t+1}))$ iff $\vee_{k=i}^{j-1}(g_{i,k}^{t}\ \wedge\ g_{k+1,j}^{t+1})=1$. Therefore, $g_{i,j}^{t}=1$ iff $(g_{i,j}^{t+1}\vee\vee_{k=i}^{j-1}(g_{i,k}^{t}\ \wedge\ g_{k+1,j}^{t+1}))=1$. Therefore, $g_{i,j}^{t}=g_{i,j}^{t+1}\vee\vee_{k=i}^{j-1}(g_{i,k}^{t}\ \wedge\ g_{k+1,j}^{t+1})$ is correct. 5. 5. $r^{t}=r^{t+1}\cdot r^{\eta^{t}}$. $g_{i,j}^{t}=1$ iff $r^{t+1}\cdot r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$, which is also equivalent to that $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ and $r^{\eta^{t}}\models\epsilon$, or $r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ and $r^{t+1}\models\epsilon$, or $\exists i\leq k<j,r^{t+1}\models filter(s_{i,k},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ and $r^{\eta^{t}}\models filter(s_{k+1,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$. $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ iff $filter(s_{i,j},\alpha(r^{t+1}))=filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ and $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$ (Lemma 2). $r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ iff $filter(s_{i,j},\alpha(r^{\eta^{t}}))=filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ and $r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{\eta^{t}}))$ (Lemma 2). Therefore, $r^{t+1}\cdot r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ iff $filter(s_{i,j},\alpha(r^{t+1}))=filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ and $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$ and $r^{\eta^{t}}\models\epsilon$, or $filter(s_{i,j},\alpha(r^{\eta^{t}}))=filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ and $r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{\eta^{t}}))$ and $r^{t+1}\models\epsilon$, or $\exists i\leq k<j,\ filter(s_{i,k},\alpha(r^{t+1}))=filter(s_{i,k},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ and $r^{t+1}\models filter(s_{i,k},\alpha(r^{t+1}))$ and $filter(s_{k+1,j},\alpha(r^{\eta^{t}}))=filter(s_{k+1,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ and $r^{\eta^{t}}\models filter(s_{k+1,j},\alpha(r^{\eta^{t}}))$. $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$ iff $g_{i,j}^{t+1}=1$ (Structural inductive hypothesis). $r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{\eta^{t}}))$ iff $g_{i,j}^{\eta^{t}}=1$ (Structural inductive hypothesis). $r^{t+1}$ (reps. $r^{\eta^{t}}$) matches $\epsilon$ iff $g_{1,0}^{t+1}=1$ (reps. $g_{1,0}^{\eta^{t}}=1$) (Mathematical inductive hypothesis). Because $|s_{i,k}|=k-i+1<l$, $r^{t+1}\models filter(s_{i,k},\alpha(r^{t+1}))$ iff $g_{i,k}^{t+1}=1$ (Mathematical inductive hypothesis). Because $|s_{k+1,j}|=j-(k+1)+1<l$, $r^{\eta^{t}}\models filter(s_{k+1,j},\alpha(r^{\eta^{t}}))$ iff $g_{k+1,j}^{\eta^{t}}=1$ (Mathematical inductive hypothesis). Therefore, $\exists i\leq k<j,\ filter(s_{i,k},\alpha(r^{t+1}))=filter(s_{i,k},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ and $r^{t+1}\models filter(s_{i,k},\alpha(r^{t+1}))$ and $filter(s_{k+1,j},\alpha(r^{\eta^{t}}))=filter(s_{k+1,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))$ and $r^{\eta^{t}}\models filter(s_{k+1,j},\alpha(r^{\eta^{t}}))$ iff $\vee_{k=i}^{j-1}(1[filter(s_{i,k},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{i,k},\alpha(r^{t+1}))]\ \wedge\ g_{i,k}^{t+1}\ \wedge\ 1[filter(s_{k+1,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{k+1,j},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{k+1,j}^{\eta^{t}})=1$. Therefore, $g_{i,j}^{t}=1$ iff $(1[filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{i,j},\alpha(r^{t+1}))]\ \wedge\ g_{i,j}^{t+1}\ \wedge\ g_{1,0}^{\eta^{t}})\ \vee\ (1[filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{i,j},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{i,j}^{\eta^{t}}\ \wedge\ g_{1,0}^{t+1})\vee\vee_{k=i}^{j-1}(1[filter(s_{i,k},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{i,k},\alpha(r^{t+1}))]\ \wedge\ g_{i,k}^{t+1}\ \wedge\ 1[filter(s_{k+1,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{k+1,j},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{k+1,j}^{\eta^{t}})=1$. Therefore, $g_{i,j}^{t}=(1[filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{i,j},\alpha(r^{t+1}))]\ \wedge\ g_{i,j}^{t+1}\ \wedge\ g_{1,0}^{\eta^{t}})\ \vee\ (1[filter(s_{i,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{i,j},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{i,j}^{\eta^{t}}\ \wedge\ g_{1,0}^{t+1})\vee\vee_{k=i}^{j-1}(1[filter(s_{i,k},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{i,k},\alpha(r^{t+1}))]\ \wedge\ g_{i,k}^{t+1}\ \wedge\ 1[filter(s_{k+1,j},\alpha(r^{t+1}\cdot r^{\eta^{t}}))=filter(s_{k+1,j},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{k+1,j}^{\eta^{t}})$ is correct. 6. 6. $r^{t}=r^{t+1}\&r^{\eta^{t}}$. $g_{i,j}^{t}=1$ iff $r^{t+1}\&r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{t+1}\&r^{\eta^{t}}))$. Because each symbol only occurs once in an SOIRE and $r^{t},r^{t+1},r^{\eta^{t}}$ are SOIREs, $\alpha(r^{t})=\alpha(r^{t+1})\cup\alpha(r^{\eta^{t}}),\alpha(r^{t+1})\cap\alpha(r^{\eta^{t}})=\emptyset$. Therefore, each symbol in $filter(s_{i,j},\alpha(r^{t+1}\&r^{\eta^{t}}))$ is either in $\alpha(r^{t+1})$ or $\alpha(r^{\eta^{t}})$. $r^{t+1}\&r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{t+1}\&r^{\eta^{t}}))$ iff $filter(s_{i,j},\alpha(r^{t+1}\&r^{\eta^{t}}))$ can be devided into two subsequences and $r^{t+1}$ matches one subsequence and $r^{\eta^{t}}$ matches the other. Because each symbol in $filter(s_{i,j},\alpha(r^{t+1}\&r^{\eta^{t}}))$ is either in $\alpha(r^{t+1})$ or $\alpha(r^{\eta^{t}})$, one subsequence consists of all the symbols in $\alpha(r^{t+1})$, i.e. $filter(s_{i,j},\alpha(r^{t+1}))$, which is matched by $r^{t+1}$, and the other subsequence consists of all the symbols in $\alpha(r^{\eta^{t}})$, i.e. $filter(s_{i,j},\alpha(r^{\eta^{t}}))$, which is matched by $r^{\eta^{t}}$. Therefore, $r^{t+1}\&r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{t+1}\&r^{\eta^{t}}))$ iff $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$ and $r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{\eta^{t}}))$. $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$ iff $g_{i,j}^{t+1}=1$ (Structural inductive hypothesis). $r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{\eta^{t}}))$ iff $g_{i,j}^{\eta^{t}}=1$ (Structural inductive hypothesis). Therefore $g_{i,j}^{t}=1$ iff $(g_{i,j}^{t+1}\ \wedge\ g_{i,j}^{\eta^{t}})=1$. Therefore $g_{i,j}^{t}=g_{i,j}^{t+1}\ \wedge\ g_{i,j}^{\eta^{t}}$ is correct. 7. 7. $r^{t}=r^{t+1}|r^{\eta^{t}}$. $g_{i,j}^{t}=1$ iff $r^{t+1}|r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))$, which is equivalent to that $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))$ or $r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))$. $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))$ iff $filter(s_{i,j},\alpha(r^{t+1}))=filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))$ and $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$ (Lemma 2). $r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))$ iff $filter(s_{i,j},\alpha(r^{\eta^{t}}))=filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))$ and $r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{\eta^{t}}))$ (Lemma 2). Therefore, $r^{t+1}|r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))$ iff $filter(s_{i,j},\alpha(r^{t+1}))=filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))$ and $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$, or $filter(s_{i,j},\alpha(r^{\eta^{t}}))=filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))$ and $r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{\eta^{t}}))$. $r^{t+1}\models filter(s_{i,j},\alpha(r^{t+1}))$ iff $g_{i,j}^{t+1}=1$ (Structural inductive hypothesis). $r^{\eta^{t}}\models filter(s_{i,j},\alpha(r^{\eta^{t}}))$ iff $g_{i,j}^{\eta^{t}}=1$ (Structural inductive hypothesis). Therefore, $g_{i,j}^{t}=1$ iff $(1[filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))=filter(s_{i,j},\alpha(r^{t+1}))]\ \wedge\ g_{i,j}^{t+1})\ \vee\ (1[filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))=filter(s_{i,j},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{i,j}^{\eta^{t}})=1$. Therefore, $g_{i,j}^{t}=(1[filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))=filter(s_{i,j},\alpha(r^{t+1}))]\ \wedge\ g_{i,j}^{t+1})\ \vee\ (1[filter(s_{i,j},\alpha(r^{t+1}|r^{\eta^{t}}))=filter(s_{i,j},\alpha(r^{\eta^{t}}))]\ \wedge\ g_{i,j}^{\eta^{t}})$ is correct. Theorem 3 has been proved. ### Proof of Proposition 4 Proof: Algorithm 3 shows the properties of the prefix notation of an SOIRE. Consider the number of elements of the stack, denoted as $|\Gamma|$. If $o^{i}\in\Sigma$, $|\Gamma|$ adds $1$. If $o^{i}\in\\{?,*,+\\}$, $|\Gamma|\geq 1$ and after that $|\Gamma|$ stays unchanged. If $o^{i}\in\\{\cdot,\&,|\\}$, $|\Gamma|\geq 2$ and after that $|\Gamma|$ minus $1$. Finally, $|\Gamma|=1$. Therefore, $\displaystyle\forall 1\leq i\leq|r_{p}|,(\sum_{j=i}^{|r_{p}|}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])\geq 1$ $\displaystyle\wedge(\sum_{i=1}^{|r_{p}|}1[o^{i}\in\Sigma]-1[o^{i}\in\\{\cdot,\&,|\\}])=1$ (15) We shows that $\texttt{Enc2Pre}(\theta)$ satisfies Equation Proof of Proposition 4, given the faithful encoding $\theta=(w,u)$. Let $last$ be the biggest number that $w^{last}_{\text{none}}=0$ ($|r_{p}|=last$) and $T$ be the bounded size. We first use mathematical induction to prove $\forall 1\leq i\leq|r_{p}|,(\sum_{j=i}^{|r_{p}|}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])\geq 1$. Mathematical induction base: $i=last$. If $i=last=T$, then $\sum_{a\in\Sigma}w^{i}_{a}=1$ (the last vertex can only select a symbol), i.e. $o^{i}\in\Sigma$. Therefore, $(\sum_{j=i}^{last}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])\geq 1$ is correct. If $i=last<T$, because $w^{i+1}_{\text{none}}=1$, $\forall a\in\\{?,*,+,\cdot,\&,|\\},w^{i}_{a}=0$ (Definition 2 Condition 5). Therefore, $\sum_{a\in\Sigma}w^{i}_{a}=1$ (Definition 2 Condition 1), i.e. $o^{i}\in\Sigma$. Therefore, $(\sum_{j=i}^{last}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])\geq 1$ is correct. Mathematical induction step: $1\leq i<last$. Suppose for $i+1$, $(\sum_{j=i+1}^{last}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])\geq 1$. $(\sum_{j=i+1}^{last}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])\geq 1$ (Mathematical inductive hypothesis). If $o^{i}\in\Sigma\cup\\{?,*,+\\}$, $(\sum_{j=i}^{last}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])\geq 1$ is correct. If $o^{i}\in\\{\cdot,\&,|\\}$, consider two cases: 1. 1. $(\sum_{j=i+1}^{last}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])>1$ $(\sum_{j=i+1}^{last}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])+1[o^{i}\in\Sigma]-1[o^{i}\in\\{\cdot,\&,|\\}]>1-1$. Therefore, $(\sum_{j=i}^{last}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])>1$ is correct. 2. 2. $(\sum_{j=i+1}^{last}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])=1$ Let $B=\\{j|i\leq j\leq last\wedge o^{j}\in\\{\cdot,\&,|\\}\\}$. For all $j\in B$, $\exists j+2\leq k\leq last,u^{j}_{k}=1$, denoted as $u^{j}=k$ (Definition 2 Condition 3). For any $j,k\in B$, if $j\neq k$, $u^{j}\neq u^{k}$ (Definition 2 Condition 2 and 5). For all $j\in B$, $o^{u^{j}-1}\in\Sigma$ (Definition 2 Condition 5 and 1). Because $o^{i}\in\\{\cdot,\&,|\\}$, $(\sum_{j=i}^{last}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])=0$. Therefore, $\\{u^{j}-1|j\in B\\}$ is the set of all $o\in\Sigma$. Therefore, $o^{last}\notin\Sigma$, which is a contradiction. Therefore, this case does not exist. Therefore, $\texttt{Enc2Pre}(\theta)$ satisfies $\forall 1\leq i\leq|r_{p}|,(\sum_{j=i}^{|r_{p}|}1[o^{j}\in\Sigma]-1[o^{j}\in\\{\cdot,\&,|\\}])\geq 1$. Let $B=\\{i|1\leq i\leq last\wedge o^{i}\in\\{\cdot,\&,|\\}\\}$. For all $i\in B$, $\exists i+2\leq j\leq last,u^{i}_{j}=1$, denoted as $u^{i}=j$ (Definition 2 Condition 3). For any $i,j\in B$, if $i\neq j$, $u^{i}\neq u^{j}$ (Definition 2 Condition 2 and 5). For all $i\in B$, $o^{u^{i}-1}\in\Sigma$ (Definition 2 Condition 5 and 1). In addition, $last\notin\\{u^{i}-1|i\in B\\}$ and $o^{last}\in\Sigma$. Therefore, $\texttt{Enc2Pre}(\theta)$ satisfies $(\sum_{i=1}^{|r_{p}|}1[o^{i}\in\Sigma]-1[o^{i}\in\\{\cdot,\&,|\\}])=1$. Therefore, $\texttt{Enc2Pre}(\theta)$ satisfies Equation Proof of Proposition 4 and $\texttt{Enc2Pre}(\theta)$ returns the prefix notation of an SOIRE. Proposition 4 has been proved. ### Proof of Proposition 5 Proof: We construct the faithful encoding $\theta=(w,u)$, given $T\in\mathbb{Z}^{+}$, and an SOIRE $r(|r|\leq T)$ written in prefix notation. Construct the syntax tree of $r$, and $o^{t}(1\leq t\leq|r|)$ denotes the symbol or operator represented by vectex $t$ and vertex $v^{t}$ is the right son of vertex $t$. Let $w=\mathbf{0}$ and $u=\mathbf{0}$ initially. $\forall|r|<t\leq T,w^{t}_{\text{none}}\leftarrow 1$. $\forall 1\leq t\leq|r|,w^{t}_{o^{t}}\leftarrow 1$. $\forall 1\leq t\leq|r|\wedge o^{t}\in\\{\cdot,\&,|\\},u^{t}_{v^{t}}\leftarrow 1$. Because $\theta=(w,u)$ is constructed based on the syntax tree of $r$, $\theta=(w,u)$ are faithful encoding and $\texttt{Enc2Pre}(\theta)=\texttt{PreForm}(r)$. Proposition 5 has been proved. ### Proof of Proposition 6 Proof: We apply proof by contradiction to prove this theorem. Given the bouned size of SOIREs $T\in\mathbb{Z}^{+}$, for any two different faithful encodings $\theta_{1},\theta_{2}(\theta_{1}\neq\theta_{2})$, $\texttt{Enc2Pre}(\theta_{1})=\texttt{Enc2Pre}(\theta_{2})$. Let $\theta_{1}=(w^{(1)},u^{(1)}),\theta_{2}=(w^{(2)},u^{(2)})$. Because $\texttt{Enc2Pre}(\theta_{1})=\texttt{Enc2Pre}(\theta_{2})$, $w^{(1)}=w^{(2)}$ (Algorithm 4 and Definition 2 Condition 4). Let $p$ be the biggest number that $u^{(1)p}\neq u^{(2)p}$. Because $w^{(1)}=w^{(2)}$, $u^{(1)p}=u^{(2)p}=\mathbf{0}$ or $\exists p+2\leq t_{1},t_{2}\leq T,u^{(1)p}_{t_{1}}=1\wedge u^{(2)p}_{t_{2}}=1$. In addition, $u^{(1)p}\neq u^{(2)p}$. Therefore, $\exists p+2\leq t_{1},t_{2}\leq T,u^{(1)p}_{t_{1}}=1\wedge u^{(2)p}_{t_{2}}=1$. Without loss of generality, let $t_{1}<t_{2}$. Because $u^{(1)p}_{t_{1}}=1$, $\forall a\in\\{?,*,+,\cdot,\&,|\\},w^{(1)t_{1}-1}_{a}=0$ (Definition 2 Condition 5). Because $w^{(1)}=w^{(2)}$, $\forall a\in\\{?,*,+,\cdot,\&,|\\},w^{(2)t_{1}-1}_{a}=0$ (Definition 2 Condition 5). Therefore, $\exists 1\leq p_{2}\leq t_{1}-2,u^{(2)p_{2}}_{t_{1}}=1$ (Definition 2 Condition 5). Because $p$ be the biggest number that $u^{(1)p}\neq u^{(2)p}$, $1\leq p_{2}<p$. For $\theta_{2}$, because vertex $p_{2}$ has a right son, vertex $t_{1}$, then for all vertex $p^{\prime}(p_{2}<p^{\prime}<t_{1})$, if vertex $p^{\prime}$ has a right son, vertex $t^{\prime}$, then $t^{\prime}<t_{1}$ (Definition 2 Condition 6). However, $p_{2}<p<t_{1}$ and vertex $p$ has a right son, vertex $t_{2}$, and $t_{2}>t_{1}$, which is a contradiction. Therefore, there does not exist two different faithful encoding $\theta_{1},\theta_{2}(\theta_{1}\neq\theta_{2})$, where the size of $\theta_{1}$ and $\theta_{2}$ is $T$, $\texttt{Enc2Pre}(\theta_{1})=\texttt{Enc2Pre}(\theta_{2})$. Proposition 6 has been proved. ### Proof of Theorem 7 Proof: Theorem 7 can be proved by combining Proposition 4, Proposition 6 and Proposition 5. Therefore, Faithful encoding is one-to-one corresponding to the SOIRE in prefix notation for a certain size of the parameters. ### Proof of Proposition 8 Proof: Because $(r^{?})^{?}=r^{?},(r^{?})^{*}=r^{*},(r^{?})^{+}=r^{*},(r^{*})^{*}=r^{*},(r^{*})^{?}=r^{*},(r^{*})^{+}=r^{*},(r^{+})^{+}=r^{+},(r^{+})^{?}=r^{*},(r^{+})^{*}=r^{*}$, the consecutive unary operators can be replaced by one. For an SOIRE $r$ over $\Sigma$, we use $r^{\prime}$ ($r^{\prime}\equiv r$) to denote the SOIRE without consecutive unary operators. For the syntax tree of $r^{\prime}$, there are no consecutive unary operators in $r^{\prime}$, so only the parent of a vertex representing a symbol or a binary operator can represent a unary operator. The Equation Proof of Proposition 4 shows that the number of the binary operators is $|\alpha(r^{\prime})|-1$ for $r^{\prime}$. Therefore, the size of the syntax tree of $r^{\prime}$ is $(|\alpha(r^{\prime})|+|\alpha(r^{\prime})|-1)*2=4|\alpha(r^{\prime})|-2\leq 4|\Sigma|-2$. Proposition 5 shows that given $T=4|\Sigma|-2$, there exists a faithful encoding $\theta$ that $\texttt{Enc2Pre}(\theta)=\texttt{PreForm}(r^{\prime})$, because $r^{\prime}\leq T$. Therefore, given $T=4|\Sigma|-2$, for each SOIRE $r$ over $\Sigma$, there exists a faithful encoding $\theta$ that $\texttt{Enc2Pre}(\theta)=\texttt{PreForm}(r^{\prime})$ and $\\{s|r^{\prime}\models s\\}=\\{s|r\models s\\}$. Proposition 8 has been proved. ## Appendix C ### Regularization To make the set of parameters more faithful, we can add one regularization for each condition when training the neural network. We use Equation 16-22 as regularization for each condition. $\text{Onehotloss}(x)=\text{Mean}_{i}(1-x_{i})x_{i}+(1-\sum_{i}x_{i})^{2}$ for a vector $x$. $\text{ReLU}(x)=\max(x,0)$ for a scalar $x$. $\displaystyle\text{Mean}_{t=1}^{T}(\text{Onehotloss}(w^{t}))$ (16) $\displaystyle\text{Mean}_{t=1}^{T}(\text{Onehotloss}(u^{t}))$ (17) $\displaystyle\text{Mean}_{t=1}^{T}(\text{Onehotloss}([u^{t}_{t+2},u^{t}_{t+3},\dots,u^{t}_{T},$ $\displaystyle w^{t}_{a\in\Sigma},w^{t}_{?},w^{t}_{*},w^{t}_{+},w^{t}_{\text{none}}]))$ (18) $\displaystyle\text{Mean}_{t=1}^{T-1}(\text{ReLU}(w^{t}_{\text{none}}-w^{t+1}_{\text{none}}))$ (19) $\displaystyle\text{Mean}_{t=2}^{T}(\text{Onehotloss}([w^{t-1}_{?},w^{t-1}_{+},w^{t-1}_{*},w^{t-1}_{\cdot},$ $\displaystyle w^{t-1}_{\&},w^{t-1}_{|},u^{1}_{t},u^{2}_{t},\dots,u^{t-2}_{t},w^{t}_{\text{none}}]))$ (20) $\displaystyle\text{Mean}_{t=3}^{T}(\text{Mean}_{p=1}^{t-2}(\text{ReLU}((t-1-p)u^{p}_{t}+$ $\displaystyle\sum_{p^{\prime}=p+1}^{t-1}\sum_{t^{\prime}=t+1}^{T}u^{p^{\prime}}_{t^{\prime}}-(t-1-p))))$ (21) $\displaystyle\text{Mean}_{a\in\Sigma}(\text{ReLU}(\sum_{t=1}^{T}w^{t}_{a}-1))$ (22) ### Results on noise-free data The SOIREs learnt by SOIREDL on each dataset with noise-free data are show in Table 5. Dataset | Ground-truth SOIREs | SOIREDL ---|---|--- 1 | $((a|b)c^{*})^{+}d$ | $((a|b)c^{*})^{+}d$ 2 | $(a|b|c|d|e^{?})^{*}$ | $(b^{*}e^{*}a^{*}c^{*}d^{*})^{*}$ 3 | $a^{+}|(b|c)^{*}|d^{+}$ | $(b^{+}|c)^{*}|a^{*}|d^{*}$ 4 | $(ab^{*})^{+}|c^{+}$ | $(a^{+}b^{*})^{*}$ 5 | $ab(c|d|e|f|g)^{*}$ | $((c|g^{*}|a\&b|d^{*})e^{*})^{*}$ 6 | $a(b|c|d|e)^{+}f^{*}$ | $(b^{*}\&a\&d^{*}\&e^{*}\&c^{*})f^{*}$ 7 | $a^{*}|(b|c|d|e)^{*}$ | $(c^{*}d^{*}e^{*})^{*}\&b^{*}$ 8 | $(a|b|c|d|e)^{+}\&f^{*}\&g^{*}$ | $(f^{*}g^{*}d^{*}a^{*}e^{*}c^{*}b^{*})^{+}$ 9 | $a^{+}|b|(c|d)^{+}$ | $(c^{+}|d^{+})^{+}$ 10 | $(a|b)^{+}c$ | $((a^{*}\&b^{*})^{+}c)^{?}$ 11 | $a^{?}(b|c|d)^{*}e$ | $(((b^{*}|d^{*})a^{*}c^{*})^{+}e)^{?}$ 12 | $(a|b|c|d^{?})^{*}$ | $(a^{*}b^{*}d^{*}c^{*})^{*}$ 13 | $a^{?}\&b^{*}\&c^{?}$ | $a^{?}\&c^{?}\&b^{*}$ 14 | $(a(bc^{?})^{*})^{*}d^{?}e^{?}$ | $(a(b^{+}\&c)^{*}d^{*}e^{?})^{*}$ 15 | $a^{?}\&b^{*}\&c^{*}\&d^{?}$ | $c^{*}\&a^{?}\&d\&b^{*}$ 16 | $a^{*}\&b^{?}\&c^{*}$ | $a^{*}\&c^{*}\&b$ 17 | $a^{+}\&(b|c|d|e)^{+}$ | $(c^{*}|e^{*}\&(a^{*}|b^{+})|d^{*})^{*}$ 18 | $a^{?}(b|c)^{*}(d|e)$ | $a^{*}(b|c)^{*}d^{?}e^{?}$ 19 | $a(b|c|d)^{*}e^{*}$ | $a((c|d)^{?}|b)^{*}e^{*}$ 20 | $(a^{+}|b^{?}|c^{?}|(d|e)^{*})fgh^{*}$ | $(d^{*}\&f\&g^{*}\&e^{*})^{?}h^{*}$ 21 | $(a^{*}b)^{+}$ | $(a^{*}b)^{*}$ 22 | $(a|b)^{*}c^{*}d^{*}$ | $(a^{*}\&b^{*})c^{*}d^{*}$ 23 | $((ab^{*})^{+}|c^{+})d$ | $(ab^{*}d^{?})^{+}$ 24 | $(a|b|c)^{+}|(de^{*}f^{*})^{+}$ | $((c^{*}a^{*})^{*}b^{*})^{+}$ 25 | $a|(b^{?}c)^{+}$ | $(b^{?}c^{+})^{+}$ 26 | $a^{*}(b|c|d)^{*}$ | $((b^{*}|a^{*}d)c^{*})^{*}$ 27 | $a^{?}(b^{+}\&c^{*}\&d^{*}\&e^{*}\&f^{*})$ | $((c^{*}\&d^{*}\&e^{*})f^{*}b^{*})^{*}$ 28 | $(a^{?}b)^{+}$ | $(a^{?}b)^{*}$ 29 | $(a^{+}|b^{+}|(c|d)^{*})e$ | $(d^{*}c^{*})^{*}e$ 30 | $a^{?}(b^{?}c^{+}d)^{+}$ | $((a|b)^{*}cd^{?})^{*}$ Table 5: The ground-truth SOIREs and the SOIREs learnt by SOIREDL on each dataset with noise-free data. ### Hyperparameters All hyperparameters are tuned on the first noise-free dataset. Hyperparameters of RE2RNN Max-state and $\gamma$ are two important hyperparameters of RE2RNN. We use grid search to choose max-state from $\\{$$20$, $40$, $60$, $80$, $100$, $120$, $140$, $160$, $180$, $200$$\\}$ and choose $\gamma$ from $\\{$$0$, $0.02$, $0.04$, $0.06$, $0.08$, $0.1$, $0.12$, $0.14$, $0.16$, $0.18$, $0.2$$\\}$. Max-state is the number of state in the automaton and Figure 5 shows that max- state should be set as $100$ to get the best performance. Figure 5: Accuracy(%) of RE2RNN on the first dataset with different number of states max-state. $\gamma$ is the threshold in the interpretation. The values above $\gamma$ are set as $1$s and the values below $\gamma$ are set as $0$s. Figure 6 shows that $\gamma$ should be set as $0.12$ to get the best performance. Figure 6: Accuracy(%) of RE2RNN on the first dataset with different thresholds in the interpretation $\gamma$. Hyperparameters of SOIREDL The beam width $\beta$ and the coefficient of regularization $\lambda$ are two important hyperparameters of SOIREDL. We use grid search to choose $\beta$ from $\\{$$10$, $50$, $100$, $300$, $500$, $1000$$\\}$ and choose $\lambda$ from $\\{$$0$, $10^{-3}$, $10^{-2}$, $10^{-1}$, $1$, $10$$\\}$. Figure 7 shows that SOIREDL gets the best result when $\beta\geq 500$, so $\beta$ should be set as $500$ to make a balance between accuracy and running time. Figure 7: Accuracy(%) of SOIREDL on the first dataset with different beam width $\beta$. Figure 8 and 9 show that the accuracy and faithfulness decrease when $\lambda>10^{-3}$. The larger coefficient of regularization makes the neural network more difficult to train and the results show that SOIREDL can also gets a well performance without regularization, so we set $\lambda$ as $0$. Figure 8: Accuracy(%) of SOIREDL on the first dataset with different coefficient of regularization $\lambda$. Figure 9: Faithfulness(%) of SOIREDL on the first dataset with different coefficient of regularization $\lambda$.
# Emerging Diversity in a Population of Evolving Intransitive Dice Julius B. Kirkegaard Kim Sneppen Niels Bohr Institute, University of Copenhagen, 2100 Copenhagen, Denmark ###### Abstract Exploiting the mathematical curiosity of intransitive dice, we present a simple theoretical model for co-evolution that captures scales ranging from the genome of the individual to the system-wide emergence of species diversity. We study a set of evolving agents that interact competitively in a closed system, in which both the dynamics of mutations and competitive advantage emerge directly from interpreting a genome as the sides of a die. The model demonstrates sympatric speciation where new species evolve from existing ones while in contact with the entire ecosystem. Allowing free mutations both in the genomes and the mutation rates, we find, in contrast to hierarchical models of fitness, the emergence of a metastable state of finite mutation rate and diversity. ## I Introduction Evolution is the optimization scheme of the biological realm: with the correct initial conditions and ample time, random mutations and natural selection are sufficient to ensure the emergence of highly complex organisms. But the nature of what is being optimized is context-dependent. Even for a fixed environment, evolution does not necessarily have a single goal. If we were to rerun the “experiment” of evolution, it is virtually guaranteed this would result in species distinct from those that are alive today. Evolution of specific biological features of a species depend on properties of the other species in its environment ehrlich1964butterflies ; woolhouse2002biological . Such co-evolution is believed to be central to describe evolution on the large scale KaufffmanBook ; Bak1992 ; Sneppen1993 ; Weitz2005 ; Xue2017 as implied by the Red Queen hypothesis of Van Valen van1973new ; liow2011red . Further, species interactions are not necessarily ranked, as observed among corals, plants and microbes jackson1975alleopathy ; taylor1990complex ; cameron2009parasite ; kirkup2004antibiotic ; Kerr2002 . Many studies have been devoted to understanding and evolving such intransitive interactions, ranging from molecular scale autocatalytic network eigen1978hypercycle ; jain1998autocatalytic ; segre2000composing to extensions of the competitive game of rock, paper and scissors Kerr2002 ; Laird2006 ; Reichenbach2006 ; Reichenbach2007 ; Mathiesen2011 ; Mitarai2012 ; Dobrinevski2012 ; Ulrich2014 ; szolnoki2014cyclic ; Levine2017 ; Soliveres2018 . Non-hierarchical species dynamics can readily be studied for a set of species whose interactions are fixed. Here, in contrast, we are interested in evolving systems where intransitive interactions emerge ex nihilo. We suggest a minimal model for such a system consisting of individuals that interact by rolling dice. Figure 1: The intransitive interaction of five 9-sided proper dice. Direction of the arrows indicate domination. The graph formed from the interactions has two Hamiltonian paths, one of which is indicated in the background. In other words, out of the $4!=24$ orderings of the dice, there are two that form an intransitive loop of dominance. ## II Model We define the characteristics of an individual in our system by its genome as given by a list of $n$ integers. We only consider competitive interactions between species and settle these by interpreting the genome of integers as the sides of dice that are rolled. The outcome of a fight is stochastic, but certain dice will tend to out-compete other dice. For instance, a fight between $A=(3,3,3,3,3,6)$ and $B=(2,2,2,5,5,5)$ will typically be won by $A$ despite $\sum_{i}A_{i}=\sum_{i}B_{i}$. The probability of $A$ winning in the present example is $n^{-2}\sum_{i}\sum_{j}[A_{i}>B_{j}]=7/12$. What makes this particular interaction interesting in the context of competing species is the fact that we can introduce a species $C$ such that both $B\succ C$ and $C\succ A$, or succinctly: $A\succ B\succ C\succ A$. This is for instance the case for $C=(1,4,4,4,4,4)$. This intransitive behavior of dice is well-known Finkelstein2006 ; Conrey2016a , but its applicability and simplicity for modeling co-evolution are unexplored. A plethora of systems could be designed around the above interaction rule. We choose to consider one of the simplest and study $k$ individuals that interact in a well-mixed scenario. At each time step of our simulation, we let each individual randomly attack another. Two individuals, $X$ and $Y$, are considered to belong to the same species if $\sum_{i}|X_{i}-Y_{i}|<\delta$, in which case they will not fight. Otherwise, the losing individual of the competition will be replaced by a copy of the winner. On ties, a random winner is chosen. Figure 2: Probability of intransitive (Hamiltonian) $k$-loops in sets of $k$ proper dice as a function of the number of sides $n$ on the dice. Here, proper dice are sorted, has $\sum_{i}X_{i}=n(n+1)/2$ and $X_{i}\leq n$. A set of $k$ dice $\\{X_{i}\\}$ has a $k$-loop if an ordering $\sigma$ exists such that $X_{\sigma_{1}}\prec X_{\sigma_{2}}\prec\cdots\prec X_{\sigma_{k}}\prec X_{\sigma_{1}}$. Each point is the result of averaging over $100,\\!000$ Monte Carlo samples with uniform probability for each valid die. Figure 3: Speciation. (a–b) show $1,\\!500$ individuals of the system with links between them if they consider one another the same species. (a) has $n=9$ and (b) $n=14$, and in both cases $\mu=0.1$. Nodes are colored, for visualization purposes only, using the Leiden community detection algorithm Traag2019 on the undirected, unweighted graph. (c–e) show species richness $S$ (solid lines) and diversity $D$ (dashed) stabilizing in a simulations of varying genomic complexity $n$ and mutation rate $\mu$ as indicated by the legend. Curves are averages over simulations with initial conditions of all dice equaling the standard die. All simulation were run with $\delta=3$. $k=10,\\!000$. A crucial novelty of our model is that the dice interpretation not only sets the rules for interactions but also naturally provides a genome space in which mutations may occur. In our model, at each time step, each individual mutates with probability $\mu$. A mutation event is the random change ($\pm 1$) of one of its genome digits. As dice that are permutations of one another have identical competitive advantages, we restrict our genome space to that of ordered dice. Thus we disallow mutations that break the sorted nature of a genome. For instance, $A$ may mutate to $(2,3,3,3,3,6)$ but not to $(3,2,3,3,3,6)$. In effect this accelerates the dynamics of our system, as most neutral competitions are avoided that would otherwise have to be settled by stochastic extinction. Further, naturally, it is universally better for a species to mutate up in sum rather than down. We set a fitness ceiling by only allowing mutations that keep $\sum_{i}X_{i}\leq n\,(n+1)/2$ (the sum of the standard die) and any $X_{i}\leq n$; the latter of which excludes a large of set of dice that are typically less competitive and allows efficient enumeration of the set of allowed dice. In total, we have thus created a mutation model where universal fitness can be measured, but where, occasionally, universally unfavorable mutations are preferred to adapt to coexistent competitors. Instead of defining the interaction as a single roll of the dice, one might consider competitions of $r$ rounds. As $r\rightarrow\infty$, any slight competitive advantage will result in certain overall wins. In this way, $r$ can be chosen to control the ruggedness of the fitness landscape KaufffmanBook . Here, we limit our attention to $r=1$. ## III Probability of intransitivity Before delving into the dynamics of the model, it is useful to have a feeling for the prevalence of intransitivity in random dice. Consider all dice such that $\sum X_{i}=n(n+1)/2$ and $1\leq X_{i}\leq n$. Fig. 2 shows the probability that $k$ such dice contains at least one intransitive (Hamiltonian) $k$-loop as function of the number of sides of the dice $n$. The plot shows the fact that if you choose $k=3$ random dice (with $n<15$ sides), the probability that these dice interact intransitively is less than $\sim 20\%$. With a larger set of dice, not only do the probability of $k$-loops increase, but so does probability of smaller sub-cycles [see SI paperSI for a version of Fig. 2 for loops of any size]. In Fig. 2 we only consider a small number of dice, but very long intransitive cycles can also be found. For instance, in the set of $k=910$ allowed $9$-sided dice, the longest possible cycle is at least $891$ (finding the precise length is NP-hard). Thus, intransitivity is by no means rare, but (short) loops are not the norm either. However, in the dynamics of our model, it is much more unlikely for a species to go extinct in an intransitive loop than when species interact in a dominant manner. Thus we expect one of two things to happen in the long run: the system will be taken over by one species or will be inhabited by a number of species that interact intransitively and show oscillations. There is also a large heterogeneity in the advantages of the different dice. For instance, for $n=6$, the die that has an advantage over most other dice is $(1,3,3,4,5,5)$, beating on average $\sim 60\,\%$ of the other dice. The worst is $(1,2,2,4,6,6)$, which is better than only $\sim 40\,\%$. In contrast, the standard die $(1,2,3,4,5,6)$ has precisely a $50\,\%$ chance of beating any other proper die. Similar conclusions can be made for all $n$. We note that the precise statistics of Fig. 2 would be different if we considered not only proper dice, as some dice have many more permutations of their sides than others. The qualitative conclusions drawn would remain similar, nonetheless. ## IV Finite mutation rate upholds species diversity Random, well-mixed ecosystems of many competing species interacting under demographic noise are unstable may1972will , and competitive exclusion often leads to a collapse to only a single or a few surviving species hardin1960competitive ; levin1970community ; haerter2016food . This, naturally, also applies to the present dynamical system. However, since the genome space of our model is not hierarchically organized, a finite mutation rate can induce a perpetual co-evolutionary arms race. One complication in counting the number of species in a system is due to the fact that the network of individuals is very unlikely to organize into fully connected components. This is a complication that is not unique to our system, but indeed any problem related to speciation tucker2017guide . For each pair of individuals, our genome distance rule specifies if they belong to the same species. Denote by $C_{i}$ the number of individuals that individual $i$ is considered the same species as (including $i$ itself). An effective measure for species richness is then given by $S=\sum_{i}C_{i}^{-1}$. In the case of fully connected species, with no overlap between them, this measure coincides with simply counting the number of species. Likewise, we can define a species diversity measure that also accounts for evenness as $D=k^{2}\left(\sum_{i}C_{i}\right)^{-1}$. This is equal to the number of species only if each species occupy the same fraction of the entire system and thus small species contribute only negligibly to its value. Fig. 3 shows that both the mutation rate $\mu$ and the genome complexity $n$ (dice size) set the number of species that a system of a certain size can maintain. At low mutation rates, the system is dominated by a small cloud of individuals that form a quasispecies eigen1989molecular , since, in this case, a single species can be locally dominant and no individuals can escape this local optimum at the low mutation rate. At higher mutation rates, however, intransitive interactions appear and oscillatory dynamics of a high diversity system emerges. The systems have intransitive loops of many lengths, but the dynamics are dominated by short cycles ($\lesssim 5$ in the systems studied here). This is demonstrated and studied in the SI by considering the mean-field Lotka–Volterra equations of the system. In detail, the system behaves oscillatory with a frequency that remain relatively constant for multiple oscillation periods. On long time scales, however, stochastic events can change the dynamics, such as when a species in an intransitive loop stochastically goes extinct or when an individual suddenly mutates to dominate the existent intransitive interactions initiating a “punctuated equilibrium” event causing a sudden shift in oscillation frequency. Fig. 3(a–b) visualizes species connectivity in the steady-state ecosystems that evolve from the dynamics of the model. In these graphs, an edge is drawn between two dice if they consider each other to belong to the same species. Despite starting with a single species, the system can evolve to one that has many species that are genomically disconnected. While no single measure can capture the complexity of these inter-species connections, running community detection algorithms on these graphs tend to find a number of clusters in the same order of magnitude as our richness $S$ and diversity $D$ measures. We note that for comparisons with other system sizes $k$, the values of the mutation rates should be rescaled accordingly, as the number of attempted mutation events per step is $\sim\mu k$. Thus systems with a higher number of individuals can support high diversity in spite of having a low mutation rate per individual. ## V Heterogeneous mutation rates Since a finite mutation rate is needed to maintain a finite diversity, we have an ecosystem collapse if the mutation rate is taken to zero. For instance, in a system of purely hierarchically interacting species, the dominant species will prefer a low mutation rate thus leading to a collapse of ecosystem diversity. In the present system, however, there is no global optimum, and a high mutations rate means quick adaptability and increases the chance of an individual to out-compete others by an evolutionary advantage. A high mutation rate is not strictly an advantage though, since it also means a high rate of bad mutation events towards either locally or globally worse genomes. In Fig. 4(inset), we show the competition between two populations with $\mu_{A}=0.2$ and $\mu_{B}=0.001$, respectively. For early times, a low mutation rate gives an advantage because there is a low rate of genomic decay and thus we see population $B$ winning initially. This reflects the advantage of localizing a population around a local fitness maximum over more diffuse quasispecies at higher mutation rates eigen1989molecular . However, at some point population $A$ finds a competitive advantage over the slowly adapting population $B$ and annihilates the latter completely. Varying $\mu_{A}$ and $\mu_{B}$, the average outcome of these scenarios is shown in the main part of Fig. 4. The exact results depend on the initialization of the dice, but in this case, we see that $\mu\approx 0.1$ is generally advantageous. Figure 4: Annihilation statistics of two populations with different mutation rates. Inset shows a simulation of $2\times 10,\\!000$ dice with one half having $\mu_{A}=0.2$ and the other $\mu_{B}=0.001$. The time of annihilation of population $B$ is $T=130$ time steps. Main plot shows the average of $1/T$ measured with a negative sign if $B$ wins. High values thus indicate that $B$ tends to be annihilated and negative values that $A$ tends to be annihilated. Values near zero indicate a system dominated by stochastic extinction or one where annihilation takes a very long time. All dice were initialized to $(1,1,1,6,6,6)$ at the beginning of the simulations. Figure 5: Mutating the mutation rate. Simulations for which each individual mutates its mutation rate with a 1 % chance of a 1 % change per time step (Gaussian multiplicative noise). (abc) Simulations for $n=6,\,9,\,14$ showing the system average mutation rate $\langle\mu\rangle$ as a function of time steps. Orange lines show average over all simulations, whereas pink line shows an average over only those simulations that at $t=10^{7}$ have $\langle\mu\rangle>10^{-3}$. The full distribution of $\mu$ is narrowly peaked around $\langle\mu\rangle$ (see SI paperSI ). Inset in (c) shows, for each realization, the average mutation rate versus the system diversity $D$ (dark curves) for $t>5\cdot 10^{5}$, with two trajectories emphasized, one of which collapses towards $\langle\mu\rangle=0$. All simulations use $\delta=3$, but obtain similar results for other $\delta$. The fraction of systems in the high mutation state depends greatly on initial mutation rate, even for the $n=6$ die (see SI paperSI ). We complete the design of our model by, finally, also permitting mutations in the mutation rate itself. In a hierarchical setting, this would lead to the immediate collapse of both the mutation rates and diversity. Fig. 4 indicates, however, that in the present system there is also an advantage to having a finite mutation rate. Each panel in Fig. 5 show realizations of a system of individuals, all initialized with the same genome and an initial mutation rate of $\mu=0.01$. We observe two distinct outcomes: most trajectories reach a metastable state with a high mutations rate (die size $n=6$ never reaches this state, but would do so if we instead started with initial $\mu=0.1$, see SI paperSI ), and some that decay towards zero mutation rates. A characteristic of the metastable state is the rare but sudden decay events of both mutation rate and diversity of the entire system [Fig. 5(b–c)]. As the mutation rate decreases, it becomes less and less likely to escape the low diversity situation thereby creating a positive feedback loop for decreasing the mutation rate even further. Once collapsed, very large perturbations are needed to bring the system out of this situation. Even changing the mutation rate of half of the collapsed system to $\mu=0.1$ is typically not enough to return to the high diversity state [see SI paperSI ]. In contrast, for the surviving high diversity ecosystems the average number of mutations that separate two random individuals is large. For instance, $\langle\sum_{i}|X_{i}-Y_{i}|\rangle_{X,Y}\approx 12$ for $n=14$, which is about $2/3$ of the average obtained between random dice. ## VI Perspective We have presented a theoretical model that, on one hand, is exceedingly simple to define, and at the same time successfully describes a host of complex phenomena related to co-evolution. At fixed, finite mutation rates, the model permits a state of finite diversity in co-evolutionary balance. For hierarchically interacting systems, allowing mutations in the mutation rates themselves, will lead to an ecosystem collapse. In contrast, we find a metastable state of finite diversity, whose stability increases quickly with genomic complexity, measured by the number of sides of the dice $n$. We only considered a well-mixed system, meaning that at all times each individual could meet any other individual. Introducing space to the model, e.g. putting the agents on a lattice, will most likely stabilize the observed effects even further; with spatial dynamics, intransitive relations will decay very slowly Mitarai2012 , and thus the rate of species extinctions decreases. Furthermore, speciation events should increase in frequency as space allows for transient allopatric speciation. Precise quantification of intransitivity in the dynamical system is another interesting avenue for further research: despite being dominated by few intransitive cycles, a static view of the systems at any given time will not reveal the dominance of these cycles as at least one species will have a low population count due to the oscillatory Lotka–Volterra-like dynamics imposed by the dynamics of the intransitive loops. In conclusion, from the simple rules of competing dice emerge a natural balance of mutation rates and diversity. Too high a mutation rate risks genomic decay and the disintegration of quasispecies: “mutate and die”. Too few mutations are disfavoured in analogy to the Red Queen hypothesis: “mutate or die”. ###### Acknowledgements. This project has received funding from the Novo Nordisk Foundation Grant Agreement NNF20OC0062047 and from the European Research Council (ERC) under the European Union’s Horizon 2020 Research and Innovation Programme, Grant Agreement No. 740704. ## References * (1) Paul R Ehrlich and Peter H Raven. Butterflies and plants: a study in coevolution. Evolution, pages 586–608, 1964. * (2) Mark EJ Woolhouse, Joanne P Webster, Esteban Domingo, Brian Charlesworth, and Bruce R Levin. Biological and biomedical implications of the co-evolution of pathogens and their hosts. Nature genetics, 32(4):569–577, 2002. * (3) Stuart A Kauffmann. The Origins of Order. Oxford University Press, 1993. * (4) Per Bak, Henrik Flyvbjerg, and Benny Lautrup. Coevolution in a rugged fitness landscape. Physical Review A, 46(10):6724–6730, 1992. * (5) Per Bak and Kim Sneppen. Punctuated Equilibribum and Criticality in a simple model of evolution. Physical Review Letters, 71(24):4083–4086, 1993. * (6) J S Weitz, H Hartman, and S A Levin. Coevolutionary arms races between bacteria and bacteriophage. Proceedings of the National Academy of Sciences of the United States of America, 102(27):9535–9540, 2005. * (7) Chi Xue and Nigel Goldenfeld. Coevolution Maintains Diversity in the Stochastic ”kill the Winner” Model. Physical Review Letters, 119(26):1–5, 2017. * (8) Leigh Van Valen. A new evolutionary law. Evol theory, 1:1–30, 1973. * (9) Lee Hsiang Liow, Leigh Van Valen, and Nils Chr Stenseth. Red queen: from populations to taxa and communities. Trends in ecology & evolution, 26(7):349–358, 2011. * (10) JBC Jackson and LEO Buss. Alleopathy and spatial competition among coral reef invertebrates. Proceedings of the National Academy of Sciences, 72(12):5160–5163, 1975. * (11) Douglas R Taylor and Lonnie W Aarssen. Complex competitive relationships among genotypes of three perennial grasses: implications for species coexistence. The American Naturalist, 136(3):305–327, 1990. * (12) Duncan D Cameron, Andy White, and Janis Antonovics. Parasite–grass–forb interactions and rock–paper–scissor dynamics: predicting the effects of the parasitic plant rhinanthus minor on host plant communities. Journal of Ecology, 97(6):1311–1319, 2009. * (13) Benjamin C Kirkup and Margaret A Riley. Antibiotic-mediated antagonism leads to a bacterial game of rock–paper–scissors in vivo. Nature, 428(6981):412–414, 2004. * (14) Benjamin Kerr, Margaret A. Riley, Marcus W. Feldman, and Brendan J.M. Bohannan. Local dispersal promotes biodiversity in a real-life game of rock-paper-scissors. Nature, 418(6894):171–174, 2002. * (15) Manfred Eigen and Peter Schuster. The hypercycle. Naturwissenschaften, 65(1):7–41, 1978. * (16) Sanjay Jain and Sandeep Krishna. Autocatalytic sets and the growth of complexity in an evolutionary model. Physical Review Letters, 81(25):5684, 1998. * (17) Daniel Segré and Doron Lancet. Composing life. EMBO reports, 1(3):217–222, 2000. * (18) Robert A. Laird and Brandon S. Schamp. Competitive intransitivity promotes species coexistence. American Naturalist, 168(2):182–193, 2006. * (19) Tobias Reichenbach, Mauro Mobilia, and Erwin Frey. Coexistence versus extinction in the stochastic cyclic Lotka-Volterra model. Physical Review E - Statistical, Nonlinear, and Soft Matter Physics, 74(5):1–11, 2006. * (20) Tobias Reichenbach, Mauro Mobilia, and Erwin Frey. Mobility promotes and jeopardizes biodiversity in rock-paper-scissors games. Nature, 448(7157):1046–1049, 2007. * (21) Joachim Mathiesen, Namiko Mitarai, Kim Sneppen, and Ala Trusina. Ecosystems with mutually exclusive interactions self-organize to a state of high diversity. Physical Review Letters, 107(18):1–5, 2011. * (22) Namiko Mitarai, Joachim Mathiesen, and Kim Sneppen. Emergence of diversity in a model ecosystem. Physical Review E - Statistical, Nonlinear, and Soft Matter Physics, 86(1):1–6, 2012. * (23) Alexander Dobrinevski and Erwin Frey. Extinction in neutrally stable stochastic Lotka-Volterra models. Physical Review E - Statistical, Nonlinear, and Soft Matter Physics, 85(5):1–12, 2012. * (24) Werner Ulrich, Santiago Soliveres, Wojciech Kryszewski, Fernando T. Maestre, and Nicholas J. Gotelli. Matrix models for quantifying competitive intransitivity from species abundance data. Oikos, 123(9):1057–1070, 2014. * (25) Attila Szolnoki, Mauro Mobilia, Luo-Luo Jiang, Bartosz Szczesny, Alastair M Rucklidge, and Matjaž Perc. Cyclic dominance in evolutionary games: a review. Journal of the Royal Society Interface, 11(100):20140735, 2014. * (26) Jonathan M. Levine, Jordi Bascompte, Peter B. Adler, and Stefano Allesina. Beyond pairwise mechanisms of species coexistence in complex communities. Nature, 546(7656):56–64, 2017. * (27) Santiago Soliveres and Eric Allan. Everything you always wanted to know about intransitive competition but were afraid to ask. Journal of Ecology, 106(3):807–814, 2018. * (28) Mark Finkelstein and Edward O Thorp. Nontransitive Dice With Equal Means *. pages 1–24, 2006. * (29) Brian Conrey, James Gabbard, Katie Grant, Andrew Liu, and Kent E. Morrison. Intransitive Dice. Mathematics Magazine, 89(2):133–143, 2016. * (30) V. A. Traag, L. Waltman, and N. J. van Eck. From Louvain to Leiden: guaranteeing well-connected communities. Scientific Reports, 9(1):1–12, 2019. * (31) See supplemental material at [url] for more details. * (32) Robert M May. Will a large complex system be stable? Nature, 238(5364):413–414, 1972. * (33) Garrett Hardin. The competitive exclusion principle: An idea that took a century to be born has implications in ecology, economics, and genetics. science, 131(3409):1292–1297, 1960. * (34) Simon A Levin. Community equilibria and stability, and an extension of the competitive exclusion principle. The American Naturalist, 104(939):413–423, 1970. * (35) Jan O Haerter, Namiko Mitarai, and Kim Sneppen. Food web assembly rules for generalized lotka-volterra equations. PLoS computational biology, 12(2):e1004727, 2016. * (36) Caroline M Tucker, Marc W Cadotte, Silvia B Carvalho, T Jonathan Davies, Simon Ferrier, Susanne A Fritz, Rich Grenyer, Matthew R Helmus, Lanna S Jin, Arne O Mooers, et al. A guide to phylogenetic metrics for conservation, community ecology and macroecology. Biological Reviews, 92(2):698–715, 2017. * (37) Manfred Eigen, John McCaskill, and Peter Schuster. The molecular quasi-species. Advances in chemical physics, 75:149–263, 1989.
# Domain Adaptive Scene Text Detection via Subcategorization Zichen Tian1, Chuhui Xue2, Jingyi Zhang1, Shijian Lu1 1Nanyang Technological University, 2ByteDance Inc. <EMAIL_ADDRESS><EMAIL_ADDRESS> ###### Abstract Most existing scene text detectors require large-scale training data which cannot scale well due to two major factors: 1) scene text images often have domain-specific distributions; 2) collecting large-scale annotated scene text images is laborious. We study domain adaptive scene text detection, a largely neglected yet very meaningful task that aims for optimal transfer of labelled scene text images while handling unlabelled images in various new domains. Specifically, we design SCAST, a subcategory-aware self-training technique that mitigates the network overfitting and noisy pseudo labels in domain adaptive scene text detection effectively. SCAST consists of two novel designs. For labelled source data, it introduces pseudo subcategories for both foreground texts and background stuff which helps train more generalizable source models with multi-class detection objectives. For unlabelled target data, it mitigates the network overfitting by co-regularizing the binary and subcategory classifiers trained in the source domain. Extensive experiments show that SCAST achieves superior detection performance consistently across multiple public benchmarks, and it also generalizes well to other domain adaptive detection tasks such as vehicle detection. Codes are available at https://github.com/doem97/SCAST. ## 1 Introduction Figure 1: The proposed subcategorization technique mitigates the overfitting in domain adaptive scene text detection: Existing domain adaptive scene text detector suffers from clear overfitting with over-confident predictions with most prediction probabilities lying around $0$ or $1$ (blue curves). The proposed subcategorization mitigates the overfitting effectively (orange curves), leading to higher entropy and lower pseudo-labeling error. The studies adopt scene text detector EAST [71] over domain adaptive scene text detection task SynthText $\to$ IC15. Scene text detection has been studied intensely for years, largely due to its wide applications in many real-world tasks such as scene understanding [1, 46], image retrieval [55, 17], autonomous indoor and outdoor navigation [39], etc. With the recent advances of deep neural networks, it has achieved significant progress under the presence of large-scale annotated training images [29, 50, 71, 31, 40, 68, 56, 9, 30, 67]. However, collecting large- scale annotated scene text images is laborious which has become a bottleneck while facing various scene text detection tasks that often have different camera setups, environmental parameters, etc. Unsupervised domain adaptation (UDA), which aims to learn from a labeled source domain for a well-performing model in an unlabelled target domain, has achieved great success in different computer vision tasks such as semantic segmentation [18, 64, 54], person re-identification [14, 34, 66], object detection [19, 6, 58, 70, 69], etc. However, existing UDA methods often suffer from a clear overfitting problem while applied to the scene text detection task. Take the prevalent self-training [57, 18, 72, 73] as an example. It tackles UDA by pseudo-labeling target samples for network retraining and has demonstrated superb performance while handling multi-class data as in general semantic segmentation and object detection tasks. Scene text detection instead involves a bi-class pseudo-label prediction task (i.e., foreground text and background stuff) where the prediction scores often have an extreme bimodal pattern with two sharp peaks at probabilities 0 and 1 as illustrated in Fig. 1. As a result, the predictions tend to be over-confidently wrong which introduces clear overfitting together with very high model likelihood and severe noises in pseudo-labelling. We design SCAST, a subcategory-aware self-training technique that introduces subcategorization for robust UDA for the scene text detection task. For the labelled source-data, SCAST identifies multiple pseudo subcategories for both foreground texts and background stuff via clustering which converts a bi-class prediction task to a multi-class prediction task, leading to more generalizable source models via regularization with a multi-class learning objective. For unlabelled target data, SCAST predicts pseudo labels with both bi-class and subcategory-aware classifiers which mitigates overfitting and allows transferring more diverse and informative source knowledge to the target domain. In addition, it exploits the prediction consistency between the bi-class and subcategory-aware classifiers over target samples which leads to better domain adaptation and more accurate pseudo-labelling of target images. The contributions of this work are threefold. First, we identify the overfitting issue in domain adaptive scene text detection and propose a subcategorization method that mitigates the overfitting effectively. To the best of our knowledge, this is the first work that explores subcategorization in domain adaptive scene text detection tasks. Second, we design SCAST that can learn more generalizable source models by identifying subcategories for model regularization. SCAST can transfer more diverse source knowledge and improve pseudo-labeling of target samples with prediction consistency between the bi-class and subcategory-aware classifiers learnt from source data. Third, extensive experiments show that SCAST achieve superior detection performance and the subcategorization idea can generalize to other bi-class detection tasks with little adaptation. Figure 2: The framework of the proposed SCAST: SCAST mitigates the overfitting of text detectors via subcategorization in the source and the target domains: 1) In the labeled source domain, we conduct feature clustering to discover fine-grained subcategories learning with which relieves the overfitting in bi- class classification greatly; 2) In the unlabeled target domain, we co- regularize pseudo-labelling with the bi-class and multi-class classifiers learnt from source samples, which reduces pseudo labelling noises in self- training effectively. ## 2 Related Work Scene text detection aims to locate texts in scene images and it has been widely explored via box regression and image segmentation. The regression approach exploits different object detection frameworks and text-specific shapes and orientations [71, 28, 27, 41, 68, 49, 60, 62, 61, 63]. For example, RRD [31] extracts rotation-sensitive features with SSD [35] for rotation- invariant detection. Textboxes++ [28] modifies convolutional kernels and anchor boxes to capture various text shapes effectively. EAST [71] directly infers quadrangles without proposal mechanism. However, the regression approach often struggle while handling scene texts with irregular shapes. The segmentation approach addresses this issue effectively by predicting a semantic label for each image pixel [10, 56, 9, 51, 30]. For example, PixelLink [10] locates text regions by associating neighborhood image pixels. PSENet [56] generates text bounding boxes from multi-scale segmentation maps. DB [30] introduces differentiable binarization for adaptive thresholding while predicting bounding regions from segmentation maps. Many existing methods can achieve very impressive detection performance under the presence of large-scale labelled training images that are often laborious to collect. Domain adaptive scene text detection, which aims to exploit previous collection of annotated scene text images for handling various new data, is largely neglected. Several recent studies [4, 57, 7, 66] attempt to tackle this challenge but they focus on adapting existing UDA methods without addressing text-specific problems. Unsupervised domain adaptation (UDA) has been studied extensively via three typical approaches, namely, adversarial learning, image translation, and self- training. Adversarial learning aligns source and target domains by minimizing certain distribution discrepancy in feature spaces [3, 37, 38, 15, 16]. Image translation works by translating source images to have similar target styles [48, 22, 2]. The two approaches cannot handle scene text detection well as they focus on adapting low-frequency appearance information. Self-training works by pseudo-labelling target data, which has been widely explored for various multi-class recognition tasks such as sign recognition [44], semantic segmentation [72, 73] and panoptic segmentation [24, 33]. We adopt self-training that pseudo-labels target images with a source-trained model. However, we focus on a more challenging bi-class self-training task that suffers from severe overfitting. We address the challenge by introducing subcategory pseudo-labeling that mitigates the overfitting by learning from multiple subcategories beyond the original background and foreground texts. ## 3 Method The proposed subcategory-aware self-training explores subcategorization to mitigate the overfitting in domain adaptive scene text detection task as illustrated in Fig. 2. In the source domain, it aims for Subcategory Discovery and Learning that first performs feature clustering to determine multiple subcategories beyond the original two classes and then employs the subcategory labels as pseudo labels to learn multi-class classification model. In the target domain, it aims for Subcategory Regularized Self-Training that employs the learnt multi-class classifier to label target samples and retrain the network iteratively, more details to be described in the ensuing subsections. ### 3.1 Problem Definition Given scene text images in source and target domains $\left\\{X^{s},X^{t}\right\\}\subset\mathbb{R}^{H\times W\times 3}$ where only the source data $X^{s}$ are labelled, we aim to learn a scene text detection model that performs well over the target samples $X^{t}$. The baseline model $\mathbf{G}$ is trained on labeled source data with loss: $\mathcal{L}_{\mathrm{bi}}\left(X^{s},Y^{s};\mathbf{G}=\mathbf{C}(\mathbf{E})\right),$ (1) where $Y^{s}$ is bi-class annotations of the source data and $\mathcal{L}_{\mathrm{bi}}$ denotes a binary classification loss (e.g. dice loss [56] or binary cross-entropy loss [30, 71]). The baseline model $\mathbf{G}$ consists of a bi-class classifier $\mathbf{C}$ and a feature extractor $\mathbf{E}$. ### 3.2 Subcategory Discovery and Learning In the source domain, we perform feature clustering to discover multiple subcategories on top of the original two-class labels to address the overfitting problem. Besides a bi-class classifier trained with the original scene text bounding boxes, a multi-class classifier is trained by using the discovered subcategories as pseudo labels as illustrated in Fig. 2. The multi- class classifier mitigates the overconfident predictions effectively which can be observed in the Histograms of Prediction Scores that are produced by the bi-class and multi-class classifiers. Subcategory Discovery in Source Domain. SCAST discovers subcategories by clustering features instead of spatial relationships among image patches or pixels as in [20, 12, 65, 52, 23, 5]. Feature clustering can produce fine- grained sub-clusters of scene texts that are distinguishable by their high- level features such as textures, strokes, etc. Given $X^{s}$, we first obtain their feature maps $f^{s}$ with a pre-trained feature extractor $\mathbf{E}$ of the baseline model $\mathbf{G}$. DBSCAN [12] clustering is then applied to $f^{s}$ to discover sub-clusters for texts and image background (obtained with the source annotations) separately. The indexes of the identified sub-clusters are then used as pseudo labels $\hat{Y}^{s}_{\mathrm{sub}}$ to train a multi-class classifier. The subcategory pseudo-labeling process can be formulated by: $\hat{Y}_{\mathrm{sub}}^{s}=\mathbf{\Gamma}\left(f^{s}\right),f^{s}=\mathbf{E}\left(X^{s}\right),$ (2) where $\mathbf{\Gamma}$ denotes the subcategory pseudo labelling operation. For efficiency, we perform clustering on down-sampled feature maps with a down-sample ratio $d$, where each feature point corresponds to a text-height area. Figure 3: The framework of the proposed SCAST: SCAST consists of a multi- class subcategory classifier $\mathbf{C}_{\mathrm{sub}}$, a source-domain subcategorizer $\mathbf{\Gamma}$, and a target-domain co-regularizer $\mathbf{\Phi}_{\mathrm{reg}}$. In the source domain, $\mathbf{\Gamma}(\cdot)$ clusters features of source samples to produce multi-class subcategory pseudo labels $\hat{Y}_{\mathrm{sub}}^{s}$ and applies them to re-train a network model that suffers from much less overfitting than that trained with the original bi-class labels $Y^{s}$. In the target domain, $\mathbf{\Phi}_{\mathrm{reg}}(\cdot)$ co-regularizes between the bi-class and multi-class subcategory pseudo labels $\hat{Y}_{\mathrm{bi}}^{t}$ and $\hat{Y}_{\mathrm{sub}}^{t}$ to filter out noisy pseudo labels which further mitigates overfitting and improves domain adaptation effectively. Subcategory Learning in Source Domain. To learn from the pseudo-labeled subcategories in the source domain, we include a multi-class classifier $\mathbf{C}_{\mathrm{sub}}$ on top of the feature extractor $\mathbf{E}$. The multi-class subcategory classifier is learnt with a cross-entropy loss: $\mathcal{L}_{\mathrm{sub}}\left(P^{s}_{\mathrm{sub}},\hat{Y}^{s}_{\mathrm{sub}}\right)=\\\ -\sum_{h,w}\sum_{k\in K}\left(\hat{Y}_{\mathrm{sub}}^{s}\right)^{(h,w,k)}\log\left(\left(P^{s}_{\mathrm{sub}}\right)^{(h,w,k)}\right),$ (3) where $P^{s}_{\mathrm{sub}}$ is predictions of multi-class classifier and $\hat{Y}_{\mathrm{sub}}^{s}$ refers to multi-class pseudo labels (Eq. 2). We denote the subcategory class number as $K$, which is decided automatically by DBSCAN algorithm. The learned model $\mathbf{G}^{s}$ is then applied to target domain through a subcategory regularized self-training, introduced next. ### 3.3 Subcategory Regularized Self-Training In the target domain, we employ bi-class and subcategory classifiers to predict pseudo labels and co-regularize them for effective knowledge transfer. As illustrated in Fig. 2, the subcategory predictions with less overfitting co-regularize the bi-class predictions, which help filter out noisy pseudo labels as ‘Untrain’ effectively. Subcategory learning in target domain. We select the most confident predictions by $\mathbf{G}^{s}$ as the pseudo labels for target images. Algorithm 1 Determination of threshold $\theta$ 0: Predictions $P_{\mathrm{sub}}^{t}$; Selection portion $\rho\%$ 1: for $k=1$ to $K$ do 2: $M^{k}$ = sort($(P^{t}_{\mathrm{sub}})^{k}$, order = descending) 3: $l^{k}$ = length$(M^{k})\times\rho\%$ 4: $\theta^{k}$ = $M^{k}[l^{k}]$ 5: end for 6: return $\theta=\\{\theta^{k}|k\in K\\}$ As summarized in Algorithm. 1, for each subcategory, we employ a selection proportion $\rho$ [72, 73] to select the top $\rho\%$ most confident predictions as pseudo labels. With the threshold value $\theta$ at the top $\rho\%$-th position, the pseudo-labeling process can be formulated as: $\hat{Y}^{t}_{\mathrm{sub}}=\mathbf{\Theta}\left(\theta,P^{t}_{\mathrm{sub}}\right)=\begin{cases}1,&\text{if}\ (P^{t}_{\mathrm{sub}})^{k}>\theta^{k},\\\ 0,&\text{otherwise},\\\ \end{cases}$ (4) where $k\in K$ is the subcategory index, and $\hat{Y}^{t}_{\mathrm{sub}}$ is the subcategory pseudo labels. The bi-class pseudo labelling follows the same formula: $\hat{Y}^{t}_{\mathrm{bi}}=\mathbf{\Theta}\left(\theta,P^{t}_{\mathrm{bi}}\right)$ where $k=0,1$ refers to texts or image background, respectively. The network can then be re-trained with the target data with the obtained pseudo labels $\left\\{\hat{Y}^{t}_{\mathrm{bi}},\hat{Y}^{t}_{\mathrm{sub}}\right\\}$. The training losses are the same as the supervised losses $\mathcal{L}_{\mathrm{bi}}$ and $\mathcal{L}_{\mathrm{sub}}$ (in Eq. 1 and Eq. 3) in the source domain. The learning objective in the target domain is thus a weighted combination of bi-class loss and subcategory loss: $\displaystyle\begin{split}\mathcal{L}^{t}=&\lambda_{\mathrm{sub}}\mathcal{L}_{\mathrm{sub}}\left(P_{\mathrm{sub}}^{t},\hat{Y}_{\mathrm{sub}}^{t};\mathbf{C}_{\mathrm{sub}},\mathbf{E}\right)\ +\\\ &\lambda_{\mathrm{bi}}\mathcal{L}_{\mathrm{bi}}\left(P_{\mathrm{bi}}^{t},\hat{Y}_{\mathrm{bi}}^{t};\mathbf{C}_{\mathrm{bi}},\mathbf{E}\right).\end{split}$ (5) where $\lambda_{\mathrm{bi}}$ is a weight parameter and we set it empirically at 1 in our implementation. Eq. 4 and Eq. 5 define a single iteration of self- training that generates pseudo labels and retrains the network in single- round. In practice, we implement multi-round optimization as described in Sec 3.4. Subcategory regularization in target domain. As shown in Fig. 2, we discard inconsistent predictions between the bi-class and subcategory classifiers, which effectively reduces false positives and pseudo-label noises and leads to more robust pseudo-labeling. Specifically, we measure the distance between bi- class predictions and subcategory predictions by cross-entropy and drop out predictions with large distances (Eq. 7). To control the proportion of dropped predictions, we set a ratio $\rho_{\mathrm{reg}}$ to drop out positions with top $\rho_{\mathrm{reg}}\%$ distance (we set $\rho_{\mathrm{reg}}\%=10\%$ empirically). We denote such co-regularization process as $\mathbf{\Phi}_{\mathrm{reg}}$ and formulate it as a loss minimization problem: $\displaystyle\mathbf{\Phi}_{\mathrm{reg}}=\operatorname*{arg\,min}_{\left\\{P^{t}_{\mathrm{bi}},P^{t}_{\mathrm{sub}}\right\\}}\mathcal{L}^{t}_{\mathrm{reg}}\left(P^{t}_{\mathrm{bi}},P^{t}_{\mathrm{sub}};\rho_{\mathrm{reg}}\right),$ (6) where $\mathcal{L}^{t}_{\mathrm{reg}}\left(P^{t}_{\mathrm{bi}},P^{t}_{\mathrm{sub}}\right)=-\left(P^{t}_{\mathrm{sub}}\right)\log\left(P^{t}_{\mathrm{bi}}\right)$ is the cross-entropy distance between bi- and multi-class predictions. We can get regularized pseudo labels by solving Eq. 6: $\left\\{\hat{Y}^{t}_{\mathrm{bi}},\hat{Y}^{t}_{\mathrm{sub}}\right\\}=\begin{cases}1,&\text{if}\ \mathcal{L}^{t}_{\mathrm{reg}}\left(P^{t}_{\mathrm{bi}},P^{t}_{\mathrm{sub}}\right)<\theta_{\mathrm{reg}},\\\ 0,&\text{otherwise},\\\ \end{cases}$ (7) where $\theta_{\mathrm{reg}}$ is the threshold value at top $\rho_{\mathrm{reg}}\%$-th position. ### 3.4 Network Optimization Method | Arch. | ICDAR13 [26] | ICDAR15 [25] | COCO-Text [53] | Total-Text (Reg) [45] | Mean ---|---|---|---|---|---|--- Precision | Recall | F-Score | Precision | Recall | F-Score | Precision | Recall | F-Score | Precision | Recall | F-Score | Precision | Recall | F-Score Supervised | - | 92.64 | 82.67 | 87.37 | 84.36 | 81.27 | 82.79 | 50.39 | 32.40 | 39.45 | 50.00 | 36.20 | 42.00 | 69.35 | 58.14 | 62.90 Baseline (EAST [71]) | - | 65.33 | 68.85 | 67.05 | 69.63 | 53.44 | 60.47 | 38.23 | 20.73 | 26.89 | 40.35 | 36.61 | 38.39 | 53.38 | 44.90 | 48.20 TST [57] | ST | 71.50 | 70.70 | 71.10 | 69.30 | 60.50 | 64.60 | 53.00 | 22.76 | 31.85 | 43.48 | 37.56 | 40.30 | 59.32 | 47.88 | 51.96 EntMin [18] | ST | 68.29 | 67.81 | 68.05 | 70.41 | 53.16 | 60.58 | 57.12 | 18.52 | 27.97 | 43.99 | 34.50 | 38.67 | 59.95 | 43.50 | 48.82 CBST [72] | ST | 70.28 | 71.08 | 70.68 | 71.36 | 58.40 | 64.23 | 55.17 | 20.96 | 30.38 | 43.17 | 38.23 | 40.55 | 60.00 | 47.17 | 51.46 CRST [73] | ST | 72.21 | 73.05 | 72.63 | 71.85 | 62.19 | 66.67 | 51.44 | 22.28 | 31.09 | 46.31 | 39.13 | 42.42 | 60.45 | 49.16 | 53.20 FDA [64] | Tran. | 47.28 | 73.88 | 57.66 | 59.13 | 55.18 | 57.09 | 44.69 | 19.79 | 27.43 | 37.24 | 40.15 | 38.64 | 43.86 | 47.02 | 43.35 ADVENT [54] | Adv. | 62.15 | 66.30 | 64.16 | 63.00 | 61.00 | 61.99 | 38.60 | 21.06 | 27.25 | 38.44 | 37.51 | 37.97 | 50.55 | 46.47 | 47.84 Ours (SCAST) | ST | 79.88 | 76.65 | 78.23 | 75.80 | 65.27 | 70.14 | 49.28 | 26.07 | 34.10 | 52.86 | 42.97 | 47.40 | 64.46 | 52.74 | 57.47 Table 1: Regular scene text detection: The experiments are conducted over domain adaptive scene text detection tasks SynthText $\to$ {ICDAR13, ICDAR15, COCO-Text and Total-Text (Reg)} using EAST [71]. ST stands for Self-training, Adv. for Adversarial training, and Tran. for Image-to image translation. Supervised refers to supervised learning from the labeled target data. Method | Arch. | Total-Text (Curve) [45] | CTW1500 [36] | TextSeg [59] | Mean ---|---|---|---|---|--- Precision | Recall | F-Score | Precision | Recall | F-Score | Precision | Recall | F-Score | Precision | Recall | F-Score Supervised | - | 81.77 | 75.11 | 78.30 | 80.57 | 75.55 | 78.00 | 89.02 | 81.55 | 85.12 | 83.79 | 77.40 | 80.47 Baseline (PSENet [56]) | - | 72.41 | 46.12 | 56.35 | 44.03 | 44.26 | 44.15 | 77.24 | 49.33 | 60.20 | 64.56 | 46.57 | 53.57 TST [57] | ST | 68.94 | 51.03 | 58.65 | 42.07 | 50.98 | 46.09 | 82.91 | 51.03 | 63.17 | 64.64 | 51.01 | 55.97 EntMin [18] | ST | 72.20 | 46.33 | 56.44 | 43.38 | 48.50 | 45.80 | 82.23 | 48.34 | 60.88 | 65.94 | 47.72 | 54.37 CBST [72] | ST | 71.39 | 47.57 | 57.09 | 44.58 | 49.19 | 46.77 | 79.17 | 52.15 | 62.88 | 65.05 | 49.64 | 55.58 CRST [73] | ST | 73.64 | 50.31 | 59.78 | 46.35 | 47.75 | 47.04 | 82.30 | 51.41 | 63.29 | 67.43 | 49.82 | 56.70 FDA [64] | Tran. | 71.60 | 43.50 | 54.12 | 45.02 | 44.38 | 44.70 | 78.21 | 50.02 | 61.02 | 64.94 | 45.97 | 53.28 ADVENT [54] | Adv. | 76.27 | 34.34 | 47.36 | 41.37 | 46.97 | 43.99 | 80.83 | 50.04 | 61.81 | 66.16 | 43.78 | 51.05 Ours (SCAST) | ST | 71.82 | 56.73 | 63.39 | 51.02 | 56.10 | 53.44 | 81.90 | 58.78 | 68.44 | 68.25 | 57.20 | 61.76 Table 2: Irregular scene text detection: The experiments are conducted over domain adaptive scene text detection tasks SynthText $\to$ {Total-Text (Curve), CTW1500, TextSeg} using PSENet [56]. ST stands for Self-training, Adv. for Adversarial training, and Tran. for Image translation architecture. Supervised refers to supervised learning from the labeled target data. Figure 4: Samples of clustered subcategories: Our method captures high-level features ($e.g.$ curvature, texture, stroke shapes, hue, etc.) and clusters text instances and background into different subcategories with different features. The experiments were conducted on ICDAR13 with EAST detector. $k$ is the subcategory index and we show five subcategories ($k=1$, 12 and 19 for text, 37 and 40 for background) for illustration. Method | Arch. | ICDAR13 [26] | ICDAR15 [53] | Total-Text (Reg) [45] | Mean ---|---|---|---|---|--- Precision | Recall | F-Score | Precision | Recall | F-Score | Precision | Recall | F-Score | Precision | Recall | F-Score Supervised | - | 92.64 | 82.67 | 87.37 | 84.36 | 81.27 | 82.79 | 50.00 | 36.20 | 42.00 | 75.67 | 66.71 | 70.72 Baseline (EAST [71]) | - | 62.31 | 70.31 | 66.07 | 71.29 | 55.20 | 62.22 | 41.03 | 35.28 | 37.94 | 58.21 | 53.60 | 55.41 TST [57] | ST | 66.27 | 72.05 | 69.04 | 73.46 | 57.21 | 64.32 | 42.23 | 38.10 | 40.06 | 60.65 | 55.79 | 57.81 EntMin [18] | ST | 64.13 | 72.98 | 68.27 | 70.68 | 56.08 | 62.54 | 44.06 | 36.19 | 39.74 | 59.62 | 55.08 | 56.85 CBST [72] | ST | 68.68 | 71.36 | 69.99 | 69.05 | 59.30 | 63.80 | 42.28 | 37.92 | 39.98 | 60.00 | 56.19 | 57.93 CRST [73] | ST | 71.33 | 74.42 | 72.84 | 73.89 | 63.09 | 68.06 | 43.60 | 39.33 | 41.36 | 62.94 | 58.95 | 60.75 FDA [64] | Tran. | 43.62 | 61.50 | 51.04 | 62.07 | 52.75 | 57.03 | 37.92 | 33.40 | 35.52 | 47.87 | 49.22 | 47.86 ADVENT [54] | Adv. | 57.42 | 61.96 | 59.60 | 65.83 | 51.47 | 57.77 | 43.15 | 33.29 | 37.58 | 55.47 | 48.91 | 51.65 Ours (SCAST) | ST | 76.48 | 78.26 | 77.36 | 77.24 | 67.40 | 71.99 | 47.06 | 41.75 | 44.25 | 66.93 | 62.47 | 64.53 Table 3: Real to real adaptation: The experiments are conducted over real-to- real domain adaptive scene text detection tasks COCO-Text $\to$ {ICDAR13, ICDAR15 and Total-Text (Reg)} using EAST [71]. ST stands for Self-training, Adv. for Adversarial training, and Tran. for Image translation architecture. Supervised refers to supervised learning from the labeled target data. We optimize SCAST by multiple rounds of self-training. Each round consists of two alternate steps including updating of pseudo labels $\\{\hat{Y}^{t}_{\mathrm{bi}},\hat{Y}^{t}_{\mathrm{sub}}\\}$ and retraining network $\mathbf{G}$ with the updated pseudo labels. The optimization can be formulated as a unified loss minimization problem: $\operatorname*{arg\,min}_{\mathbf{G},\hat{Y}^{t}_{\mathrm{bi}},\hat{Y}^{t}_{\mathrm{sub}}}\mathcal{L}^{t}\left(P^{t}_{\mathrm{bi}},P^{t}_{\mathrm{sub}},\hat{Y}^{t}_{\mathrm{bi}},\hat{Y}^{t}_{\mathrm{sub}};\mathbf{G},\theta\right),$ (8) where $\mathbf{G}(\cdot)=\left\\{\mathbf{C}_{\mathrm{bi}}\left(\mathbf{E}(\cdot)\right),\mathbf{C}_{\mathrm{sub}}\left(\mathbf{E}(\cdot)\right)\right\\}$ and $\mathcal{L}^{t}$ is the weighted cross-entropy loss (Eq. 5). Pseudo label prediction. We fix network $\mathbf{G}$ while predicting pseudo labels $\hat{Y}^{t}_{\mathrm{bi}}$ and $\hat{Y}^{t}_{\mathrm{sub}}$ with thresholds $\theta$ in each round. We predict pseudo labels in an “easy-to- hard” manner so that the network can learn from confident pseudo labels first. This strategy helps reduce the impact of noisy pseudo labels for both bi-class multi-class objectives effectively. In implementation, we increase the selection proportion $\rho\%$ gradually after each training round which reduces the corresponding thresholds $\theta$ accordingly. The optimization in this step can be formulated by: $\operatorname*{arg\,min}_{\hat{Y}_{\mathrm{sub}}^{t},\hat{Y}_{\mathrm{bi}}^{t}}\mathcal{L}^{t}\left(P^{t}_{\mathrm{bi}},P^{t}_{\mathrm{sub}},\hat{Y}^{t}_{\mathrm{bi}},\hat{Y}^{t}_{\mathrm{sub}};\mathbf{G},\theta_{(i)}\right),$ (9) where $i$ is the optimization round. Update the network. We fix pseudo labels while retraining $\mathbf{G}$. As more pseudo labels are selected round by round, model $\mathbf{G}$ learns and mitigates the domain gap gradually. The optimization can be formulated by: $\operatorname*{arg\,min}_{\mathbf{G}}\mathcal{L}^{t}\left(P^{t}_{\mathrm{bi}},P^{t}_{\mathrm{sub}},\hat{Y}^{t}_{\mathrm{bi}},\hat{Y}^{t}_{\mathrm{sub}};\mathbf{G},\theta_{(i)}\right),$ (10) where $i$ is the optimization round. ## 4 Experiments ### 4.1 Datasets and Evaluation Datasets. Our experiments involve several domain adaptive scene text detection tasks including SynthText $\to$ {ICDAR13, ICDAR15, COCO-Text17, Total-Text (Reg)} for regular scene text and SynthText $\to$ {Total-Text (Curve), CTW1500, TextSeg} for irregular scene text. More dataset details are provided in the appendix. Evaluation. We evaluate using Precision, Recall and F-Score as in [25, 26]. The evaluations are based on the Intersection-over-Union (IoU) criterion in PASCAL [13], with a widely adopted threshold of $50\%$. During training and evaluations, we ignore unreadable text regions that are labeled by either “do not care” or “illegal” in all datasets. Following previous UDA methods [72, 73, 54], we adopt dense evaluation in the experiments. ### 4.2 Implementation Details Subcategorizor. We perform DBSCAN clustering on feature maps to discover subcategories. Specifically, we down-sample the feature maps to make each feature point correspond to a text-height region. We conduct DBSCAN clustering on these feature maps for both text and background (the distance $\epsilon$ in DBSCAN is discussed in Sec. 4.3). Network architectures. We evaluate our method with regular text detector EAST [71] and irregular text detector PSENet [56]. The two detectors have simple detection strategies without complex post-processing, which reduce interference in evaluations. For EAST, we adopt the settings of the officially released code, which use VGG-16 pre-trained model on FPN [32] as the backbone, and dice-loss [47] as classification loss. During post-processing, we use Local-Aware NMS with a threshold $0.5$, same as the original paper. For PSENet, we follow PSENet-1s from the original paper, which uses ResNet [21] pre-trained on ImageNet [11] as backbone. In addition, we train the network from scratch on SynthText without pretraining on the ICDAR17-MLT [42] to avoid introducing extra data. Training details. We use SGD optimizer with momentum $0.9$ and a weight decay $5e-4$ in training. The initial learning rate is $1e-3$ and then decays by a polynomial policy of power $0.9$. For training with source data, we adopted all data augmentation strategies as used in the original papers. For training with target data, we remove all data augmentations and resize training images to $512\times 512$. We set the training batch-size at $12$, including $6$ source images and $6$ target images during domain adaption. Self-Training details. The pseudo label proportion $\rho\%$ and related threshold $\theta$ are important for pseudo label prediction. To avoid heuristic setups, we adopt linearly increased $\rho\in\\{20,40,60,80,100\\}$ to select the top $\\{20\%,40\%,60\%,80\%,100\%\\}$ confident predictions as pseudo labels. ### 4.3 Ablation Study and Analysis We study the contributions of each design in the proposed SCAST, including a $K$-Subcategorizor $SC_{\mathrm{K}}$ for source data, a subcategory self- training $ST_{\mathrm{K}}$ and a subcategory regularized self-training $ST_{\mathrm{K,reg}}$ over the target data. As Table 4 shows, including subcategorization in $+SC_{\mathrm{K}}$ in the source domain outperforms the Baseline by a large margin, showing that the proposed subcategorization mitigates the overfitting effectively. For target data, including the bi-class self-training with $+ST_{\mathrm{2}}$ improves the scene text detection greatly, and combining the source-domain subcategorization with the target-domain bi-class self-training in $+ST_{\mathrm{2}}+SC_{\mathrm{K}}$ further improves the detection by a large margin. In addition, including the multi-class subcategory self-training on target data in $+ST_{\mathrm{2}}+SC_{\mathrm{K}}+ST_{\mathrm{K}}$ also improves clearly. Finally, the inclusion of co-regularization between bi-class and multi-class pseudo labels of target data in $+ST_{\mathrm{2}}+SC_{\mathrm{K}}+ST_{K,reg}$ produces the best detection. The ablation studies show the effectiveness of the proposed subcategorization and subcategory regularized self-training. ### 4.4 Comparison with SOTA Method | $\mathcal{L}^{s}_{\mathrm{bi}}$ | $\mathcal{L}^{s}_{\mathrm{sub}}$ | $\mathcal{L}^{t}_{\mathrm{bi}}$ | $\mathcal{L}^{t}_{\mathrm{sub}}$ | $\mathcal{L}^{t}_{\mathrm{reg}}$ | Prec. | Rec. | F-sco. ---|---|---|---|---|---|---|---|--- Baseline | ✓ | | | | | 69.63 | 53.44 | 60.47 \+ $SC_{\mathrm{K}}$ | ✓ | ✓ | | | | 67.94 | 62.93 | 65.34 \+ $ST_{\mathrm{2}}$ | ✓ | | ✓ | | | 71.36 | 58.40 | 64.23 \+ $ST_{\mathrm{2}}$ \+ $SC_{\mathrm{K}}$ | ✓ | ✓ | ✓ | | | 75.12 | 63.01 | 68.53 \+ $ST_{\mathrm{2}}$ \+ $SC_{\mathrm{K}}$ \+ $ST_{\mathrm{K}}$ | ✓ | ✓ | ✓ | ✓ | | 73.09 | 65.98 | 69.35 \+ $ST_{\mathrm{2}}$ \+ $SC_{\mathrm{K}}$ \+ $ST_{K,reg}$ | ✓ | ✓ | ✓ | ✓ | ✓ | 75.80 | 65.27 | 70.14 Table 4: Ablation study of SCAST over domain adaptive scene text detection task SynthText $\to$ ICDAR15 using EAST. $SC_{\mathrm{K}}$, $ST_{\mathrm{2}}$, $ST_{\mathrm{K}}$, and $ST_{K,reg}$ refer to subcategorization (via $\mathcal{L}^{s}_{\mathrm{sub}}$), bi-class CBST [72] (via $\mathcal{L}^{t}_{\mathrm{bi}}$), subcategory self-training (via $\mathcal{L}^{t}_{\mathrm{sub}}$), and subcategory regularized self-training (via $\mathcal{L}^{t}_{\mathrm{reg}}$), respectively. Method | $\epsilon$ | Subcategory Number | Prec. | Rec. | F-sc. ---|---|---|---|---|--- Text | Background Baseline | - | - | - | 69.63 | 53.44 | 60.47 $+SC_{K=15}$ | $0.5$ | 6 | 9 | 64.28 | 60.82 | 62.50 $+SC_{K=21}$ | $0.1$ | 10 | 11 | 69.48 | 59.30 | 63.99 $+SC_{K=33}$ | $0.05$ | 19 | 14 | 65.66 | 62.27 | 63.92 $+SC_{K=54}$ | $0.01$ | 35 | 19 | 67.94 | 62.93 | 65.34 Table 5: Maximum cluster distance $\epsilon$. With the decrease of the maximum cluster distance $\epsilon$, DBSCAN clustering discovers more subcategories. $SC_{\mathrm{K}}$ stands for $K$-classes SubCategorization in the source domain, where $K$ equals to the number of subcategories in both Text and Background. The experiments were performed over task SynthText $\to$ ICDAR15 with EAST backbone. Network | ICDAR13 | ICDAR15 | COCO-Text | Total-Text ---|---|---|---|--- Entropy | Err. Rate | Likelihood | Entropy | Err. Rate | Likelihood | Entropy | Err. Rate | Likelihood | Entropy | Err. Rate | Likelihood Baseline (Text) | 0.0092 | 38.49% | 844.79 | 0.0099 | 63.35% | 833.52 | 0.0101 | 67.83% | 832.02 | 0.0091 | 68.49% | 844.80 \+ $SC_{\mathrm{K}}$ | 0.1553 | 33.16% | 587.70 | 0.2305 | 58.90% | 537.10 | 0.2724 | 67.92% | 511.69 | 0.2222 | 63.65% | 587.70 Baseline (Back) | 0.0007 | 0.78% | 1101.30 | 0.0001 | 0.15% | 1262.55 | 0.0002 | 0.21% | 1101.30 | 0.0006 | 0.83% | 1116.67 \+ $SC_{\mathrm{K}}$ | 0.1186 | 0.70% | 653.80 | 0.1013 | 0.11% | 770.34 | 0.1451 | 0.15% | 653.80 | 0.1290 | 0.96% | 648.00 Table 6: Overfitting mitigation. Experiments are conducted on SynthText $\to$ ICDAR15 with EAST [71]. Baseline (Text) and Baseline (Back) measure overfitting of text and background with the baseline model. $SC_{\mathrm{K}}$ denotes the proposed $K$-class subcategorization. ‘Likelihood’ is described in page 2 footnote and ‘Err. Rate’ is classification error rate. Figure 5: Comparison with the state-of-the-art. The Baseline detector and the compared methods produce over-confident predictions with an extreme bimodal distribution (i.e., most prediction probabilities are around $0$ or $1$), indicating overfitting with degraded detection (with more false negatives). The proposed subcategory-aware detector instead generates prediction probabilities that smoothly distribute between $0$ and $1$, indicating less overfitting. The evaluation is performed on task SynthText $\to$ TotalText with EAST detector. We compare SCAST with several state-of-the-art UDA methods that adopt self- training with EntMin [18], TST [57], CBST [72], and CRST [73]), image translation with FDA [64] and adversarial with ADVENT [54]. Tables 1 and 2 show experimental results on regular and irregular scene text under the synthetic-to-real setup, respectively. We also evaluate real-to-real adaptation to show the robustness of our method in Table 3. We can see that SCAST outperforms all compared methods in F-score. Specifically, methods using image translation and adversarial learning do not perform well as image translation tends to smooth text strokes while adversarial learning focuses on aligning image background. For UDA using self-training, TST [57] and CRST [73] outperform the Baseline clearly as retraining using pseudo-labeled target data often produces stronger models. However, the improvements are limited as their predicted pseudo labels are noisy due to the overfitted classifiers as shown in Table 6. As a comparison, SCAST mitigates the overfitting and reduces pseudo label errors and produces clearly better detection as illustrated in Fig. 5 (more samples available in appendix). Note TST combines adversarial learning with self-training while SCAST performs self-training alone. Note SCAST even outperforms supervised models over the Total-Text (Reg) as shown in Tables 1 and 3. The outstanding performance is largely because the subcategory-aware classifier predicts text regions as multiple fragments instead of a single quadrangle. Such predictions are much more compatible for the detection of curved text instances which cannot be located well with quadrangles. ### 4.5 Discussion Maximum cluster distance $\epsilon$. Table 5 shows the impact of the maximum cluster distance $\epsilon$ in DBSCAN clustering, where the baseline is EAST detector. We can observe a trade-off between the recall and precision, showing that our method greatly improves the recall ($53.44$ to $62.93$) while introducing marginal decrease in precision ($69.63$ to $67.94$) which can be compensated with self-training over target data. We adopt $\epsilon=0.01$ in our implemented system. Subcategorization. Fig. 4 shows subcategorized foreground texts and image backgrounds. We can see that DBSCAN clustering can group image patches into subcategories with similar features in polygonal lines, traffic signs, curved corners and thin strokes (e.g., subcategories 1, 12, 19, 37, and 40 for illustration). Fig. 5 further shows subcategory predictions where over- confident predictions are mitigated and False Negatives are reduced effectively. Overfitting mitigation. We adopt entropy, error rate, and likelihood metrics to evaluate how the proposed subcategorization helps mitigate the overfitting in domain adaptive scene text detection. Table 6 shows experimental results over four widely adopted datasets. It can be observed that including the proposed subcategorization (i.e., via +$SC_{k}$) helps mitigate the overfitting effectively with higher entropy, fewer pseudo-labeling errors, and lower likelihood. Other bi-class tasks The proposed SCAST can generalize well to other bi-class detection tasks. We evaluate another two bi-class detection tasks on vehicle detection and pedestrian detection. Experiments over the adaptation task GTA5 [43] $\rightarrow$ Cityscapes [8] show that SCAST outperforms the state-of- the-art by 2.9% and 3.1% (in foreground IoU), respectively. Please refer to the appendix for details. ## 5 Conclusion This paper presents SCAST, a subcategory-aware self-training technique for mitigating overfitting in domain adaptive scene text detection. We subcategorize source data by feature clustering which mitigates the overfitting substantially. For target data, we design a subcategory regularized self-training that employs the source-learned subcategory priors to co-regularizes bi-class and multi-class subcategory pseudo labels which further mitigates the overfitting effectively. Extensive experiments show that SCAST mitigates the overfitting in scene text detection task with superior detection performance. The idea of subcategorization is generic and can adapt to other bi-class detection and recognition tasks. We will explore it in our future research. ## References * [1] Ali Furkan Biten, Ruben Tito, Andres Mafla, Lluis Gomez, Marçal Rusinol, Ernest Valveny, CV Jawahar, and Dimosthenis Karatzas. Scene text visual question answering. In Proceedings of the IEEE/CVF international conference on computer vision, pages 4291–4301, 2019. * [2] Konstantinos Bousmalis, Nathan Silberman, David Dohan, Dumitru Erhan, and Dilip Krishnan. Unsupervised pixel-level domain adaptation with generative adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3722–3731, 2017. * [3] Konstantinos Bousmalis, George Trigeorgis, Nathan Silberman, Dilip Krishnan, and Dumitru Erhan. Domain separation networks. Advances in neural information processing systems, 29:343–351, 2016\. * [4] Danlu Chen, Lihua Lu, Yao Lu, Ruizhe Yu, Shunzhou Wang, Lin Zhang, and Tingxi Liu. Cross-domain scene text detection via pixel and image-level adaptation. In International Conference on Neural Information Processing, pages 135–143. Springer, 2019. * [5] Tao Chen, Shijian Lu, and Jiayuan Fan. S-cnn: Subcategory-aware convolutional networks for object detection. IEEE transactions on pattern analysis and machine intelligence, 40(10):2522–2528, 2017. * [6] Yuhua Chen, Wen Li, Christos Sakaridis, Dengxin Dai, and Luc Van Gool. Domain adaptive faster r-cnn for object detection in the wild. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3339–3348, 2018. * [7] Yudi Chen, Wei Wang, Yu Zhou, Fei Yang, Dongbao Yang, and Weiping Wang. Self-training for domain adaptive scene text detection. In 2020 25th International Conference on Pattern Recognition (ICPR), pages 850–857. IEEE, 2021. * [8] Marius Cordts, Mohamed Omran, Sebastian Ramos, Timo Rehfeld, Markus Enzweiler, Rodrigo Benenson, Uwe Franke, Stefan Roth, and Bernt Schiele. The cityscapes dataset for semantic urban scene understanding. In Proc. of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016. * [9] Pengwen Dai, Sanyi Zhang, Hua Zhang, and Xiaochun Cao. Progressive contour regression for arbitrary-shape scene text detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7393–7402, 2021. * [10] Dan Deng, Haifeng Liu, Xuelong Li, and Deng Cai. Pixellink: Detecting scene text via instance segmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32, 2018. * [11] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009. * [12] Martin Ester, Hans-Peter Kriegel, Jörg Sander, Xiaowei Xu, et al. A density-based algorithm for discovering clusters in large spatial databases with noise. In kdd, volume 96, pages 226–231, 1996. * [13] Mark Everingham, SM Ali Eslami, Luc Van Gool, Christopher KI Williams, John Winn, and Andrew Zisserman. The pascal visual object classes challenge: A retrospective. International journal of computer vision, 111(1):98–136, 2015. * [14] Yang Fu, Yunchao Wei, Guanshuo Wang, Yuqian Zhou, Honghui Shi, and Thomas S Huang. Self-similarity grouping: A simple unsupervised cross domain adaptation approach for person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6112–6121, 2019. * [15] Yaroslav Ganin and Victor Lempitsky. Unsupervised domain adaptation by backpropagation. In International conference on machine learning, pages 1180–1189. PMLR, 2015. * [16] Muhammad Ghifary, W Bastiaan Kleijn, Mengjie Zhang, David Balduzzi, and Wen Li. Deep reconstruction-classification networks for unsupervised domain adaptation. In European conference on computer vision, pages 597–613. Springer, 2016. * [17] Lluis Gomez, Andrés Mafla, Marçal Rusinol, and Dimosthenis Karatzas. Single shot scene text retrieval. In Proceedings of the European Conference on Computer Vision (ECCV), pages 700–715, 2018. * [18] Yves Grandvalet, Yoshua Bengio, et al. Semi-supervised learning by entropy minimization. CAP, 367:281–296, 2005. * [19] Dayan Guan, Jiaxing Huang, Aoran Xiao, Shijian Lu, and Yanpeng Cao. Uncertainty-aware unsupervised domain adaptation in object detection. IEEE Transactions on Multimedia, 2021. * [20] John A Hartigan and Manchek A Wong. Algorithm as 136: A k-means clustering algorithm. Journal of the royal statistical society. series c (applied statistics), 28(1):100–108, 1979. * [21] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016. * [22] Judy Hoffman, Eric Tzeng, Taesung Park, Jun-Yan Zhu, Phillip Isola, Kate Saenko, Alexei Efros, and Trevor Darrell. Cycada: Cycle-consistent adversarial domain adaptation. In International conference on machine learning, pages 1989–1998. PMLR, 2018. * [23] Jiaxing Huang, Dayan Guan, Shijian Lu, and Aoran Xiao. Mlan: Multi-level adversarial network for domain adaptive semantic segmentation. arXiv preprint arXiv:2103.12991, 2021. * [24] Jiaxing Huang, Dayan Guan, Aoran Xiao, and Shijian Lu. Cross-view regularization for domain adaptive panoptic segmentation. arXiv preprint arXiv:2103.02584, 2021. * [25] Dimosthenis Karatzas, Lluis Gomez-Bigorda, Anguelos Nicolaou, Suman Ghosh, Andrew Bagdanov, Masakazu Iwamura, Jiri Matas, Lukas Neumann, Vijay Ramaseshan Chandrasekhar, Shijian Lu, et al. Icdar 2015 competition on robust reading. In 2015 13th International Conference on Document Analysis and Recognition (ICDAR), pages 1156–1160. IEEE, 2015. * [26] Dimosthenis Karatzas, Faisal Shafait, Seiichi Uchida, Masakazu Iwamura, Lluis Gomez i Bigorda, Sergi Robles Mestre, Joan Mas, David Fernandez Mota, Jon Almazan Almazan, and Lluis Pere De Las Heras. Icdar 2013 robust reading competition. In 2013 12th International Conference on Document Analysis and Recognition, pages 1484–1493. IEEE, 2013. * [27] Minghui Liao, Guan Pang, Jing Huang, Tal Hassner, and Xiang Bai. Mask textspotter v3: Segmentation proposal network for robust scene text spotting. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XI 16, pages 706–722. Springer, 2020. * [28] Minghui Liao, Baoguang Shi, and Xiang Bai. Textboxes++: A single-shot oriented scene text detector. IEEE transactions on image processing, 27(8):3676–3690, 2018. * [29] Minghui Liao, Baoguang Shi, Xiang Bai, Xinggang Wang, and Wenyu Liu. Textboxes: A fast text detector with a single deep neural network. In Thirty-first AAAI conference on artificial intelligence, 2017\. * [30] Minghui Liao, Zhaoyi Wan, Cong Yao, Kai Chen, and Xiang Bai. Real-time scene text detection with differentiable binarization. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 11474–11481, 2020. * [31] Minghui Liao, Zhen Zhu, Baoguang Shi, Gui-song Xia, and Xiang Bai. Rotation-sensitive regression for oriented scene text detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5909–5918, 2018. * [32] Tsung-Yi Lin, Piotr Dollár, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. Feature pyramid networks for object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2117–2125, 2017. * [33] Dongnan Liu, Donghao Zhang, Yang Song, Fan Zhang, Lauren O’Donnell, Heng Huang, Mei Chen, and Weidong Cai. Unsupervised instance segmentation in microscopy images via panoptic domain adaptation and task re-weighting. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4243–4252, 2020. * [34] Jiawei Liu, Zheng-Jun Zha, Di Chen, Richang Hong, and Meng Wang. Adaptive transfer network for cross-domain person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7202–7211, 2019. * [35] Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott Reed, Cheng-Yang Fu, and Alexander C Berg. Ssd: Single shot multibox detector. In European conference on computer vision, pages 21–37. Springer, 2016. * [36] Yuliang Liu, Lianwen Jin, Shuaitao Zhang, Canjie Luo, and Sheng Zhang. Curved scene text detection via transverse and longitudinal sequence connection. Pattern Recognition, 90:337–345, 2019. * [37] Mingsheng Long, Yue Cao, Jianmin Wang, and Michael Jordan. Learning transferable features with deep adaptation networks. In International conference on machine learning, pages 97–105. PMLR, 2015. * [38] Mingsheng Long, Han Zhu, Jianmin Wang, and Michael I Jordan. Deep transfer learning with joint adaptation networks. In International conference on machine learning, pages 2208–2217. PMLR, 2017. * [39] Shangbang Long, Xin He, and Cong Yao. Scene text detection and recognition: The deep learning era. International Journal of Computer Vision, 129(1):161–184, 2021\. * [40] Pengyuan Lyu, Minghui Liao, Cong Yao, Wenhao Wu, and Xiang Bai. Mask textspotter: An end-to-end trainable neural network for spotting text with arbitrary shapes. In Proceedings of the European Conference on Computer Vision (ECCV), pages 67–83, 2018. * [41] Pengyuan Lyu, Cong Yao, Wenhao Wu, Shuicheng Yan, and Xiang Bai. Multi-oriented scene text detection via corner localization and region segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7553–7563, 2018. * [42] Nibal Nayef, Fei Yin, Imen Bizid, Hyunsoo Choi, Yuan Feng, Dimosthenis Karatzas, Zhenbo Luo, Umapada Pal, Christophe Rigaud, Joseph Chazalon, et al. Icdar2017 robust reading challenge on multi-lingual scene text detection and script identification-rrc-mlt. In 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), volume 1, pages 1454–1459. IEEE, 2017. * [43] Stephan R. Richter, Vibhav Vineet, Stefan Roth, and Vladlen Koltun. Playing for data: Ground truth from computer games. In Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling, editors, European Conference on Computer Vision (ECCV), volume 9906 of LNCS, pages 102–118. Springer International Publishing, 2016. * [44] Kuniaki Saito, Yoshitaka Ushiku, and Tatsuya Harada. Asymmetric tri-training for unsupervised domain adaptation. In International Conference on Machine Learning, pages 2988–2997. PMLR, 2017. * [45] Baoguang Shi, Cong Yao, Minghui Liao, Mingkun Yang, Pei Xu, Linyan Cui, Serge Belongie, Shijian Lu, and Xiang Bai. Icdar2017 competition on reading chinese text in the wild (rctw-17). In 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), volume 1, pages 1429–1434. IEEE, 2017. * [46] Amanpreet Singh, Vivek Natarajan, Meet Shah, Yu Jiang, Xinlei Chen, Dhruv Batra, Devi Parikh, and Marcus Rohrbach. Towards vqa models that can read. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8317–8326, 2019. * [47] Carole H Sudre, Wenqi Li, Tom Vercauteren, Sebastien Ourselin, and M Jorge Cardoso. Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations. In Deep learning in medical image analysis and multimodal learning for clinical decision support, pages 240–248. Springer, 2017. * [48] Yaniv Taigman, Adam Polyak, and Lior Wolf. Unsupervised cross-domain image generation. arXiv preprint arXiv:1611.02200, 2016. * [49] Shangxuan Tian, Shijian Lu, and Chongshou Li. Wetext: Scene text detection under weak supervision. In Proceedings of the IEEE International Conference on Computer Vision, pages 1492–1500, 2017. * [50] Zhi Tian, Weilin Huang, Tong He, Pan He, and Yu Qiao. Detecting text in natural image with connectionist text proposal network. In European conference on computer vision, pages 56–72. Springer, 2016. * [51] Zhuotao Tian, Michelle Shu, Pengyuan Lyu, Ruiyu Li, Chao Zhou, Xiaoyong Shen, and Jiaya Jia. Learning shape-aware embedding for scene text detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4234–4243, 2019. * [52] Yi-Hsuan Tsai, Kihyuk Sohn, Samuel Schulter, and Manmohan Chandraker. Domain adaptation for structured output via discriminative patch representations. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 1456–1465, 2019. * [53] Andreas Veit, Tomas Matera, Lukas Neumann, Jiri Matas, and Serge Belongie. Coco-text: Dataset and benchmark for text detection and recognition in natural images. arXiv preprint arXiv:1601.07140, 2016. * [54] Tuan-Hung Vu, Himalaya Jain, Maxime Bucher, Matthieu Cord, and Patrick Pérez. Advent: Adversarial entropy minimization for domain adaptation in semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2517–2526, 2019. * [55] Hao Wang, Xiang Bai, Mingkun Yang, Shenggao Zhu, Jing Wang, and Wenyu Liu. Scene text retrieval via joint text detection and similarity learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4558–4567, 2021. * [56] Wenhai Wang, Enze Xie, Xiang Li, Wenbo Hou, Tong Lu, Gang Yu, and Shuai Shao. Shape robust text detection with progressive scale expansion network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9336–9345, 2019. * [57] Weijia Wu, Ning Lu, Enze Xie, Yuxing Wang, Wenwen Yu, Cheng Yang, and Hong Zhou. Synthetic-to-real unsupervised domain adaptation for scene text detection in the wild. In Proceedings of the Asian Conference on Computer Vision, 2020\. * [58] Chang-Dong Xu, Xing-Ran Zhao, Xin Jin, and Xiu-Shen Wei. Exploring categorical regularization for domain adaptive object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11724–11733, 2020. * [59] Xingqian Xu, Zhifei Zhang, Zhaowen Wang, Brian Price, Zhonghao Wang, and Humphrey Shi. Rethinking text segmentation: A novel dataset and a text-specific refinement approach. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12045–12055, 2021. * [60] Chuhui Xue, Shijian Lu, and Steven Hoi. Detection and rectification of arbitrary shaped scene texts by using text keypoints and links. Pattern Recognition, 124:108494, 2022. * [61] Chuhui Xue, Shijian Lu, and Fangneng Zhan. Accurate scene text detection through border semantics awareness and bootstrapping. In Proceedings of the European conference on computer vision (ECCV), pages 355–372, 2018. * [62] Chuhui Xue, Shijian Lu, and Wei Zhang. Msr: multi-scale shape regression for scene text detection. arXiv preprint arXiv:1901.02596, 2019. * [63] Chuhui Xue, Wenqing Zhang, Yu Hao, Shijian Lu, Philip HS Torr, and Song Bai. Language matters: A weakly supervised vision-language pre-training approach for scene text detection and spotting. In European Conference on Computer Vision, pages 284–302. Springer, 2022. * [64] Yanchao Yang and Stefano Soatto. Fda: Fourier domain adaptation for semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4085–4095, 2020. * [65] Xiangyu Yue, Yang Zhang, Sicheng Zhao, Alberto Sangiovanni-Vincentelli, Kurt Keutzer, and Boqing Gong. Domain randomization and pyramid consistency: Simulation-to-real generalization without accessing target domain data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2100–2110, 2019. * [66] Kaiwei Zeng, Munan Ning, Yaohua Wang, and Yang Guo. Hierarchical clustering with hard-batch triplet loss for person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13657–13665, 2020. * [67] Fangneng Zhan, Chuhui Xue, and Shijian Lu. Ga-dan: Geometry-aware domain adaptation network for scene text detection and recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 9105–9115, 2019. * [68] Chengquan Zhang, Borong Liang, Zuming Huang, Mengyi En, Junyu Han, Errui Ding, and Xinghao Ding. Look more than once: An accurate detector for text of arbitrary shapes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10552–10561, 2019. * [69] Gongjie Zhang, Kaiwen Cui, Rongliang Wu, Shijian Lu, and Yonghong Tian. PNPDet: Efficient few-shot detection without forgetting via plug-and-play sub-networks. In WACV, 2021. * [70] Gongjie Zhang, Zhipeng Luo, Kaiwen Cui, Shijian Lu, and Eric P. Xing. Meta-DETR: Image-level few-shot detection with inter-class correlation exploitation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022\. * [71] Xinyu Zhou, Cong Yao, He Wen, Yuzhi Wang, Shuchang Zhou, Weiran He, and Jiajun Liang. East: an efficient and accurate scene text detector. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pages 5551–5560, 2017. * [72] Yang Zou, Zhiding Yu, BVK Kumar, and Jinsong Wang. Unsupervised domain adaptation for semantic segmentation via class-balanced self-training. In Proceedings of the European conference on computer vision (ECCV), pages 289–305, 2018. * [73] Yang Zou, Zhiding Yu, Xiaofeng Liu, BVK Kumar, and Jinsong Wang. Confidence regularized self-training. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5982–5991, 2019.
> Proceedings of the > 3rd International Workshop on > Reading Music Systems > > 23rd July, 2021 > Alicante, Spain ## Organization General Chairs Jorge Calvo-Zaragoza | | University of Alicante, Spain ---|---|--- Alexander Pacha | | TU Wien, Austria Program Committee Ichiro Fujinaga | | McGill University, Canada ---|---|--- Jose M. Iñesta | | University of Alicante, Spain Heinz Roggenkemper | | Canamus, United States of America George Fazekas | | Queen Mary University of London, United Kingdom Gabriel Vigliensoni | | Goldsmiths University of London, United Kingdom Elona Shatri | | Queen Mary University of London, United Kingdom Alicia Fornés | | Computer Vision Center, Spain Proceedings of the 3rd International Workshop on Reading Music Systems, Alicante, 2021 Edited by Jorge Calvo-Zaragoza and Alexander Pacha © The respective authors. Licensed under a Creative Commons Attribution 4.0 International License (CC- BY-4.0). Logo made by Freepik from www.flaticon.com. Adapted by Alexander Pacha. ## Preface Dear colleagues! We are more than pleased to present to you the proceedings of the 3rd International Workshop on Reading Music Systems (WoRMS). With the pandemic hopefully slowly fading, we can finally resume our tradition of having an annual workshop that brings together researchers and practitioners that work on music reading systems. For us, it was always important to create an interactive workshop that brings together people that share a common interest in music reading systems, allowing them to exchange ideas and form relationships with one another. 2020 was a difficult year for many of us. Personal situations changed rapidly, research projects were canceled, and not knowing whether an online-only edition of WoRMS would be desirable under these circumstances, we decided, heavy-hearted, to skip WoRMS 2020. Nevertheless, we are looking forward to this year’s edition, which will take place in a hybrid mode with some participants being on-site, while others joining remotely via Zoom. We believe that this way, we can find the balance between enabling interaction and keeping everyone safe. Nothing can replace in-person communication, so we hope that future editions will be fully in- person again. However, we also want to highlight the benefits of this format: offering an online option allows people to join the workshop that could not participate otherwise. This year’s edition features 11 contributions, reaching from exciting new datasets to multi-modal methods that might change the way how we think about processing written music. We noted that machine-learning remains a common theme throughout most papers—a trend that we expect to resume in the future. Finally, we want to thank University of Alicante Polytechnic School for providing the room and the TU Wien for providing Zoom conferencing facilities. Jorge Calvo-Zaragoza and Alexander Pacha ###### Contents 1. Preface See pages - of papers/WoRMS_2021_paper_1.pdf See pages - of papers/WoRMS_2021_paper_2.pdf See pages - of papers/WoRMS_2021_paper_3.pdf See pages - of papers/WoRMS_2021_paper_4.pdf See pages - of papers/WoRMS_2021_paper_5.pdf See pages - of papers/WoRMS_2021_paper_6.pdf See pages - of papers/WoRMS_2021_paper_7.pdf See pages - of papers/WoRMS_2021_paper_8.pdf See pages - of papers/WoRMS_2021_paper_11.pdf See pages - of papers/WoRMS_2021_paper_12.pdf See pages - of papers/WoRMS_2021_paper_13.pdf
# The effect of linear background rotational flows on magnetoacoustic modes of a photospheric magnetic flux tube S. J. Skirvin1,2, V. Fedun2, S. S. A. Silva2, T. Van Doorsselaere1, N. Claes1, M. Goossens1 and G. Verth3 1Centre for mathematical Plasma Astrophysics, Mathematics Department, KU Leuven, Celestijnenlaan 200B bus 2400, B-3001 Leuven, Belgium. 2Plasma Dynamics Group, Department of Automatic Control & Systems Engineering, The University of Sheffield, Sheffield, S3 7RH, UK 3Plasma Dynamics Group, Department of Mathematics and Statistics, The University of Sheffield, Sheffield, S3 7RH, UK E-mail<EMAIL_ADDRESS> (Accepted XXX. Received YYY; in original form ZZZ) ###### Abstract Magnetoacoustic waves in solar magnetic flux tubes may be affected by the presence of background rotational flows. Here, we investigate the behaviour of $m=0$ and $m=\pm 1$ modes of a magnetic flux tube in the presence of linear background rotational flows embedded in a photospheric environment. We show that the inclusion of a background rotational flow is found to have little effect on the obtained eigensolutions for the axisymmetric $m=0$ sausage mode. However, solutions for the kink mode are dependent on the location of the flow resonance modified by the slow frequency. A background rotational flow causes the modified flow resonances to possess faster phase speeds in the thin-tube (TT) limit for the case $m=1$. This results in solutions for the slow body and slow surface kink modes to follow this trajectory, changing their dispersive behaviour. For a photospheric flux tube in the TT limit, we show that it becomes difficult to distinguish between the slow surface and fast surface kink ($m=1$) modes upon comparison of their eigenfunctions. 2D velocity field plots demonstrate how these waves, in the presence of background rotational flows, may appear in observational data. For slow body kink modes, a swirling pattern can be seen in the total pressure perturbation. Furthermore, the tube boundary undergoes a helical motion from the breaking of azimuthal symmetry, where the $m=1$ and $m=-1$ modes become out of phase, suggesting the resulting kink wave is circularly polarised. These results may have implications for seismology of magnetohydrodynamic waves in solar magnetic vortices. ###### keywords: magnetohydrodynamics (MHD) – waves ††pubyear: 2022††pagerange: The effect of linear background rotational flows on magnetoacoustic modes of a photospheric magnetic flux tube–The effect of linear background rotational flows on magnetoacoustic modes of a photospheric magnetic flux tube ## 1 Introduction High resolution observations have revealed, over the last few decades, that magnetohydrodynamic (MHD) waves are ubiquitous throughout the Sun’s atmosphere (Nakariakov et al., 1999; Aschwanden et al., 1999; De Pontieu et al., 2007; Tomczyk et al., 2007; Morton et al., 2015; Grant et al., 2015; Keys et al., 2018; Stangalini et al., 2022; Bate et al., 2022). Understanding MHD wave properties from a theoretical point of view is of the utmost importance in solar physics as these waves could contribute to the heating of local plasma and may also be used as a proxy to determine sub-resolution plasma properties. Additionally, observed MHD wave behaviour can be used to provide an estimate of the properties of local plasma that cannot be measured directly, for example the magnetic field in the corona. Furthermore, MHD waves may also be responsible in the formation of some observed phenomena, e.g. jets, in the solar atmosphere (De Pontieu et al., 2004; Rouppe van der Voort et al., 2007; Scullion et al., 2011). The uniform cylindrical waveguide model (see e.g. Wilson, 1979; Spruit & Zweibel, 1979; Edwin & Roberts, 1983) has provided a foundation for analytical studies into MHD wave investigations under solar atmospheric conditions. These studies have shown that the trapped wave modes of a magnetic flux tube can be described by the number of nodes present in each geometrical direction. The axisymmetric sausage mode has zero nodes in the azimuthal direction $m=0$, whereas the (linearly polarised) non-axisymmetric kink mode has a single azimuthal node $m=1$ or $m=-1$. Whilst this uniform model is simplistic in nature, due to the uniform plasma considered inside and outside the waveguide, it has been modified over recent decades to model specific configurations which better match observational data (see e.g. Van Doorsselaere et al., 2004; Verth et al., 2007; Erdélyi & Fedun, 2010; Aldhafeeri et al., 2021; Ruderman & Petrukhin, 2022). The cylinder model can be further extended to incorporate additional physical environments that may be common configurations in magnetic flux tubes, such as including a background rotational flow. Magnetic flux tubes with background rotational flows are a common configuration observed in, e.g., flux tubes rooted in intergranular lanes, solar tornadoes and spicules (Bonet et al., 2010; Wedemeyer-Böhm et al., 2012; Tziotziou et al., 2018; Shetye et al., 2019). Such structures display dynamic characteristics and excite a wide range of MHD waves which couple different layers of the solar atmosphere. As a result, these structures act as a natural conduit for the transfer of mass, momentum and energy throughout the solar atmosphere. Furthermore, rotational flows naturally appear in numerical MHD simulations of regions in the solar atmosphere with vortex drivers (see, e.g. Fedun et al., 2011a; Fedun et al., 2011b; Shelyag et al., 2011, 2012, 2013; González-Avilés et al., 2017; González-Avilés et al., 2018; Snow et al., 2018) and also in magnetoconvection simulations (Yadav et al., 2020, 2021; Silva et al., 2021). The presence of a background rotational flow would manifest itself as a non- zero azimuthal component of the background velocity field vector. Considering $m=\pm 1$ transverse kink modes, the effect of introducing a background rotational flow into the model would break the symmetry of the system with respect to the direction of azimuthal wave propagation. Similarly, it has been shown that the inclusion of a steady vertical background plasma flow, aligned with the magnetic field, can obtain solutions which also break the symmetry of the forward and backward propagating waves (Nakariakov & Roberts, 1995; Terra- Homem et al., 2003; Soler et al., 2009; Skirvin et al., 2022). The presence of the longitudinal flow results in a Doppler-shifted wave frequency, depending upon the amplitude of the flow and the vertical wavenumber, which may shift certain wave modes into different physical regimes. Furthermore, a background longitudinal flow may also affect the wavelength and damping length of resonant absorption for kink modes, causing more efficient damping of backward propagating modes compared to forward propagating modes (Soler et al., 2011). Circularly polarised kink modes, previously observed in chromospheric magnetic elements (Stangalini et al., 2017) and sunspots (Jess et al., 2017), have also been studied recently by Magyar et al. (2022) where the authors discuss the differences that resonant absorption and phase mixing have on linearly and circularly polarised kink waves. Magyar et al. (2022) also conclude, upon analysis of the Doppler signatures for both polarisation states, that there is very little difference between the two polarisation’s in Doppler observations. Circularly polarised kink waves have been shown to develop in coronal loops with twisted magnetic fields (Terradas & Goossens, 2012; Ruderman & Terradas, 2015). Previous analytical studies have investigated the stability status of rotating flux tubes, as an azimuthal velocity shear across the waveguide boundary may be susceptible to the Kelvin-Helmholtz instability (KHI) (Soler et al., 2010; Zaqarashvili et al., 2015; Zhelyazkov & Chandra, 2019). However, these studies assume zero plasma-$\beta$ and focus on coronal conditions only, ignoring the slow magnetoacoustic modes entirely. The stability of a magnetic flux tube with a linear background magnetic twist and rotational flow component was studied by Cheremnykh et al. (2018) who found that the $m=0$ sausage mode becomes unstable for azimuthal flow speeds that create a centrifugal force which can overcome the magnetic tension, whereas, the $m=1$ kink mode can only become unstable for sufficiently large values of longitudinal (axial) flow speed. Whilst the stability studies mentioned above investigate the susceptibility of the KHI to various wave modes in magnetic flux tubes with background flows, we stress that it is not the goal of this study. The primary aim of the present work, however, is to provide a description of how $m=0$ and $m=\pm 1$ wave modes would manifest themselves in solar magnetic flux tubes, under photospheric conditions, in the presence of background rotational flows and to aid mode interpretations of observational and numerical data. This paper is presented as follows, in Section 2 the governing equations describing a magnetic flux tube in the presence of a background rotational flow is presented, along with a brief description of the numerical eigensolver implemented in this work. Section 3 presents the results of an investigation into the effect of a linear profile of background rotational flow on the properties of magnetoacoutsic waves under photospheric conditions. Section 3 looks closely at the obtained eigenvalues, one dimensional eigenfunctions and two/three dimensional visualisations of the kink mode in the presence of a background rotational flow. Finally, in Section 4, we summarise the findings presented in this work and discuss avenues of future research. ## 2 Method The ideal MHD equations adopted in this study are: $\displaystyle\frac{d\rho}{dt}+\rho\nabla\cdot\textbf{v}$ $\displaystyle=0,$ (1) $\displaystyle\rho\left(\frac{d\textbf{v}}{dt}\right)$ $\displaystyle=-\nabla p+\frac{1}{\mu_{0}}\left(\nabla\times\textbf{B}\right)\times\textbf{B},$ (2) $\displaystyle\frac{d}{dt}\left(\frac{p}{\rho^{\gamma}}\right)$ $\displaystyle=0,$ (3) $\displaystyle\frac{\partial\textbf{B}}{\partial t}$ $\displaystyle=\nabla\times\left(\textbf{v}\times\textbf{B}\right),$ (4) $\displaystyle\nabla\cdot\textbf{B}$ $\displaystyle=0,$ (5) where $\rho$, v, $p$, B, $\gamma$ and $\mu$ denote plasma density, plasma velocity, plasma pressure, magnetic field, ratio of specific heats (taken $\gamma=5/3$) and the magnetic permeability respectively. We consider a cylindrical geometry $(r,\varphi,z)$ where the initial equilibrium has a radially spatially dependent form of the velocity field vector $\textbf{v}_{0}=(0,v_{\varphi}(r),0)$. In our model the magnetic field is taken to be straight and uniform with no background axial plasma flow. In the lower solar atmosphere it should be noted that magnetic flux tubes possess significant non-vertical magnetic field as a result of flux tube expansion to maintain pressure balance, however, this effect can be considered to be negligible in the current study due to the analysis in the local plasma environment. Since the equilibrium quantities depend on $r$ only, the perturbed quantities can be Fourier-analysed with respect to the ignorable coordinates $\varphi,z$ and time $t$ and put proportional to: $\text{exp}\left[i\left(m\varphi+kz-\omega t\right)\right],$ where $m$ is the azimuthal wave number, $k$ the vertical wavenumber and $\omega$ the wave frequency. After linearising Equations (1)-(5), we arrive at a system of two differential equations containing the total pressure perturbation $\hat{P}_{T}$ and the radial displacement perturbation $r\hat{\xi}_{r}$ (see e.g. Sakurai et al., 1991; Goossens et al., 1992), which can be written as: $\displaystyle D\frac{d}{dr}\left(r\hat{\xi}_{r}\right)$ $\displaystyle=C_{1}r\hat{\xi}_{r}-C_{2}r\hat{P}_{T},$ (6) $\displaystyle D\frac{d\hat{P}_{T}}{dr}$ $\displaystyle=C_{3}\hat{\xi}_{r}-C_{1}\hat{P}_{T},$ (7) where, $\displaystyle D$ $\displaystyle=\rho\left(c^{2}+v_{A}^{2}\right)\left(\Omega^{2}-k^{2}v_{A}^{2}\right)\left(\Omega^{2}-k^{2}c_{T}^{2}\right),$ (8) $\displaystyle\Omega$ $\displaystyle=\omega-\frac{m}{r}v_{\varphi},$ (9) $\displaystyle c^{2}$ $\displaystyle=\frac{\gamma p}{\rho},\ \ \ \ v_{A}^{2}=\frac{B_{z}^{2}}{\mu\rho},\ \ \ \ c_{T}^{2}=\frac{v_{A}^{2}c^{2}}{\left(c^{2}+v_{A}^{2}\right)},$ (10) $\displaystyle C_{1}$ $\displaystyle=Q\Omega^{2}-2m\left(c^{2}+v_{A}^{2}\right)\left(\Omega^{2}-k^{2}c_{T}^{2}\right)\frac{T^{2}}{r^{2}},$ (11) $\displaystyle C_{2}$ $\displaystyle=\Omega^{4}-\left(c^{2}+v_{A}^{2}\right)\left(\frac{m^{2}}{r^{2}}+k^{2}\right)\left(\Omega^{2}-k^{2}c_{T}^{2}\right),$ (12) $\displaystyle C_{3}$ $\displaystyle=D\left\\{\rho\left(\Omega^{2}-k^{2}v_{A}^{2}\right)+r\frac{d}{dr}\left[-\rho\left(\frac{v_{\varphi}}{r}\right)^{2}\right]\right\\}+$ (13) $\displaystyle+Q^{2}-4\left(c^{2}+v_{A}^{2}\right)\left(\Omega^{2}-k^{2}c_{T}^{2}\right)\frac{T^{2}}{r^{2}},$ $\displaystyle Q$ $\displaystyle=-\left(\Omega^{2}-k^{2}v_{A}^{2}\right)\frac{\rho v_{\varphi}^{2}}{r},$ (14) $\displaystyle T$ $\displaystyle=\rho\Omega v_{\varphi}.$ (15) It should be noted here that we assume $B_{\varphi}=v_{z}=0$, which simplifies the fully inclusive ($B_{\varphi}=B_{\varphi}(r),B_{z}=B_{z}(r),v_{z}=v_{z}(r),\rho=\rho(r)$) set of equations previously noted in literature by Goossens et al. (1992). Quantities $c^{2}$, $v_{A}^{2}$, and $c_{T}^{2}$ define the squares of the local sound speed, Alfvén speed and cusp (tube) speed respectively. The quantity $\Omega$ represents the Doppler shifted frequency as a result of the background rotational plasma flow. Equations (6)-(7) can be combined to create a single differential equation in either $r\hat{\xi}_{r}$: $\frac{d}{dr}\left[f(r)\frac{d}{dr}\left(r\hat{\xi}_{r}\right)\right]-g(r)\left(r\hat{\xi}_{r}\right)=0,$ (16) where, $f(r)=\frac{D}{rC_{2}},$ (17) $g(r)=\frac{d}{dr}\left(\frac{C_{1}}{rC_{2}}\right)-\frac{1}{rD}\left(C_{3}-\frac{C_{1}^{2}}{C_{2}}\right),$ (18) or $\hat{P}_{T}$: $\frac{d}{dr}\left[\tilde{f}(r)\frac{d\hat{P}_{T}}{dr}\right]-\tilde{g}(r)\hat{P}_{T}=0,$ (19) where, $\tilde{f}(r)=\frac{rD}{C_{3}},$ (20) $\tilde{g}(r)=-\frac{d}{dr}\left(\frac{rC_{1}}{C_{3}}\right)-\frac{r}{D}\left(C_{2}-\frac{C_{1}^{2}}{C_{3}}\right).$ (21) For a non-uniform plasma, the governing Equations (6)-(7) possess regular singularities where the wave frequency matches the local characteristic frequencies at: $\omega=\frac{m}{r}v_{\varphi}(r)\pm kv_{A},$ (22) $\omega=\frac{m}{r}v_{\varphi}(r)\pm kc_{T}.$ (23) Equations (22) and (23) define the flow continua modified by the local Alfvén ($kv_{A}$) and slow ($kc_{T}$) frequencies, respectively. In ideal MHD, the wave solutions existing inside the continua, with positions given by Equations (22) and (23), are known as ‘quasi-modes’ where the wave frequency becomes a complex quantity (De Groof & Goossens, 2000; Goedbloed & Poedts, 2004; Geeraerts et al., 2022). The nature of the solutions lying inside the continua is not discussed in the present study and instead will be the focus of future work. Both Equations (16) and (19) have no known closed form analytical solutions, without making assumptions that somehow reduce the mathematical complexity. Therefore, investigating the properties of wave modes propagating within an equilibrium which is non-uniform must be done numerically. The numerical approach used in this study is based on the eigensolver applied in Skirvin et al. (2021); Skirvin et al. (2022) for non-uniform magnetic slabs and non- uniform flux tubes, respectively. The eigensolver implements the numerical shooting and bisection methods whilst also relying on fundamental properties of the sausage and kink modes. The numerical shooting method solves Equations (16) and (19) ensuring continuity of $\hat{P}_{T}$ and $\hat{\xi}_{r}$ across the boundary of the flux tube. This technique has been applied before in solar physics by, e.g. Tirry & Goossens (1996); Pinter et al. (1998); Andries et al. (2000); Taroyan & Erdélyi (2002, 2003). These studies also utilise the jump conditions introduced by Sakurai et al. (1991); Goossens et al. (1992) to deal with the regular singularities that appear in the governing equations, allowing the authors to study the resulting quasi-modes. However, following the primary objective of the present study, we only consider eigenmodes with real valued wave frequency and wavenumber. ## 3 Magnetic flux tube in the presence of a linear background rotational flow In this section, a magnetic flux tube in the presence of a linear rotational background flow is investigated. For all cases the magnetic flux tube is otherwise uniform such that the equilibrium plasma density and magnetic field is constant across the flux tube. A profile comparable to the magnetic twist profile incorporated by Erdélyi & Fedun (2007) and Erdélyi & Fedun (2010) is chosen but applied to the azimuthal velocity field component $v_{\varphi}$ instead. A rotational flow can be either clockwise or counter-clockwise in the reference frame relative to the observer. The only difference between a clockwise rotational flow and an anti-clockwise rotational flow will be the sign in front of $v_{\varphi}$ and the direction of shifted wave frequency relative to the flow. In the following sections, a magnetic flux tube is presented with an equilibrium azimuthal flow component, acting in a counter- clockwise direction, which takes the form: $v_{\varphi}(r)=A\left(\frac{r}{a}\right)^{\alpha},$ (24) where $A$ is the amplitude of the rotational flow, $\alpha$ is the parameter (exponent) dictating the radial profile of the rotational flow and $a$ indicates the location of the boundary of the flux tube, taken to be $a=1$ in our study. The case when $\alpha=1$ for example corresponds to a linear rotational flow, which will be the focus of this study, (see e.g. Figure 1). It can be shown that when $v_{\varphi}$ is linear, Equations (6)-(15) simplify, and in many cases the dependence on the radial coordinate is removed. The rotational flow is constant with height $z$ in all cases considered in this work. Obtaining an equilibrium in a magnetic cylinder with a background rotational velocity component is not as mathematically simple as the scenario of a uniform magnetic cylinder. In order to maintain total pressure balance across the waveguide the following expression must be satisfied (Goossens et al., 2011): $\frac{d}{dr}\left(p+\frac{B_{0z}^{2}}{2\mu}\right)=\frac{\rho v_{\varphi}^{2}(r)}{r}=\rho A^{2}r.$ (25) Integration of Equation (25) yields: $p+\frac{B_{0z}^{2}}{2\mu}=\rho A^{2}\frac{r^{2}}{2},$ (26) where the constant of integration is absorbed into the gas pressure term, $p$, and corresponds to the plasma pressure on the axis of the cylinder where the amplitude of the flow is zero (see e.g. Cheremnykh et al., 2018). Under the photospheric conditions considered in this work, the total pressure balance is achieved by an increase in temperature to balance the increase in azimuthal flow amplitude towards the boundary of the flux tube. For configurations where the amplitude of the rotational flow is weak (e.g. $A<0.5c_{i}$), then the change in spatial behaviour of the plasma pressure and temperature is small, but must not be dismissed. The presence of a background rotational flow not only modifies the equilibrium pressure balance relationship, but also affects the continuity conditions on the boundary of the waveguide. Considering a magnetic flux tube in the presence of a background rotational flow, the resulting boundary continuity conditions state: $\displaystyle\hat{\xi}_{re}\Bigr{\rvert}_{r=a}$ $\displaystyle=\hat{\xi}_{ri}\Bigr{\rvert}_{r=a},$ (27) $\displaystyle\hat{P}_{Te}\Bigr{\rvert}_{r=a}$ $\displaystyle=\left(\hat{P}_{Ti}+\frac{\rho_{0i}v_{\varphi}^{2}}{a}\hat{\xi}_{ri}\right)\Bigr{\rvert}_{r=a}.$ (28) The change in boundary conditions are accounted for in the numerical eigensolver, and a pair of eigenvalues will only be retrieved for values satisfying the above conditions for each respective case study. Finally, when the background rotational flow is taken to be linear with respect to the radial direction, Equations (22) and (23) become: $\displaystyle\omega$ $\displaystyle=Am\pm kv_{A},$ (29) $\displaystyle\omega$ $\displaystyle=Am\pm kc_{T}.$ (30) Equations (29) and (30) no longer define continuum regions as the resonant positions no longer cover a range of frequencies, rather they define singular resonant locations which correspond to the flow resonance position modified by the Alfvén and slow frequencies respectively. The locations of these resonant positions depend heavily on the amplitude of the rotational flow in the setup presented in this work. ### 3.1 Linear rotating magnetic flux tube under photospheric conditions Figure 1: Equilibrium background rotational flow profiles for cases with increasing amplitude for a photospheric cylinder. In all cases the profiles are linear with respect to spatial coordinate $r$ (i.e. $\alpha=1$). The amplitude of the rotational flow increases linearly up to a boundary value of $A=0.01$ (green line), $A=0.05$ (black line), $A=0.1$ (yellow line), $A=0.15$ (red line) and $A=0.25$ (blue line). The boundary of the flux tube is located at $r=1$. In this section, a magnetic flux tube under photospheric conditions ($v_{Ae}<c_{i}<c_{e}<v_{Ai}$) in the presence of a linear background rotational flow is investigated. For all photospheric cases in this work, the numerical values correspond to $c_{i}=1$, $c_{e}=1.5c_{i}$, $v_{Ai}=2c_{i}$, $v_{Ae}=0.5c_{i}$ and $\rho_{i}=1$. This choice of equilibrium parameters results in a density contrast between the internal and external plasma to be roughly $\rho_{i}/\rho_{e}=0.567$. The background velocity vector inside the waveguide can be written as $\mathbf{v}_{0i}=(0,Ar,0)$. The flow outside the cylinder is zero which results in a velocity shear across the cylinder boundary at $r=a$, however the value of $A$ is chosen to be small and both sub-Alfvénic, sub-sonic and below the threshold for the Kelvin-Helmholtz instability (Soler et al., 2010). This choice of amplitude $A$ also agrees with observed values of photospheric flows when compared with the local sound/Alfvén speed (Bonet et al., 2008). Shown in Figure 1 are the linear profiles of background rotational flow considered in this section. In all cases the flow amplitude is proportional to the radial distance from the center of the flux tube, at $r=0$, up to the boundary at $r=1$, however the amplitude is allowed to vary. #### 3.1.1 Eigenvalues (a) (b) (c) (d) Figure 2: Dispersion diagrams for a photospheric cylinder with a linear background rotational flow of varying flow amplitudes. The different cases with varying amplitude, displayed on the top of each panel, are shown corresponding to those in Figure 1. The red curves indicate solutions for the $m=0$ sausage mode and the blue curves show the $m=+1$ kink mode solutions. The green curve in all cases shows the flow resonance locations modified by the slow frequency as given by Equation (30). Dash-dotted lines indicate the equilibrium characteristic kink speed, $c_{k}$ sound speed, $c$, and tube speed, $c_{T}$, internal (subscript ‘i’) and external (subscript ‘e’) to the flux tube. The dashed box region shown in Figure 2(c) is investigated in more detail in Figure 5. Figure 2 highlights the change in eigenvalues for the different cases of flow profiles considered in Figure 1. For now, we focus on the forward propagating wave modes ($\omega/k>0$). The axisymmetric $m=0$ sausage mode appears unaffected, or at least not significantly affected, by the presence of the background flow, which is a result of the $m=0$ mode being the axisymmetric mode in the azimuthal direction. As a result of this, the rotational flow does not break the symmetry of this mode. This can also be seen analytically, by setting $m=0$ in Equations (6)-(15), the majority of terms containing the presence of the (linear) background flow are removed when $m=0$. Furthermore, when $m=0$, the resonant locations described by Equations (29) and (30) reduce to the resonant positions corresponding to a uniform magnetic flux tube (Edwin & Roberts, 1983). Whilst Equations (6)-(15) do contain some terms, for example in the variables $Q$ and $T$, which are dependent on the background rotational flow and that remain for the $m=0$ solution, we do not observe any significant modification to the obtained eigenvalues for the sausage mode in Figure 2 when compared to the analytical solutions for the uniform magnetic flux tube, although very minor effects can be seen when the amplitude of the rotational flow is sufficiently large ($A>0.25$). Investigating this result both analytically and numerically may be a focus of future work. Conversely, there is a considerable effect on the $m=1$ kink mode solutions due to the presence of a background rotational flow. In the long wavelength (thin-tube) limit, the phase speeds of the slow body and slow surface kink modes tend to an infinite phase speed, similar to the case study of a linear background magnetic twist (Erdélyi & Fedun, 2010), where they may even enter the leaky regime. As the amplitude of the azimuthal flow is increased, the corresponding phase speeds of the slow kink modes also increases for all values of wavenumber. The flow resonance locations modified by the slow frequency given by Equation (30) is shown by the green line in Figure 2. The slow body and slow surface kink modes follow this resonance curve in the long wavelength limit and even undergo an avoided crossing (Abdelatif, 1990; Mather & Erdélyi, 2016; Allcock & Erdélyi, 2017) where the slow surface mode approaches the fast surface mode. This avoided crossing implies a transfer of properties between the fast and slow surface modes. In the long wavelength limit it may not be appropriate to refer to the slow surface mode as a slow mode anymore as, in the reference frame of the observer, it possesses phase speeds similar to that of the fast surface mode. Therefore, using the observed phase speeds alone may introduce some difficulty when distinguishing between the fast and slow surface kink modes in rotating photospheric flux tubes. It may be more appropriate to differentiate between the two modes by comparing their eigenfunctions both parallel and perpendicular to the magnetic field, which is discussed in more detail in Section 3.1.2. In the case of a uniform photospheric cylinder, the slow body modes tend toward $c_{Ti}$ in the long wavelength limit (Edwin & Roberts, 1983; Priest, 2014), however with the inclusion of a background rotational flow, these modes now encounter the modified flow resonance point where they may become resonantly damped. Furthermore, the dispersive nature of the fast surface kink mode is also modified by the presence of a background rotational flow. In a uniform photospheric flux tube, the phase speed of the fast surface mode tends to the kink speed in the long wavelength limit (Edwin & Roberts, 1983). However, similar to the slow modes, the fast surface mode encounters the modified flow resonance in the long-wavelength limit, where the phase speed of this mode also appears to increase sharply depending on the amplitude of the flow. The fast mode may also enter the leaky regime in the long-wavelength limit under the assumed equilibrium configuration, which may be important in the context of solar observations. Of course, the kink mode possesses an azimuthal wavenumber which can be either positive or negative. Granular buffeting in the lower solar photosphere due to convective motions beneath the solar surface, may excite both $m=1$ and $m=-1$ kink modes. In the uniform cylinder model of a solar waveguide, the ‘traditional’ kink mode is considered to display signatures that resemble a periodic transverse displacement of the waveguide. This is because, due to the symmetry of the model, the opposite rates of rotation set up a standing wave in the azimuthal direction. However, introducing a background rotational flow breaks this azimuthal symmetry and, as a result, the $m=1$ and $m=-1$ modes are expected to behave differently and the resulting observed mode will no longer be a standing wave in the azimuthal direction. In Figure 3, we show the obtained wave solutions for the $m=1$ and $m=-1$ modes in a photospheric flux tube with a linear background rotational flow of amplitude $A=0.1$. For the case when the kink wave is rotating with the flow, the phase speed of the fast surface kink mode increases in the long-wavelength limit and enters the leaky regime for small $ka$. This suggests that the fast kink surface mode, when propagating in a thin waveguide with the background rotational flow, should not be seen in observations of rotating photospheric structures, for example in magnetic bright points or flux bundles rooted within intergranular lanes. However, the obtained solutions for the fast surface mode are different for the two cases when $m=1$ and $m=-1$. This is due to the $m=1$ mode rotating in the same direction as the background flow, in our specific configuration. Therefore, the $m=1$ mode constructively interferes with the background rotational flow, increasing its phase velocity, whereas the $m=-1$ mode, rotating against the flow, destructively superimposes with the flow, hence lowering its phase velocity. The most notable difference between the case when $m=1$ and $m=-1$ can be seen by the dispersive behaviour of the slow surface and body modes. In both cases, the slow modes follow the trajectory of the flow resonance locations modified by the slow frequency as given by Equation (30), however for the case when $m=-1$, this curve decreases with decreasing $ka$. As a result, the difference in phase speed for a given wavenumber between $m=1$ and $m=-1$ is greater for the slow modes, with a more dramatic difference seen as $ka$ approaches zero. This example of the phase speed difference between the forward propagating slow body and surface modes, for $m=1$ and $m=-1$, demonstrates how the background flow breaks the symmetry of the kink mode. Figure 3: The dispersion diagrams showing the obtained solutions for the $m=1$ and $m=-1$ modes denoted by the blue curves. The green curve shows the behaviour of Equation (30) for both cases of $m=\pm 1$, respectively. The respective zoom in plots highlight the difference in the behaviour of the various category of modes for different values of $m$. #### 3.1.2 Eigenfunctions Figure 4: The resulting eigenfunctions for the slow surface kink mode ($m=1$) for all linear cases of rotational flow with the colour scheme consistent with Figure 1. A wavenumber value of $k=0.6$ was chosen for all plots. Shown in Figure 4 are the spatial eigenfunctions for the slow surface kink mode at a fixed wavenumber $ka=0.6$ for varying rotational flow amplitudes with $m=1$. The colour scheme shown in the eigenfunctions is consistent with that for the rotational flow profiles shown in Figure 1. It can be seen that increasing the amplitude of the equilibrium linear rotational flow, changes the spatial behaviour of the observable eigenfunctions. For the case of $A=0.01$ which corresponds to a very small rotational flow parameter, the eigenfunctions still obey a ‘surface-like’ structure, that is, the amplitude of the radial displacement and velocity perturbations possesses a maximum at the boundary of the flux tube and decays away from the boundary. However, increasing the amplitude of the background rotational flow causes the radial displacement perturbation to increase towards the centre of the flux tube, such that the maximum displacement perturbation is no longer at the point where $r=a$. This results in an eigenfunction that shares striking similarities to that of the fundamental kink mode, and may therefore be misinterpreted in observational data. To further emphasise this point, it is possible to plot the eigenfunctions of $\hat{P}_{T}$ and $\hat{\xi}_{r}$ for eigenvalues of a similar phase speed on either side of the modified flow resonance point. One of these solutions corresponds to the slow surface kink mode and the other is the fast surface kink mode. (a) (b) Figure 5: Panels showing the obtained wave solutions in a zoomed region denoted by the dashed box in Figure 2(c). The eigenfunctions $\hat{P}_{T}$, $\hat{\xi}_{r}$, $\hat{\xi}_{z}$ and $\hat{\xi}_{\varphi}$ for the (a) fast and (b) slow magnetoacoustic surface kink mode solutions in a photospheric flux tube with a background rotational flow given by $v_{\varphi}=0.1r$. Both panels (a) and (b) are for the same $v_{ph}=\omega/k=1.3$ indicated by the blue dot on the dispersion diagram. All plots are normalised such that the external value of each eigenfunction is equal to unity at the boundary. Shown in Figure 5 are these eigenfunctions for the slow and fast magnetoacoustic kink modes at a similar phase speed. It can be seen that the normalised eigenfunctions for $\hat{P}_{T}$, $\hat{\xi}_{r}$ and $\hat{\xi}_{\varphi}$ are difficult to distinguish between the slow and the fast surface modes. Furthermore, the modes no longer display the typical characteristics of the surface mode anymore. In particular, the main characteristic of a surface mode from the uniform cylinder model is that it possesses a maximum amplitude of radial displacement perturbation at the boundary of the waveguide, which is no longer the case when a rotational background plasma flow is present. Investigating the nature of perturbations parallel and perpendicular to the magnetic field may aid in distinguishing between the slow and fast surface kink modes. Slow modes tend to propagate (mainly) along the magnetic field lines, which in our study are straight and vertical, so we should expect $\hat{\xi}_{z}$ to dominate for the slow mode when compared to the fast mode, which may propagate at an angle across the magnetic field. In Figure 5, we also show the normalised $z$ component of the displacement perturbation for both the fast and slow surface kink modes. As expected, $\hat{\xi}_{z}$ is dominant over $\hat{\xi}_{r}$ for the slow surface mode (see e.g. Moreels & Van Doorsselaere, 2013), however, the same is also true for the fast surface mode. Although, the absolute magnitude of the normalised $z$ component is greater for the slow surface mode compared to the fast surface modes, by roughly a factor of $2$, suggesting that the component of displacement is still more dominant for the slow surface kink mode in a rotating photospheric flux tube. It should be noted that, displayed in some plots of the eigenfunctions, the amplitude of the resulting eigenfunctions increases as we approach the center of the cylinder. This is a feature which relates to the fact that there is a singularity in the set of Equations (16) and (19) at $r=0$. However, this is purely an artifact of the plotting technique and does not affect the eigensolver obtaining the solutions. #### 3.1.3 2D and 3D velocity fields Following the analysis from Section 2, the wave solutions are put proportional to $\text{exp}\left[i\left(m\varphi+kz-\omega t\right)\right]$, therefore, it is possible to convert the one dimensional radial eigenfunctions, for example those shown in Figures 4 and 5, into two and three dimensional plots, to better represent an observer’s perspective. Figure 6: Snapshots of the $2$D velocity field for the scenario of a slow body mode in a photospheric flux in the presence of a background flow given by $v_{\varphi}=0.1r$. The plots are arranged in a $3\times 2$ configuration where the left hand column shows the velocity perturbation only, whereas the right hand column shows the perturbed velocity field plus the background velocity field. The top row corresponds to the solution for the $m=1$ mode, the middle row shows the solution for the $m=-1$ mode and the bottom row shows the resulting velocity field and total pressure perturbation for the sum of the $m=1$ and $m=-1$ modes. In all panels the velocity vectors are normalised by their maximum values. The colour contour denotes the normalised total pressure perturbation, $\hat{P}_{T}$, which is the same for both the left and right columns. The boundary of the flux tube is highlighted by the solid blue line. An animated movie of this Figure is available online. In Figure 6, we show a snapshot of the two dimensional (converted into Cartesian $x$ and $y$) velocity field for the perturbations alongside the addition of the background flow onto the perturbations. This snapshot is chosen specifically at the time when the $m=1$ and the $m=-1$ perturbations effectively cancel each other such that the sum of the perturbed velocity field is zero. Of course, due to the breaking of symmetry from the presence of the background rotational flow, the boundary of the flux tube is no longer at equilibrium, as the $m=1$ and $m=-1$ modes are slightly out of phase with one another. In this snapshot, the perturbed total pressure, displayed by the colour contour, can be seen to display a swirling characteristic, albeit with a small magnitude, and is clearly visible in the bottom panels of Figure 6. This swirling behaviour is present purely in the perturbations, signifying the effect that a rotating waveguide has on the perturbed eigenfunctions. This swirling behaviour of the total pressure perturbation may represent an observational signature of the slow body kink mode propagating in a rotating photospheric flux tube, although detecting this may be a challenge for observers due to the small absolute value of magnitude. In the plots in the right hand column of Figure 6, we add the background velocity field to the perturbed velocity field. The background velocity field is given by $v_{\varphi}=0.1r$ and acts counter-clockwise in the $xy$ plane. The addition of the background velocity field completely changes the observed distribution of the velocity field for the slow body kink mode. In the case of a uniform cylinder, the velocity field for the slow body kink mode can be seen emanating from two ‘islands’, which correspond to two anti-nodes in the total pressure eigenfunction. This can be seen in the left hand column showing the perturbations only for the case of $m=1$ and $m=-1$, however, when the background flow is added, the velocity field no longer displays these typical characteristics. It should be noted that both the perturbed velocity field and the background velocity field are normalised separately to their respective maximum values to aid these visualisations, as it is to be expected that the perturbed components will be significantly smaller in absolute value than the background quantities. Figure 7: Same as Figure 6 but a snapshot at a later time. An animated movie of this Figure is available online. In Figure 7, we show the same quantities as those displayed in Figure 6 at a later time. As we have shown in Figure 3, there is a significant difference in the phase speed for the $m=1$ and $m=-1$ forward propagating slow body waves. Therefore, we would expect these perturbations to become increasingly out of phase with one another as time progresses, a behaviour which is displayed in Figure 7. The two modes becoming out of phase with one another can be seen by comparing the top row and the middle rows of Figure 7 (which display the $m=1$ and $m=-1$ modes respectively). For the case of a uniform cylinder with no background rotational flow, we would expect that the $m=1$ and $m=-1$ modes would rotate in opposite directions but perfectly in phase with one another, however, this is no longer true when a rotational background flow is added. As a result of one mode rotating with the background flow (in our case this is the $m=1$ mode), and the other mode propagating against the background flow ($m=-1$ mode), the sum of the two modes results in a total pressure perturbation which is also propagating (rotating) in the azimuthal direction. Furthermore, the boundary of the flux tube can be seen to rotate (see associated online movie). The observed rotating motion when combining the individual $m=1$ and $m=-1$ wave modes presents another observational signature that can be sought to identify the kink mode in the presence of a background rotational flow. (a) (b) Figure 8: This figure shows the three dimensional structure of the slow body kink mode for (a) a uniform magnetic flux tube without the presence of a background flow and (b) a magnetic flux tube in the presence of a background rotational flow, by visualising the normalised total pressure perturbation $\hat{P}_{T}$. On the left panels, the magnetic flux tube is immersed in the volume rendering of $\hat{P}_{T}$. The three cross-sectional cuts (shown as coloured rings at three different heights, $z=0.0,2.5$ and $5.0$) correspond to the corresponding right subplots. These three subplots show the LIC visualisation at the same heights. The white rings represent the boundary of the flux tube. An animated version of this Figure is available online. Figure 8 shows a three dimensional visualisation of how the sum of the $m=\pm 1$ slow body kink mode manifests itself in a uniform magnetic flux tube and also for the case of a magnetic flux tube in the presence of a background rotational flow. Figure 8 also displays a Line Integral Contour (LIC) plot at three different heights in the flux tube which are denoted by white rings in the 3D plot. A LIC visualisation aids imaging the velocity field of the plasma (see, e.g. Cabral & Leedom, 1993). In other words, LIC visualisations display the path lines that an object would follow if it were placed into the fluid with a stationary velocity field. It can be seen that the LIC for the slow body kink mode in a uniform magnetic flux tube (Figure 8(a)) remains unchanged with height, as the wave propagates and the structure oscillates in one plane, in other words it is linearly polarised. In addition, the LIC also displays a ‘double X-point’ close to $x=0$ and $y=\pm 1$, near the boundary in the plane in which the structure does not oscillate. This behaviour is due to the rotational motions that are inherently associated with the non-axisymmetric kink mode (Goossens et al., 2014). However, once a rotational flow is present, the observed evolution of the flux tube motion is affected. Firstly, we can observe that the three dimensional behaviour of the total pressure (density) perturbation begins to represent a helical structure, in contrast to the sinusoidal oscillation in the uniform case, as the perturbation now also displays a rotational behaviour. This suggests that the resulting kink wave is circularly polarised, as a result of the $\pm 1$ modes becoming out of phase with one another. Therefore, we show that, in the presence of a background rotational flow, the kink mode may manifest itself as a circularly polarised mode, similar to a twisted magnetic flux tube (Terradas & Goossens, 2012; Ruderman & Terradas, 2015). Furthermore, the LIC visualisations show that the velocity field evolution is modified in the presence of a background rotational flow. Firstly, the ‘double X-point’ feature seen in the LIC for the uniform flux tube no longer displays the same characteristics in the scenario including a background rotational flow. Instead, the ‘X-points’ can be seen at different locations at certain times as the wave propagates upwards. Secondly, the evolution of the velocity field is no longer constant and does not appear to oscillate in a plane on a single axis anymore. Similar to the behaviour seen in Figures 6 and 7, the LIC visualisation shows that the velocity field also rotates as it evolves, and the traditional signatures of the kink mode in a uniform flux tube are no longer present. These features could possibly be retrieved from numerical simulations of kink waves in rotating structures. ## 4 Conclusions In the present study, we have extended previous studies investigating the properties of MHD waves in photospheric waveguides, by introducing a linear azimuthal component to the background velocity field. In order to conduct this investigation, we utilised a previously developed numerical eigensolver (Skirvin et al., 2021; Skirvin et al., 2022) to model a rotating flux tube in a photospheric environment with a background $v_{\varphi}$ component. For the inclusion of a linear rotational flow, very similar results to those of the linear magnetic twist were recovered (Erdélyi & Fedun, 2010). We find that the obtained kink mode solutions possess phase speeds along the magnetic field which tend to infinite values in the long wavelength limit where they may become leaky. We have shown this to be a result of the modes being guided by the trajectory of the flow resonance location modified by the slow frequency. As a result of considering a linear background flow, the equations describing the continua regions reduce to single point locations. Therefore, there becomes a point where, under photospheric conditions, the slow surface kink mode and the slow body kink mode approach the same solution at the resonant point and undergo an avoided crossing where their properties become mixed. Comparison of the eigenfunctions for the slow surface kink mode and fast surface kink mode, in the presence of a linear background rotational flow at similar phase speeds, indicated that identification of the two modes becomes extremely difficult in the long wavelength limit using the total pressure and radial displacement perturbations. However, comparison of $\hat{\xi}_{z}$ shows that this component is still dominant for slow modes, comparable to uniform theory. Furthermore, we find that the axisymmetric $m=0$ sausage mode remains unaffected by any background azimuthal component. Analytically this can be understood by examining the governing set of Equations (6)-(15) when setting $m=0$ and noticing many terms either simplifying or disappearing altogether. In addition, it suggests that sausage mode observations in the lower solar atmosphere in e.g. pores and sunspots, may not be a suitable wave mode to conduct atmospheric-seismology, if the structure is in the presence of any magnetic twist or background rotational flows. We have also presented 2D plots showing the velocity field with and without the inclusion of the background flow (Figures 6-7) and 3D plots showing the perturbation of the normalised total pressure $\hat{P}_{T}$ (Figure 8). These plots clearly show the breaking of symmetry between the $m=1$ and $m=-1$ kink modes when a flux tube is in the presence of a rotational background flow. Furthermore, inspection of the total pressure perturbations shows that it displays signatures of a swirling motion when the flux tube experiences its maximum displacement. It is hoped that these visualisations will aid future interpretations of observational data from high resolution instruments on state of the art telescopes such as DKIST. Finally, we have also produced three dimensional visualisations of the kink mode propagating in a photospheric flux tube with a linear background rotational flow incorporated. We have also shown, in Figure 8, the LIC visualisations at different heights displaying the evolution of the velocity field as the wave propagates along the magnetic field, which may be retrieved from numerical simulations. In this study, we have suggested potential observational signatures of magnetoacoustic kink modes in photospheric flux tubes with background rotational flows. These signatures share striking characteristics with previous observations of circularly polarised kink modes in the lower solar atmosphere (Jess et al., 2017; Stangalini et al., 2017). However, naturally, the magnitude of the background flows are much greater than those of the perturbations arising from the waves, which then raises the important question of how to actually observe this phenomena in the solar atmosphere. Both the rotational motion of the waveguide combined with the swirling pattern seen in the total pressure perturbation are observational signatures that may be detected with current and future telescopes, however distinguishing between the contribution of the background and perturbed components may be challenging. Furthermore, we have focused only on a magnetic flux tube in a photospheric environment. Under the photospheric conditions adopted in this work, the Alfvén continuum given by Equation (22) exists in the leaky regime at phase speeds above the cut-off speed at $v_{ph}=c_{e}$. Therefore this resonant region is not discussed in detail in this work, however, it will become important for studies under coronal conditions, where contributions from both Equations (22) and (23) become important. In the solar atmosphere, vortices can be buffeted from multiple directions, for example convective motion due to convection cells. In this study, we have considered just one fixed driver that oscillates from side to side, however in reality the picture will be more complicated due to drivers in additional directions. These perturbations along different axis will manifest themselves as apparent rotation, regardless of a background rotational flow, this is work to be investigated in the future. Additional future work includes a closer inspection on the effect that rotational flows have on the sausage mode. We have shown that the sausage mode may be affected by the presence of background rotational flow in magnetic flux tubes, however, more insight is required to quantify exactly how the sausage mode is affected and how it manifests itself in rotating flux tubes. Finally, future work to be investigated includes modelling a background rotational flow with a realistic radial profile. For example Silva et al. (2020) have shown, using information from MURAM simulations, that the radial profile of the azimuthal velocity component in solar vortex tubes can be accurately modelled with a cubic polynomial. In this case, the background rotational flow becomes non-linear, such that the value of parameter $\alpha$ in Equation (24) no longer equals $1$. We expect that in this example, the MHD spectrum is densely occupied by the (modified) flow continua, consequently, trapped modes may find it difficult to exist. Obtaining the eigenvalues that lie inside the continua can be achieved by modifying the numerical eigensolver employed in this work or by using an alternative 1D eigensolver such as Legolas (Claes et al., 2020). Modelling a non-linear background rotational flow such as this is an objective for future studies. ## Acknowledgements This work has been supported by STFC (UK). SJS is grateful to STFC for the PhD studentship project reference (2135820). VF, GV, and SSAS are grateful to Science and Technology Facilities Council (STFC) grant ST/V000977/1, and The Royal Society, International Exchanges Scheme, collaboration with Brazil (IES191114) and Chile (IE170301). TVD was supported by the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No 724326) and the C1 grant TRACEspace of Internal Funds KU Leuven. The results received support from the FWO senior research project with number G088021N. SJS and VF would like to thank the International Space Science Institute (ISSI) in Bern, Switzerland, for the hospitality provided to the members of the team on ‘The Nature and Physics of Vortex Flows in Solar Plasmas’. SJS and GV wish to acknowledge scientific discussions with the Waves in the Lower Solar Atmosphere (WaLSA; https://www.WaLSA.team) team, which is supported by the Research Council of Norway (project no. 262622). This research has also received financial support from the ISEE, International Joint Research Program (Nagoya University, Japan) and the European Union’s Horizon 2020 research and innovation program under grant agreement No. 824135 (SOLARNET). Software: Numpy (Harris et al., 2020), matplotlib (Hunter, 2007), SciPy (Virtanen et al., 2020). The numerical code, Sheffield Dispersion Diagram Code (SDDC) introduced and applied in this work is available on the Plasma Dynamics Group (PDG) website 111https://sites.google.com/sheffield.ac.uk/pdg/solar- codes?authuser=0 along with the user manual which explains some cases shown in this work. This code and the accompanying tools have been developed using Python an open-source and community-developed programming language. ## Data Availability Accessibility of data used in this research is available upon request from the authors. ## References * Abdelatif (1990) Abdelatif T. E., 1990, Sol. Phys., 129, 201 * Aldhafeeri et al. (2021) Aldhafeeri A. A., Verth G., Brevis W., Jess D. B., McMurdo M., Fedun V., 2021, ApJ, 912, 50 * Allcock & Erdélyi (2017) Allcock M., Erdélyi R., 2017, Sol. Phys., 292, 35 * Andries et al. (2000) Andries J., Tirry W. J., Goossens M., 2000, ApJ, 531, 561 * Aschwanden et al. (1999) Aschwanden M. J., Fletcher L., Schrijver C. J., Alexander D., 1999, ApJ, 520, 880 * Bate et al. (2022) Bate W., et al., 2022, ApJ, 930, 129 * Bonet et al. (2008) Bonet J. A., Márquez I., Sánchez Almeida J., Cabello I., Domingo V., 2008, ApJ, 687, L131 * Bonet et al. (2010) Bonet J. A., et al., 2010, ApJ, 723, L139 * Cabral & Leedom (1993) Cabral B., Leedom L. C., 1993, in Proceedings of the 20th Annual Conference on Computer Graphics and Interactive Techniques. SIGGRAPH ’93. Association for Computing Machinery, New York, NY, USA, p. 263–270, doi:10.1145/166117.166151, https://doi.org/10.1145/166117.166151 * Cheremnykh et al. (2018) Cheremnykh O., Fedun V., Ladikov-Roev Y., Verth G., 2018, ApJ, 866, 86 * Claes et al. (2020) Claes N., De Jonghe J., Keppens R., 2020, ApJS, 251, 25 * De Groof & Goossens (2000) De Groof A., Goossens M., 2000, A&A, 356, 724 * De Pontieu et al. (2004) De Pontieu B., Erdélyi R., James S. P., 2004, Nature, 430, 536 * De Pontieu et al. (2007) De Pontieu B., et al., 2007, Science, 318, 1574 * Edwin & Roberts (1983) Edwin P. M., Roberts B., 1983, Sol. Phys., 88, 179 * Erdélyi & Fedun (2007) Erdélyi R., Fedun V., 2007, Sol. Phys., 246, 101 * Erdélyi & Fedun (2010) Erdélyi R., Fedun V., 2010, Sol. Phys., 263, 63 * Fedun et al. (2011a) Fedun V., Shelyag S., Verth G., Mathioudakis M., Erdélyi R., 2011a, Annales Geophysicae, 29, 1029 * Fedun et al. (2011b) Fedun V., Shelyag S., Erdélyi R., 2011b, ApJ, 727, 17 * Geeraerts et al. (2022) Geeraerts M., Vanmechelen P., Van Doorsselaere T., Soler R., 2022, arXiv e-prints, p. arXiv:2203.02359 * Goedbloed & Poedts (2004) Goedbloed J. P. H., Poedts S., 2004, Principles of Magnetohydrodynamics * González-Avilés et al. (2017) González-Avilés J. J., Guzmán F. S., Fedun V., 2017, ApJ, 836, 24 * González-Avilés et al. (2018) González-Avilés J. J., Guzmán F. S., Fedun V., Verth G., Shelyag S., Regnier S., 2018, ApJ, 856, 176 * Goossens et al. (1992) Goossens M., Hollweg J. V., Sakurai T., 1992, Sol. Phys., 138, 233 * Goossens et al. (2011) Goossens M., Erdélyi R., Ruderman M. S., 2011, Space Sci. Rev., 158, 289 * Goossens et al. (2014) Goossens M., Soler R., Terradas J., Van Doorsselaere T., Verth G., 2014, ApJ, 788, 9 * Grant et al. (2015) Grant S. D. T., et al., 2015, ApJ, 806, 132 * Harris et al. (2020) Harris C. R., et al., 2020, Nature, 585, 357 * Hunter (2007) Hunter J. D., 2007, Computing in Science and Engineering, 9, 90 * Jess et al. (2017) Jess D. B., et al., 2017, ApJ, 842, 59 * Keys et al. (2018) Keys P. H., et al., 2018, ApJ, 857, 28 * Magyar et al. (2022) Magyar N., Duckenfield T., Van Doorsselaere T., Nakariakov V. M., 2022, A&A, 659, A73 * Mather & Erdélyi (2016) Mather J. F., Erdélyi R., 2016, ApJ, 822, 116 * Moreels & Van Doorsselaere (2013) Moreels M. G., Van Doorsselaere T., 2013, A&A, 551, A137 * Morton et al. (2015) Morton R. J., Tomczyk S., Pinto R., 2015, Nature Communications, 6, 7813 * Nakariakov & Roberts (1995) Nakariakov V. M., Roberts B., 1995, Sol. Phys., 159, 213 * Nakariakov et al. (1999) Nakariakov V. M., Ofman L., Deluca E. E., Roberts B., Davila J. M., 1999, Science, 285, 862 * Pinter et al. (1998) Pinter B., Cadez V. M., Goossens M., 1998, A&A, 332, 775 * Priest (2014) Priest E. R., 2014, Magnetohydrodynamics of the Sun * Rouppe van der Voort et al. (2007) Rouppe van der Voort L. H. M., De Pontieu B., Hansteen V. H., Carlsson M., van Noort M., 2007, ApJ, 660, L169 * Ruderman & Petrukhin (2022) Ruderman M. S., Petrukhin N. S., 2022, Sol. Phys., 297, 72 * Ruderman & Terradas (2015) Ruderman M. S., Terradas J., 2015, A&A, 580, A57 * Sakurai et al. (1991) Sakurai T., Goossens M., Hollweg J. V., 1991, Sol. Phys., 133, 227 * Scullion et al. (2011) Scullion E., Erdélyi R., Fedun V., Doyle J. G., 2011, ApJ, 743, 14 * Shelyag et al. (2011) Shelyag S., Fedun V., Keenan F. P., Erdélyi R., Mathioudakis M., 2011, Annales Geophysicae, 29, 883 * Shelyag et al. (2012) Shelyag S., Mathioudakis M., Keenan F. P., 2012, ApJ, 753, L22 * Shelyag et al. (2013) Shelyag S., Cally P. S., Reid A., Mathioudakis M., 2013, ApJ, 776, L4 * Shetye et al. (2019) Shetye J., Verwichte E., Stangalini M., Judge P. G., Doyle J. G., Arber T., Scullion E., Wedemeyer S., 2019, ApJ, 881, 83 * Silva et al. (2020) Silva S. S. A., Fedun V., Verth G., Rempel E. L., Shelyag S., 2020, ApJ, 898, 137 * Silva et al. (2021) Silva S. S. A., Verth G., Rempel E. L., Shelyag S., Schiavo L. A. C. A., Fedun V., 2021, ApJ, 915, 24 * Skirvin et al. (2021) Skirvin S. J., Fedun V., Verth G., 2021, MNRAS, 504, 4077 * Skirvin et al. (2022) Skirvin S. J., Fedun V., Silva S. S. A., Verth G., 2022, MNRAS, 510, 2689 * Snow et al. (2018) Snow B., Fedun V., Gent F. A., Verth G., Erdélyi R., 2018, ApJ, 857, 125 * Soler et al. (2009) Soler R., Oliver R., Ballester J. L., 2009, ApJ, 693, 1601 * Soler et al. (2010) Soler R., Terradas J., Oliver R., Ballester J. L., Goossens M., 2010, ApJ, 712, 875 * Soler et al. (2011) Soler R., Terradas J., Goossens M., 2011, ApJ, 734, 80 * Spruit & Zweibel (1979) Spruit H. C., Zweibel E. G., 1979, Sol. Phys., 62, 15 * Stangalini et al. (2017) Stangalini M., et al., 2017, ApJ, 840, 19 * Stangalini et al. (2022) Stangalini M., et al., 2022, Nature Communications, 13, 479 * Taroyan & Erdélyi (2002) Taroyan Y., Erdélyi R., 2002, Physics of Plasmas, 9, 3121 * Taroyan & Erdélyi (2003) Taroyan Y., Erdélyi R., 2003, Physics of Plasmas, 10, 266 * Terra-Homem et al. (2003) Terra-Homem M., Erdélyi R., Ballai I., 2003, Sol. Phys., 217, 199 * Terradas & Goossens (2012) Terradas J., Goossens M., 2012, A&A, 548, A112 * Tirry & Goossens (1996) Tirry W. J., Goossens M., 1996, ApJ, 471, 501 * Tomczyk et al. (2007) Tomczyk S., McIntosh S. W., Keil S. L., Judge P. G., Schad T., Seeley D. H., Edmondson J., 2007, Science, 317, 1192 * Tziotziou et al. (2018) Tziotziou K., Tsiropoula G., Kontogiannis I., Scullion E., Doyle J. G., 2018, A&A, 618, A51 * Van Doorsselaere et al. (2004) Van Doorsselaere T., Debosscher A., Andries J., Poedts S., 2004, A&A, 424, 1065 * Verth et al. (2007) Verth G., Van Doorsselaere T., Erdélyi R., Goossens M., 2007, A&A, 475, 341 * Virtanen et al. (2020) Virtanen P., et al., 2020, Nature Methods, 17, 261 * Wedemeyer-Böhm et al. (2012) Wedemeyer-Böhm S., Scullion E., Steiner O., Rouppe van der Voort L., de La Cruz Rodriguez J., Fedun V., Erdélyi R., 2012, Nature, 486, 505 * Wilson (1979) Wilson P. R., 1979, A&A, 71, 9 * Yadav et al. (2020) Yadav N., Cameron R. H., Solanki S. K., 2020, ApJ, 894, L17 * Yadav et al. (2021) Yadav N., Cameron R. H., Solanki S. K., 2021, A&A, 645, A3 * Zaqarashvili et al. (2015) Zaqarashvili T. V., Zhelyazkov I., Ofman L., 2015, ApJ, 813, 123 * Zhelyazkov & Chandra (2019) Zhelyazkov I., Chandra R., 2019, Sol. Phys., 294, 20
> Proceedings of the > 2nd International Workshop on > Reading Music Systems > > 2nd November, 2019 > Delft, The Netherlands ## Organization General Chairs Jorge Calvo-Zaragoza | | University of Alicante, Spain ---|---|--- Alexander Pacha | | TU Wien, Austria Heinz Roggenkemper | | Canemus, United States Program Committee Andreas Arzt | | Johannes Kepler Universität, Austria ---|---|--- Jürgen Diet | | Bavarian State Library, Germany Ichiro Fujinaga | | McGill University, Canada Jose M. Iñesta | | University of Alicante, Spain Gabriel Vigliensoni | | McGill University, Canada Proceedings of the 2nd International Workshop on Reading Music Systems, Delft, 2019 Edited by Jorge Calvo-Zaragoza and Alexander Pacha © The respective authors. Licensed under a Creative Commons Attribution 4.0 International License (CC- BY-4.0). Logo made by Freepik from www.flaticon.com. Adapted by Alexander Pacha. ## Preface Dear colleagues, it is our greatest pleasure to introduce the proceedings of the 2nd International Workshop on Music Reading Systems (WoRMS). After the first successful installation last year, we are more than happy to have another edition of a workshop that brings together researchers and practitioners that work on reading music systems. This workshop has gained global interest with submissions from all around the world. Similar to last year, we have sessions on both technical approaches as well as applications that can be used by end-users or other researchers to build upon. What is even more pleasant is the increase in open-source software that seems to become the scientific norm. Following the suggestions from last year, the workshop will provide the opportunity for discussions right after each presentation, as well as in a panel after each session. This way we hope to drive the interactive character of WoRMS, while giving sufficient time to present ongoing research. Finally, we want to thank Jaehun Kim and Ginny Ruiter who assisted us in organizing the venue. Jorge Calvo-Zaragoza and Alexander Pacha ###### Contents 1. Preface See pages - of papers/WoRMS_2019_paper_7.pdf See pages - of papers/WoRMS_2019_paper_1.pdf See pages - of papers/WoRMS_2019_paper_8.pdf See pages - of papers/WoRMS_2019_paper_9.pdf See pages - of papers/WoRMS_2019_paper_6.pdf See pages - of papers/WoRMS_2019_paper_4.pdf See pages - of papers/WoRMS_2019_paper_3.pdf
# SPOT: Secure and Privacy-preserving prOximiTy protocol for e-healthcare systems Souha Masmoudi Samovar, Telecom SudParis, Institut Polytechnique de Paris Palaiseau, France Member of the Chair Values and Policies of Personal Information ORCID: 0000-0002-7194-8240 Nesrine Kaaniche Samovar, Telecom SudParis, Institut Polytechnique de Paris Palaiseau, France Member of the Chair Values and Policies of Personal Information ORCID: 0000-0002-1045-6445 Maryline Laurent Samovar, Telecom SudParis, Institut Polytechnique de Paris Palaiseau, France Member of the Chair Values and Policies of Personal Information ORCID: 0000-0002-7256-3721 ###### Abstract This paper introduces SPOT, a Secure and Privacy-preserving prOximity based protocol for e-healthcare systems. It relies on a distributed proxy-based approach to preserve users’ privacy and a semi-trusted computing server to ensure data consistency and integrity. The proposed protocol ensures a balance between security, privacy and scalability. As far as we know, in terms of security, SPOT is the first one to prevent malicious users from colluding and generating false positives. In terms of privacy, SPOT supports both anonymity of users being in proximity of infected people and unlinkability of contact information issued by the same user. A concrete construction based on structure-preserving signatures and NIWI proofs is proposed and a detailed security and privacy analysis proves that SPOT is secure under standard assumptions. In terms of scalability, SPOT’s procedures and algorithms are implemented to show its efficiency and practical usability with acceptable computation and communication overhead. ###### Index Terms: Anonymity, e-healthcare, NIWI proofs, Privacy, Structure-preserving signature, Unlinkability. ## I Introduction The recent health crisis has led governments to apply different tracing solutions to control the contamination chain among the population. These solutions are aimed at sharing valuable data while preserving users’ privacy. However, there are still privacy threats and robustness challenges as long as users are required to disclose and share correct sensitive and personal information with different third parties with various levels of trust. Most of the solutions rely on the Bluetooth technology, namely Bluetooth Low Energy (BLE), to exchange contact information, thanks to its efficiency in active communications [17]. Among these governmental solutions, the TraceTogether application [11] has been launched by Singapore. TraceTogether enables to collect, via the Bluetooth technology, temporary IDs (generated by a central trusted server) of users in close proximity. Collected IDs are stored in an encrypted form using the server’s public key, at users’ devices, and in case of infection, they are shared with the server. The COVIDSafe application [6] from the Australian government is another Bluetooth-based solution. It also logs encrypted users’ contact information, and share them once an infection is detected. The server is required to alert users at risk without revealing the identity of the infected user. Both TraceTogether and COVIDSafe applications are set upon a centralized architecture. Many other applications like Stop COVID-19 (Croatia) [3], CA Notify (California) [1] rely on the Google and Apple Exposure Notification (GAEN) service [5], which is set upon a decentralized architecture. Although contact tracing applications have helped governments to alleviate the widespread of the pandemic by automating the manual contact tracing done by health authorities, they raise critical privacy concerns, namely users tracking and identification [15]. Academic solutions have been also proposed to support both centralized [13] and decentralized [7, 8, 18, 14, 16] architectures. However, each architecture has his merits and limits in terms of security and privacy. Indeed, using centralized solutions, users guarantee the reception of correct alerts as long as the generation of users’ contact tokens and the verification in case of infection are performed by a centralized server. This guarantee is compensated with threats to users’ privacy, i.e., users are exposed to tracking and identification of their contact lists. Decentralized solutions have been proposed to mitigate these privacy threats. Users are responsible for generating their contact tokens in order to ensure their privacy and anonymity, but, they are not prevented from forging contact information, which results in high level of false alerts. To get the best of both architectures, hybrid architecture based solutions [9, 12] have been proposed. However, security and privacy requirements, like the correctness of contact information and the anonymity of contacted users, are not yet handled and ensured together. In this paper, we present SPOT, a novel hybrid Secure and Privacy-preserving prOximity-based protocol for e-healthcare systems. It combines a decentralized proxy front-end architecture, ensuring both users’ anonymity and contact information integrity, and a centralized back-end computing server, guaranteeing a real time verification of contact information integrity. SPOT assumes that two users in close proximity rely on their Ephemeral Bluetooth IDentifiers (EBID) to compute a common contact message. This message is relayed to a central server through a group of proxies. With the help of the computing server and relying on a proof-based group signature, SPOT prevents users from forging their contact lists. The signed contact messages are given to the user to be locally stored. In case of a detected infection, the user consents to share his contact list, i.e., a set of signed contact messages, with the health authority. This latter checks the correctness of the received list and shared it back with all the involved users, if the verification holds. The contributions of this paper are summarized as follows: * • we design a proximity protocol for e-health services that prevents the injection of false positives, i.e., alert users to be at risk when they are not. SPOT enforces the verification of the correctness and the integrity of users’ contact information by health authorities, thanks to the support of both a computing server and a group of proxies. * • we guarantee strong privacy properties namely the anonymity of users being in contact with infected people, and the unlinkability of users’ transactions when relying on random EBIDs that can neither be linked to each other nor be linked to their issuer. * • we propose a concrete construction of the SPOT protocol relying on a structure-preserving signature scheme [4] that supports securely signing group’s elements, i.e., contacts’ information. This signature scheme is coupled with Groth-Sahai Non-Interactive Witness-Indistinguishable (NIWI) proof [10] in order to ensure integrity of proxies’ keys. Indeed, NIWI proofs guarantee the anonymity of proxies while the health authority is still able to verify that proxies are trustful by verifying the validity of their keys without having access to them. * • we evaluate the performances of SPOT through the full implementation of different procedures and algorithms. The conducted experiments have shown acceptable computation times proving the efficiency and practical usability of the proposed solution. This paper is organized as follows. Section II introduces and compares most closely-related proximity tracing algorithms and solutions to SPOT. Section III gives an overview of SPOT. After introducing the underlying building blocks in Section IV The concrete construction of the proposed SPOT protocol is presented in Section V. Section VI and Section VII provide security and privacy properties and a detailed discussion of SPOT’s conducted experiments, before concluding in Section VIII. ## II Related Work Recently, several industrial and research contact tracing solutions have been proposed for e-health applications [19, 17]. These solutions aim at ensuring security properties, namely anti-replay mitigating the multi-submission of the same contact information, and unforgeability preventing malicious entities111Malicious entities involve either a single malicious adversary or colluding adversaries. from threatening data integrity. Privacy properties have been significantly addressed, including the anonymity of end-users and the unlinkability between their different transactions (i.e., a formal definition of security and privacy requirements is given in Section III-D). Indeed, researchers from Inria, France, and Fraunhofer, Germany proposed Robert [13], a contact tracing protocol that relies on a centralized architecture, where a central server delivers pseudonyms to users. Each user collects pseudonyms of users in close proximity and shares them with the server when being infected. In such centralized solution, users are sure that they receive correct alerts (i.e., collected pseudonyms are neither replayed nor falsified by malicious entities), however, their privacy is threatened as long as the server is able to identify users’ contacts and to track them. In [7], Troncoso _et al._ introduced the Decentralized Privacy-Preserving Proximity Tracing (DP-3T) solution which is one of the most popular contact- tracing protocols. DP-3T has been proved to mitigate the privacy threats of centralized solutions as there is no need for a central entity which collects users’ contact information with the risk of tracking them. However, it does not prevent relay and replay attacks and gives no mean to verify the correctness of contact information. Thus, users are exposed to false alerts from malicious entities either by creating falsified information or replaying information of previous sessions. Afterwards, Castelluccia _et al._ proposed Desire [9], a proximity tracing protocol that leverages the advantages of the centralized and decentralized solutions. However, some security and privacy issues have not been considered in this solution. First malicious users are able to collude and merge their contact lists, which leads to false positive injection attacks. Second, the server requested to compute the exposure status and risk, is able to link users’ requests, and to de-anonymize them. Two very similar proposals called PACTs are also introduced. The east coast PACT [18] and the west coast PACT [8] are very close to DP-3T. The two solutions rely on random pseudonyms derived from a private seed, that are broadcasted to users in proximity via Bluetooth. The pseudonyms are generated using cryptographic pseudorandom generators and pseudorandom functions. Apart from the non- resistance against replay attacks, these two proposals give no mean to check the correctness of the contact information before being broadcasted. Pietrzak [16] proposed a decentralized contact-tracing solution to mitigate replay attacks against DP-3T. However, privacy concerns are raised, namely tracking users, as geo-location and time of contacts are shared within the Bluetooth message. In [12], Hoepman proposed two tracing protocols, the first one relies on an interactive session between two users in proximity to register contact information. If the interaction fails, the contact is not registered. Thus, the second protocol comes to mitigate this risk of failure and relies on an authority that relays information between users. In both protocols, the identities of users who have been in contact with an infected person, are revealed to a central entity, which contradicts the defined anonymity requirement. Liu et al. [14] use zero-knowledge proofs and group signatures in order to preserve users’ privacy for their proposed tracing protocol. Zero- knowledge proofs are generated by users over the contact information they collected. Indeed, users prove the contacts to their doctors without revealing the information. Afterwards, doctors, being members of a group, sign the proofs and publish them in a public board. Then, relying on their secrets, users can check if they were in contact with infected people. As such, no entity can identify the contacts of an infected user. However, based on a long interactive protocol between two devices, the collection of contacts’ information may result in a failed interaction, thus causing the non- registration of the contact. Furthermore, the authors only consider honest but curious adversaries, which leads to possible false alerts due to malicious colluding users. It is also worth noticing that, unlike SPOT and other related work [9, 7, 16], [14] assume that all the computations (handshake, zero- knowledge, verification) are performed by the user’s device, which leads to device’s battery depletion. A new contact tracing strategy based on online social networking is proposed in [20] but does not provide privacy guarantees. Table I provides a comparative summary between SPOT and related works in terms of architecture settings and properties. As shown, SPOT is the only solution which supports strong security and privacy requirements. TABLE I: Comparison between SPOT and related works | | SPOT | [14] | [12] | [7] | [18] [8] | [13] | [9] | [16] ---|---|---|---|---|---|---|---|---|--- Architecture | Centralized | - | - | - | - | - | ✓ | - | - Decentralized | - | ✓ | - | ✓ | ✓ | - | - | ✓ Semi-centralized | ✓ | - | ✓ | - | - | - | ✓ | - Properties | Unforgeability a | ✓ | ✓b | ✓b | ✗ | ✗ | N.A. | ✗ | ✗ Anti-replay | ✓ | ✓ | ✓ | ✗ | ✗ | ✓ | ✓ | ✓ Unlinkability | ✓ | N.A. | ✗ | ✓ | ✓ | ✗ | ✓ | ✗ Anonymity | ✓ | ✓ | ✗ | ✗ | ✗ | ✗ | ✓ | ✗ NOTE: N.A. is the abbreviation for Not Applicable; a indicates that the unforgeability implies the integrity of users’ contact information and the prevention of false positives injection; a indicates that unforgeability is partially satisfied while not considering malicious colluding entities. ## III SPOT Overview This section first presents the involved entities and gives an overview of SPOT. Then, it details the system model with its procedures and algorithms and defines the threat model through formal security games. ### III-A Entities Figure 1: Overview of the SPOT protocol Figure 1 depicts the four main actors involved in SPOT with their interactions according to the different phases. Four actors are defined as follows: * • The user ($\mathcal{U}$) represents the entity that owns the device where the proximity-tracing application is installed. During the Generation phase, $\mathcal{U}$ broadcasts his EBID (Ephemeral Bluetooth IDentifier), collects the EBIDs of other users in proximity, and computes a common contact message shared between each two users being in contact. $\mathcal{U}$ wants to receive alerts if he was in contact with confirmed cases. * • The Health Authority ($\mathcal{HA}$) is responsible for issuing users’ identifiers during the Sys_Init phase, and for checking the correctness of the contact messages provided by an infected user during the Verification phase. * • The Server ($\mathcal{S}$) is responsible for anonymously collecting and storing users’ contact messages relayed by proxies during the Generation phase222The Server can be distributed by considering one or several servers per geographical area, each server participating in locally storing part of users’ contact messages databases. All the parts are then collected on offline in a centralized server. Thus, for ease of presentation, we consider only one server.. $\mathcal{S}$ performs a real-time verification of the received contacts during the Generation phase, in order to help $\mathcal{HA}$ to verify the correctness and integrity of the contact messages. * • The Proxy ($\mathcal{P}$) is considered as a member of a group of proxies managed by the group manager ($\mathcal{GM}$)333Proxies are distributed over several geographical areas. We assume that a load-balancing is established between at least two proxies in the same geographical area to ensure the system availability in case of failure or overload. More precisely, proxies in a same geographical area are separated into two subsets - a primary and a secondary - and two users in a contact interaction must contact proxies belonging to different subsets in order to prevent a proxy from gaining too much knowledge about users’ interactions.. Proxies form an intermediate layer by relaying the common contact messages of users to $\mathcal{S}$ in order to ensure the anonymity of involved users towards the server during the Generation phase. Proxies also play an important role in ensuring data integrity and user geolocation privacy thanks to group signatures. ### III-B Overview SPOT is set upon an hybrid architecture that leverages the best of the centralized and decentralized settings in proximity-tracing protocols. It relies on a proxy-based solution to preserve users’ privacy (i.e., users remain anonymous towards the server, thus preventing users’ tracking) and is based on a semi-trusted computing server to ensure data consistency and integrity (i.e., users are ensured that the received alerts are correct). The architecture of the proposed protocol is depicted in Figure 1. SPOT involves three main phases: Sys_Init, Generation and Verification presented hereafter. The Sys_Init phase consists of initializing the whole system. It relies on seven algorithms, referred to as $\mathsf{Set\\_params}$, $\mathsf{HA\\_keygen}$, $\mathsf{S\\_keygen}$, $\mathsf{Setup\\_ProxyGr}_{\mathcal{GM}}$ and $\mathsf{Join\\_ProxyGr}_{\mathcal{P}/\mathcal{GM}}$, $\mathsf{Set\\_UserID}_{\mathcal{HA}}$ and $\mathsf{Userkeygen}_{\mathcal{U}}$. During the Sys_Init phase, a trusted authority444For ease of presentation, the trusted authority is neither presented in Figure 1 nor in the system’s model entities. generates the system public parameters published to all involved entities and the pair of keys of both $\mathcal{HA}$ and $\mathcal{S}$, relying on $\mathsf{Set\\_params}$, $\mathsf{HA\\_keygen}$ and $\mathsf{S\\_keygen}$ algorithms. During this phase, the group manager defines the group of proxies. It generates the group signature parameters using the $\mathsf{Setup\\_ProxyGr}_{\mathcal{GM}}$ algorithm and it interacts with each group member to derive the associated keys relying on the $\mathsf{Join\\_ProxyGr}_{\mathcal{P}/\mathcal{GM}}$ algorithm. The Health Authority is also involved in this phase to register a user when installing the proximity-tracing application. $\mathcal{HA}$ generates a specific secret value $t_{\mathcal{U}}$ (only known by $\mathcal{HA}$) and a unique identifier $\mathtt{ID}_{\mathcal{U}}$ for each user ($\mathcal{U}$), using the $\mathsf{Set\\_UserID}_{\mathcal{HA}}$ algorithm. Finally, $\mathcal{U}$ uses his identifier to generate his pair of keys relying on the $\mathsf{Userkeygen_{U}}$ algorithm. The user’s identifier $\mathtt{ID}_{\mathcal{U}}$, secret value $t_{\mathcal{U}}$ and public key are stored in a database $DB_{USER}$ owned by $\mathcal{HA}$. We note that the trusted authority, the group manager and proxies are involved only once in the Sys_Init phase, while the health authority must intervene every time a user wants to register. The Generation phase occurs when two users $\mathcal{U}_{A}$ and $\mathcal{U}_{B}$ are in contact. It represents the process of generating contact messages and contact lists for users. Three main entities participate in this phase relying on three different algorithms, referred to as $\mathsf{Set\\_CCM}_{\mathcal{U}}$, $\mathsf{S\\_PSign}_{\mathcal{S}}$ and $\mathsf{P\\_Sign}_{\mathcal{P}}$. At first, $\mathcal{U}_{A}$ and $\mathcal{U}_{B}$ execute the $\mathsf{Set\\_CCM}_{\mathcal{U}}$ algorithm to generate a common contact message relying on their random $EBID$s (denoted by $\mathtt{D}^{e}_{A}$ and $\mathtt{D}^{e}_{B}$) for an epoch $e$555An epoch $e$ denotes a period of time in which the Bluetooth identifier (EBID) remains unchanged.. $\mathcal{U}_{A}$ and $\mathcal{U}_{B}$ choose two different proxies to relay their common contact message to the server. For this purpose, they compare their $EBID$s, i.e., if $\mathtt{D}^{e}_{A}$ $>$ $\mathtt{D}^{e}_{B}$, $\mathcal{U}_{A}$ chooses the first proxy and $\mathcal{U}_{B}$ selects the second one, and vice versa. Each of the two proxies relays the common contact message to the server. $\mathcal{S}$ checks if the two copies are similar. If so, $\mathcal{S}$ executes the $\mathsf{S\\_PSign}_{\mathcal{S}}$ algorithm to partially sign the common contact message, thus proving that the contact message correctly reached the server. Afterwards, given back only a correct message, each proxy executes the $\mathsf{P\\_Sign}_{\mathcal{P}}$ algorithm. Indeed, each proxy extends the message, given by $\mathcal{S}$, with the corresponding user’s identifier and it signs the resulting message on behalf of the group. He, finally, sends back the message and the corresponding group signature to the user and closes the communication session, while removing all the exchanged and generated contact information. The user adds the group signature, along with the common contact message, the date, time and duration of contact, in his contact list $CL_{\mathcal{U}}$. Note that each contact information is stored for $\Delta$ days. The Verification phase is run by the health authority to check the correctness of a contact list $CL_{\mathcal{U}}$ provided by $\mathcal{U}$ during a period of time $t$. To this end, $\mathcal{HA}$ performs three successive verifications relying on two main algorithms, referred to as $\mathsf{Sig\\_Verify}_{\mathcal{HA}}$ and $\mathsf{CCM\\_Verify}_{\mathcal{HA}}$. (i) $\mathcal{HA}$ checks if, in his $DB_{USER}$ database, $\mathcal{U}$ is infected666We suppose that the health status of a user is updated when being tested. Indeed, to be tested, $\mathcal{U}$ has to provide an encrypted form of his identifier $\mathtt{ID}_{\mathcal{U}}$ (i.e., $\mathtt{ID}_{\mathcal{U}}$ is encrypted meaning the $\mathcal{HA}$ public key). Afterwards, the analysis’ result is sent with the encrypted identifier to $\mathcal{HA}$, that extracts the identifier and updates the user’s health status in the $DB_{USER}$ database.. (ii) $\mathcal{HA}$ checks the validity of the group signatures relying on the $\mathsf{Sig\\_Verify}_{\mathcal{HA}}$ algorithm, w.r.t. the messages contained in the contact list $CL_{\mathcal{U}}$. (iii) $\mathcal{HA}$ verifies that the contact messages have been correctly generated and have successfully reached $\mathcal{S}$, using the $\mathsf{CCM\\_Verify}_{\mathcal{HA}}$ algorithm. It is worth mentioning that if one of the verifications given above fails, the contact message is rejected. Otherwise, $\mathcal{HA}$ collects all verified messages of all infected users in a set $S_{\mathtt{CCM}}$ that she signs. Note that for each period of time $t$, $\mathcal{HA}$ removes users’ contact lists after verifications. $S_{\mathtt{CCM}}$ and the corresponding signature are sent to the server that shares them with all users. To compute the risk score, each user compares the set $S_{\mathtt{CCM}}$ with his contact list, taking into account the number of infected users being contacted and the contact duration. For ease of presentation, the different notations used in this paper are depicted in Table II. TABLE II: Notations used in this paper Notation | Description ---|--- $\mathcal{U}$ | User $\mathcal{HA}$ | Health Authority $\mathcal{S}$ | Server $\mathcal{P}$ | Proxy $\mathcal{GM}$ | Group Manager $\mathcal{TA}$ | Trusted Authority $\mathtt{ID}_{\mathcal{U}}$ | An identifier of a user $\mathcal{U}$ $t_{\mathcal{U}}$ | A secret value associated to $\mathcal{U}$ $DB_{USER}$ | The users’ database at $\mathcal{HA}$ $\mathtt{D}^{e}$ | An ephemeral Bluetooth Identifier during an epoch $e$ $CL_{\mathcal{U}}$ | A user’s contact list $\lambda$ | A security parameter $pp$ | The system’s public parameters $\mathtt{sk}$ | A private key $\mathtt{pk}$ | A public key $\mathtt{vk}_{g}$ | The group public parameters $\mathtt{CCM}$ | A common contact message $(\mathtt{PS},\mathtt{PS}^{\prime})$ | A partial signature $\sigma$ | A signature $\mathtt{M}$ | A message derived from $\mathtt{PS}$ and $\mathtt{ID}_{\mathcal{U}}$ $\pi$ | A NIWI proof $S_{\mathtt{CCM}}$ | A set of verified $\mathtt{CMM}$s of infected users ### III-C System Model Based on the three phases, Figure 2 presents the chronological sequence of twelve PPT algorithms, defined below. For ease of presentation, we consider only one user and one proxy in the sequence diagram. For the Generation phase, we suppose that two users have been in contact and exchanged their EBIDs. The Verification phase occurs only if the user receives a negative analysis’ result. Figure 2: Workflow of the SPOT protocol * • Sys_Init phase: $\mathsf{Set\\_params}(\lambda)\rightarrow pp$ – run by a trusted authority. Given the security parameter $\lambda$, this algorithm generates the system public parameters $pp$ that will be considered as a default input for all the following algorithms. $\mathsf{K\\_keygen}()\rightarrow(\mathtt{sk}_{j},\mathtt{pk}_{j})$ – performed by a trusted authority. It returns the pair of keys $(\mathtt{sk}_{j},\mathtt{pk}_{j})$ of $j$ where $j=\\{\mathcal{HA},\mathcal{S}\\}$. $\mathsf{Setup\\_ProxyGr}_{\mathcal{GM}}()\rightarrow(\mathtt{sk}_{g},\mathtt{vk}_{g})$ – this algorithm is performed by the group manager to set up the group signature. It returns the proxies’ group verification key $\mathtt{vk}_{g}$ represented as $(\mathtt{pk}_{g},\Sigma_{\mathsf{NIWI}})$, where $\mathtt{pk}_{g}$ is the public key of the group manager and $\Sigma_{\mathsf{NIWI}}$ is the Common Reference String CRS of a NIWI proof [10]. The $\mathsf{Setup\\_ProxyGr}_{\mathcal{GM}}$ algorithm also outputs the secret key $\mathtt{sk}_{g}$ that is privately stored by $\mathcal{GM}$. $\mathsf{Join\\_ProxyGr_{\mathcal{P}/\mathcal{GM}}}(\mathtt{sk}_{g})\rightarrow(\mathtt{sk}_{p},\mathtt{pk}_{p},\sigma_{p})$ – this algorithm is performed through an interactive session between the proxy and the group manager. It takes as input the secret key $\mathtt{sk}_{g}$, and outputs the pair of keys $(\mathtt{sk}_{p},\mathtt{pk}_{p})$ of $\mathcal{P}$ belonging to the group (i.e., $\mathcal{P}$ is responsible for generating his pair of keys), and a signature $\sigma_{p}$ over $\mathcal{P}$’s public key $\mathtt{pk}_{p}$ (i.e., $\sigma_{p}$ is generated by $\mathcal{GM}$). $\mathsf{Set\\_UserID_{\mathcal{HA}}}()\rightarrow(t_{\mathcal{U}},\mathtt{ID}_{\mathcal{U}})$ – this algorithm is run by $\mathcal{HA}$ and returns a secret value $t_{\mathcal{U}}$ specific for $\mathcal{U}$ and the identifier $\mathtt{ID}_{\mathcal{U}}$ of $\mathcal{U}$. $\mathsf{Userkeygen_{U}}(\mathtt{ID}_{\mathcal{U}})\rightarrow(\mathtt{sk}_{\mathcal{U}},\mathtt{pk}_{\mathcal{U}})$ – performed by $\mathcal{U}$ to set his pair of keys $(\mathtt{sk}_{\mathcal{U}},\mathtt{pk}_{\mathcal{U}})$ relying on the identifier $\mathtt{ID}_{\mathcal{U}}$. * • Generation phase: $\mathsf{Set\\_CCM_{\mathcal{U}}}(\mathtt{D}^{e}_{A},\mathtt{D}^{e}_{B})\rightarrow\mathtt{CCM}^{e}_{AB}$ – run by each of two users $\mathcal{U}_{A}$ and $\mathcal{U}_{B}$ being in contact during an epoch $e$. Given two Bluetooth identifiers $\mathtt{D}^{e}_{A}$ and $\mathtt{D}^{e}_{B}$, this algorithm generates a common contact message $\mathtt{CCM}^{e}_{AB}$. $\mathsf{S\\_PSign_{\mathcal{S}}}(\mathtt{CCM}^{e}_{AB},\mathtt{sk}_{\mathcal{S}})\rightarrow(\mathtt{PS}^{e}_{AB},\mathtt{PS^{\prime}}^{e}_{AB})$ – run by $\mathcal{S}$. Given a common contact message $\mathtt{CCM}^{e}_{AB}$ sent by $\mathcal{U}_{A}$ and $\mathcal{U}_{B}$ through two different proxies $\mathcal{P}_{1}$ and $\mathcal{P}_{2}$, this algorithm outputs the couple ($\mathtt{PS}^{e}_{AB}$, $\mathtt{PS^{\prime}}^{e}_{AB}$) that is stored with $\mathtt{CCM}^{e}_{AB}$ at $\mathcal{S}$, for $\Delta$ days. Note that only $\mathtt{PS}^{e}_{AB}$ is given back to $\mathcal{P}_{1}$ and $\mathcal{P}_{2}$ to prove that $\mathtt{CCM}^{e}_{AB}$ has been successfully received and verified by $\mathcal{S}$ (i.e., a real contact took place), while $\mathtt{PS^{\prime}}^{e}_{AB}$ is kept secret at $\mathcal{S}$ and is sent only to $\mathcal{HA}$ to check the correctness of a contact message provided by a infected user. $\mathsf{P\\_Sign}_{\mathcal{P}}(\mathtt{vk}_{g},\mathtt{sk}_{p},\mathtt{pk}_{p},\sigma_{p},\mathtt{ID}_{\mathcal{U}_{A}},\mathtt{PS}^{e}_{AB})\rightarrow(\mathtt{M}^{e}_{AB}$, $\sigma_{m}$, $\pi)$777In this algorithm, we only consider user $\mathcal{U}_{A}$ with $\mathtt{ID}_{\mathcal{U}_{A}}$. The same operations are performed for user $\mathcal{U}_{B}$ with $\mathtt{ID}_{\mathcal{U}_{B}}$. – performed by the proxy $\mathcal{P}$ ($\mathcal{P}_{1}$ or $\mathcal{P}_{2}$). This algorithm takes as input the proxies’ group public parameters $\mathtt{vk}_{g}$, the pair of keys $(\mathtt{sk}_{p},\mathtt{pk}_{p})$ of $\mathcal{P}$, the signature $\sigma_{p}$ over $\mathcal{P}$’s public key, the identifier $\mathtt{ID}_{\mathcal{U}_{A}}$ of user $\mathcal{U}_{A}$ and the message $\mathtt{PS}^{e}_{AB}$. It returns a signature $\sigma_{m}$ over a new message $\mathtt{M}^{e}_{AB}$ and a group signature represented by a NIWI proof $\pi$ over the two signatures $\sigma_{p}$ and $\sigma_{m}$. The couple ($\mathtt{M}^{e}_{AB}$, $\pi$) is sent to user $\mathcal{U}_{A}$ to be stored with the contact message $\mathtt{CCM}^{e}_{AB}$ in his contact list. Note that each input of the contact list is stored for $\Delta$ days. * • Verification phase: $\mathsf{Sig\\_Verify_{\mathcal{HA}}}(\mathtt{vk}_{g},\mathtt{M}^{e}_{AB},\pi)\rightarrow b$ – performed by $\mathcal{HA}$. Given the public parameters $\mathtt{vk}_{g}$, a message $\mathtt{M}^{e}_{AB}$ from the contact list of an infected user, and the corresponding NIWI proof $\pi$, the $\mathsf{Sig\\_Verify_{\mathcal{HA}}}$ algorithm returns a bit $b\in\\{0,1\\}$ stating whether the proof is valid. $\mathsf{CCM\\_Verify_{\mathcal{HA}}}(\mathtt{M}^{e}_{AB},\mathtt{PS^{\prime}}^{e}_{AB},\mathtt{pk}_{\mathcal{S}},t_{\mathcal{U}_{A}})\rightarrow b$ – run by $\mathcal{HA}$. This algorithm takes as input the message $\mathtt{M}^{e}_{AB}$, the message $\mathtt{PS^{\prime}}^{e}_{AB}$ requested from $\mathcal{S}$, the server’s public key $\mathtt{pk}_{\mathcal{S}}$ and the secret value $t_{\mathcal{U}_{A}}$, and outputs a bit $b\in\\{0,1\\}$, i.e., $\mathtt{CCM}^{e}_{AB}$ is correctly generated or not. ### III-D Threat Model In this section, we first present the adversaries considered in SPOT, and then, the formal definitions of the different security and privacy properties. * • A malicious user ($\mathcal{U}$): this adversary attempts to inject false contact messages or contact messages of other users in his contact list. He may also collude with corrupted proxies or malicious users. * • A honest but curious health authority ($\mathcal{HA}$): given a valid group signature, $\mathcal{HA}$ tries to identify the signer (i.e., proxy) of a particular message, hence for identifying the appropriate geographical area and for tracking the user’s movements. She may also attempt to link two signatures issued by the same group member. A curious $\mathcal{HA}$ may also try to identify, from a contact list of a particular user, the list of users he had been in contact with. * • A honest but curious server ($\mathcal{S}$): he attempts to link several common contact messages generated by the same user, to trace users’ movements. * • A malicious proxy ($\mathcal{P}$): this adversary, either colluding with a malicious user or with $n$ other proxies, attempts to forge the partial signature of the server and to generate a valid signature over a false contact message. Both malicious users and malicious proxies are considered against security properties, i.e., unforgeability, anti-replay, while the curious health authority and server are considered against privacy requirements, i.e., unlinkability and anonymity. The different adversaries are involved in different phases. Note that the anti-replay property which aims at mitigating the multi- submission of the same contact information is not formally presented below, but is informally discussed in Section VI. The following properties are defined w.r.t the corresponding phases and the involved adversaries. ###### Remark 1. We do not deeply analyze the case of a malicious $\mathcal{GM}$ although our scheme is resistant against this adversary. Indeed, proxies are responsible for generating their key pair and only their public keys are shared with $\mathcal{GM}$. Thus, unless holding a proxy’s secret key, $\mathcal{GM}$ is not able to generate a valid signature on behalf of $\mathcal{P}$ thanks to the unforgeability of the signature scheme. #### III-D1 Unforgeability The unforgeability property ensures the security of SPOT for the different phases. It states that a malicious user is not able to forge his contact list (i.e., forging either the group signature or the server’s partial signature when colluding with a malicious proxy)888We assume that (i) malicious user refers to either a single user or colluding users, and (ii) the group signature scheme used in SPOT is unforgeable as proven in [4], thus in the security model and analysis, we will only consider the unforgeability of the server’s partial signature.. Formally, this is defined in a game $\textbf{Exp}_{\mathcal{A}}^{unforg}$ where an adversary $\mathcal{A}$, playing the role of a corrupted proxy colluding with a malicious user, has access to a $\mathsf{S\\_PSign}$ oracle. We note that, for each session $i$, $\mathcal{A}$ only gets $\mathtt{PS}^{i}$ from the $\mathsf{S\\_PSign}$ oracle, while $\mathtt{PS^{\prime}}^{i}$ is kept secret from the adversary. Then, given a valid message $\mathtt{PS^{\prime}}$ that cannot be obtained by combining either a part of or all messages $\mathtt{PS}^{i}$, $\mathcal{A}$ succeeds if it outputs a valid message $\mathtt{PS}^{*}$ to be signed using $\mathsf{P\\_Sign}$, such that the $\mathsf{CCM\\_Verify}$ verification holds. ###### Definition 1. Unforgeability – We say that SPOT satisfies the unforgeability property, if for every _PPT_ adversary $\mathcal{A}$, there exists a negligible function $\kappa$ such that: $Pr[\textbf{Exp}_{\mathcal{A}}^{unforg}(1^{\lambda})=1]\leq\kappa(\lambda)$, where $\textbf{Exp}_{\mathcal{A}}^{unforg}$ is given below. $\textbf{{Exp}}_{\mathcal{A}}^{unforg}(\lambda)$ $pp\leftarrow\mathsf{Set\\_params}(\lambda)$ $(\mathtt{sk_{\mathcal{HA}}},\mathtt{pk_{\mathcal{HA}}})\leftarrow\mathsf{HA\\_keygen}(pp)$ $(\mathtt{sk_{\mathcal{S}}},\mathtt{pk_{\mathcal{S}}})\leftarrow\mathsf{S\\_keygen}(pp)$ $(\mathtt{sk}_{g},\mathtt{vk}_{g})\leftarrow\mathsf{Setup\\_ProxyGr}(pp)$ $(\mathtt{sk}_{p},\mathtt{pk}_{p},\sigma_{p})\leftarrow\mathsf{Join\\_ProxyGr}(pp,\mathtt{sk}_{g})$ $\mathtt{ID_{\mathcal{U}}}\leftarrow\mathsf{Set\\_UserID}(pp)$ $(\mathtt{sk}_{\mathcal{U}},\mathtt{pk}_{\mathcal{U}})\leftarrow\mathsf{Userkeygen}(pp,\mathtt{ID_{\mathcal{U}}})$ $\mathtt{CCM}\leftarrow\mathsf{Set\\_CCM}(\mathtt{D}_{1},\mathtt{D}_{2})$ $(\mathtt{PS},\mathtt{PS^{\prime}})$ $\leftarrow$ $\\{\mathsf{S\\_PSign}(\mathtt{CCM},\mathtt{sk_{\mathcal{S}}})\\}$ $\mathcal{O}$ $\leftarrow$ $\\{\mathsf{S\\_PSign}(\cdot,\mathtt{sk_{\mathcal{S}}})\\}$ $\mathtt{PS}^{*}\leftarrow\mathcal{A}^{\mathcal{O}}(\mathtt{vk_{g}},\mathtt{sk_{p}},\mathtt{pk_{p}},\sigma_{p},ID_{\mathcal{U}},pp,\mathtt{PS^{\prime}})$ letting $\mathtt{CCM}$ and $\mathtt{PS}^{i}$ denote the queries and answers to and from oracle $\mathsf{S\\_PSign}$ $(\mathtt{M}^{*},\sigma^{*},\pi^{*})\leftarrow\mathsf{P\\_Sign}(\mathtt{vk}_{g},\mathtt{sk}_{p},\mathtt{pk}_{p},\sigma_{p},ID_{\mathcal{U}},\mathtt{PS}^{*})$ If $\mathsf{CCM\\_Verify}(\mathtt{M}^{*},\mathtt{PS^{\prime}}^{*},\mathtt{pk_{\mathcal{S}}},t_{\mathcal{U}})=1$ return 1 Else return 0 #### III-D2 Unlinkability The unlinkability property can be divided into two sub-properties. The first one constitutes the _group-signature unlinkability_ stating that a curious health authority is not able to link two or several group signatures issued by the same proxy during the Verification phase. The second sub-property _multi- CCM unlinkability_ ensures that a curious server is not able to link two or several common contact messages to the same user during the Generation phase 999The collusion between the health authority and the server does not pose additional and plausible threats to the different procedures of the whole framework. Indeed, during the Generation phase, contact messages are anonymous to the server (and a possible colluding health authority); during the Verification phase, the health authority knows the true identity of the confirmed cases with their contact information; as such, a collusion between the server and the authority does not bring extra knowledge.. We note that the _multi-CCM unlinkability_ property will be informally discussed in Section VI. In this section, we only focus on the _group- signature unlinkability_. Formally, this property is defined in a game $\textbf{Exp}_{\mathcal{A}}^{unlink}$ where an adversary $\mathcal{A}$ acting as a curious $\mathcal{HA}$ has access to a $\mathsf{P\\_Sign}$ oracle. The adversary may query this oracle on the same message $\mathtt{PS}^{*}$ and on a tuple $((\mathtt{sk}_{p_{j}},\mathtt{pk}_{p_{j}},\sigma_{p_{j}})$, where $j\in\\{0,1\\}$ (i.e., the tuple belongs either to proxy $\mathcal{P}_{0}$ or proxy $\mathcal{P}_{1}$). A left-or-right oracle $\mathsf{LoRSig}$ is initialized with a secret random bit $b$ and returns to $\mathcal{A}$ $\mathsf{P\\_Sign}$ on message $\mathtt{PS}^{*}$ and respectively on tuples $(\mathtt{sk}_{p_{0}},\mathtt{pk}_{p_{0}},\sigma_{p_{0}})$ and $(\mathtt{sk}_{p_{b}},\mathtt{pk}_{p_{b}},\sigma_{p_{b}})$. The adversary wins the game if he successfully predicts the bit $b$ (i.e., the guessing probability should be greater than $\frac{1}{2}$). ###### Definition 2. Unlinkability – We say that SPOT satisfies the unlinkability property, if for every _PPT_ adversary $\mathcal{A}$, there exists a negligible function $\kappa$ such that: $Pr[\textbf{Exp}_{\mathcal{A}}^{unlink}(\lambda)=1]=\frac{1}{2}\pm\kappa(\lambda)$, where $\textbf{Exp}_{\mathcal{A}}^{unlink}$ is defined below. $\textbf{{Exp}}_{\mathcal{A}}^{unlink}(\lambda)$ $pp\leftarrow\mathsf{Set\\_params}(\lambda)$ $(\mathtt{sk_{\mathcal{HA}}},\mathtt{pk_{\mathcal{HA}}})\leftarrow\mathsf{HA\\_keygen}(pp)$ $(\mathtt{sk_{\mathcal{S}}},\mathtt{pk_{\mathcal{S}}})\leftarrow\mathsf{S\\_keygen}(pp)$ $(\mathtt{sk}_{g},\mathtt{vk}_{g})\leftarrow\mathsf{Setup\\_ProxyGr}(pp)$ $(\mathtt{sk}_{p_{i}}$, $\mathtt{pk}_{p_{i}}$, $\sigma_{p_{i}})$ $\leftarrow$ $\mathsf{Join\\_ProxyGr}$ ($pp$, $\mathtt{sk}_{g}$), $i\in\\{0,1\\}$ $\mathtt{ID_{\mathcal{U}}}\leftarrow\mathsf{Set\\_UserID}(pp)$ $(\mathtt{sk}_{\mathcal{U}},\mathtt{pk}_{\mathcal{U}})\leftarrow\mathsf{Userkeygen}(pp,\mathtt{ID_{\mathcal{U}}})$ $m^{*}\leftarrow\mathsf{Set\\_CCM}(\mathtt{D}_{1},\mathtt{D}_{2})$ $(\mathtt{PS}^{*},\mathtt{PS^{\prime}}^{*})\leftarrow\mathsf{S\\_PSign}(m^{*},\mathtt{sk_{S}})$ $b\leftarrow\\{0,1\\}$ $\mathcal{O}$ $\leftarrow$ $\\{\mathsf{P\\_Sign}$($\cdot$,$\mathtt{sk}_{p_{j}}$,$\mathtt{pk}_{p_{j}}$,$\sigma_{p_{j}}$,$\cdot$,$\cdot$), $\mathsf{LoRSig}$($\cdot$,$\cdot$,$b)\\}$ $b^{\prime}$ $\leftarrow$ $\mathcal{A}^{\mathcal{O}}$ $(\mathtt{sk_{\mathcal{HA}}}$,$\mathtt{pk_{\mathcal{HA}}}$, $\mathtt{vk}_{g}$,$\mathtt{pk}_{p_{0}}$,$\mathtt{pk}_{p_{1}}$,$pp$,$\mathtt{PS}^{*})$ If $b=b^{\prime}$ return 1 Else return 0 $\mathsf{LoRSig}$ $(\mathtt{vk}_{g}$, (($\mathtt{sk}_{p_{0}}$, $\mathtt{pk}_{p_{0}}$, $\sigma_{p_{0}})$, ($\mathtt{sk}_{p_{1}}$, $\mathtt{pk}_{p_{1}}$, $\sigma_{p_{1}}$)), $\mathtt{ID_{\mathcal{U}}}$, $\mathtt{PS}^{*},b)$ $({\mathtt{M}}^{*},\sigma^{*},\pi^{*})$ $\leftarrow$ $\mathsf{P\\_Sign}$($\mathtt{vk}_{g}$, $\mathtt{sk}_{p_{0}}$, $\mathtt{pk}_{p_{0}}$, $\sigma_{p_{0}}$, $\mathtt{ID_{\mathcal{U}}}$ , $\mathtt{PS}^{*})$ $({\mathtt{M}}^{*},\sigma^{*}_{b},\pi^{*}_{b})$ $\leftarrow$ $\mathsf{P\\_Sign}$($\mathtt{vk}_{g}$, $\mathtt{sk}_{p_{b}}$, $\mathtt{pk}_{p_{b}}$, $\sigma_{p_{b}}$, $\mathtt{ID_{\mathcal{U}}}$ , $\mathtt{PS}^{*})$ return $(({\mathtt{M}}^{*},\pi^{*}),({\mathtt{M}}^{*},\pi^{*}_{b}))$ #### III-D3 Anonymity This property guarantees that no entity is able to identify users involved in a contact list (i.e., the owner and the contacted users), during the Verification phase, and is described through the game $\textbf{Exp}_{\mathcal{A}}^{anon}$. The anonymity property implies that even if $\mathcal{HA}$ knows that a contact list belongs to a user ($\mathcal{U}$), $\mathcal{HA}$ is not able to identify users being in contact with $\mathcal{U}$ 101010We assume that the probability of two confirmed users being in contact and submitting their respective contact lists to $\mathcal{HA}$ at the same period, is low.. This should hold even if an efficient adversary, playing the role of the curious health authority, is given access to $\mathsf{Set\\_CCM}$, $\mathsf{S\\_PSign}$, $\mathsf{P\\_Sign}$ oracles. $\mathcal{A}$ can learn contact messages and signatures associated to the selected users’ identifiers. $\mathcal{A}$ also gets access to a left-or-right oracle $\mathsf{LoRCU}$ which is initialized with a secret random bit $b\in\\{0,1\\}$. $\mathcal{A}$ may query this oracle on $\mathtt{ID_{\mathcal{U}_{0}}}$ and $\mathtt{ID_{\mathcal{U}_{1}}}$ referred to as the identifiers of respectively user $\mathcal{U}_{0}$ and user $\mathcal{U}_{1}$. Observe that user $\mathcal{U}_{A}$ is involved in all queries. $\mathtt{D}^{*}_{\mathcal{U}_{A}}$ and $\mathtt{D}^{*}_{\mathcal{U}_{b}}$, respectively belonging to user $\mathcal{U}_{A}$ and user $\mathcal{U}_{b}$, are randomly selected in order to execute the $\mathsf{LoRCU}$ oracle. To win the proposed anonymity game, the adversary should predict the bit $b$ (i.e., which one of users $\mathcal{U}_{0}$ and $\mathcal{U}_{1}$ is involved in the contact with user $\mathcal{U}_{A}$) with a probability greater than $\frac{1}{2}$. ###### Definition 3. Anonymity – We say that SPOT fulfills the anonymity requirement, if for every _PPT_ adversary $\mathcal{A}$, there exists a negligible function $\kappa$ such that: $Pr[\textbf{Exp}_{\mathcal{A}}^{anon}(1^{\lambda})=1]=\frac{1}{2}\pm\kappa(\lambda)$, where $\textbf{Exp}_{\mathcal{A}}^{anon}$ is defined as follows. $\textbf{{Exp}}_{\mathcal{A}}^{anon}(\lambda)$ $pp\leftarrow\mathsf{Set\\_params}(\lambda)$ $(\mathtt{sk_{\mathcal{HA}}},\mathtt{pk_{\mathcal{HA}}})$ $\leftarrow$ $\mathsf{HA\\_keygen}(pp)$ $(\mathtt{sk_{\mathcal{S}}},\mathtt{pk_{\mathcal{S}}})$ $\leftarrow$ $\mathsf{S\\_keygen}(pp)$ $(\mathtt{sk}_{g},\mathtt{vk}_{g})$ $\leftarrow$ $\mathsf{Setup\\_ProxyGr}(pp)$ $(\mathtt{sk}_{p},\mathtt{pk}_{p},\sigma_{p})$ $\leftarrow$ $\mathsf{Join\\_ProxyGr}(pp,\mathtt{sk}_{g})$ $\mathtt{ID}_{\mathcal{U}_{A}}$ $\leftarrow$ $\mathsf{Set\\_UserID}(pp)$ $(\mathtt{sk}_{\mathcal{U}_{A}},\mathtt{pk}_{\mathcal{U}_{A}})$ $\leftarrow$ $\mathsf{Userkeygen}(pp,\mathtt{ID}_{\mathcal{U}_{A}})$ $\mathtt{ID}_{\mathcal{U}_{i}}$ $\leftarrow$ $\mathsf{Set\\_UserID}(pp)$, $i\in\\{0..N\\}$ $(\mathtt{sk}_{\mathcal{U}_{i}},\mathtt{pk}_{\mathcal{U}_{i}})$ $\leftarrow$ $\mathsf{Userkeygen}(pp,\mathtt{ID}_{\mathcal{U}_{i}})$, $i\in\\{0..N\\}$ $b\leftarrow\\{0,1\\}$ $\mathcal{O}$ $\leftarrow$ $\\{\mathsf{Set\\_CCM}$($\cdot$,$\cdot$), $\mathsf{S\\_PSign}$($\cdot$,$\mathtt{sk_{\mathcal{S}}}$), $\mathsf{P\\_Sign}$($\cdot$,$\mathtt{sk_{p}}$, $\cdot$,$\sigma_{p}$,$\cdot$,$\cdot$), $\mathsf{LoRCU}$($\cdot$,$\cdot$,$b$,$\cdot)$ $b^{\prime}$ $\leftarrow$ $\mathcal{A}^{\mathcal{O}}$ $(\mathtt{sk_{\mathcal{HA}}}$, $\mathtt{pk_{\mathcal{HA}}}$, $pp$, $\mathtt{ID}_{\mathcal{U}_{A}}$, $\\{\mathtt{ID}_{\mathcal{U}_{i}}\\}_{i=0}^{N})$ If $b=b^{\prime}$ return 1 Else return 0 $\mathsf{LoRCU}$ $(\mathtt{D}^{*}_{\mathcal{U}_{A}}$, $\mathtt{D}^{*}_{\mathcal{U}_{b}}$, $b$, $\mathtt{vk}_{g}$, $\mathtt{sk_{\mathcal{S}}}$, $\mathtt{sk}_{p}$, $\mathtt{pk}_{p}$, $\sigma_{p}$, $\mathtt{ID}_{\mathcal{U}_{A}}$, $\mathtt{ID}_{\mathcal{U}_{b}})$ ${\mathsf{CCM}}^{*}_{b}$ $\leftarrow$ $\mathsf{Set\\_CCM_{\mathcal{U}_{A}}}$ $(\mathtt{D}^{*}_{\mathcal{U}_{A}}$, $\mathtt{D}^{*}_{\mathcal{U}_{b}})$ $({\mathsf{PS}}^{*}_{b},{\mathsf{PS}^{\prime}}^{*}_{b})$ $\leftarrow$ $\mathsf{S\\_PSign}$ $({\mathsf{CCM}}^{*}_{b}$, $\mathtt{sk_{\mathcal{S}}})$ $(\mathtt{M}^{*}_{b},\sigma^{*}_{b},\pi^{*}_{b})$ $\leftarrow$ $\mathsf{P\\_Sign}$ $(\mathtt{vk}_{g}$, $\mathtt{sk}_{p}$, $\mathtt{pk}_{p}$, $\sigma_{p}$, $\mathtt{ID}_{\mathcal{U}_{A}}$, ${\mathsf{PS}}^{*}_{b})$ return $({\mathsf{CCM}}^{*}_{b},\mathtt{M}^{*}_{b},\pi^{*}_{b})$ ## IV Building Blocks After introducing bilinear maps and security standard assumptions in Section IV-A, the section next presents structure-preserving signatures [4] with their different variants as main building blocks of the SPOT protocol. Sections IV-B and IV-C describe respectively constant-size signatures and signatures on mixed-group messages that are instantiated in Appendix IV-D to build a group signature scheme on group element messages. ### IV-A Mathematical Background and Cryptographic Assumptions Hereafter, we define bilinear maps and we present the computational indistinguishability property and the CDH assumption. #### IV-A1 Bilinear Maps Let $\mathbb{G}_{1}$=$\langle{g_{1}}\rangle$ and $\mathbb{G}_{2}$=$\langle{g_{2}}\rangle$ be two cyclic groups of order $n$ so there exists a bilinear map $e:\mathbb{G}_{1}\times\mathbb{G}_{2}\rightarrow\mathbb{G}_{3}$ that satisfies the following properties: (i) bilinearity for all $g_{1}\in\mathbb{G}_{1}$, $g_{2}\in\mathbb{G}_{2}$, (ii) non-degeneracy: $e(g_{1},g_{2})\neq 1$ and (iii) $e(g_{1},g_{2})$ is efficiently computable for any $g_{1}\in\mathbb{G}_{1}$ and $g_{2}\in\mathbb{G}_{2}$. #### IV-A2 Computational Witness-Indistinguishability The Computational Witness-Indistinguishability property is defined as follows: Let $L\in\mathcal{NP}$ be a language and let $(\mathcal{P},\mathcal{V})$ be an interactive proof system for $L$. We say that $(\mathcal{P},\mathcal{V})$ is _witness-indistinguishable_ (WI) if for every _PPT_ algorithm $\mathcal{V}^{*}$ and every two sequences $\\{w^{1}_{x}\\}_{x\in L}$ and $\\{w^{2}_{x}\\}_{x\in L}$ such that $w^{1}_{x}$ and $w^{2}_{x}$ are both witnesses for $x$, the following ensembles are computationally indistinguishable, where $z$ is an auxiliary input to $\mathcal{V}^{*}$: 1. 1. $\\{\langle\mathcal{P}(w^{1}_{x}),\mathcal{V}^{*}\rangle(x)\\}_{x\in L,z\in\\{0,1\\}^{*}}$ 2. 2. $\\{\langle\mathcal{P}(w^{2}_{x}),\mathcal{V}^{*}\rangle(x)\\}_{x\in L,z\in\\{0,1\\}^{*}}$ #### IV-A3 Computational Diffie Hellman Assumption (CDH) The CDH assumption is defined as follows: Let $\mathbb{G}$ be a group of prime order $n$, and $g$ is a generator of $\mathbb{G}$. The CDH problem is defined as: Given the tuple of elements $(g,g^{x},g^{y})$, where $\\{x,y\\}\leftarrow\mathbb{Z}_{n}$, there is no efficient algorithm $\mathcal{A}_{CDH}$ that can compute $g^{xy}$. ### IV-B Structure-preserving Constant-size Signature Structure-preserving constant-size signature was defined by Abe _et al._ [4] as the main scheme of structure-preserving signatures used to sign a message $\vec{m}=(m_{1},...,m_{k})\in{\mathbb{G}_{2}}^{k}$, considering an asymmetric bilinear group $(n,\mathbb{G}_{1},\mathbb{G}_{2},$ $\mathbb{G}_{3},g_{1},g_{2},e)$. A constant-size signature scheme $\mathsf{CSIG}$ [4] relies on the following three PPT algorithms ($\mathsf{CSIG}.\mathsf{Key}$, $\mathsf{CSIG}.\mathsf{Sign}$, $\mathsf{CSIG}.\mathsf{Verify}$): ${\mathsf{CSIG}}.\mathsf{Key}(1^{\lambda})$: This algorithm takes as input the security parameter $(1^{\lambda})$ and outputs the pair of public and secret keys $(\mathtt{sk},\mathtt{pk})$ of the signer. It chooses two random generators $g_{r},h_{u}\leftarrow\mathbb{G}^{*}_{1}$ and random values $\gamma_{i},\delta_{i}\leftarrow\mathbb{Z}^{*}_{n}$ and computes $g_{i}={g_{r}}^{\gamma_{i}}$ and $h_{i}={h_{u}}^{\delta_{i}}$, for $i=1,...,k$. It then selects $\gamma_{z},\delta_{z}\leftarrow\mathbb{Z}^{*}_{n}$ and computes $g_{z}={g_{r}}^{\gamma_{z}}$ and $h_{z}={h_{u}}^{\delta_{z}}$. It also chooses $\alpha,\beta\leftarrow\mathbb{Z}^{*}_{n}$ and sets the couples $(g_{r},g_{2}^{\alpha})$ and $(h_{u},g_{2}^{\beta})$. The public key is set as $\mathtt{pk}=(g_{z},h_{z},g_{r},h_{u},g_{2}^{\alpha},g_{2}^{\beta},\\{g_{i},h_{i}\\}^{k}_{i=1})$ and the secret key is set as $\mathtt{sk}=(\mathtt{pk},\alpha,\beta,\gamma_{z},\delta_{z},\\{\gamma_{i},\delta_{i}\\}^{k}_{i=1})$. $\mathsf{CSIG}.\mathsf{Sign}(\mathtt{sk},\vec{m})$: This algorithm generates a signature $\sigma$ over a message $\vec{m}$ using the secret key $\mathtt{sk}$. That is, the signer randomly selects $\zeta,\rho,\tau,\varphi,\omega\leftarrow\mathbb{Z}^{*}_{n}$ and computes $z=g_{2}^{\zeta},r={g_{2}}^{\alpha-\rho\tau-{\gamma_{z}}\zeta}{\prod}^{k}_{i=1}{m_{i}}^{-\gamma_{i}},s={g_{r}}^{\rho},t={g_{2}}^{\tau},$ $u={g_{2}}^{\beta-\varphi\omega-{\delta_{z}}\zeta}{\prod}^{k}_{i=1}{m_{i}}^{-\delta_{i}},v={h_{u}}^{\varphi},w={g_{2}}^{\omega}$ The signature is set as $\sigma=(z,r,s,t,u,v,w)$. $\mathsf{CSIG}.\mathsf{Verify}(\mathtt{pk},\vec{m},\sigma)$: This algorithm checks the validity of the signature $\sigma$ on the message $m$ relying on the signer’s public key $\mathtt{pk}$. It outputs 1 if the signature is valid and 0 otherwise. The verifier checks if the following equations hold: $A=e(g_{z},z)e(g_{r},r)e(s,t){\prod}^{k}_{i=1}e(g_{i},m_{i})$ (1) $B=e(h_{z},z)e(h_{u},u)e(v,w){\prod}^{k}_{i=1}e(h_{i},m_{i})$ (2) where $A=e(g_{r},g_{2}^{\alpha})$ and $B=e(h_{u},g_{2}^{\beta})$ ### IV-C Structure-preserving signature on mixed-group messages A structure-preserving signature on mixed-group messages $\mathsf{XSIG}$ [4] represents a signature scheme where the message space is a mixture of the two groups $\mathbb{G}_{1}$ and $\mathbb{G}_{2}$. We consider two constant-size signature schemes $\mathsf{CSIG}1$ and $\mathsf{CSIG}2$. $\mathsf{CSIG}2$ is the same scheme as in Section IV-B where the message space is ${\mathbb{G}_{2}}^{k_{2}}$, while $\mathsf{CSIG}1$ is a ’dual’ scheme obtained by exchanging $\mathbb{G}_{1}$ and $\mathbb{G}_{2}$ in the same scheme, where the message space is ${\mathbb{G}_{1}}^{k_{1}}$. The message space for the $\mathsf{XSIG}$ is then ${\mathbb{G}_{1}}^{k_{1}}\times{\mathbb{G}_{2}}^{k_{2}}$. Let $(\vec{m},\vec{\tilde{m}})$ be a message in ${\mathbb{G}_{1}}^{k_{1}}\times{\mathbb{G}_{2}}^{k_{2}}$. For a vector $\vec{\tilde{m}}\in{\mathbb{G}_{1}}^{k_{1}}$ and a single element $s\in\mathbb{G}_{1}$, let $\vec{m}||s$ denote a vector in ${\mathbb{G}_{1}}^{k_{1}+1}$ obtained by appending $s$ to the end of $\vec{m}$. A mixed-group messages signature scheme $\mathsf{XSIG}$ relies on the following three PPT algorithms ($\mathsf{XSIG}.\mathsf{Key}$, $\mathsf{XSIG}.\mathsf{Sign}$, $\mathsf{XSIG}.\mathsf{Verify}$): $\mathsf{XSIG}.\mathsf{Key}(1^{\lambda})$: This algorithm runs $(\mathtt{sk}_{1}$, $\mathtt{pk}_{1})$ $\leftarrow$ $\mathsf{CSIG}1.\mathsf{Key}(1^{\lambda})$ and $(\mathtt{sk}_{2}$, $\mathtt{pk}_{2})$ $\leftarrow$ $\mathsf{CSIG}2.\mathsf{Key}(1^{\lambda})$ and sets $(\mathtt{sk},\mathtt{pk})=((\mathtt{sk}_{1},\mathtt{sk}_{2}),(\mathtt{pk}_{1},\mathtt{pk}_{2}))$. $\mathsf{XSIG}.\mathsf{Sign}(\mathtt{sk},(\vec{m},\vec{\tilde{m}}))$: This algorithm runs $\sigma_{2}$ = $(z,r,s,t,u,v,w)$ $\leftarrow$ $\mathsf{CSIG}2.\mathsf{Si}$\- $\mathsf{gn}(\mathtt{sk_{2}}$, $\vec{\tilde{m}})$ and $\sigma_{1}=(z^{\prime},r^{\prime},s^{\prime},t^{\prime},u^{\prime},v^{\prime},w^{\prime})\leftarrow\mathsf{CSIG}1.\mathsf{Sign}(\mathtt{sk_{1}},\vec{m}||s)$, and outputs $\sigma=(\sigma_{1},\sigma_{2})$. $\mathsf{XSIG}.\mathsf{Verify}(\mathtt{pk},(\vec{m},\vec{\tilde{m}}),(\sigma_{1},\sigma_{2}))$: This algorithm takes $s\in\mathbb{G}_{1}$ from $\sigma_{2}$, runs $b_{2}=\mathsf{CSIG}2.\mathsf{Verify}(\mathtt{pk_{2}},\vec{\tilde{m}},\sigma_{2})$ and $b_{1}=\mathsf{CSIG}1.\mathsf{Verify}(\mathtt{pk_{1}},\vec{m}||s,\sigma_{1})$. If $b_{1}=b_{2}=1$, the algorithm outputs 1, otherwise it outputs 0. ### IV-D Group signatures drawn from structure-preserving signatures We present hereafter an instantiation of a group signature scheme that allows to sign a group element message relying on a constant-size signature scheme $\mathsf{CSIG}$, a mixed-group messages signature scheme $\mathsf{XSIG}$ and a witness indistinguishable proof of knowledge system $\mathsf{NIWI}$ [10] (cf. Appendix A). A group signature scheme $\mathsf{GSIG}$ relies on the four following algorithms ($\mathsf{GSIG}.\mathsf{Setup}$, $\mathsf{GSIG}.\mathsf{Join}$, $\mathsf{GSIG}.\mathsf{Sign}$, $\mathsf{GSIG}.\mathsf{Verify}$): $\mathsf{GSIG}.\mathsf{Setup}$ : represents the setup algorithm. It runs $\mathsf{XSIG}.\mathsf{Key}$ algorithm that generates the key pair $(\mathtt{sk_{g}},\mathtt{pk}_{g})$ of the group manager and sets up a CRS $\Sigma_{\mathsf{NIWI}}$ for the $\mathsf{NIWI}$ proof. The group verification key is set as $\mathtt{vk}_{g}=(\mathtt{pk}_{g},\Sigma_{\mathsf{NIWI}})$, while the certification secret key $\mathtt{sk_{g}}$ is privately stored by the group manager. $\mathsf{GSIG}.\mathsf{Join}$: represents the join algorithm. It is composed of two steps. In the first one, the group member generates his key-pair $(\mathtt{sk_{p}},\mathtt{pk}_{p})$ while running the $\mathsf{CSIG}.\mathsf{Key}$ algorithm. Only the public key $\mathtt{pk}_{p}$ is sent to the group manager. This latter generates a signature $\sigma_{p}$ over $\mathtt{pk}_{p}$, using the $\mathsf{XSIG}.\mathsf{Sign}$ algorithm, and sends it to the group member. $\mathsf{GSIG}.\mathsf{Sign}$: represents the signing algorithm run by a group member on a message $m\in\mathbb{G}_{2}$. The group member generates, over the message $m$, a signature $\sigma_{m}\leftarrow\mathsf{CSIG}.\mathsf{Sign}(\mathtt{sk}_{p},m)$ and a non-interactive witness indistinguishable proof of knowledge $\pi\leftarrow\mathsf{NIWI}.\mathsf{Proof}(\Sigma_{\mathsf{NIWI}},pub,wit)$ that proves $1=\mathsf{XSIG}.\mathsf{Verify}(\mathtt{pk}_{g},\mathtt{pk}_{p},\sigma_{p})$ and $1=\mathsf{CSIG}.\mathsf{Verify}(\mathtt{pk}_{p},m,\sigma_{m})$ with respect to the witness $wit=(\mathtt{pk}_{p},\sigma_{p},\sigma_{m})$ and the public information $pub=(\mathtt{pk}_{g},m)$. The signing algorithm outputs the group signature $\pi$. $\mathsf{GSIG}.\mathsf{Verify}$: represents the group signature verification algorithm run by a verifier. It takes $(\mathtt{vk}_{g},m,\pi)$ as input and verifies the correctness of the $\mathsf{NIWI}$ proof $\pi$ w.r.t. $pub=(\mathtt{pk}_{g},m)$ and the CRS $\Sigma_{\mathsf{NIWI}}$. ## V SPOT Algorithms This section gives a concrete construction of the different phases and algorithms of SPOT, introduced in Section III-A. SPOT relies on the different variants of structure-preserving signatures represented in Appendix IV. ### V-A Sys_Init phase * • $\mathsf{Set\\_params}$ – a trusted authority sets an asymmetric bilinear group $(n$, $\mathbb{G}_{1}$, $\mathbb{G}_{2}$, $\mathbb{G}_{3}$, $g_{1}$, $g_{2}$, $e)$ relying on the security parameter $\lambda$, where $\mathbb{G}_{1}$ and $\mathbb{G}_{2}$ are two cyclic groups of prime order $n$, $g_{1}$ and $g_{2}$ are generators of respectively $\mathbb{G}_{1}$ and $\mathbb{G}_{2}$ and $e$ is a bilinear map such that $e:\mathbb{G}_{1}\times\mathbb{G}_{2}\rightarrow\mathbb{G}_{3}$. The trusted authority also considers a cryptographic hash function $\mathbf{H}:\\{0,1\\}^{*}\rightarrow\mathbb{Z}_{n}$. The output of the $\mathsf{Set\\_params}$ algorithm represents the system global parameters that are known by all the system entities. The tuple $(n,\mathbb{G}_{1},\mathbb{G}_{2},\mathbb{G}_{3},g_{1},g_{2},e,\mathbf{H})$ is denoted by $pp$, and is considered as a default input of all algorithms. * • $\mathsf{HA\\_keygen}$ – a trusted authority takes as input the public parameters $pp$, selects a random $x\in\mathbb{Z}^{*}_{n}$ and generates the pair of secret and public keys $(\mathtt{sk}_{\mathcal{HA}},\mathtt{pk}_{\mathcal{HA}})$ of the health authority as follows: $\mathtt{sk}_{\mathcal{HA}}=x\quad;\quad\mathtt{pk}_{\mathcal{HA}}=g_{2}^{x}$ * • $\mathsf{S\\_keygen}$ – a trusted authority generates the pair of secret and public keys $(\mathtt{sk}_{\mathcal{S}},\mathtt{pk}_{\mathcal{S}})$ of the server as given below, relying on the system public parameters $pp$ and two selected randoms $y_{1},y_{2}\in\mathbb{Z}^{*}_{n}$. $\mathtt{sk}_{\mathcal{S}}=(y_{1},y_{2})\quad;\quad\mathtt{pk}_{\mathcal{S}}=(Y_{1},Y_{2})=(g_{2}^{y_{1}},g_{2}^{y_{2}})$ * • $\mathsf{Setup\\_ProxyGr_{\mathcal{GM}}}$ – $\mathcal{GM}$ sets up the group of proxies by generating a group public key $\mathtt{vk}_{g}$ and a certification secret key $\mathtt{sk}_{g}$ as shown in Algorithm 1. * • $\mathsf{Join\\_ProxyGr_{\mathcal{P}/\mathcal{GM}}}$ – $\mathcal{P}$ first generates his pair of keys $(\mathtt{sk}_{p},\mathtt{pk}_{p})$ w.r.t. the ${\mathsf{CSIG}}.\mathsf{Key}$ algorithm (cf. Section IV-B). Afterwards, $\mathcal{GM}$ generates a signature $\sigma_{p}$ over the public key $\mathtt{pk}_{p}$ w.r.t. the ${\mathsf{XSIG}}.\mathsf{Sign}$ algorithm (cf. Section IV-C). The $\mathsf{Join\\_ProxyGr_{\mathcal{P}/\mathcal{GM}}}$ algorithm is detailed in Algorithm 2. Algorithm 1 $\mathsf{Setup\\_ProxyGr_{\mathcal{GM}}}$ algorithm 1:Input: the system public parameters $pp$ 2:Output: the public parameters $\mathtt{vk}_{g}$ of the proxies’ group and the secret key $\mathtt{sk}_{g}$ 3:// The next iterations are executed to generate the pair of keys of $\mathcal{GM}$ 4:pick at random $g_{r1},h_{u1}\leftarrow\mathbb{G}^{*}_{1}$, $g_{r2},h_{u2}\leftarrow\mathbb{G}^{*}_{2}$ 5:for $i=1$ to $2$ do pick at random $\gamma_{1i},\delta_{1i}\leftarrow\mathbb{Z}^{*}_{n}$ compute $g_{1i}\leftarrow{g_{r1}}^{\gamma_{1i}}$, $h_{1i}\leftarrow{h_{u1}}^{\delta_{1i}}$ 6:end for 7:for $j=1$ to $7$ do pick at random $\gamma_{2j},\delta_{2j}\leftarrow\mathbb{Z}^{*}_{n}$ compute $g_{2i}\leftarrow{g_{r2}}^{\gamma_{2j}}$ and $h_{2j}\leftarrow{h_{u2}}^{\delta_{2j}}$ 8:end for 9:pick at random $\gamma_{1z},\delta_{1z},\gamma_{2z},\delta_{2z}\leftarrow\mathbb{Z}^{*}_{n}$ ; 10:compute $g_{1z}\leftarrow{g_{r1}}^{\gamma_{1z}}$, $h_{1z}\leftarrow{h_{u1}}^{\delta_{1z}}$, $g_{2z}\leftarrow{g_{r2}}^{\gamma_{2z}}$ and $h_{2z}\leftarrow{h_{u2}}^{\delta_{2z}}$ ; 11:pick at random $\alpha_{1},\alpha_{2},\beta_{1},\beta_{2}\leftarrow\mathbb{Z}^{*}_{n}$ ; 12:$\mathtt{pk}_{1}\leftarrow(g_{2z},h_{2z},g_{2r},h_{2u},g_{1}^{\alpha_{2}},g_{1}^{\beta_{2}},\\{g_{2j},h_{2j}\\}^{7}_{j=1})$ and $\mathtt{sk}_{1}\leftarrow(\mathtt{pk}_{1},\alpha_{2},\beta_{2},\gamma_{2z},\delta_{2z},\\{\gamma_{2j},\delta_{2j}\\}^{7}_{j=1})$ ; 13:$\mathtt{pk}_{2}\leftarrow(g_{1z},h_{1z},g_{1r},h_{1u},g_{2}^{\alpha_{1}},g_{2}^{\beta_{1}},\\{g_{1i},h_{1i}\\}^{2}_{i=1})$ and $\mathtt{sk}_{2}\leftarrow(\mathtt{pk}_{2},\alpha_{1},\beta_{1},\gamma_{1z},\delta_{1z},\\{\gamma_{1i},\delta_{1i}\\}^{2}_{i=1})$ ; 14:set $\mathtt{pk}_{g}\leftarrow(\mathtt{pk}_{1},\mathtt{pk}_{2})$ and $\mathtt{sk}_{g}\leftarrow(\mathtt{sk}_{1},\mathtt{sk}_{2})$ ; 15:// The next iterations are executed to generate the CRS $\Sigma_{\mathsf{NIWI}}$ 16:pick at random $r,s\leftarrow\mathbb{Z}^{*}_{n}$ and set $\mathcal{U}=rg_{1}$ and $\mathcal{V}=sg_{2}$ ; 17:set $\Sigma_{\mathsf{NIWI}}=(\mathbb{G}_{1},\mathbb{G}_{2},\mathbb{G}_{3},e,\iota_{1},p_{1},\iota_{2},p_{2},\iota_{3},\mathcal{U},\mathcal{V})$ ; 18:$\mathtt{vk}_{g}\leftarrow(\mathtt{pk}_{g},\Sigma_{\mathsf{NIWI}})$ ; 19:return $(\mathtt{sk}_{g},\mathtt{vk}_{g})$ * • $\mathsf{Set\\_UserID_{\mathcal{HA}}}$ – every time, a user ($\mathcal{U}$) installs the application and wants to register, $\mathcal{HA}$ picks a secret $t_{\mathcal{U}}\in\mathbb{Z}^{*}_{n}$ and sets the user’s identifier $\mathtt{ID}_{\mathcal{U}}$ as $\mathtt{ID}_{\mathcal{U}}=h_{\mathcal{U}}=g_{2}^{t_{\mathcal{U}}}$ * • $\mathsf{Userkeygen_{\mathcal{U}}}$ – After receiving his identifier $\mathtt{ID}_{\mathcal{U}}$, a user generates his pair of secret and private keys $(\mathtt{sk}_{\mathcal{U}},\mathtt{pk}_{\mathcal{U}})$. Indeed, $\mathcal{U}$ randomly selects $q_{\mathcal{U}}\in\mathbb{Z}^{*}_{n}$ and sets $(\mathtt{sk}_{\mathcal{U}},\mathtt{pk}_{\mathcal{U}})$ as $\mathtt{sk}_{\mathcal{U}}=q_{\mathcal{U}}\quad;\quad\mathtt{pk}_{\mathcal{U}}={h_{\mathcal{U}}}^{q_{\mathcal{U}}}$ ### V-B Generation phase * • $\mathsf{Set\\_CCM_{\mathcal{U}}}$ – For each epoch $e$, $\mathcal{U}_{A}$ and $\mathcal{U}_{B}$ generate random EBIDs $\mathtt{D}^{e}_{\mathcal{U}_{A}}$ and $\mathtt{D}^{e}_{\mathcal{U}_{B}}$, respectively. $\mathcal{U}_{A}$ and $\mathcal{U}_{B}$ exchange their EBIDs and each of them executes the $\mathsf{Set\\_CCM}$ algorithm. $\mathcal{U}_{A}$ (resp. $\mathcal{U}_{B}$) computes $m^{e}_{AB}=\mathtt{D}^{e}_{\mathcal{U}_{A}}*\mathtt{D}^{e}_{\mathcal{U}_{B}}$ and sets the common contact element between $\mathcal{U}_{A}$ and $\mathcal{U}_{B}$ as $\mathtt{CCM}^{e}_{AB}=\mathbf{H}(m^{e}_{AB})$. * • $\mathsf{S\\_PSign_{\mathcal{S}}}$ – After checking that he receives two copies of $\mathtt{CCM}^{e}_{AB}$, the server picks at random $r_{s}\leftarrow\mathbb{Z}^{*}_{n}$ and, relying on his secret key $\mathtt{sk}_{\mathcal{S}}$, he computes the two messages $\mathtt{PS}^{e}_{AB}$ and $\mathtt{PS^{\prime}}^{e}_{AB}$ such that $\mathtt{PS}^{e}_{AB}=\mathtt{CCM}^{e}_{AB}y_{1}r_{s}+y_{2}\quad and\quad\mathtt{PS^{\prime}}^{e}_{AB}=\mathtt{CCM}^{e}_{AB}r_{s}$ * • $\mathsf{P\\_Sign_{\mathcal{P}}}$ – We consider that when being requested by a user $\mathcal{U}_{A}$, the proxy opens a session and saves the user’s identifier $\mathtt{ID}_{\mathcal{U}_{A}}$. This latter is used when executing the $\mathsf{P\\_Sign_{\mathcal{P}}}$ algorithm (c.f. Algorithm 3) to generate a new message $\mathtt{M}^{e}_{AB}$ (Line 4). The proxy then signs $\mathtt{M}^{e}_{AB}$ (Line 6 – Line 8) following the ${\mathsf{CSIG}}.\mathsf{Sign}$ algorithm and finally generates a proof $\pi$ (Line 10 – Line 16) w.r.t. the ${\mathsf{GSIG}}.\mathsf{Sign}$ algorithm. Algorithm 2 $\mathsf{Join\\_ProxyGr_{\mathcal{P}/\mathcal{GM}}}$ algorithm 1:Input: the security parameter $\lambda$ and the secret key of the group manager $\mathtt{sk}_{g}$ 2:Output: the pair of keys of a proxy group member $(\mathtt{sk}_{p},\mathtt{pk}_{p})$ and the signature $\sigma_{p}$ over the public key the public $\mathtt{pk}_{p}$ 3:// The next is set by $\mathcal{P}$ 4:pick at random $g_{r},h_{u}\leftarrow\mathbb{G}^{*}_{1}$, $\gamma,\delta\leftarrow\mathbb{Z}^{*}_{n}$ ; 5:compute $g_{\gamma}\leftarrow{g_{r}}^{\gamma}$ and $h_{\delta}\leftarrow{h_{u}}^{\delta}$ ; 6:pick at random $\gamma_{z},\delta_{z}\leftarrow\mathbb{Z}^{*}_{n}$ ; 7:compute $g_{z}\leftarrow{g_{r}}^{\gamma_{z}}$ and $h_{z}\leftarrow{h_{u}}^{\delta_{z}}$ ; 8:pick at random $\alpha,\beta\leftarrow\mathbb{Z}^{*}_{n}$ ; 9:set $\mathtt{pk}_{p}=(g_{z},h_{z},g_{r},h_{u},g_{2}^{\alpha},g_{2}^{\beta},g_{\gamma},h_{\delta})$ and $\mathtt{sk}_{p}=(\mathtt{pk}_{p},\alpha,\beta,\gamma_{z},\delta_{z},\gamma,\delta)$ ; 10:// The next is set by $\mathcal{GM}$ 11:$\sigma_{p}\leftarrow\mathsf{XSIG}.\mathsf{Sign}(\mathtt{sk}_{g},\mathtt{pk}_{p})$ ; 12:return $(\mathtt{sk}_{p},\mathtt{pk}_{p},\sigma_{p})$ Algorithm 3 $\mathsf{P\\_Sign_{\mathcal{P}}}$ algorithm 1:Input: the public parameters of the proxies’ group $\mathtt{vk}_{g}$, the secret key $\mathtt{sk}_{p}$, the signature $\sigma_{p}$ over the proxy’s public key, the identifier $\mathtt{ID}_{\mathcal{U}_{A}}$ of user $\mathcal{U}_{A}$ and the message $\mathtt{PS}$ 2:Output: a message $\mathtt{M}$, the corresponding signature $\sigma_{m}$ and a proof $\pi$ 3:// The next is executed by $\mathcal{P}$ to generate $\mathtt{M}$ 4:compute $\mathtt{M}={\mathtt{ID}_{\mathcal{U}_{A}}}^{\mathtt{PS}}$; 5:// The next is executed by $\mathcal{P}$ to sign $\mathtt{M}$ 6:pick at random $\zeta,\rho,\tau,\varphi,\omega\leftarrow\mathbb{Z}^{*}_{n}$ ; 7:run $z=g_{2}^{\zeta}$, $r={g_{2}}^{\alpha-\rho\tau-{\gamma_{z}}\zeta}{\mathtt{M}}^{-\gamma}$, $s={g_{r}}^{\rho}$, $t={g_{2}}^{\tau}$, $u={g_{2}}^{\beta-\varphi\omega-{\delta_{z}}\zeta}{\mathtt{M}}^{-\delta}$, $v={h_{u}}^{\varphi}$, $w={g_{2}}^{\omega}$ ; 8:set $\sigma_{m}=(z,r,s,t,u,v,w)$ ; 9:// The next is set to generate a proof on equations $\\{(\vec{\mathcal{A}_{im}},\vec{\mathcal{B}_{im}},\Gamma_{im},t_{im})\\}^{2}_{i=1}$ where $\vec{\mathcal{A}_{im}}$ = $\vec{\mathcal{B}_{im}}$ = $\vec{0}$, $\Gamma_{im}$ = $\mathcal{MAT}_{3\times 3}(1)$ for $i=1,2$, $t_{1m}$ = $t_{2m}$ = $1_{\mathbb{G}_{3}}$ 10:$\vec{\mathcal{X}}_{1m}=(g_{z},g_{r},s)$, $\vec{\mathcal{X}}_{2m}=(h_{z},h_{u},v)$, $\vec{\mathcal{Y}}_{1m}=(z,{g_{2}}^{\alpha-\rho\tau-{\gamma_{z}}\zeta},t)$ and $\vec{\mathcal{Y}}_{2m}=(z,{g_{2}}^{\beta-\rho\tau-{\delta{z}}\zeta},w)$ ; 11:$\pi_{m}=\\{(\vec{\mathcal{C}_{im}},\vec{\mathcal{D}_{im}},\pi_{im},\theta_{im})\\}^{2}_{i=1}\leftarrow\mathsf{NIWI}.\mathsf{Proof}(\mathtt{vk}_{g}$, $\\{(\vec{\mathcal{A}_{im}},\vec{\mathcal{B}_{im}},\Gamma_{im},t_{im})\\}^{2}_{i=1},$ $\\{(\vec{\mathcal{X}_{im}},\vec{\mathcal{Y}_{im}})\\}^{2}_{i=1})$ ; 12:// The next is set to generate a proof on equations $\\{(\vec{\mathcal{A}_{ip}},\vec{\mathcal{B}_{ip}},\Gamma_{ip},t_{ip})\\}^{4}_{i=1}$ where $\vec{\mathcal{A}}_{1p}=(g_{1}^{\alpha_{2}})$, $\vec{\mathcal{A}}_{2p}=(g_{1}^{\beta_{2}})$, $\vec{\mathcal{A}}_{3p}=(g_{1z},g_{1r})$, $\vec{\mathcal{A}}_{4p}=(h_{1z},h_{1u})$, $\vec{\mathcal{B}}_{1p}=(g_{2z},g_{2r})$, $\vec{\mathcal{B}}_{2p}=(h_{2z},h_{2u})$, $\vec{\mathcal{B}}_{3p}=(g_{2}^{\alpha_{1}})$, $\vec{\mathcal{B}}_{4p}=(g_{2}^{\beta_{1}})$, $\Gamma_{1p}=(\gamma_{2z},-1)$, $\Gamma_{2p}=(\delta_{2z},-1)$, $\Gamma_{3p}=(\gamma_{1z},-1)$, $\Gamma_{4p}=(\delta_{1z},-1)$, $t_{1p}=e(g_{1}^{\alpha_{2}},g_{2r})$, $t_{2p}=e(g_{1}^{\beta_{2}},h_{2u})$, $t_{3p}=e(g_{1r},g_{2}^{\alpha_{1}})$ and $t_{4p}=e(h_{1u},g_{2}^{\beta_{1}})$ 13:$\vec{\mathcal{X}}_{1p}=(z_{1},{g_{1}}^{\alpha_{2}-\rho_{1}\tau_{1}-{\gamma_{2z}}\zeta_{1}})$, $\vec{\mathcal{X}}_{2p}=(z_{1},{g_{1}}^{\beta_{2}-\rho_{1}\tau_{1}-{\delta_{2z}}\zeta_{1}})$, $\vec{\mathcal{X}}_{3p}=(g_{1r})$, $\vec{\mathcal{X}}_{4p}=(h_{1u})$, $\vec{\mathcal{Y}}_{1p}=(g_{2r})$, $\vec{\mathcal{Y}}_{2p}=(h_{2u})$, $\vec{\mathcal{Y}}_{3p}=(z_{2},{g_{2}}^{\alpha_{1}-\rho_{2}\tau_{2}-{\gamma_{1z}}\zeta_{2}})$, and $\vec{\mathcal{Y}}_{4p}=(z_{2},{g_{2}}^{\beta_{1}-\rho_{2}\tau_{2}-{\delta_{1z}}\zeta_{2}})$ ; 14:$\pi_{p}=\\{(\vec{\mathcal{C}_{ip}},\vec{\mathcal{D}_{ip}},\pi_{ip},\theta_{ip})\\}^{4}_{i=1}\leftarrow\mathsf{NIWI}.\mathsf{Proof}(\mathtt{vk}_{g},\\{(\vec{\mathcal{A}_{ip}},\vec{\mathcal{B}_{ip}},\Gamma_{ip},t_{ip})\\}^{4}_{i=1},$ $\\{(\vec{\mathcal{X}_{ip}},\vec{\mathcal{Y}_{ip}})\\}^{4}_{i=1})$ ; 15:set $\pi_{p}=((\pi_{ip},\theta_{ip})_{i=1}^{4})$ ; 16:set $\pi=(\pi_{p},\pi_{m})$ ; 17:return $(\mathtt{M},\sigma_{m},\pi)$ ### V-C Verification phase * • $\mathsf{Sig\\_Verify_{\mathcal{HA}}}$ – Given a contact list of user $\mathcal{U}_{A}$ (a list of tuples $(\mathtt{CCM}$, $\mathtt{M}$, $\pi)$ such that $\pi$ can be parsed as $\\{(\vec{\mathcal{A}_{i}},\vec{\mathcal{B}_{i}},\Gamma_{i},t_{i})\\}^{N}_{i=1}$, $\\{(\vec{\mathcal{C}_{i}},\vec{\mathcal{D}_{i}},\pi_{i},\theta_{i})\\}^{N}_{i=1}$), $\mathcal{HA}$ verifies the validity of the group signature of each message, w.r.t. ${\mathsf{GSIG}}.\mathsf{Verify}$ algorithm (cf. Appendix IV-D). * • $\mathsf{CCM\\_Verify_{\mathcal{HA}}}$ – We consider that $\mathcal{HA}$ requests from $\mathcal{S}$ the message $\mathtt{PS^{\prime}}$ corresponding to a contact message $\mathtt{CCM}$ contained in the contact list of user $\mathcal{U}_{A}$. The message $\mathtt{PS^{\prime}}$ is taken as input with the message $\mathtt{M}$ (corresponding to $\mathtt{CCM}$), the server’s public key $\mathtt{pk}_{\mathcal{S}}$ and the secret value $t_{\mathcal{U}_{A}}$ specific to user $\mathcal{U}_{A}$, to the $\mathsf{CCM\\_Verify_{HA}}$ algorithm that checks if the equation 3 holds: $\mathtt{M}={Y_{1}}^{t_{\mathcal{U}_{A}}\mathtt{PS^{\prime}}}{Y_{2}}^{t_{\mathcal{U}_{A}}}$ (3) ## VI Security and Privacy Analysis In this section, we prove that SPOT achieves the defined security and privacy requirements with respect to the threat models defined in Section III-D, by relying on the following theorems and lemmas. ###### Theorem 1 (Unforgeability). If a probabilistic-polynomial time (PPT) adversary $\mathcal{A}$ wins $\textbf{Exp}_{\mathcal{A}}^{unforg}$, as defined in Section III-D1, with a non-negligible advantage $\epsilon$, then a PPT simulator $\mathcal{B}$ can be constructed to break the CDH assumption with a non-negligible advantage $\epsilon$. ###### Proof. In this proof, we show that a simulator $\mathcal{B}$ can be constructed with the help of an adversary $\mathcal{A}$ having advantage $\epsilon$ against SPOT scheme. The CDH challenger $\mathcal{C}$ sends to $\mathcal{B}$ the tuple $(g_{2},g_{2}^{a},g_{2}^{b})$, where $a,b\leftarrow\mathbb{Z}^{*}_{n}$ are randomly selected. $\mathcal{C}$ asks $\mathcal{B}$ to compute $g_{2}^{ab}$. Then, $\mathcal{B}$ sets $g_{2}^{t_{\mathcal{U}}}$ to $g_{2}^{a}$ and $g_{2}^{y_{2}}$ to $g_{2}^{b}$. During the challenge phase, $\mathcal{B}$ randomly selects $y_{1}\in\mathbb{Z}^{*}_{n}$ and sends $g_{2}^{y_{1}}$ to $\mathcal{A}$ as part of the server’s public key. $\mathcal{A}$ forges the partial signature over the message $\mathtt{PS^{\prime}}$ and generates the message $\mathtt{M}^{*}$ with advantage $\epsilon$: $\mathtt{M}^{*}$ = ${\mathtt{ID}_{\mathcal{U}}}^{\mathtt{PS}^{*}}$ = ${g_{2}}^{t_{\mathcal{U}}(\mathtt{PS^{\prime}}y_{1}+y_{2})}$. The tuple $({g_{2}}^{t_{\mathcal{U}}},\mathtt{PS^{\prime}},\mathtt{M}^{*})$ is sent back to $\mathcal{B}$. Upon receiving this tuple and knowing $y_{1}$, $\mathcal{B}$ can compute the value of $g_{2}^{t_{\mathcal{U}}y_{2}}$ which is the same as $g_{2}^{ab}$ and can then send the result to the CDH challenger. As such, $\mathcal{B}$ succeeds the forgery against the CDH assumption with advantage $\epsilon$. ∎ ###### Theorem 2 (Unlinkability). Our SPOT system achieves the unlinkability requirement with respect to the _group-signature unlinkability_ and _multi-CCM unlinkability_ properties. We prove Theorem 2 through Lemma 3 and Lemma 4 with respect to _group- signature unlinkability_ and _multi-CCM unlinkability_ properties, respectively. ###### Lemma 3 (Group-signature unlinkability). SPOT satisfies the group signature unlinkability requirement with respect to the computational witness indistinguishability property of the NIWI proof. ###### Proof. In this proof, the objective is to show that the adversary is not able to distinguish group signatures issued by the same proxy. For this purpose, we suppose that, for each session $i$, the adversary receives the message $M^{*}$ (i.e., the same message $M^{*}$ is returned by each oracle) and the NIWI proof $\pi^{i}$ = $(\pi_{m}^{i},\pi_{p}^{i})$ = $((\pi^{i}_{jm},\theta^{i}_{jm})_{j=1}^{2}$, $(\pi^{i}_{jp},\theta^{i}_{jp})_{j=1}^{4})$. To simplify the proof, we will only consider the NIWI proof $\pi_{m}^{i}$, as the statements used to generate the proofs $\pi_{p}^{i}$ do not give any information about the proxy generating the proof (i.e., statements do not include the proxy’s public key). Thus, for each session $i$, the adversary is given the tuples $(\vec{\mathcal{C}^{i}_{k1}},\vec{\mathcal{D}^{i}_{k1}},\pi^{i}_{k1},\theta^{i}_{k1})$ and $(\vec{\mathcal{C}^{i}_{k2}},\vec{\mathcal{D}^{i}_{k2}},\pi^{i}_{k2},\theta^{i}_{k2})$ referred to as the group signature generated by a proxy $P_{k}$, where $k\in\\{0,1\\}$. During the challenge phase, the adversary is also given two group signatures. The first signature is represented by the tuples $(\vec{\mathcal{C}^{*}_{1}},\vec{\mathcal{D}^{*}_{1}},\pi^{*}_{1},\theta^{*}_{1})$ and $(\vec{\mathcal{C}^{*}_{2}},\vec{\mathcal{D}^{*}_{2}},\pi^{*}_{2},\theta^{*}_{2})$ generated by proxy $\mathcal{P}_{0}$, while the second one is represented by the tuples $(\vec{\mathcal{C}^{*}_{b1}},\vec{\mathcal{D}^{*}_{b1}},\pi^{*}_{b1},\theta^{*}_{b1})$ and $(\vec{\mathcal{C}^{*}_{b2}},\vec{\mathcal{D}^{*}_{b2}},\pi^{*}_{b2},\theta^{*}_{b2})$ and is generated by a proxy $\mathcal{P}_{b}$ ($b\in\\{0,1\\}).$ Let us consider a simulator $\mathcal{B}$ that can be constructed with the help of an adversary $\mathcal{A}$ having advantage $\epsilon$ against SPOT scheme. A challenger $\mathcal{C}$ selects two couples of witnesses $(X_{0},Y_{0})$ and $(X_{1},Y_{1})$. $\mathcal{C}$ computes a commitment $(C,D)$ over $(X_{0},Y_{0})$, and then selects a bit $b\in\\{0,1\\}$ and computes a commitment $(C^{\prime}_{b},D^{\prime}_{b})$ over $(X_{b},Y_{b})$. $\mathcal{C}$ asks $\mathcal{B}$ to guess the bit $b$. Then, $\mathcal{B}$ selects the tuples $(A,B,\Gamma,t)$ and $(A^{\prime}_{b},B^{\prime}_{b},\Gamma^{\prime}_{b},t^{\prime}_{b})$ and computes the proofs $(\pi,\theta)$ and $(\pi^{\prime}_{b},\theta^{\prime}_{b})$. $\mathcal{B}$ returns the two proofs to $\mathcal{A}$. Finally, $\mathcal{A}$ outputs a bit $b^{\prime}$ that it sends to $\mathcal{B}$. This latter outputs the same bit $b^{\prime}$ to its own challenger $\mathcal{C}$. As such, $\mathcal{A}$ succeeds in breaking the group-signature unlinkability with advantage $\epsilon$, which is the same as breaking the computational witness-indistinguishability property. ∎ ###### Corollary 3.1. If SPOT satisfies the unlinkability property, then the proxies’ group signature (i.e., NIWI proof) fulfills the anonymity requirement stating that it is not possible to identify the proxy that issued a particular group signature. ###### Lemma 4 (Multi-CCM unlinkability). SPOT satisfies the multi-CCM unlinkability requirement with respect to the common contact message structure. ###### Sketch of proof. Let $\mathcal{A}$ be a successful adversary against the _multi-CCM unlinkability_ property. Assume that $\mathcal{A}$ receives two messages $\mathtt{CCM_{1}}=\mathbf{H}(\mathtt{D}_{i}*\mathtt{D}_{j})$ and $\mathtt{CCM_{2}}=\mathbf{H}(\mathtt{D}_{i}*\mathtt{D}_{k})$ (with $j\neq k$) meaning that user $U_{i}$ met $U_{j}$ and $U_{k}$, then $\mathcal{A}$ is not able to link $\mathtt{CCM_{1}}$ and $\mathtt{CCM_{2}}$ to the same user $\mathcal{U}_{i}$ as all EBIDs are randomly generated in each epoch $e$, and the hashing function $\mathbf{H}$ behaves as a pseudo-random function. ∎ ###### Theorem 5 (Anonymity). SPOT satisfies the anonymity property, in the sense of Definition 3, if and only if, the CCM-unlinkability requirement is fulfilled. ###### Sketch of proof. We prove that our proximity-based protocol SPOT satisfies the anonymity property using an _absurdum_ reasoning. We suppose that an adversary $\mathcal{A}$ can break the anonymity of SPOT, in the sense of Definition 3, by reaching the advantage $Pr[\textbf{Exp}_{\mathcal{A}}^{anon}(1^{\lambda})=1]\geq\frac{1}{2}\pm\kappa(\lambda)$. $\mathcal{A}$ is given the pair of public-private keys $(\mathtt{pk}_{\mathcal{HA}},\mathtt{sk}_{\mathcal{HA}})$ of the health authority, the identifiers $\mathtt{ID}_{\mathcal{U}}$ of all users and a contact list of a particular user $\mathcal{U}_{A}$, obtained when relying on several sessions. Then, relying on the _left-or-right_ $\mathsf{LoRCU}$ oracle, $\mathcal{A}$ tries to distinguish the user $\mathcal{U}_{b}$ being in contact with $\mathcal{U}_{A}$, better than a flipping coin. That is, given the tuple $({\mathsf{CCM}}^{*}_{b},\mathtt{M}^{*}_{b},\pi^{*}_{b})$, $\mathcal{A}$ successfully predicts the identifier $\mathtt{ID}_{\mathcal{U}_{b}}$. Obviously, $\mathcal{A}$ tries to identify user $\mathcal{U}_{b}$ relying on the message ${\mathsf{CCM}}^{*}_{b}$, since both $\mathtt{M}^{*}_{b}$ and $\pi^{*}_{b}$ are generated based on ${\mathsf{CCM}}^{*}_{b}$ and give no further information about the user $\mathcal{U}_{b}$. This refers to link the message ${\mathsf{CCM}}^{*}_{b}$ to its issuer $\mathcal{U}_{b}$. Thus, if $\mathcal{A}$ succeeds, this means that $\mathcal{A}$ is able to link two or several common contact messages to the same user, which contradicts the _multi-CCM unlinkability_ property previously discussed. As such, we prove that the adversary succeeds $\textbf{Exp}_{\mathcal{A}}^{anon}(1^{\lambda})$ with a probability $Pr=\frac{1}{2}\pm\kappa(\lambda)$, where $\kappa(\lambda)$ is negligible. Thus, SPOT satisfies anonymity. ∎ ###### Theorem 6 (Anti-replay). SPOT satisfies the anti-replay requirement and supports false positive hindrance, if the proposed scheme is unforgeable. ###### Sketch of proof. To successfully replay a common contact message generated in an epoch $e$, in another epoch $e^{\prime}\neq e$, a malicious user can perform in two ways. (i) The user reinserts, in his contact list, the tuple $(\mathtt{CCM}^{e}$, $\mathtt{M}^{e}$, $\pi^{e})$ generated in an epoch $e$. The reinsertion is performed in an epoch $e^{\prime}>e+\Delta$111111It makes no sense to reinsert an element in an epoch $e^{\prime}<e+\Delta$, as duplicated messages will be deleted either by the server or at the user’s end-device.. Afterwards, the contact list is sent to $\mathcal{HA}$ when the user is infected. $\mathcal{HA}$ asks the server to provide the message $\mathtt{PS}^{\prime}$ corresponding to $\mathtt{CCM}^{e}$. As the server has no entry corresponding to $\mathtt{CCM}^{e}$ in the last $\Delta$ days, the second verification performed by $\mathcal{HA}$ does not hold and the tuple is rejected. (ii) We assume that, in an epoch $e^{\prime}>e+\Delta$, the user is able to replay a message $\mathtt{CCM}^{e}$ with two different proxies and he successfully receives the corresponding message $\mathtt{M}$ and the group signature $\pi$. Thus, when the user is infected, the health authority validates false positives, but this has no impact on the computation of the risk score, as no user has the same entry in his contact list. As such, we can prove the resistance of SPOT against replay attacks. ∎ ## VII Performance Analysis This section introduces SPOT test-bed, discusses the experimental results, presented in Table III, and demonstrates the usability of the proposed construction for real world scenarios. ### VII-A SPOT Test-bed For our experiments, we developed a prototype of the SPOT protocol that implements the three phases Sys_Init, Generation and Verification including the twelve algorithms121212The source code is available at https://github.com/privteam/SPOT. The tests are made on an Ubuntu $18.04.3$ machine - with an _Intel_ Core<EMAIL_ADDRESS>\- 4 cores processor and $8GB$ memory. The twelve algorithms were implemented based on JAVA version $11$, and the cryptographic library JPBC131313http://gas.dia.unisa.it/projects/jpbc/. We evaluate the computation time of each algorithm relying on two types of bilinear pairings, i.e., type A and type F. The pairing type A is the fastest symmetric pairing type in the JPBC library constructed on the curve $y^{2}=x^{3}+x$ with an embedding degree equal to 2. The pairing type F is an asymmetric pairing type introduced by Barreto and Naehrig [2]. It has an embedding degree equal to 12. For the two types of pairing, we consider two different levels of security i.e., 112-bits and 128-bits security levels recommended by the US National Institute of Standards and Technology141414http://keylength.com (NIST). Based on the selected cryptographic library and the implementation of Groth- Sahai proofs151515https://github.com/gijsvl/groth-sahai, the SPOT test-bed is built with six main java classes, w.r.t. to the different entities of SPOT, referred to as _TrustedAuthority.java_ , _GroupManager.java_ , _Proxy.java_ , _HealthAuthority.java_ , _User.java_ and _Server.java_. Each class encompasses the algorithms that are performed by the relevant entity as described in Section III-A. In order to obtain accurate measurements of the computation time, each algorithm is run 100 times. Thus, the computation times represent the mean of the 100 runs while considering a standard deviation of an order $10^{-2}$. ### VII-B Communication and Computation Performances of SPOT This section first proposes a theoretical analysis of the communication cost. Then, it presents the experimental results of the implementation of SPOT algorithms. TABLE III: Computation time and communication overhead of SPOT algorithms Algorithm | Entity | Synch/Asynch | Communication cost | Computation time (ms) ---|---|---|---|--- A/112-bits | A/128-bits | F/112-bits | F/128-bits $\mathsf{Set\\_params}$ | $\mathcal{TA}$ | Asynch. | $|\mathbb{Z}_{n}|+|\mathbb{G}_{1}|+|\mathbb{G}_{2}|+|\mathbb{G}_{3}|$ | 874 | 2521 | 1230 | 1364 $\mathsf{HA\\_Keygen}$ | $\mathcal{TA}$ | Asynch. | $|\mathbb{G}_{1}|$ | 59 | 123 | 12 | 16 $\mathsf{S\\_Keygen}$ | $\mathcal{TA}$ | Asynch. | $2|\mathbb{G}_{2}|$ | 119 | 244 | 24 | 31 $\mathsf{Setup\\_ProxyGr}$ | $\mathcal{GM}$ | Asynch. | $21(|\mathbb{G}_{1}|+|\mathbb{G}_{2}|)$ | 1955 | 4075 | 346 | 451 $\mathsf{Join\\_ProxyGr}$ | $\mathcal{P}$/$\mathcal{GM}$ | Synch. | $\mathcal{P}:8|\mathbb{G}_{1}|+2|\mathbb{G}_{2}|$ / $\mathcal{GM}:7(|\mathbb{G}_{1}|+|\mathbb{G}_{2}|)$ | 2861 | 6014 | 1159 | 1409 $\mathsf{Set\\_UserID}$ | $\mathcal{HA}$ | Synch. | $|\mathbb{G}_{2}|$ | 58 | 121 | 12 | 16 $\mathsf{Userkeygen}$ | $\mathcal{U}$ | Asynch. | $|\mathbb{G}_{2}|$ | 117 | 242 | 24 | 31 $\mathsf{Set\\_CCM}$ | $\mathcal{U}$ | Synch. | $|\mathbb{Z}_{n}|$ | 0.1 | 0.1 | 0.1 | 0.1 $\mathsf{S\\_PSign}$ a | $\mathcal{S}$ | Synch. | $|\mathbb{Z}_{n}|$ | 0.1 | 0.08 | 0.02 | 0.02 $\mathsf{P\\_Sign}$ a | $\mathcal{P}$ | Synch. | $6|\mathbb{G}_{1}|+7|\mathbb{G}_{2}|$ | 19353 | 40371 | 3164 | 4170 $\mathsf{Sig\\_Verify}$ a | $\mathcal{HA}$ | Asynch. | N.A. | 6541 | 15406 | 31637 | 36892 $\mathsf{CCM\\_Verify}$ a | $\mathcal{HA}$ | Asynch. | N.A. | 174 | 360 | 148 | 190 NOTE: Synch./Asynch. indicates whether the algorithm must be run online (i.e., in real time) or offline (i.e., later); a indicates that the algorithm is performed on a single contact message that is generated by the $\mathsf{Set\\_CCM}$ algorithm; $|\mathbb{G}_{1}|$ (resp. $|\mathbb{G}_{2}|$, $|\mathbb{G}_{3}|$ and $|\mathbb{Z}_{n}|$) indicates the size of an element in ${\mathbb{G}_{1}}$ (resp. ${\mathbb{G}_{2}}$, ${\mathbb{G}_{3}}$ and ${\mathbb{Z}_{n}}$); N.A. is the abbreviation for Not Applicable. As presented in Table III, the communication cost is measured according to the size of the elements ${\mathbb{G}_{1}}$, ${\mathbb{G}_{2}}$, ${\mathbb{G}_{3}}$ and ${\mathbb{Z}_{n}}$ exchanged between entities. $\mathsf{Setup\\_ProxyGr}$ and $\mathsf{Join\\_ProxyGr}$ are the most bandwidth consuming algorithms, however this result must be put into perspective as both algorithms are performed once. Other algorithms have acceptable communication overhead, in particular those performed repeatedly by the user, which proves the efficiency of SPOT. From Table III, it is worth noticing that the computation time depends on the selected pairings types and is strongly related to the security level. Some algorithms of the Sys_Init phase are consuming but they are limited to only one execution from a powerful trusted authority. For the Generation phase, the most consuming algorithm is $\mathsf{P\\_Sign}$ which requires 19 seconds (resp. 40 seconds) for pairing type A and 3 seconds (resp. 4 seconds) for pairing type F. The computation time of $\mathsf{Set\\_CCM}$ and $\mathsf{S\\_PSign}$ are negligible, which means that the user and the server are not required to have important computation capacities. Finally, to verify the correctness of a single contact message, the Verification phase requires approximately 7 seconds (resp. 15 seconds) for pairing type A and 32 seconds (resp. 37 seconds) for pairing type F. It is clear that $\mathsf{P\\_Sign}$ and $\mathsf{Sig\\_Verify}$ are the most consuming algorithms in terms of computation time as they include a large number of exponentiations and pairing functions, however this result must be put into perspective as both the proxy and the health authority are assumed to have advanced hardware features. Table III also shows that the $\mathsf{Sig\\_Verify}$ algorithm run with pairing type A is faster than with pairing type F, as this latter requires excessive memory allocation and deallocation. The $\mathsf{P\\_Sign}$ algorithm has an opposite behavior where the execution with pairing type F is faster than pairing type A, which is compliant to the JPBC library benchmark161616http://gas.dia.unisa.it/projects/jpbc/benchmark.html showing that elementary functions of multiplication and exponentiation require less computation time for pairing type F. From Table III, we can deduce that the algorithms executed at the user’s side, have very low computation and communication overhead, which confirms the usability of SPOT, even when being run on a smartphone with low capacities. For both consuming algorithms ($\mathsf{P\\_Sign}$ and $\mathsf{Sig\\_Verify}$) that are repeatedly run, some performance improvement means are proposed in the next subsection. ### VII-C Improved Performances with Multithreading and Preprocessing For both computation consuming algorithms $\mathsf{P\\_Sign}$ and $\mathsf{Sig\\_Verify}$ (see Section VII-B), in an effort to make the computation time as efficient as possible, although they run on powerful SPOT entities, we rely on a two step improvement: * • Multithreading: applied to both $\mathsf{P\\_Sign}$ and $\mathsf{Sig\\_Verify}$ algorithms. It enables simultaneous multiple threads execution (e.g., the computation of the different parts of the NIWI proof for the $\mathsf{P\\_Sign}$ algorithm, the computation of either the different verification equations of the NIWI proof, or the two sides of each equation, for the $\mathsf{Sig\\_Verify}$ algorithm). * • Preprocessing: applied only to $\mathsf{Sig\\_Verify}$ algorithm. It enables to prepare in advance a value to be later paired several times, like the variables $\mathcal{U}$ and $\mathcal{V}$ which are used as input to pairing functions for each verification equation. (a) Influence of multithreading on $\mathsf{P\\_Sign}$ algorithm (b) Influence of preprocessing or/and multithreading on $\mathsf{Sig\\_Verify}$ algorithm Figure 3: Influence of improvements on $\mathsf{P\\_Sign}$ and $\mathsf{Sig\\_Verify}$ algorithms Figure 3 exposes the impacts of one or two combined improvements applied to $\mathsf{P\\_Sign}$ and $\mathsf{Sig\\_Verify}$. From Figure 3(a), we notice that multithreading reduces the computation time for $\mathsf{P\\_Sign}$ of approximately $35\%$, for the two types of pairing and the two levels of security. For $\mathsf{Sig\\_Verify}$, Figure 3(b) shows that multithreading has a greater impact on the computation time (i.e., approximately $40\%$ for pairing type A and $28\%$ for pairing type F) than preprocessing (i.e., approximately $10\%$ for pairing type A and $5\%$ for pairing type F 171717For type F - 128 bits, the preprocessing decreases the performances. This is due to the excessive memory allocation and deallocation required by the pairing type F.). The two combined improvements ensure a gain of almost $50\%$ for pairing type A and $30\%$ for pairing type F. ## VIII Conclusion In this paper, a novel secure and privacy-preserving proximity-based SPOT protocol for e-healthcare systems is introduced. The objective of SPOT is to help governments and healthcare systems to deal with pandemics by automating the process of contact tracing, with security guarantees against fake contacts injection and privacy preservation for users. Thanks to the underlying network architecture relying on a centralized computing server and decentralized proxies, SPOT enables users to determine whether they were in close proximity with infected people, with no risk of false positive alerts. The strength of the paper is to provide a full concrete construction of SPOT which is proven to be secure and to support several privacy properties under standard assumptions. Another strength of the contribution is a PoC of SPOT including a full implementation of the different algorithms, where practical computation costs measurements demonstrate the feasibility of our proposed protocol. Further research will consider aggregating the verification of multiple contact messages in an effort to improve verification performances. ## References * [1] Ca notify. https://canotify.uchealth.edu/ucberkeley/. * [2] Jpbc library: Bilinear pairing parameters generators. http://gas.dia.unisa.it/projects/jpbc/docs/ecpg.html. * [3] Stop covid-19 app. https://github.com/Stop-COVID-19-Croatia/stopcovid19-docs, [Online accessed June 2022]. * [4] Masayuki Abe, Kristiyan Haralambiev, and Miyako Ohkubo. Signing on elements in bilinear groups for modular protocol design. IACR Cryptology ePrint Archive, 2010. * [5] Apple Inc. and Google Inc. Contact tracing, bluetooth specification. https://www.blog.google/documents/58/Contact_Tracing_-_Bluetooth_, 2020\. * [6] Australian Government Department of Health. Covidsafe app. https://www.health.gov.au/resources/apps-and-tools/covidsafe-app, 2020. * [7] Troncoso Carmela, Payer Mathias, Hubaux Jean-Pierre, Salathé Marcel, Larus James, Bugnion Edouard, Lueks Wouter, Stadler Theresa, Pyrgelis Apostolos, Antonioli Daniele, Barman Ludovic, Chatel Sylvain, Paterson Kenneth, Capkun Srdjan, Basin David, Beutel Jan, Jackson Dennis, Roeschlin Marc, Leu Patrick, Preneel Bart, Smart Nigel, Abidin Aysajan, and et al. Decentralized privacy-preserving proximity tracing. https://github.com/DP-3T/documents/blob/master/DP3T%20White%20Paper.pdf, 2020\. * [8] Justin Chan, D. Foster, Shyam Gollakota, E. Horvitz, J. Jaeger, Sham M. Kakade, T. Kohno, J. Langford, J. Larson, S. Singanamalla, J. Sunshine, and S. Tessaro. Pact: Privacy-sensitive protocols and mechanisms for mobile contact tracing. 2020\. * [9] Castelluccia Claude, Bielova Nataliia, Boutet Antoine, Cunche Mathieu, Lauradoux Cedric, Le Metayer Daniel, and Roca Vincent. Desire: A third way for a european exposure notification system. https://github.com/3rd-ways-for-EU-exposure-notification/project-DESIRE/blob/master/DESIRE-specification-EN-v1_0.pdf, 2020\. * [10] Jens Groth and Amit Sahai. Efficient non-interactive proof systems for bilinear groups. In Nigel Smart, editor, Advances in Cryptology – EUROCRYPT 2008, pages 415–432, Berlin, Heidelberg, 2008. Springer Berlin Heidelberg. * [11] Asghar Hassan, Farokhi Farhad, Kaafar Dali, and Rubinstein Ben. On the privacy of tracetogether, the singaporean covid-19 contact tracing app and recommendations for australia, 2020. * [12] Jaap-Henk Hoepman. Hansel and gretel and the virus: Privacy conscious contact tracing. arXiv preprint arXiv:2101.03241, 2021. * [13] Inria and Fraunhofer AISEC. Robert: Robust and privacy-preserving proximity tracing. https://github.com/ROBERT-proximity-tracing/documents/blob/master/ROBERT-specification-EN-v1_1.pdf, 2020, [Online accessed June 2022]. * [14] K. Liu Joseph, Au Man Ho, Yuen Tsz Hon, Zuo Cong, Wang Jiawei, Sakzad Amin, Luo Xiapu, and Li Li. Privacy-preserving covid-19 contact tracing app: A zero-knowledge proof approach. IACR Cryptol. ePrint Arch, 2020, (528), 2020. * [15] Douglas J Leith and Stephen Farrell. Contact tracing app privacy: What data is shared by europe’s gaen contact tracing apps. In IEEE INFOCOM 2021, pages 1–10. IEEE, 2021. * [16] Krzysztof Pietrzak. Delayed authentication: Preventing replay and relay attacks in private contact tracing. In Progress in Cryptology – INDOCRYPT 2020, pages 3–15, Cham, 2020\. Springer International Publishing. * [17] Leonie Reichert, Samuel Brack, and BjÖRN Scheuermann. A survey of automatic contact tracing approaches using bluetooth low energy. ACM Trans. Comput. Healthcare, 2021. * [18] Rivest Ronald L., Callas Jon, Canetti Ran, Esvelt Kevin, Gillmor Daniel Kahn, Kalai Yael Tauman, Lysyanskaya Anna, Norige Adam, Raskar Ramesh, Shamir Adi, Shen Emily, Soibelman Israel, Specter Michael, Teague Vanessa, Trachtenberg Ari, Varia Mayank, Viera Marc, Weitzner Daniel, Wilkinson John, and Zissman Marc. The pact protocol specification, 2020. * [19] Viktoriia Shubina, Sylvia Holcer, Michael Gould, and Elena Simona Lohan. Survey of decentralized solutions with mobile devices for user location tracking, proximity detection, and contact tracing in the covid-19 era. Data, 5(4), 2020. * [20] Sahraoui Yesin, Ludovica Lucia, Anna Maria Vegni, Chaker Kerrache, Marica Amadeo, and Ahmed Korichi. Traceme: Real-time contact tracing and early prevention of covid-19 based on online social networks. pages 893–896, 2022. ## Appendix A Non-Interactive Witness Indistinguishable Proof In this section we represent the Groth-Sahai NIWI proof scheme applied on pairing product equations with an asymmetric bilinear map. Witness- indistinguishability implies that the verifier of a group signature is not able to find the group member that has generated the signature. The $\mathsf{NIWI}$ scheme we consider, involves four PPT algorithms ($\mathsf{NIWI}.\mathsf{Setup}$, $\mathsf{NIWI}.\mathsf{CRS}$, $\mathsf{NIWI}.\mathsf{Proof}$, $\mathsf{NIWI}.\mathsf{Verify}$): $\mathsf{NIWI}.\mathsf{Setup}$: This algorithm outputs a setup $(\mathtt{gk},\mathtt{sk})$ such that $\mathtt{gk}=(n$, $\mathbb{G}_{1}$, $\mathbb{G}_{2}$, $\mathbb{G}_{3}$, $g_{1}$, $g_{2}$, $e)$ and $\mathtt{sk}=(p,q)$ where $n=pq$. $\mathsf{NIWI}.\mathsf{CRS}$: This algorithm generates a common reference string $\mathtt{CRS}$. It takes $(\mathtt{gk},\mathtt{sk})$ as inputs and produces $\mathtt{CRS}$ = $(\mathbb{G}_{1},\mathbb{G}_{2},\mathbb{G}_{3},e,\iota_{1},p_{1},\iota_{2},p_{2},\iota_{3},p_{3},\mathcal{U},\mathcal{V})$, where $\mathcal{U}=rg_{1}$, $\mathcal{V}=sg_{2}$ ; $r,s\in\mathbb{Z}^{*}_{n}$ and $\iota_{1}$: | $\mathbb{G}_{1}\longrightarrow\mathbb{G}_{1}$ | $\iota_{2}$: | $\mathbb{G}_{2}\longrightarrow\mathbb{G}_{2}$ | $\iota_{3}$: | $\mathbb{G}_{3}\longrightarrow\mathbb{G}_{3}$ ---|---|---|---|---|--- | $x\longmapsto x$ | | $y\longmapsto y$ | | $z\longmapsto z$ $p_{1}$: | $\mathbb{G}_{1}\longrightarrow\mathbb{G}_{1}$ | $p_{2}$: | $\mathbb{G}_{2}\longrightarrow\mathbb{G}_{2}$ | $p_{3}$: | $\mathbb{G}_{3}\longrightarrow\mathbb{G}_{3}$ | $x\longmapsto\lambda x$ | | $y\longmapsto\lambda y$ | | $z\longmapsto z^{\lambda}$ $\mathsf{NIWI}.\mathsf{Proof}$: This algorithm generates a NIWI proof for satisfiability of a set of pairing product equations of the form of ${\prod}^{l}_{i=1}e(\mathcal{A}_{i},\mathcal{Y}_{i}){\prod}^{k}_{i=1}e(\mathcal{X}_{i},\mathcal{B}_{i}){\prod}^{k}_{i=1}{\prod}^{l}_{j=1}e(\mathcal{X}_{i},\mathcal{Y}_{j})^{\gamma_{ij}}=t$ also written as $(\vec{\mathcal{A}}\cdot\vec{\mathcal{Y}})(\vec{\mathcal{X}}\cdot\vec{\mathcal{B}})(\vec{\mathcal{X}}\cdot\Gamma\vec{\mathcal{Y}})=t$ It takes as input $\mathtt{gk}$, $\mathtt{CRS}$ and a list of pairing product equations $\\{(\vec{\mathcal{A}_{i}},\vec{\mathcal{B}_{i}},\Gamma_{i},t_{i})\\}^{N}_{i=1}$ and a satisfying witness $\vec{\mathcal{X}}\in\mathbb{G}_{1}^{k}$, $\vec{\mathcal{Y}}\in\mathbb{G}_{2}^{l}$. To generate a proof over a pairing product equation, the algorithm, first, picks at random $\mathcal{R}\leftarrow Vec_{k}(\mathbb{Z}_{n})$ and $\mathcal{S}\leftarrow Vec_{l}(\mathbb{Z}_{n})$, commits to all variables as $\vec{\mathcal{C}}:=\vec{\mathcal{X}}+\mathcal{R}\mathcal{U}$ and $\vec{\mathcal{D}}:=\vec{\mathcal{Y}}+\mathcal{S}\mathcal{V}$, and computes ${\pi}=\mathcal{R}^{\top}\iota_{2}(\vec{\mathcal{B}})+\mathcal{R}^{\top}\Gamma\iota_{2}(\vec{\mathcal{Y}})+\mathcal{R}^{\top}\Gamma\mathcal{S}\mathcal{V}$ ${\theta}=\mathcal{S}^{\top}\iota_{1}(\vec{\mathcal{A}})+\mathcal{S}^{\top}\Gamma^{\top}\iota_{1}(\vec{\mathcal{X}})$ The algorithm outputs the proof $(\pi,\theta)$. $\mathsf{NIWI}.\mathsf{Verify}$: This algorithm checks if the proof is valid. It takes $\mathtt{gk}$, $\mathtt{CRS}$, $\\{(\vec{\mathcal{A}_{i}},\vec{\mathcal{B}_{i}},\Gamma_{i},t_{i})\\}^{N}_{i=1}$ and $(\vec{\mathcal{C}_{i}},\vec{\mathcal{D}_{i}},\\{(\pi_{i},\theta_{i})\\}^{N}_{i=1})$ as inputs and for each equation, checks the following equation: $e(\iota_{1}(\vec{\mathcal{A}_{i}}),\vec{\mathcal{D}_{i}})e(\vec{\mathcal{C}_{i}},\iota_{2}(\vec{\mathcal{B}_{i}}))e(\vec{\mathcal{C}_{i}},\Gamma_{i}\vec{\mathcal{D}_{i}})=\iota_{3}(t_{i})e(\mathcal{U},\pi_{i})e(\theta_{i},\mathcal{V})$ (4) The algorithm outputs 1 if the equation holds, else it outputs 0.
# Tetrahedron of flavors: One Three to rule them all Aharon Davidson<EMAIL_ADDRESS>https://physics.bgu.ac.il/~davidson Physics Department, Ben-Gurion University of the Negev, Beer-Sheva 84105, Israel (October10, 2022) ###### Abstract Under the conception that the total number three of fermion families must have the one and the same gauge theoretical origin as all other threes which accompany the single family grand unifiable group structure, we trade the trinification $SU(3)$ symmetry building block for its semi-simple Vertical- Horizontal symmetric $SU(3)_{V}\otimes SU(3)_{H}$ extension. The anomaly free flavor-chiral fermion representation is then constructed, solely out of threes and without any superfluous replication, by treating each standard $V$-triplet (and anti triplet) as an $H$-singlet, and conversely, by letting each standard $V$-singlet transform as an $H$-triplet (or anti triplet). The model can be schematically described by a tetrahedron of flavors. In its bi-trinification phase, the model exhibits two coupled trinification cycles, sharing the same color group, and is furthermore accompanied by a built-in dark sector. In its isomorphic quartification phase, the model presents a novel non-Pati-Salam version of quark/lepton correspondence, where quarks are paired with anti- leptons to cancel horizontal anomalies, and all fermion masses stem from one and the same Yukawa source. ## I Introduction Decoding the mass spectrum of quarks and leptons is without any doubt the holy grail of theoretical particle physics. While the mass generating Higgs mechanism, mediated by Yukawa couplings, is rooted in the heart of the standard electro/nuclear theory, we still do not have the slightest idea what physics actually determines the eigenmasses and mixings. At this stage, even an empirical formula to account for the observed mass hierarchy is certainly welcome. The structure of the Fermi mass matrix is however just the tip of the flavor puzzle iceberg. The various fermionic pieces which make a single family (or a generation) are well organized, up to some electro/nuclear neutral members, within the framework of the by now standard $SU(3)_{C}\otimes SU(2)_{L}\otimes U(1)_{Y}$ gauge theory. Anomaly cancellation is then a necessary self consistency condition, but obviously not a sufficient one for shedding light on the full picture. The consequent grand unification procedure to combine the otherwise separated fermionic pieces into a single representation has been exclusively realized for the $SU(5)\subset SO(10)\subset E(6)$ group sequence, and is most welcome, but leaves a fundamental question unanswered. Namely, why can/must grand unification be realized (at least partially) already at the single family level, while apparently leaving the observed multiplicity of fermion families field/group theoretically out of the game. At the time, embedding the single family GUT within a larger multi family GUT seemed to be a logical step forwards. Unfortunately, such a potentially promising direction has been practically exhausted, and so far failed to deliver. On the short list of theoretical (on top of experimental) obstacles encountered one can find: (i) The magic yet theoretically challenging number of exactly _three_ standard fermion families. (ii) The restrictive chiral flavor structure required at the low energy group theoretical reduction level. (iii) The ever growing sizes (counting generators) of the candidate GUT groups involved. (iv) Heavy field theoretical artillery, such as super-symmetry and dimensional reduction, while being attractive from various other reasons, has not contributed so far its part to the flavor puzzle. The more so superstring theory, including its string phenomenology outer branch, has not been proven too useful either. It may well be that the tempting strategy of passing through the single family grand unification stage may have counter intuitively blocked us from unveiling the full flavor picture. ## II Vertical $\leftrightarrow$ Horizontal symmetry Equipped with no compelling answers, the flavor puzzle has sourced plenty of imaginative ideas, ranging from atomic style isotopes all the way to compositeness, involving a variety of field theoretical techniques. In this paper, however, with the focus solely on group theory, we humbly follow the hypothesis that the total number _three_ of fermion families shares the one and the same gauge theoretical origin as all other _threes_ which govern the single family unifiable (not grand unified) group structure. Adopting this line of thought, let our starting point be trinification 333 ; trinification , based on the semi-simple gauge group $G=SU(3)_{C}\otimes SU(3)_{L}\otimes SU(3)_{R}~{}.$ (1) An accompanying $C_{3}$ discrete symmetry is optional, and becomes mandatory if a common gauge coupling constant is in order. Associated with is the familiar anomaly free flavor chiral representation of left handed fermions $\psi_{L}=\begin{array}[]{|c ||c|c|c|}\hline\cr q&3&3^{\ast}&1\\\ \hline\cr q^{c}&3^{\ast}&1&3\\\ \hline\cr\ell+\ell^{c}&1&3&3^{\ast}\\\ \hline\cr\end{array}$ (2) which can always be supplemented, if so required, by a bunch of dark $G$-singlets $\begin{array}[]{|c ||c|c|c|}\hline\cr~{}~{}\chi{}{}&~{}1{}&~{}1{}&~{}1{}\\\ \hline\cr\end{array}$ (3) Schematically, the model can be neatly represented by a triangle, as depicted in Fig.1, where the vertices represent fermions and the edges stand for the various $SU(3)$ group factors involved. Figure 1: The trinification model: The vertices are associated with the various fermions involved., and the three edges stand for the three $SU(3)$ group factors involved. An incoming/outgoing arrow signals a triplet/anti- triplet of the respective $SU(3)$. Schematically, blue and red circles represent triplets and anti-triplets respectively Already at the trinification level, prior to $E(6)$ unification E6 , there is a heavy price (not everyone is willing) to pay, namely extending the standard single fermion family, composed of 16 members, to include extra 27-16=11 non- standard members. The latter supplement, easily classified by means of the electric charge operator $Q=T_{3L}+T_{3R}+\frac{1}{2}(Y_{L}+Y_{R})~{},$ (4) decouples however once $SU(3)_{L}\otimes SU(3)_{R}$ parent symmetry breaks down to its left-right symmetric sub-group $SU(2)_{L}\otimes SU(2)_{R}\otimes U(1)_{B-L}$. Furthermore, notice that the full quark/lepton correspondence, i.e. treating lepton number as the forth color, PS is not restored until (and if) $E(6)$ embedding is realized. Multiplying the trinification group $G$ by extra $SU(3)$ factors is then quite a conservative step forwards. $SU(3)^{N}$ models, with $N>3$, mostly single family models, have been extensively studied by Ma SUnk and by others SU34 . Horizontal symmetries horizontal , primarily invoked to classify the otherwise electro/nuclear degenerate fermionic families, start from minimal $U(1)$, local or else global (incorporating the Peccei-Quinn mechanism familon ) and in particular include an $SU(3)$ example HSU3 . Here, we revive the idea of Vertical $\leftrightarrow$ Horizontal symmetry VH , and trade the single family group, now referred to as $G_{V}$, for its $G_{V}\otimes G_{H}$ semi- simple extension. Note that models incorporating semi-simple group structure, for example Pati-Salam’s $SU(4)^{4}$ PS , first ever unification scheme (originally designed for two families), the hybrid single family left-right symmetric $SU(5)_{L}\otimes SU(5)_{R}$ model 5x5 hosting chiral color chiralcolor , and multi family models of the $SO(10)\otimes SO(10)$ type 10x10 , have already been discussed in the literature. To construct the tenable $G_{V}\otimes G_{H}$ fermion representation, we adopt the following prescription: _Treat each standard $V$-triplet (and anti-triplet) as an $H$-singlet, and conversely, let each standard $V$-singlet transform as an $H$-triplet (or anti-triplet)._ This prescription generalizes Eq.(2) into $\begin{array}[]{ccc}&~{}~{}\quad\quad V\quad\quad\quad\quad H&\\\ &\psi_{L}=\begin{array}[]{|c |c |c ||c |c |c |}\hline\cr 3&3^{\ast}&1&1&1&x\\\ \hline\cr 3^{\ast}&1&3&1&y&1\\\ \hline\cr 1&3&3^{\ast}&z&1&1\\\ \hline\cr\hline\cr 1&1&1&z^{\ast}&y^{\ast}&x^{\ast}\\\ \hline\cr\end{array}&\end{array}$ (5) The cancelation of ABJ anomalies is done pairwise, while maintaining overall flavor chirality, so that each of the $SU(3)$ representations $x,y,z$ can be either a $3$ or else a $3^{\ast}$. While the particular choice seems irrelevant from any individual $SU(3)$ point of view, a matter of definition, it does become relevant in the presence of an accompanying discrete symmetry. Altogether, with only $SU(3)$ triplets and anti-triplets at our disposal, only two independent configurations exist. The manifestly symmetric configuration $x=y=z\quad\textrm{}~{},$ (6) could have been our naive preference. But unfortunately, as demonstrated in Fig.(2), its associated discrete symmetry does not really go much beyond the original trinity model, offering the same three ways to identify $\\{q,q^{c},\ell+\ell^{c}\\}$, and this without involving $V\leftrightarrow H$ interplay, and without reviving quark/lepton correspondence. Figure 2: With blue and red circles representing triplets and anti-triplets respectively, the three trinification ways to identify $\\{q,q^{c},\ell+\ell^{c}\\}$ are illustrated by $\\{a,b,c\\}$ cyclic permutations accompanied by suitable re-arrangements of the six $SU(3)$ group factors. ## III Tetrahedron of flavors In this paper, however, from reasons to be specified, we construct the so- called bi-trinification model (along with its isomorphic quartification phase) based on the alternative configuration $x=y=z^{\ast}\quad\textrm{up to permutations}~{}.$ (7) It offers six variant ways to identify $\\{q,q^{c},\ell+\ell^{c}\\}$, each of which comes with its own (practically equivalent) permuted and/or conjugate horizontal assignments. To fully appreciate this point, as demonstrated in Fig.3, one may rearrange raws as well as columns, identify an alternative trinification V-sector (three left columns), and then verify how one H-sector (three right columns) gets systematically replaced by an equivalent one, such that $(x,y,z)\rightarrow(x^{\prime},y^{\prime},z^{\prime})$, where the latter is a permuted version of the former, with or without conjugation. Figure 3: Horizontal bi-trinification: By rearranging raws (fermion assignments) and columns ($SU(3)$) group factors), while recasting the exact structure of the trinification V-sector (marked with lines), one can easily switch from one H-sector (three right columns) to another. The six variants are thus equivalent to each other. Note that triplets and anti-triplets are represented here by blue and red circles, respectively. Figure 4: The Tetrahedron model (Bi-Trinification phases): The two coupled 3-cycles (red arrows), namely $231$ (left plot) and $246$ (right plot), describe two different trinification V-sectors. Attached to each one of them is a corresponding H-sector (blue arrows). In this phase, $\ell$ and $\ell^{c}$ conventionally share a common leptonic vertex, and associated with the fourth vertex $\chi$ is a novel dark (= electro/nuclear neutral) sector. The fermion representation furnishes a tetrahedron. While the edges, numbered $k=1,2,...,6$, correspond to the individual $SU(3)_{k}$ group factors, the vertices are associated with the four fermion classes $q,q^{c},\ell+\ell^{c},\chi$. The various arrows involved, making the edges directed, stand for triplets (outgoing arrows) and anti-triplets (ingoing arrows). Notice that there are in fact two different ways to consistently identify the V-sector. Sticking to Fig.(3), it is either the 231-cycle (three upper configurations) or alternatively the 246-cycle (three lower configurations). It is this characteristic feature which justifies using the terminology bi-trinification. The two cycles correspond to the trinifications $SU(3)_{C}\otimes\left[SU(3)_{L}\otimes SU(3)_{R}\right]_{V,H}$, respectively. This in turn gives rise to two electric charges $Q_{V,H}=\left[T_{3L}+T_{3R}+\frac{1}{2}Y_{L+R}\right]_{V,H}~{},$ (8) one of which, that is $Q_{H}$, must be spontaneously broken, thereby opening the door for the Holdom effect Holdom , also known as photon mixing. The $\chi$-fermions are required on anomaly cancelation grounds. They are $V$-sector singlets by construction, and as such, have no standard electro/nuclear interactions. This makes them, by definition, candidates for dark matter particles of the WIMP kind. They do interact with ordinary matter though, with the $H$-sector serving as the tenable portal. The mechanism involved, as demonstrated in Fig.(5) for the special lepto/dark case, is the exchange of super heavy horizontal gauge bosons. Figure 5: Dark matter portal: Dark matter particles ($V$-sector singlets) interact with standard model particles ($V$-sector non singlets) by exchanging heavy horizontal gauge bosons. Depicted in this figure is the lepto/dark case. ## IV Bi-Tri/Quart isomorphism Alternative rearrangements of raws and columns can make the single family group flavor assignments depart from their built-in trinification (and thus bi-trinification) construction, and at the expense of a smaller H-sector, extend the V-sectior into a quartification phase. Depicted in Fig.(6) are the four associated quartification variant configurations. Indeed, they share a common extended V-sector (left four columns connected with lines) and admit shrank (only two, rather than three, right columns) horizontal structures. They are equivalent to each other. And most importantly, the existence of these four variants serves to establish that, as far as our model is concerned, bi-trinification and quartification are two faces of the same coin. Figure 6: Horizontal quartification: Alternative rearrangements of particle assignments (raws) and $SU(3)$ group factors (columns) depart from trinification, and extend the V-sectior (marked with lines) into quartification at the expense of a smaller (two right columns) H-sector. The equivalence of these four variants to each other is interpreted as a novel version of quark/lepton correspondence. As before, triplets and anti-triplets are represented by blue and red circles, respectively. Back to the tetrahedron of flavors, the quartification phase calls for some reshuffling of fermion assignments. To be specific, the leptonic sub- representation bifurcates, such that $\ell$ and $\ell^{c}$ belong now to two different vertices. To classify the fermions under the extended single family group $SU(3)^{4}$, with the individual subscripts being $C,L,R,N$ (with $N$ denoting the new comer), we introduce the fully symmetric electric charge formula $Q=T_{3L}+T_{3R}+T_{3N}+\frac{1}{2}\left(Y_{L}+Y_{R}+Y_{N}\right)~{}.$ (9) Apart from the manifest $LRN$ symmetry, it has the advantage that all individual electric charges involved are standard. The electric charge $Q$, which generalizes Eq.(4), should be contrasted with Eq.(8), and with a different electric charge formula recently advocated for a similar $SU(3)^{4}$ model. To be more explicit, and expose the extra nine leptons per family, introduced on top of the trinification scheme, we piecewise specify the $SU(3)_{C}\otimes SU(2)_{L}\otimes SU(2)_{R}\otimes U(1)_{B-L}$ content of the fermion representation, namely $\begin{array}[]{ccl}(3,3^{\ast},1,1)&=&(3,2,1)_{\frac{1}{3}}+(3,1,1)_{-\frac{2}{3}}~{},\\\ (3^{\ast},1,3,1)&=&(3^{\ast},1,2)_{-\frac{1}{3}}+(3^{\ast},1,1)_{\frac{2}{3}}~{},\\\ (1,3,1,3^{\ast})&=&(1,2,1)_{-1,1}+(1,2,1)_{-1}~{}+\\\ &+&(1,1,1)_{0,2}+(1,1,1)_{0}~{},\\\ (1,1,3^{\ast},3)&=&(1,1,2)_{-1,1}+(1,1,2)_{1}~{}+\\\ &+&(1,1,1)_{-2,0}+(1,1,1)_{0}~{}.\end{array}$ (10) Of special interest are those vector-like fermions which carry non-standard $B-L$ charges. From a group theory point of view, they are primarily responsible for the small value $\sin^{2}\theta_{W}\rightarrow\frac{\sum T_{L3}^{2}}{\sum Q^{2}}=\frac{1}{4}$ (11) of the Weinberg angle at the symmetry limit, which is smaller than the conventional value of $\frac{3}{8}$. However, the vector-like surplus of quarks and leptons is not protected by the $SU(3)_{C}\otimes SU(2)_{L}\otimes SU(2)_{R}\otimes U(1)_{B-L}$ symmetry subgroup, and decouples at some presumably very heavy mass scale. Under the latter residual symmetry, we are left with the exact flavor chiral family of the Left-Right symmetric model, that is $\psi=(3,2,1)_{\frac{1}{3}}+(3^{\ast},1,2)_{-\frac{1}{3}}+(1,2,1)_{-1}+(1,1,2)_{1}~{}.$ (12) Figure 7: The Tetrahedron model (quartification phase): The focus now is on the 4-cycle $2451$ (red arrows) which describes a single family $SU(3)^{4}$ fermionic representation. It accounts for the revival of a full $q,q^{c}\leftrightarrow\ell,\ell^{c}$ quark/lepton correspondence, as expressed by the four variants shown in Fig.(6). The blue arrows account for the threefold family structure under horizontal $SU(3)_{I}\otimes SU(3)_{II}$. ## V Quark/Lepton correspondence resurrection $B-L$ does not serve in our model as the forth color. The reason is obvious: There is no Pati-Salam $SU(4)$ to the rescue. Quark/lepton correspondence is absent from the trinification model as well. Unlike $q$ and $q^{c}$ which are assigned to two different vertices, see Fig.(1), $\ell$ and $\ell^{c}$ share a common third vertex. It is only in Ma’s $SU(3)^{4}$ model that quark/lepton correspondence has been revived, at least in the sense that $\ell$ and $\ell^{c}$ split vertices, see Fig.(7). But there is more to it. The H-sector is primarily in charge of the threefold family replication. Denoting the quartification horizontal group by $SU(3)_{I}\otimes SU(3)_{II}$, and choosing for definiteness one of the four variants depicted in Fig.(6), we face the horizontal assignments $q\sim(1,3)~{},~{}q^{c}\sim(3^{\ast},1)~{},~{}\ell\sim(3,1)~{},~{}\ell^{c}\sim(1,3^{\ast})~{}.$ (13) It is the pair $\\{q^{c},\ell\\}$ which cancels the $SU(3)_{I}$ anomalies, while $\\{q,\ell^{c}\\}$ takes care of the $SU(3)_{II}$ anomalies. Such an unprecedented horizontal pairing, which seems as a unified generalization of Foot-Lew quark/lepton symmetry FL , is fully dictated by the group theoretical structure of our tetrahedron model, and gives a novel perspective to the inter relations between quarks and leptons. Another attractive aspect of quartification has to do with the Higgs sector. The standard model has already taught us that quarks and leptons alike should acquire their masses via Yukawa couplings with the one and the same Higgs scalar. One complex doublet suffices to govern all masses (and may even be the only source of spontaneous symmetry breaking). Truly, this feature is forcefully shared by the original $SU(3)^{3}$ trinification model as well. Starting from the fermionic representation Eq.(2), the mass generating job is carried out by the scalar $\phi$ which exhibits the fermion bilinear quantum numbers $\phi\sim qq^{c}\sim\ell\ell^{c}\sim(1,3^{\ast},3)~{}.$ (14) This is strikingly not the case for the tetrahedron model in the bi- trinification phase, for which $\begin{array}[]{rcc}&qq^{c}\sim(1,3^{\ast},3||1,3,3^{\ast})&\\\ &\ell\ell^{c}\sim(1,3^{\ast},3||3^{\ast},1,1)&\end{array}$ (15) The more so, once $\chi$ has entered the game, we cannot avoid having $(\ell+\ell^{c})\chi\sim(1,3,3^{\ast}||1,3^{\ast},3)$ as well. This seems to suggest that $\ell$ and $\ell^{c}$ better not share a common tetrahedron vertex. Indeed, one is back on safe (and quite attractive) grounds provided the tetrahedron model in adopted in its the quartification phase. Choosing for the sake of definiteness a particular horizontal variant, say $\begin{array}[]{|c ||c|c|c|c||c|c|}\hline\cr q&3&3^{\ast}&1&1&1&3\\\ \hline\cr q^{c}&3^{\ast}&1&3&1&3^{\ast}&1\\\ \hline\cr\ell&1&3&1&3^{\ast}&3&1\\\ \hline\cr\ell^{c}&1&1&3^{\ast}&3&1&3^{\ast}\\\ \hline\cr\end{array}~{},$ (16) quartification naturally offers the conjugate fermion bilinear quantum numbers $\begin{array}[]{rcc}&qq^{c}\sim\phi\sim(1,3^{\ast},3,1||3^{\ast},3)&\\\ &\ell\ell^{c}\sim\phi^{\ast}\sim(1,3,3^{\ast},1||3,3^{\ast})&\end{array}$ (17) In turn, as dictated by their revived correspondence, quarks and leptons do share now a common Yukawa coupling origin. The associated VEV pattern needs not be simple, to say the least, reflecting the double spinorial structure (expressed by $i,j=1,2,3$) of $\langle\phi\rangle=v_{L,R}^{ij}$ under $SU(3)_{I,II}$ of each of the two (LR symmetric) Weinberg-Salam doublets. To add a complication to the list, notice that generically the two VEV matrices $v_{L}^{ij}$ and $v_{R}^{ij}$ cannot be diagonalized simultaneously. This is apparently the reason underlying the fermion mixing phenomenon, but unfortunately not even a tentative mixing formula can be derived at this preliminary stage. It is highly suggestive that $SU(3)_{I}\otimes SU(3)_{II}$ gets broken at a certain stage down to $SU(3)_{I+II}$, under which $\phi$ transforms as a singlet, with all horizontal fine details ripped off. ## VI Epilogue In this paper, unfortunately and admittedly, we have offered no detailed insight into the structure of the Fermi mass matrix. While this is undoubtedly a drawback, it should be gracefully criticized, recalling that after forty years of flavor puzzle frustration, and after so many imaginative yet unripe theoretical ideas per year, no one can so far claim even the slightest of victories. The major clues, so we would like to believe, rowing against the conventional stream, are already with us. Namely, the challenging total number three of fermion families, and the striking failure of all grand unification trails (string theory and extra dimensions included) to account for it. The naive yet powerful idea that all threes which accompany flavor physics must have the one and the same gauge theoretical origin then simply paves the way for Vertical-Horizontal symmetric trinification. We can only hope that the resulting hereby presented tetrahedron of flavors, stemming from trinification but eventually focused on its quartification (rather than bi-trinification) phase, would shed some light on the quark/lepton correspondence and elegantly account for the threefold family replication, at least at the group theoretical level. ###### Acknowledgements. ## References * (1) A. de Rujula, H. Georgi and S.L. Glashow, in Fifth Workshop on Grand Unification, edited by K. Kang, H. Fried and P. Frampton, World Scientific (1984). * (2) K.S. Babu, X.G. He and S. Pakvasa, Phys. Rev. D33, 763 (1986); J.E. Kim, Phys. Lett. B591, 119 (2004); J. Sayre, S. Wiesenfeldt and S. Willenbrock, Phys. Rev. D73, 035013 (2006); J. Hetzel and B. Stech, Phys. Rev. D91, 055026 (2015); G.M. Pelaggi, A. Strumia and S. Vignali, JHEP 08, 130 (2015); K.S. Babu, B. Bajc, M. Nemevšek and Z. Tavartkiladze, AIP Conf. Proc. 1900, 020002 (2017); Z.W. Wang, A. Al Balushi, R. Mann and H.M. Jiang, Phys. Rev. D99, 115017 (2019); K.S. Babu, S. Jana and A. Thapa, JHEP 02, 051 (2022). * (3) F. Gursey, P. Ramond and P. Sikivie, Phys. Lett. 60B, 177 (1976); Y. Achiman and B. Stech, Phys. Lett. 77B, 389 (1978); Q. Shafi, Phys. Lett. B79, 301 (1978); R. Barbieri and D.V. Nanopoulos, Phys. Lett. 91B, 369 (1980). * (4) J.C. Pati and A. Salam, Phys. Rev. D10, 275 (1974). * (5) E. Ma, Phys. Lett. B593, 198 (2004); E. Ma, Mod. Phys. Lett. A20, 1953 (2005); E. Ma, M. Mondragon and G. Zoupanos, JHEP 12, 026 (2005); Q.H. Cao, S.L. Chen and E. Ma, Phys. Rev. D73, 015009 (2006). * (6) K.S. Babu, 1, Thomas, W. Kephart and H. Pas, Phys. Rev. D77,116006 (2008); T. Banks and J.F. Fortin, JHEP 07, 046 (2009); C. Kownacki, E. Ma, N. Pollard, O. Popov and M. Zakeri, Phys. Lett. B769, 267 (2017); ibid Nucl. Phys. B928, 520 (2018); ibid Phys. Lett. B777, 121 (2018); E. Ma, Phys. Lett. B820, 136561 (2021). * (7) F. Wilczek and A. Zee, Phys. Rev. Lett. 42, 421 (1979); A. Davidson, Phys. Rev. Lett. 43, 92 (1979). * (8) A. Davidson and K.C.Wali, Phys. Rev. Lett. 48, 11 (1982); F. Wilczek and A. Zee, Phys. Rev. Lett., 49:1549 (1982). * (9) T. Yanagida, Phys. Rev. D20, 2986 (1979); Z.G. Berezhiani, Phys. Lett. B150, 177 (1985). * (10) A. Davidson and K.C. Wali, Phys. Lett. B94, 359 (1980). * (11) A. Davidson and K.C. Wali, Phys. Rev. Lett. 58, 2623 (1987). * (12) P.H. Frampton and S.L. Glashow, Phys. Lett. B190, 157(1987); A. Davidson and K.C. Wali, Phys. Lett. B194, 247 (1987). * (13) A. Davidson, P.D. Mannheim and K.C. Wali, Phys. Rev. Lett. 47, 149 (1981). * (14) B. Holdom, Phys. Lett. 166B, 196 (1986). * (15) R. Foot, and H. Lew, Phys. Rev. D41,3502 (1990); R. Foot, H. Lew and R.R. Volkas, Phys. Rev. D44, 1531 (1991).
Date: ] # Kaon condensation in skyrmion matter and compact stars Christoph Adam<EMAIL_ADDRESS>Departamento de Física de Partículas, Universidad de Santiago de Compostela Instituto Galego de Física de Altas Enerxias (IGFAE) E-15782 Santiago de Compostela, Spain Alberto García Martín- Caro<EMAIL_ADDRESS>Instituto Galego de Física de Altas Enerxias (IGFAE) E-15782 Santiago de Compostela, Spain Physics Dept., Brookhaven National Laboratory, Bldg. 510A, Upton, NY 11973, USA Miguel Huidobro <EMAIL_ADDRESS>Ricardo Vázquez<EMAIL_ADDRESS>Departamento de Física de Partículas, Universidad de Santiago de Compostela Instituto Galego de Física de Altas Enerxias (IGFAE) E-15782 Santiago de Compostela, Spain Andrzej Wereszczynski<EMAIL_ADDRESS>Institute of Physics, Jagiellonian University, Lojasiewicza 11, Kraków, Poland ([) ###### Abstract We address the possibility of the appearance of a charged kaon condensate in neutron star cores described within a generalized Skyrme model. Our treatment of strange degrees of freedom is based on the Bound State Approach by Callan and Klebanov, which allows to obtain an in-medium effective potential for the $s$-wave kaon condensate. We predict the onset of kaon condensation at a certain threshold density—whose value depends on the parameters of the model, and ranges between $1.5$ and $2.5$ times saturation density—, and obtain both the particle fractions and equation of state for dense matter in the kaon condensed phase. Finally, we discuss the effect of such condensates on the mass-radius curves and other observable properties of neutron stars with kaon condensed cores. ###### Contents 1. I Introduction 2. II Generalized Skyrme model and skyrmion crystals 3. III Kaon condensate in skyrmion crystals 1. III.1 Kaon fluctuations in the Skyrme model 2. III.2 The kaon condensate on classical crystals 1. III.3 Effect of kaon condensation on the quantum corrections to Skyrme crystals 1. IV Results 1. V Neutron stars with kaon condensed cores within the Skyrme model. 1. V.1 The Skyrme crystal EoS for (a-)symmetric nuclear matter. 2. V.2 Maxwell construction versus Gibbs construction 3. V.3 The TOV system and NS properties 1. VI Conclusions 1. A Derivation of the WZW and sextic terms contribution to $V_{K}$ 2. B Explicit expressions ## I Introduction While heavy ion collision experiments and lattice QCD simulations provide insight into the properties of hot and dense QCD, neutron stars (NS) are the only known objects in the universe that may allow us to deepen our understanding of the rich structure of cold, ultra-dense strongly interacting nuclear matter. The astrophysical inference of NS masses, radii and moments of inertia in low mass x-ray binaries and isolated NS from the NICER experiment as well as in binary NS inspirals and subsequent mergers from Gravitational- Wave (GW) observatories have helped to significantly constrain the NS equation of state (EoS) at supra-saturation densities, which cannot be reached in laboratory experiments. These constraints, however, do not yield any information on the microscopic structure of dense matter. Indeed, owing to the non-perturbative nature of QCD at energies below the confinement scale, the precise phase structure of cold, strongly interacting matter at both finite baryon and isospin chemical potential is still very speculative. Novel phases of dense baryonic matter are expected to occur in the inner core of NS, containing additional particle species such as $\Delta$ isobar resonances Li _et al._ (2018), hyperons Glendenning (1985), or pion or kaon condensates Hartle _et al._ (1975); Celenza _et al._ (1977); Kaplan and Nelson (1988); Glendenning and Schaffner- Bielich (1999); Pal _et al._ (2000). There have been theoretical proposals of even more exotic scenarios, where a transition to deconfined quark matter takes place inside the core Heiselberg _et al._ (1993); Benvenuto and Lugones (1999); Annala _et al._ (2020), or a new state of matter in which both hadronic and quark degrees of freedom coexist, the so-called quarkionic matter McLerran and Pisarski (2007); McLerran and Reddy (2019). Studies of the dynamical features of compact stars, such as the occurrence of phase transitions during mergers or the cooling rate of proto-neutron stars may produce complementary data, as they may strongly depend on the specific microscopic degrees of freedom as well as on the EoS. Many calculations for the EoS in dense matter predict that strangeness degrees of freedom may become important in the interior of compact stars, in the form of hyperons (strange baryons) or a Bose-Einstein condensate of negatively charged kaons, for densities just a few times nuclear saturation. For a recent review, see Tolos and Fabbietti (2020). Indeed, hyperons may become stable at sufficiently high isospin chemical potential, where the decay of neutrons relieve the Fermi pressure exerted by the nucleons. On the other hand, the strong attraction between $K^{-}$ mesons and baryons increases with density and lowers the energy of the zero momentum state. A condensate is formed when this energy equals the kaon chemical potential, since kaons are favored over negatively charged fermions for achieving charge neutrality, as they are bosons and can condense in the lowest energy state. It is generally assumed that hyperons should appear at densities above $\sim 2-3$ times the nuclear saturation density $n_{0}$, whereas the critical density for kaon condensation is usually predicted to be a bit larger, around $\sim(3-4)n_{0}$ (although the specific values are of course model and parameter dependent). A density of this order is smaller than the central density of a typical NS, so a kaon condensate could be present in its core. The possibility of kaon condensates in the core of neutron stars has been extensively investigated in the literature, using different approaches. Its appearance tends to soften the EoS, producing smaller values for the allowed maximum masses. Therefore, the presence of hyperons at too low densities is not compatible with the stiffness required by the existence of such massive stars. This is the so-called _Hyperon puzzle_ , presently a subject of very active research Bombaci (2017); Vidaña (2016). In this work we will make use of a Generalized Skyrme model, a phenomenological, nonlinear chiral model that, due to its nonperturbative nature, can in principle be used as a simple model to study strongly interacting matter at all scales, from single baryons and nuclei to nuclear matter in neutron stars. The model contains a rather limited number of fundamental degrees of freedom, which in the simplest version are just pions, encoded into an $SU(2)$ valued field $U$, as chiral symmetry is nonlinearly realized. Nucleons and atomic nuclei emerge as collective, topologically nontrivial excitations of the mesonic fields. Mathematically they are described by topological solitons, called skyrmions, whose topological degree can be identified with the baryon number. The most attractive feature of the model is the small number of free parameters, which implies a rather strong predictive power. The Skyrme model Skyrme (1962) and its generalizations have been applied successfully to the description of nucleon properties Adkins _et al._ (1983), Ding and Yan (2007), nuclear interaction potentials Halcrow and Harland (2020), ground and excited states of atomic nuclei Lau and Manton (2014), Halcrow _et al._ (2017), or the problem of nuclear binding energies Adam _et al._ (2013); Gillard _et al._ (2015); Gudnason (2016). Typically, the best fit to phenomenological observations requires the extension of the original Skyrme Lagrangian Skyrme (1962), either by the addition of new degrees of freedom, e.g., vector mesons Naya and Sutcliffe (2018), or additional, physically-motivated higher derivative terms, like the so-called sextic term Adam _et al._ (2010), an effective term related to two-body interactions mediated by $\omega$ mesons. Simultaneously, in the last years there has been a significant progress in the application of the Skyrme model to investigate properties of dense nuclear matter and neutron stars Naya (2019); Adam _et al._ (2020, 2022a, 2022b), where the sextic term is especially important, as this part of the action provides the leading contribution in the regime of high pressure and density Adam _et al._ (2015). Indeed, it makes the skyrmionic matter much stiffer at extreme conditions which results in physically acceptable values of the maximal mass of neutron stars. The extension of the Skyrme model for a larger number of flavors has also been discussed in the literature. In particular, for $N_{F}=3$, it has been used to describe pentaquarks and strange hyperons, both in the flavor symmetric limit, in which the full $SU(3)_{F}$ group is quantized, and in the flavor symmetry breaking limit, the so called Bound State Approach Callan and Klebanov (1985); Klebanov (1990), in which oscillations into the strange sector are treated as perturbations of the $SU(2)$ valued classical soliton. Since kaon degrees of freedom are best described in the latter approach, in this paper we will take this path and extend its application to the crystalline phases of Skyrmion matter, in order to be able to describe the phenomenon of kaon condensation in dense matter as predicted by the Skyrme model. The paper is organized as follows: in section II, we introduce the model and review the classical and quantum properties of crystal solutions. In section III we review the bound state approach to kaons in the Skyrme model, and compute the contribution to the total energy from the kaon condensate. In section IV, we obtain the system of equations from minimizing the free energy of $npe\mu\bar{K}$ matter, and solve it to find the onset of kaon condensation for different sets of parameters, and finally in section V we calculate the Equation of State of skyrmion matter including a kaon condensate, and compute the corresponding mass-radius curves for NS. ## II Generalized Skyrme model and skyrmion crystals The generalized Skyrme model we will consider is given by the following Lagrangian density $\displaystyle\mathcal{L}=-\frac{f^{2}_{\pi}}{16}$ $\displaystyle\Tr L_{\mu}L^{\mu}+\frac{1}{32e^{2}}\Tr\left[L_{\mu},L_{\nu}\right]^{2}$ $\displaystyle-\lambda^{2}\pi^{4}\mathcal{B}_{\mu}\mathcal{B}^{\mu}+\frac{m^{2}_{\pi}f^{2}_{\pi}}{8}\Tr\left(U-I\right),$ (1) where $L_{\mu}=U^{\dagger}\partial_{\mu}U$ is the left invariant Maurer-Cartan current and the Skyrme field can be written as $U=\sigma+i\pi_{a}\tau_{a}.$ (2) Here, $\pi_{a}$ ($a$ = 1, 2, 3) are the pions and $\tau_{a}$ are the Pauli matrices. The unitarity of the matrix field implies $\sigma^{2}+\pi_{a}\pi_{a}=1$. Furthermore, $\mathcal{B}^{\mu}$ is the conserved topological current which, in the standard manner, defines the topological index of maps $U$, i.e., the baryon charge $B$ $B=\int d^{3}x\mathcal{B}^{0},\hskip 5.69054pt\mathcal{B}^{\mu}=\frac{1}{24\pi^{2}}\epsilon^{\mu\nu\alpha\beta}\Tr\left\\{L_{\nu}L_{\alpha}L_{\beta}\right\\}.$ (3) The generalized Skyrme effective model contains only four terms and, therefore, four coupling constants, $f_{\pi},m_{\pi},e,\lambda$, two of which have a direct phenomenological interpretation as the pion decay constant and the pion mass. In addition, $\lambda$ can be related to a ratio between the mass and the coupling constant of the $\omega$ meson. From the very beginning, we assume the physical mass of the pions, $m_{\pi}=140$ MeV. The remaining constants are fitted to some properties of infinite nuclear matter. We remark that the third term in the action, although often omitted in the context of light nuclei, is obligatory when one studies the properties of nuclear matter at high density, which is a natural environment in the core of neutron stars. Indeed, this sextic term governs the equation of state at this regime and asymptotically leads to the maximally stiff EoS. The canonical description of NS is provided by the relativistic TOV approach where a particular model of infinite nuclear matter gives a source term of the Einstein equations. Effectively, it enters via an EoS, that is, a relation between e.g., pressure and density. In the Skyrme model, infinite skyrmionic matter is described by a periodic minimizer of the static energy ($E=-\int d^{3}x\mathcal{L}$) and, therefore, it is usually referred to as the Skyrme crystal. Obviously, while the total energy of the crystal is infinite, the energy per baryon number remains finite $\frac{E}{B}=\frac{N_{\text{cells}}\>E_{\text{cell}}}{N_{\text{cells}}\>B_{\text{cell}}}=\frac{E_{\text{cell}}}{B_{\text{cell}}}.$ (4) Here, $N_{\text{cells}}$ is the number of cells and $E_{\text{cell}}$, $B_{\text{cell}}$ are the energy and baryon charge in a single, periodic cell. As mentioned before, skyrmion crystals minimize the static energy functional, $\displaystyle E$ $\displaystyle=$ $\displaystyle\int d^{3}x\left(\mathcal{E}_{2}+\mathcal{E}_{4}+\mathcal{E}_{6}+\mathcal{E}_{0}\right)$ (5) $\displaystyle=$ $\displaystyle\frac{1}{24\pi^{2}}\int d^{3}x\left[-\frac{1}{2}\Tr\\{L_{i}L_{i}\\}-\frac{1}{4}\Tr\\{\left[L_{i},L_{j}\right]^{2}\\}\right.$ $\displaystyle+\left.4\pi^{4}c_{6}(\mathcal{B}^{0})^{2}+\frac{c_{0}}{2}\Tr\left(I-U\right)\right]$ over a finite region of space with periodic boundary conditions. Here $U$ is the SU(2) valued Skyrme field and $L_{\mu}=U^{\dagger}\partial_{\mu}U$. Further, $\mathcal{E}_{2}$ and $\mathcal{E}_{4}$ are the standard terms of the Skyrme model quadratic and quartic in derivatives, and $\mathcal{E}_{6}$ is the sextic term mentioned above. Finally, $\mathcal{E}_{0}$ is the pion mass potential. We have defined the dimensionless constants $c_{6}=2\lambda^{2}\frac{f^{2}_{\pi}e^{4}}{\hbar^{3}}$, ${c_{0}=2\frac{m^{2}_{\pi}}{f^{2}_{\pi}e^{2}}}$ and use the so-called Skyrme model units, so that our energy and length units are $E_{s}=3\pi^{2}f_{\pi}/e,\quad l_{s}=\hbar/(f_{\pi}e),$ (6) respectively. Both the size of the unit cell (characterized by the unit cell length parameter $L$) and its geometry will affect $E_{\rm cell}$. It turns out that, for our purposes, the ground state of skyrmion crystals is well described by a cubic unit cell with side length $2L$ composed of skyrmions in a face-centered cubic (FCC) arrangement, but with an additional symmetry. Concretely, it respects the following symmetries, $\displaystyle\text{S}_{1}$ $\displaystyle:(x,y,z)\rightarrow(-x,y,z),$ $\displaystyle(\sigma,\pi_{1},\pi_{2},\pi_{3})\rightarrow(\sigma,-\pi_{1},\pi_{2},\pi_{3}),$ (7) $\displaystyle\text{S}_{2}$ $\displaystyle:(x,y,z)\rightarrow(y,z,x),$ $\displaystyle(\sigma,\pi_{1},\pi_{2},\pi_{3})\rightarrow(\sigma,\pi_{2},\pi_{3},\pi_{1}),$ (8) $\displaystyle\text{S}_{3}$ $\displaystyle:(x,y,z)\rightarrow(x,z,-y),$ $\displaystyle(\sigma,\pi_{1},\pi_{2},\pi_{3})\rightarrow(\sigma,-\pi_{1},\pi_{3},-\pi_{2}),$ (9) $\displaystyle\text{S}_{4}$ $\displaystyle:(x,y,z)\rightarrow(x+L,y,z),$ $\displaystyle(\sigma,\pi_{1},\pi_{2},\pi_{3})\rightarrow(-\sigma,-\pi_{1},\pi_{2},\pi_{3}).$ (10) A detailed description of the construction of the Skyrme crystal and the comparison of different symmetries can be found in Adam _et al._ (2022a). As in that previous work, the unit cell that we will consider has size $2L$ and a baryon content of $B_{\text{cell}}=4$. Because of the additional symmetry for this crystal, the unit cell of size $2L$ decomposes into 8 cubes of side length $L$, each forming a simple cubic arrangement of half-skyrmions, where half-skyrmions are located in the corners of the cube and lead to a baryon content of $1/2$. The fields, however, are periodic only in $2L$, hence the unit cell has side length $2L$. We obtain the value of the energy for each value of $L$ and, as explained in Adam _et al._ (2022a), the energy-size curve, $E_{\text{cell}}(L)$, is a convex function which has a minimum at a certain $L_{0}$. We identify this point with the nuclear saturation point of infinite, symmetric nuclear matter, which also presents a minimum in the energy per baryon number curve as a function of the baryon density $n_{B}=(4/(2L)^{3})=(1/2)L^{-3}$. Up to now, we have considered the classical skyrmion crystal, which corresponds to symmetric nuclear matter. However, for a realistic description of nuclear matter inside a NS we need to consider an almost completely isospin-asymmetric state, where only a small amount of protons is allowed. We have already considered this scenario in Adam _et al._ (2022b) via a semiclassical quantization of the isospin degrees of freedom (DOF) of the skyrmion crystal. Indeed, it is standard in nuclear physics to define the binding energy of a nuclear system in the following way, $\frac{E}{B}(n_{B},\delta)=E_{N}(n_{B})+S_{N}(n_{B})\delta^{2}+\order{\delta^{3}},$ (11) where $\delta$ is the isospin asymmetry parameter, defined in terms of the proton fraction $\gamma$ of the system, $\delta=(1-2\gamma)$. $E_{N}(n_{B})$ denotes the binding energy of isospin-symmetric matter, and $S_{N}(n_{B})$ represents the so-called _symmetry energy_ , which is responsible for the change in the binding energy when the neutron-to-proton ratio changes for a fixed value of the baryon number. The knowledge of the symmetry energy at high densities is fundamental for a correct description of NS interiors. However, although the values of the symmetry energy at saturation is well known ($S_{0}\sim 30$ MeV) Fiorella Burgio and Fantina (2018), the difficulty in experimentally measuring its behavior at high densities forces us to express it as an expansion in powers of the baryon density around $n_{0}$, $S_{N}(n_{B})=S_{0}+\frac{n_{B}-n_{0}}{3n_{0}}L_{\rm sym}+\frac{(n_{B}-n_{0})^{2}}{18n_{0}^{2}}K_{\rm sym}+\cdots$ (12) where $L_{\rm sym}=3n_{0}\partialderivative{S_{N}}{n_{B}}\evaluated{}_{n_{B}=n_{0}},\quad K_{\rm sym}=9n_{0}^{2}\partialderivative[2]{S_{N}}{n_{B}}\evaluated{}_{n_{B}=n_{0}}$ (13) denote the slope and curvature of the symmetry energy at saturation, respectively. The values of these coefficients are still very uncertain, but recent analysis of combined astrophysical and nuclear observations made possible to constrain the symmetry energy above $n_{0}$ Essick _et al._ (2021); Tang _et al._ (2021); de Tovar _et al._ (2021); Gil _et al._ (2021); Li _et al._ (2021). Let us now review the procedure for calculating the symmetry energy of an $SU(2)$ skyrmion crystal, as it will be generalized to the 3 flavor case in the next section. First, let us rewrite the Skyrme Lagrangian (1) as $\mathcal{L}=a\Tr\left\\{L_{\mu}L^{\mu}\right\\}+b\Tr\left\\{\left[L_{\mu},L_{\nu}\right]^{2}\right\\}+c\,\mathcal{B}_{\mu}\mathcal{B}^{\mu}+d\Tr\left(U-I\right).$ (14) In our dimensionless units, we have $a=-\frac{1}{2},\quad b=\frac{1}{4},\quad c=-8\lambda^{2}\pi^{4}f_{\pi}^{2}e^{4},\quad d=\frac{m^{2}_{\pi}}{f^{2}_{\pi}e^{2}}.$ (15) and consider a (time-dependent) isospin transformation of a static Skyrme field configuration: $U(\vec{x})\rightarrow\tilde{U}(\vec{x},t)\equiv g(t)U(\vec{x})g^{\dagger}(t).$ (16) The time dependent isospin matrices $g(t)$ are collective coordinates, whose dynamics is given by a kinetic term in the energy functional, $T=\frac{1}{2}\omega_{i}\Lambda_{ij}\omega_{j}$ (17) where $\Lambda_{ij}$ is the isospin inertia tensor, given by $\displaystyle\Lambda_{ij}$ $\displaystyle=$ $\displaystyle\int\\{2a\Tr\\{T_{i}T_{j}\\}-4b\Tr\\{[T_{i},L_{k}][T_{j},L_{k}]\\}-$ (18) $\displaystyle-\frac{c}{32\pi^{4}}\varepsilon^{abc}\Tr\\{T_{i}L_{b}L_{c}\\}\varepsilon_{ars}\Tr\\{T_{j}L_{r}L_{s}\\}\\}\,d^{3}x$ $\displaystyle=$ $\displaystyle\;\Lambda\,\delta_{ij}$ being $T_{a}$ the $\mathfrak{su}(2)$-valued current $T_{a}=\frac{i}{2}U^{\dagger}[\tau_{a},U]$ and $\vec{\omega}$ the associated isospin angular velocity, defined by $g^{\dagger}\dot{g}=\tfrac{i}{2}\omega_{a}\tau_{a}$. As shown in Adam _et al._ (2022b), we may canonically quantize the isospin collective degrees of freedom and obtain a Hamiltonian, which for a cubic crystal with a number $N$ of unit cells is given by $H=\frac{\hbar^{2}}{2N\Lambda_{\rm cell}}I^{\rm{tot}}(I^{\rm{tot}}+1)$ (19) in terms of the isospin moment of inertia $\Lambda=N\Lambda_{\rm cell}$, and the total isospin angular momentum eigenvalue $I^{\rm tot}$, given by the product of the total number of unit cells times the total isospin of each unit cell, which can be obtained by composing the isospins of each of the cells. In the charge neutral case, all cells will have the highest possible value of isospin angular momentum, so that on each unit cell with baryon number $B$, the total isospin will be $\frac{1}{2}B$, and hence for the full crystal will be $I^{\rm{tot}}=\frac{1}{2}NB$. Thus, the quantum correction to the energy (per unit cell) due to the isospin degrees of freedom in the neutral (i.e. purely neutronic) limit would be (assuming $N\rightarrow\infty$): $E^{\rm{iso}}=\frac{\hbar^{2}}{8\Lambda_{\rm cell}}B^{2},$ (20) where the value of $\hbar$ is related to the value of $e$ through: $\hbar=\frac{e^{2}}{3\pi^{2}}.$ (21) The classical skyrmion crystal configurations can therefore be understood as models for isospin-symmetric nuclear matter, i.e. nuclear matter with zero total isospin. Deviations from the exact isospin symmetric case yield quantum isospin corrections to the crystal energy per baryon, which depend on the difference between protons and neutrons through the total isospin number per unit cell. Hence, by considering the effect of iso-rotations over classical solutions we are effectively breaking the isospin symmetry of the static energy functional by adding a correction of quantum origin that explicitly breaks it. Moreover, knowing the energy correction due to isospin, it is straightforward to obtain the associated isospin chemical potential for the skyrmion crystal using its thermodynamical definition: $\mu_{I}=-\partialderivative{E}{N_{I}}$, where $N_{I}$ is the (third component of) the isospin number per unit cell. Given that $(I^{\rm{tot}})^{2}=I_{1}^{2}+I_{2}^{2}+I_{3}^{2}$ and $N_{I}=I_{3}/N$, we may rewrite eq. 20 as $E^{\rm{iso}}=\frac{1}{2\Lambda_{\rm cell}}\quantity(N_{I}^{2}+\frac{I_{2}^{2}}{N^{2}}+\frac{I_{1}^{2}}{N^{2}})$ (22) and then $\mu_{I}=-\partialderivative{E^{\rm{iso}}}{N_{I}}=-\frac{N_{I}}{\Lambda_{\rm cell}}.$ (23) Let us now consider a finite chunk of the Skyrme crystal of $N$ unit cells, and let $A=N\times B_{\rm cell}$, where $B_{\rm cell}$ is the baryon number of a unit cell. We do not enforce charge neutrality at this step, and further leave unknown the quantum state of the crystal. As in Adam _et al._ (2022b), we take a mean field approximation and consider that the isospin density in an arbitrary skyrmion crystal quantum state is approximately uniform so that $\expectationvalue{I^{0}_{3}}=\frac{\expectationvalue{I_{3}}}{\int d^{3}x}=\frac{\expectationvalue{I_{3}}}{NV_{\rm cell}}\doteq\frac{N_{I}}{V_{\rm cell}}$ (24) where $N_{I}$ is the isospin charge per unit cell in this arbitrary quantum state. The effective proton fraction that would yield such an isospin charge per unit cell is $\gamma$, so we write $N_{I}=-\frac{1}{2}(1-2\gamma)B_{\rm cell}=-\frac{B_{\rm cell}}{2}\delta.$ (25) Hence, the isospin energy per unit cell of the skyrmion crystal in such a state can be written in terms of the asymmetry parameter $E^{\rm iso}=\frac{\hbar^{2}}{8\Lambda}B_{\rm cell}^{2}\delta^{2},$ (26) and thus the symmetry energy for Skyrme crystals is given by $S_{N}(n_{B})=\frac{L^{3}}{8\Lambda}n_{B}.$ (27) As argued in Klebanov (1985), any quantum state different from purely neutron matter leads to a divergent Coulomb energy term for the skyrmion crystal. Therefore, in order to allow for a nonzero positive electric charge within the unit cell we consider the existence of a neutralizing background of negatively charged leptons, namely, electrons and muons. In this scenario, the effects of the positive charge become almost completely screened, and the residual Coulomb energy is negligible, so we do not take it into account. Apart from electromagnetic forces, nuclear matter interacts with leptons via the weak force. Indeed, the exchange between leptons and nucleons inside NS is completely described by the $\beta$-decay and electron capture processes, $n\rightarrow p+l+\bar{\nu}_{l}\quad,\quad p+l\rightarrow n+\nu_{l},$ (28) which take place simultaneously, as long as the charge neutrality and $\beta$-equilibrium conditions are satisfied, $\displaystyle n_{p}$ $\displaystyle=\frac{Z}{V}=n_{e}+n_{\mu},$ (29) $\displaystyle\mu_{n}=\mu_{p}+\mu_{l}$ $\displaystyle\implies\mu_{I}=\mu_{l},\quad l=e,\mu.$ (30) Leptons inside a NS can be described as a non-interacting relativistic Fermi gas. Then the chemical potential for a given kind of lepton is, $\mu_{l}=\sqrt{(\hbar k_{F})^{2}+m_{l}^{2}},$ (31) where $k_{F}=(3\pi^{2}n_{l})^{1/3}$ is the corresponding Fermi momentum, and $m_{l}$ is the mass of the corresponding lepton. Considering the most general case, in which we include muons, we combine this last expression with the above equilibrium conditions and obtain the following system of equations, $\displaystyle n_{\mu}=\frac{1}{3\pi^{2}}$ $\displaystyle\quantity[\quantity(\frac{\hbar B_{\rm cell}(1-2\gamma)}{2\Lambda})^{2}-\left(\frac{m_{\mu}}{\hbar}\right)^{2}]^{\tfrac{3}{2}},$ (32) $\displaystyle\frac{\hbar B_{\rm cell}}{2\Lambda}$ $\displaystyle(1-2\gamma)=\quantity[3\pi^{2}\quantity(\frac{\gamma B_{\rm cell}}{8L^{3}}-n_{\mu})]^{\tfrac{1}{3}},$ (33) In order to solve the system for $\gamma$, we take the ultrarelativistic approximation $\mu_{e}\approx\hbar k_{F,e}$ for electrons. Besides we start solving the system at low densities considering only electrons, hence we drop the first equation and set $n_{\mu}=0$ until the condition $\mu_{e}=m_{\mu}$ is reached. Then, muons start to appear and we solve both equations. For each length of the unit cell we obtain the value of the proton fraction, hence we reconstruct the curve $\gamma(L)$. The total energy per unit cell in a $\beta$-equilibrated skyrmion crystal is therefore given by $E=E_{\rm class}+E_{\rm iso}(\gamma)+E_{e}(\gamma)+E_{\mu}(\gamma),$ (34) where $E_{\rm class}$ correspond to the classical energy of the Skyrme crystal, $E_{\rm iso}$ is calculated from eq. 26 and the energies of the leptons are the usual energy of a relativistic Fermi gas with mass $m_{l}$ at zero temperature, $\displaystyle E_{\rm lep}$ $\displaystyle=\int_{0}^{k_{f}}\frac{k^{2}dk}{\pi^{2}}\sqrt{k^{2}+m_{l}^{2}}=$ (35) $\displaystyle=\frac{m^{4}_{l}}{8\pi^{2}}\quantity[x_{r}(1+2x_{r}^{2})\sqrt{1+x_{r}^{2}}-\ln{x_{r}+\sqrt{1+x_{r}^{2}}}],$ where $x_{r}=k_{F}/m_{l}$. Recall that for electrons we take the approximation $x^{-1}_{r}\rightarrow 0$, which is justified for densities $n\geq n_{0}$. ## III Kaon condensate in skyrmion crystals Having obtained the proton fraction (hence the electron chemical potential) in $npe\mu$-matter as a function of density, we can now turn to the question of whether kaon fields may condense inside a Skyrme crystal for a sufficiently high density, and, if so, whether this critical density value is relevant for the description of matter inside compact stars. ### III.1 Kaon fluctuations in the Skyrme model Following the bound-state approach first proposed in Callan and Klebanov (1985) we may include strange degrees of freedom in the Skyrme model by extending the skyrmion field to a $SU(3)$-valued field $U$ through modelling kaon fluctuations on top of a $SU(2)$ skyrmion-like background $u$. With the only requirement that unitarity must be preserved, different ansätze have been proposed in the literature for the total $SU(3)$ field describing both pions and kaons. In this work, we choose the ansatz proposed by Blom et al in Blom _et al._ (1989): $U=\sqrt{U_{K}}U_{\pi}\sqrt{U_{K}}.$ (36) In this ansatz $U_{\pi}$ represents the $SU(3)$ embedding of the purely pionic part $u$, and the field $U_{K}$ are the fluctuations in the strange directions. It can be shown that this ansatz is equivalent to the one first proposed by Callan and Klebanov in Callan and Klebanov (1985) when computing static properties of hyperons, although both may differ in other predictions of the model Nyman and Riska (1990). In the simplest $SU(3)$ embedding, the $SU(2)$ field $u$ is extended to $U_{\pi}$ by filling the rest of entries with ones in the diagonal and zeros outside. On the other hand, the kaon ansatz is modelled by a $\mathfrak{su}$(3) -valued matrix $\mathcal{D}$ which is non trivial in the off-diagonal elements: $\begin{split}U_{\pi}&=\begin{pmatrix}u&0\\\ 0&1\end{pmatrix},\hskip 8.53581ptU_{K}=e^{i\frac{2\sqrt{2}}{f_{\pi}}\mathcal{D}},\\\ &u=\sigma+i\pi_{a}\tau_{a},\quad\mathcal{D}=\matrixquantity(0\hfil&K\\\ K^{\dagger}&0)\end{split}$ (37) where $K$ consists of a scalar doublet of complex fields representing charged and neutral kaons: $K=\matrixquantity(K^{+}\\\ K^{0}),\quad K^{\dagger}=(K^{-},\bar{K}^{0}).$ (38) The extension of the Generalized Skyrme Lagrangian from (1) to include strange degrees of freedom consists in the replacement of the $\mathcal{L}_{0}$ term by Nyman and Riska (1990): $\displaystyle\mathcal{L}_{0}^{\rm new}=$ $\displaystyle\frac{f^{2}_{\pi}}{48}\left(m^{2}_{\pi}+2m^{2}_{K}\right)\Tr{U+U^{\dagger}-2}+$ $\displaystyle+\frac{\sqrt{3}}{24}f^{2}_{\pi}\left(m^{2}_{\pi}-m^{2}_{K}\right)\Tr{\lambda_{8}\left(U+U^{\dagger}\right)},$ (39) where $\lambda_{8}$ is the eighth Gell-Mann matrix and $m_{K}$ is the vacuum kaon mass, and the addition of the Wess-Zumino-Witten (WZW) term, which can be expressed in terms of a 5-dimensional action: $S_{WZ}=-i\frac{N_{c}}{240\pi^{2}}\int d^{5}x\>\epsilon^{\mu\nu\alpha\beta\gamma}\Tr{L_{\mu}L_{\nu}L_{\alpha}L_{\beta}L_{\gamma}}.$ (40) ### III.2 The kaon condensate on classical crystals The onset of kaon condensation in the Skyrme model takes place at a critical density $n_{\rm cond}$ at which $\mu_{e}$ becomes greater than the energy of the kaon zero-momentum mode (s-wave condensate). Thus, for baryon densities $n\geq n_{\rm cond}$, the macroscopic contribution of the kaon condensate to the energy must be taken into account when obtaining the EoS of dense matter. To do so, we follow the standard procedure to describe Bose-Einstein condensation of a (complex) scalar field (see eg Schmitt (2010)) in which the field condensates correspond to the non-zero vacuum expectation values (vev), $\expectationvalue{K^{\pm}}$, which are assumed to be constant in space and whose time dependence is given by: $\expectationvalue{K^{\mp}}=\phi e^{\mp i\mu_{K}t}$ (41) The real constant $\phi$ corresponds to the zero-momentum component of the fields, which acquires a nonvanishing, macroscopic value after the condensation. Its exact value is determined from the minimization of the corresponding effective potential, to whose calculation we will dedicate the rest of this section. On the other hand, the phase $\mu_{K}$ is nothing but the corresponding kaon chemical potential. First, we will need an explicit form of the $SU(3)$ Skyrme field in the kaon condensed phase. Assuming the charged kaons will be the first mesons to condense 111Actually, that the charged (in particular, the negatively charged) kaons will condense first is true in our approach (whenever $\mu_{e}>0$), since the chemical potential associated to neutral kaons is zero, so that the onset of neutral kaon condensation is given by $m^{*}_{K}=0$., we can safely drop the neutral kaon contribution, and define the following matrix $\tilde{\mathcal{D}}=\matrixquantity(0&0&\phi e^{i\mu_{K}t}\\\ 0&0&0\\\ \phi e^{-i\mu_{k}t}&0&0)$ (42) which results from substituting the kaon fields in $\mathcal{D}$ as defined in (37) by their corresponding vev in the kaon condensed phase. Also, taking advantage of the property $\mathcal{D}^{3}=\phi^{2}\mathcal{D}$, we may write the $SU(3)$ element generated by $\tilde{\mathcal{D}}$ explicitly in matrix form: $\Sigma=e^{i\tfrac{\sqrt{2}}{f_{\pi}}\tilde{\mathcal{D}}}=\matrixquantity(\cos\tilde{\phi}&0&ie^{i\mu_{K}t}\sin\tilde{\phi}\\\ 0&1&0\\\ ie^{-i\mu_{K}t}\sin\tilde{\phi}&0&\cos\tilde{\phi})$ (43) where $\tilde{\phi}=\tfrac{\sqrt{2}}{f_{\pi}}\phi$ is the dimensionless condensate amplitude. Furthermore, assuming the backreaction from the kaon condensate to the skyrmion crystal is negligible, and thus the classically obtained crystal configuration will be the physically correct background even in the kaon condensed phase, we may write the $SU(3)$ field in this phase as $U=\Sigma U_{\pi}\Sigma$, where $U_{\pi}$ is the $SU(3)$ embedding of the $SU(2)$ skyrmion background as in (37). Introducing this $U$ in the total action yields the standard Skyrme action for the $SU(2)$ field plus an effective potential term for the kaon condensate: $S_{Sk}(U)+S_{\rm WZW}(U)=S_{Sk}(U_{\pi})-\int dtV_{K}(\tilde{\phi}),$ (44) where $V_{K}=\frac{1}{24\pi^{2}}\int d^{3}x\Big{[}V^{(2)}_{K}+V^{(4)}_{K}+V^{(6)}_{K}+V^{(0)}_{K}\Big{]}+V^{(WZW)}_{K}.$ (45) Let us now calculate the contribution to the effective potential $V_{K}$ of each term in the action: * • Quadratic term: Given that the crystal background is static and the kaon condensate does not depend on spatial coordinates, the kaon part of the quadratic term may be written as $\Tr{L_{0}^{2}}=-[\Tr\\{\partial_{t}\Sigma^{\dagger}\partial_{t}\Sigma\\}+\Tr\\{\Sigma^{\dagger}\partial_{t}\Sigma U_{\pi}^{\dagger}\Sigma\partial_{t}\Sigma^{\dagger}U_{\pi}\\}].$ (46) Introducing the explicit expression for $\Sigma$, (43), yields $V_{K}^{(2)}=\mu_{K}^{2}\sin^{2}\tilde{\phi}[(1+\sigma^{2}+\pi^{2}_{3})\sin^{2}\tilde{\phi}-2(1+\sigma\cos^{2}\tilde{\phi})].$ (47) * • Quartic term: In the quartic term, the kaon effective potential comes from the terms with time derivatives of the total field, $\displaystyle\Tr{[L_{0},L_{i}]^{2}}=2[\Tr\\{\partial_{t}U^{\dagger}\partial_{i}U\partial_{t}U^{\dagger}\partial_{i}U\\}-$ (48) $\displaystyle-\Tr\\{\partial_{i}U^{\dagger}\partial_{t}U\partial_{i}U^{\dagger}\partial_{t}U\\}],$ which, after substitution of the expression for $\Sigma$, gives $\displaystyle V_{K}^{(4)}$ $\displaystyle=-2\mu_{K}^{2}\sin^{2}{\tilde{\phi}}\big{\\{}(1+\sigma)\partial_{i}n^{2}\cos^{2}{\tilde{\phi}}+$ (49) $\displaystyle+$ $\displaystyle 2[\partial_{i}\sigma^{2}(1-\pi_{3}^{2})+\partial_{i}\pi_{3}^{2}(1-\sigma^{2})+2\sigma\pi_{3}\partial_{i}\sigma\partial_{i}\pi_{3}]\sin^{2}\tilde{\phi}\big{\\}}$ * • Mass term: The kaon part associated to the mass term gives the following contribution, $V_{K}^{(0)}(\tilde{\phi})=2\frac{m^{2}_{K}}{f^{2}_{\pi}e^{2}}(1+\sigma)\sin^{2}\tilde{\phi}$ (50) * • Wess-Zumino-Witten term: The WZW term is written as a 5-form integrated over an auxiliar 5-dimensional disk $D$ whose boundary is the spacetime manifold $M$, but in appendix A we show that the variation after the kaon fluctuations of the pion background yields a local term which may be written as an effective four-dimensional lagrangian. Indeed, we show that (51) * • Sextic term: The contribution from the sextic term is also obtained in appendix A to be $V_{K}^{(6)}=-\frac{\lambda^{2}f^{2}_{\pi}e^{4}}{16}\Tr\\{[R_{j},R_{k}]\xi_{0}\\}^{2}$ (52) where $\xi_{\mu}=U_{\pi}\Sigma\partial_{\mu}\Sigma^{\dagger}U_{\pi}^{\dagger}-\Sigma^{\dagger}\partial_{\mu}\Sigma$. Once the traces are evaluated, we end up with $\displaystyle V_{K}^{(6)}=$ $\displaystyle-\lambda^{2}f^{2}_{\pi}e^{4}\mu_{K}^{2}\sin^{4}(\tilde{\phi})(\partial_{i}\pi_{3}\partial_{j}\sigma-\partial_{i}\sigma\partial_{j}\pi_{3})^{2}.$ (53) ### III.3 Effect of kaon condensation on the quantum corrections to Skyrme crystals In the above calculations, we have taken separately the contributions of a kaon condensate and an isospin angular momentum of the skyrmion crystal, and the kaon condensate interacts with the skyrmion isospin only indirectly via the charge neutrality and $\beta$ equilibrium conditions, which relate their corresponding chemical potentials. However, since kaons possess an isospin quantum number, we should consider a (time-dependent) isospin transformation of the full Skyrme field + kaon condensate configuration $U=\Sigma U_{\pi}\Sigma$: $U\rightarrow\tilde{U}\equiv A(t)UA^{\dagger}(t),$ (54) where $A$ is an element of $SU(3)$ modelling an isospin rotation, $A=\matrixquantity(a&0\\\ 0&1),\quad a\in SU(2).$ (55) The Maurer-Cartan form transforms as ($\dot{A}=dA/dt$) $\tilde{U}^{\dagger}\partial_{\mu}\tilde{U}=\left\\{\begin{array}[]{ll}AU^{\dagger}\partial_{i}UA^{\dagger},\quad(\mu=i=1,2,3),&\\\\[5.69054pt] AU^{\dagger}\partial_{0}UA^{\dagger}+A(U^{\dagger}[A^{\dagger}\dot{A},U])A^{\dagger},&(\mu=0).\end{array}\right.$ (58) We now define the isospin angular velocity $\vec{\omega}$ as $A^{\dagger}\dot{A}=\tfrac{i}{2}\omega_{a}\lambda_{a}$ ($a=1,2,3$), with $\lambda_{\textsc{a}}$ the Gell-Mann matrices generating $SU(3)$ for $\textsc{a}=1,\cdots 8$. Notice that $\vec{\omega}$ is a three-vector, since $A^{\dagger}\dot{A}$ belongs to the isospin $\mathfrak{su}(2)$ subalgebra of $\mathfrak{su}(3)$ Then, we may write the time component of the Maurer-Cartan current as $\tilde{U}^{\dagger}\partial_{0}\tilde{U}=AL_{0}A^{\dagger}+AT_{a}A^{\dagger}\omega_{a}$, where $T_{a}$ is the $\mathfrak{su}(3)$-valued current: $T_{a}=\frac{i}{2}U^{\dagger}[\lambda_{a},U]\equiv iT_{a}^{\textsc{a}}\lambda_{\textsc{a}},$ (59) where we have made use of the parametrization (37). The time dependence of the new Skyrme field induces the existence of a kinetic term in the energy functional, given by 222Remember that we are using the mostly minus convention for the metric signature. $\begin{split}T=\int\\{a&\quantity(\Tr\\{L_{0}L_{0}\\}+2\Tr\\{L_{0}T_{a}\\}\omega_{a}+\Tr\\{T_{a}T_{b}\\}\omega_{a}\omega_{b})\\\ -2b\big{(}\Tr\\{[&(L_{0}+T_{a}\omega_{a}),L_{k}][(L_{0}+T_{b}\omega_{b}),L_{k}]\\}\big{)}-c\,\,\mathcal{B}^{i}\mathcal{B}_{i}\\}d^{3}x,\end{split}$ (60) with $\mathcal{B}^{i}$ the spatial components of the topological current (3): $\mathcal{B}^{i}=\frac{3}{24\pi^{2}}\varepsilon^{ijk}\Tr\\{(L_{0}+T_{a}\omega_{a})L_{j}L_{k}\\}.$ (61) We may rewrite the kinetic isorotational energy in the standard way as a quadratic form acting on the components of the isospin angular velocity, $T=\frac{1}{2}\omega_{a}\Lambda_{ab}\omega_{b}+\Delta_{a}\omega_{a}-V_{K}$ (62) where $\Lambda_{ab}$ is the isospin inertia tensor and $\Delta_{a}$ is the kaon condensate isospin current, given by $\displaystyle\Lambda_{ab}$ $\displaystyle=\int\left\\{2a\Tr\\{T_{a}T_{b}\\}-4b\Tr\\{[T_{a},L_{k}][T_{b},L_{k}]\\}-\frac{c}{32\pi^{4}}\varepsilon^{lmn}\Tr\\{T_{a}L_{m}L_{n}\\}\varepsilon_{lrs}\Tr\\{T_{j}L_{r}L_{s}\\}\right\\}\,d^{3}x,$ (63) $\displaystyle\Delta_{a}$ $\displaystyle=\int\left\\{2a\Tr\\{L_{0}T_{a}\\}-4b\Tr\\{[T_{a},L_{k}][L_{0},L_{k}]\\}-\frac{c}{32\pi^{4}}\varepsilon^{lmn}\Tr\\{L_{0}L_{m}L_{n}\\}\varepsilon_{lrs}\Tr\\{T_{a}L_{r}L_{s}\\}\right\\}\,d^{3}x,$ (64) where $a,b$ and $c$ are those in eq. 15. The symmetries of the crystalline configuration that we consider in this work, concretely the $S_{1}$ and $S_{2}$ transformations, imply that the isospin inertia tensor becomes proportional to the identity, i.e. $\Lambda^{\rm{crystal}}_{ab}=\Lambda\delta_{ab}$. However, the presence of a kaon condensate breaks this symmetry to a $U(1)$ subgroup, so that $\Lambda_{ab}$ presents two different eigenvalues in the condensate phase, $\Lambda_{\rm cond}=\text{diag}(\Lambda,\Lambda,\Lambda_{3})$. Similarly, $\Delta_{a}=0$ in the purely barionic phase, and its third component acquires a non-zero value in the condensate phase, $\Delta_{\rm cond}=(0,0,\Delta)$. The explicit expressions for $\Lambda_{3}$ and $\Delta$ in the condensed phase are written in appendix B. One can easily check that in the non-condensed phase, $\phi=0$ and the results of the previous section are recovered, namely, $\Lambda_{3}=\Lambda$, $\Delta=0$. The quantization procedure now goes along the same lines as in the first section. However, the isospin breaking due to the kaon condensate implies that the canonical momentum associated to the third component of the isospin angular velocity will now be different, and given by $I_{3}=\Lambda_{3}\omega_{3}+\Delta$. Thus, after a Legendre transformation to rewrite (62) in Hamiltonian form, and making the $N\rightarrow\infty$ approximation, one can write the quantum energy correction per unit cell of the crystal in the kaon condensed phase as $E_{\rm{quant}}=\frac{1}{2\Lambda_{3}}(I_{3}^{2}-\Delta^{2}).$ (65) The first term on the rhs is just the isospin correction, while now there is an additional second term due to the isospin of the kaons. Indeed, since the kaon field enters also in the expression of the isospin moment of inertia $\Lambda_{3}$, both terms will depend nontrivially on the kaon vev field. ## IV Results When the kaon field develops a nonzero vev, apart from the neutron decay and lepton capture processes of eq. 28, additional processes involving kaons may occur: $n\leftrightarrow p+K^{-},\qquad l\leftrightarrow K^{-}+\nu_{l}$ (66) such that the chemical equilibrium conditions $\mu_{n}=\mu_{p}+\mu_{K},\qquad\mu_{l}=\mu_{K}$ (67) are satisfied. These are the extension of eq. 30 to the condensate phase. The total energy within the unit cell may be obtained as the sum of the baryon, lepton and kaon contributions: $\begin{split}E=E_{\rm class}+&E_{\rm iso}(\gamma,\tilde{\phi})+E_{K}(\mu_{e},\tilde{\phi})+\\\ +&E_{e}(\mu_{e})+\Theta(\mu_{e}^{2}-m_{\mu}^{2})E_{\mu}(\mu_{e})\end{split}$ (68) The kaon contribution is the effective potential energy $E_{K}(\mu_{e},\tilde{\phi})=V_{K}-\frac{\Delta^{2}}{2\Lambda_{3}},$ (69) which depends on the condensate $\tilde{\phi}$ and on the lepton chemical potential through the explicit dependence on $\mu_{K}$ of both $V_{K}$ and $\Delta$, and $\mu_{K}=\mu_{e}$ due to the equilibrium conditions (67). Therefore, the energy of the full system depends on the proton fraction, the kaon vev field and the electron chemical potential. Their respective values can be obtained, for fixed $n_{B}$ (or equivalently, fixed $L$) by minimizing the free energy $\Omega=E-\mu_{e}(N_{e}+\Theta(\mu_{e}^{2}-m_{\mu}^{2})N_{\mu}-\gamma B)$ (70) with respect to $\gamma$, $\tilde{\phi}$ and $\mu_{e}$, i.e. $\partialderivative{\Omega}{\gamma}\evaluated{}_{n_{B}}\\!\\!\\!(\gamma,\tilde{\phi},\mu_{e})=\partialderivative{\Omega}{\tilde{\phi}}\evaluated{}_{n_{B}}\\!\\!\\!(\gamma,\tilde{\phi},\mu_{e})=\partialderivative{\Omega}{\mu_{e}}\evaluated{}_{n_{B}}\\!\\!\\!(\gamma,\tilde{\phi},\mu_{e})=0.$ (71) The first equation imposes the expected condition $\mu_{e}=\mu_{I}=2\hbar^{2}(1-2\gamma)/\Lambda_{3}$. Then, after substituting into the other two conditions we get: $\displaystyle\gamma n_{B}-$ $\displaystyle\frac{(\mu^{2}_{I}-m^{2}_{e})^{3/2}+(\mu^{2}_{I}-m^{2}_{\mu})^{3/2}}{3\pi^{2}\hbar^{3}}+\frac{n_{B}}{4}\partialderivative{E_{K}}{\mu_{e}}\evaluated{}_{\mu_{e}=\mu_{I}}=0,$ (72) $\displaystyle\partialderivative{V_{K}}{\tilde{\phi}}-\frac{\Delta}{\Lambda_{3}}\partialderivative{\Delta}{\tilde{\phi}}+\partialderivative{\Lambda_{3}}{\tilde{\phi}}\left(\frac{\Delta^{2}}{2\Lambda_{3}^{2}}-\frac{\mu_{I}^{2}}{2\hbar^{2}}\right)=0,$ (73) which are precisely the charge neutrality condition, and the minimization of the grand canonical potential with respect to the kaon field. We note here that we drop the ultrarrelativistic consideration for electrons since the appearance of kaons may decrease hugely the electron fraction. By solving the system of equations 72 and(73) for $\gamma$ and $\tilde{\phi}$ we obtain all the needed information for the new kaon condensed phase. Then we may compare the particle fractions and energies between both phases, which we will call $npe\mu$ and $npe\mu\overline{K}$. Before solving the full system for different values of the lattice length $L$, we may try to obtain the value of the length at which kaons condense, $L_{\rm cond}$. This value is indeed important since it will determine wether or not a condensate of kaons will appear at some point in the interior of NS This is accomplished with the same system of eqs. 72 and 73 by factoring the $\sin\tilde{\phi}$ from the second equation and setting $\tilde{\phi}=0$. Then we may see the system as a pair of equations to obtain the values of $\gamma_{\text{cond}}$ and $L_{\text{cond}}$, the values of the proton fraction and the length parameter for which the kaons condense. We show in the table below the density at which kaons condense for different values of the parameters as well as the values of some nuclear observables they yield. All the values are given in units of MeV or fm, respectively. label | $f_{\pi}$ | $e$ | $\lambda^{2}$ | $E_{0}$ | $n_{0}$ | $S_{0}$ | $L_{\text{sym}}$ | $n_{\text{cond}}/n_{0}$ ---|---|---|---|---|---|---|---|--- set 1 | 133.71 | 5.72 | 5 | 920 | 0.165 | 23.5 | 29.1 | 2.3 set 2 | 138.11 | 6.34 | 5.78 | 915 | 0.175 | 24.5 | 28.3 | 2.2 set 3 | 120.96 | 5.64 | 2.68 | 783 | 0.175 | 28.7 | 38.7 | 1.6 set 4 | 139.26 | 5.61 | 2.74 | 912 | 0.22 | 28.6 | 38.9 | 1.6 Table 1: Sets of parameter values and observables at nuclear saturation Parameter sets 1 and 2 are chosen so that the energy per baryon and baryon density at saturation are fitted to experimental values, whereas the sets 3 and 4 correctly fit the symmetry energy and slope at saturation. In fig. 1, we show the $E(L)$ curves both without and with kaon condensation, in dimensionless Skyrme units. It is clearly visible that for sufficiently small $L$ a nonzero kaon condensate is preferred. In fig. 2 we show the resulting particle fractions. In fig. 3 we plot the symmetry energy as a function of both $n_{B}$ and the kaon condensate $\phi$. Figure 1: Energy vs lattice length in dimensionless Skyrme units, for the set 1 of parameters. The energy is shown for the classical crystal without isopin contributions (green), isospin asymmetric ($npe\mu$) matter with (black) and without (blue) kaons. We also plot the completely asymmetric neutron matter (magenta) which lies slightly above the blue curve. Figure 2: Particle fractions as a function of baryon density for the set 1 of parameters, both with (solid lines) and without (discontinuous lines) kaon condensate. For the case with kaon condensate, the contribution of muons is negligible. Figure 3: Symmetry energy of nuclear matter as a function of baryon density and kaon $vev$, for the parameter set 1. The surface is obtained by treating $n_{B}$ and $\phi$ as independent variables, whereas the red curve corresponds to the energy-minimizing solution for each $L$, i.e., $n_{B}$. ## V Neutron stars with kaon condensed cores within the Skyrme model. In this section we briefly recapitulate how to obtain the EoS from the Skyrme crystal solution. Then we calculate the full $npe\mu\bar{K}$ matter EoS, and finally solve the TOV system and compare the NS properties with and without kaons. ### V.1 The Skyrme crystal EoS for (a-)symmetric nuclear matter. The energy per baryon as a function of the lattice length, $E(L)$, has a minimum at a certain value $L_{0}$. The density $n_{0}=(1/2)L_{0}^{-3}$ at which this minimum is achieved is the so called nuclear saturation density, which has been experimentally found to be $n_{0}\simeq 0.16$ fm-3. Further, the energy per baryon at nuclear saturation is $E_{0}=923\,{\rm MeV}$. On the other hand, for the crystal solutions one can define the energy density $\rho$, pressure $p$ and baryon density $n$ as $\displaystyle\rho$ $\displaystyle=\frac{E}{V}=\frac{E_{\text{cell}}}{V_{\text{cell}}},$ (79) $\displaystyle p$ $\displaystyle=-\frac{\partial E}{\partial V}=-\frac{\partial E_{\text{cell}}}{\partial V_{\text{cell}}},$ (80) $\displaystyle n_{B}$ $\displaystyle=\frac{B}{V}=\frac{B_{\text{cell}}}{V_{\text{cell}}}=\frac{1}{2L^{3}}.$ (81) We can understand the cell length parameter $L$ as labelling the different energy-minimizing configurations at different densities. Therefore, the three quantities above are related as functions of $L$. This relation is precisely the equation of state (EoS) for skyrmion crystals. It has been recently shown in Adam _et al._ (2020) that this EoS stiffens in the generalized Skyrme model, i.e. when the sextic term in (5) is included. This stiffening significantly rises the maximal NS masses that can be reached, which is a strong motivation for the inclusion of the sextic term, as it is necessary in order to reach the mass range of massive pulsars, around $2-2.5M_{\odot}$ according to recent observations. The global energy minimum of the crystal is reached at $L=L_{0}$, at which skyrmionic matter remains in equilibrium, i.e. at zero pressure. For $L<L_{0}$ the crystal is squeezed, which translates into larger values of pressure and density. In the opposite region, $L>L_{0}$, the matter content inside the unit cell spreads and we enter the unstable branch. Indeed, the pressure in this region is negative, hence we conclude that this (low-density) regime is not well described by the crystal solution. We actually expect that matter inside the unit cell will rearrange in a kind of inhomogeneous central lump surrounded by vacuum Adam _et al._ (2022a). More exotic configurations presenting non-homogeneous structures have also been constructed in the Skyrme model, Park _et al._ (2019); Canfora _et al._ (2020). However, the physical relevance of these configurations as true energy minimizers in the model still remains unclear. Hence, the main ingredient to obtain the EoS for the Skyrme crystal is the energy dependence on the unit cell size. The $npe\mu$ matter case is easy to obtain using eq. 34 for different values of $L$ after solving the $\beta-$equilibrium and charge neutrality conditions for $\gamma$. However, once we include kaons, the change in the energy curve fig. 1 may lead to a first or second order phase transition. To distinguish the order of the phase transition in our case, we need to know accurately the pressure near the condensation point. Therefore, we computed more points for the energy near the condensation value with higher accuracies, and we obtained the pressure using a numerical derivative. We conclude that the kaon condensation produces a first order phase transition for our choices of parameters in the Skyrme model. This can be seen in the right plot of fig. 4, where we show the EoS for our best accuracy and, clearly, there is a non-physical region which must be bridged by a first order phase transition. Figure 4: Left plot: energy against the side length of the crystal, calculated with more points near the condensation values for both branches and their interpolations. Right plot: pressure against the energy density (both in Skyrme units) from which we conclude that there is a first order phase transition. ### V.2 Maxwell construction versus Gibbs construction The Maxwell construction (MC) is typically used to obtain a physical equation of state when a first order transition is present. Indeed, the MC has been already studied in the Skyrme crystals context to describe the transition between crytals with different symmetries Adam _et al._ (2022a). This construction is based on a mixed phase of constant pressure which connects the two solutions. However, the MC is only correct when there is a single conserved charge (in this case, the baryon number) for which the associated chemical potential is enforced to be common for both phases in the mixed phase Glendenning (1992). If, instead, an additional charge is conserved, like the electric charge in the case of $npe\mu$ matter, the Gibbs conditions for the phase equilibrium, $p^{\rm I}=p^{\rm II},\hskip 5.69054pt\mu^{\rm I}_{i}=\mu^{\rm II}_{i},\hskip 4.2679pti=B,q$ (82) cannot be both satisfied in a standard MC. In the last expression $\mu_{B}$ and $\mu_{q}$ represent the chemical potentials associated to the conserved baryon and electric charges, respectively. Instead, one should perform a Gibbs construction (GC) Glendenning (1992); Glendenning and Schaffner-Bielich (1999). Indeed, the GC has also been proven useful in the context of a hadron- to-quark phase transition inside NS Bhattacharyya _et al._ (2010). We may write the chemical potential of each particle species as a linear combination of the chemical potentials associated to the conserved charges of our system: $\mu_{i}=B_{i}\mu_{B}+q_{i}\mu_{q},$ (83) where $B_{i}$ and $q_{i}$ are the baryon number and electric charge of the particle species $i$. Then we might identify the baryon and electric charge chemical potentials with the neutron and electron chemical potentials respectively. The main difference between MC and GC is that, in the mixed phase, the first one imposes charge neutrality locally, i.e. both phases are neutral independently, however in the GC it is imposed globally in the mixed phase. Considering a volume fraction $\chi$ of the kaon condensed phase, charge neutrality is imposed in the GC as: $n^{MP}_{q}=(1-\chi)n^{\rm I}_{q}+\chi n^{\rm II}_{q}=0.$ (84) The mixed phase in the GC is calculated by identifying first the contributions to the pressure and charge densities in each phase separately. Then we have to solve the system of equations composed by eqs. 82, 84 and 73. We use the unit cell length parameter of the first ($npe\mu$) phase $L_{\rm I}$ as the variable defining our position in the phase diagram, then the unknowns are the length in the second ($npe\mu\bar{K}$) phase $L_{\rm II}$, the proton fractions $\gamma_{\rm I}$, $\gamma_{\rm II}$, the kaon field $\tilde{\phi}$ and the volume fraction $\chi$. Figure 5: $E(L)$ curves for the two phases. The different slopes at the point of phase separation indicate a first-order phase transition. We also show the curves resulting from a Maxwell construction (MC) and a Gibbs construction (GC). We remark that we assumed in our calculations of the kaon condensate in section III that the backreaction of the condensate on the crystal is negligible, such that our two phases are always considered in the same classical crystal background, and the energies per baryon of the two phases are compared for the same length $L$. As a result, we always should have $L_{\rm I}=L_{\rm II}$ and, consequently, $n_{B,{\rm I}}=n_{B,{\rm II}}$ by construction. On the other hand, the relation between $L$ and the thermodynamical variables $p$, $\mu_{i}$ and $\phi$ used in eqs. 82, 84 and 73 is quite nontrivial in both phases. We, therefore, treat $L_{\rm II}$ as an independent variable in our numerical calculations. We find that always $L_{\rm I}=L_{\rm II}$ within our numerical precision, which provides us with an additional consistency check both for our numerics and for the thermodynamical transformations we used. We show our results in fig. 5 and in fig. 6. We find that the mixed phase of the GC, and hence the values at which the kaon field becomes non-zero, starts at a smaller density than the value obtained in table 1. This is also found in Glendenning and Schaffner-Bielich (1999), for which the GC mixed phase extends to a larger region than the one obtained from the MC, because the mixed phase in the GC no longer is for constant pressure. In our case, even the minimum of $E(L)$ is shifted to slightly lower values, see the insert in fig. 5 and, hence, the use of the GC affects the low density regime of the EoS, as can be seen in fig. 6. Figure 6: EoS for the three different cases that we have constructed. The jump in the MC due to the first order transition and the different behaviour of the GC at low densities are clearly visible. We also show the standard nuclear physics EoS of Sharma _et al._ (2015) (BCPM) and a hybrid EoS obtained by joining the BCPM EoS at low pressure with the GC EoS at high pressure. We may also calculate the particle fractions in the mixed phase of the GC using an expression equivalent to eq. 84 for each particle. We show the new particle fractions in fig. 7. Besides, during the mixed phase, we find that there are more protons in the second phase than in the first one. However, the presence of more kaons than protons in the second phase results in a partial negative charge density. That negative charge is compensated by the overall positive charge density of the first phase. In both phases the number of electrons is much less than that of protons and kaons. Figure 7: Particle fractions for the GC. The main difference with respect to fig. 2 is the earlier appearance of kaons. ### V.3 The TOV system and NS properties In order to calculate the mass and radius for a non-rotating NS we have to solve the standard TOV (Tolman-Oppenheimer-Volkoff) system of ODEs. It is obtained inserting a spherically symmetric ansatz of the spacetime metric, $ds^{2}=-A(r)dt^{2}+B(r)dr^{2}+r^{2}(d\theta^{2}+\sin^{2}\theta d\varphi^{2}),$ (85) in the Einstein equations, $R_{\mu\nu}-\frac{1}{2}g_{\mu\nu}R=8\pi GT_{\mu\nu}.$ (86) To describe matter inside the star, in the right-hand side of the equation, we use the stress-energy tensor of a perfect fluid, $T_{\mu\nu}=(\rho+p)u_{\mu}u_{\nu}+pg_{\mu\nu},$ (87) where the pressure $p$ and the energy density $\rho$ are not independent but related by the EoS. Hence the EoS describes the nuclear interactions inside the NS and different EoS lead to different observables. The resulting TOV system involves 3 differential equations for $A,B$ and $p$, which must be solved for a given value of the pressure in the centre of the NS ($p(r=0)=p_{0}$) until the condition $p(r=R)=0$ is achieved. We use a 4th order Runge-Kutta method of step $\Delta r=1$ m to solve the system and to obtain the main observables from the solutions. The radial point at which the pressure vanishes defines the radius of the NS, and the mass $M$ is obtained from the Schwarzschild metric definition outside the star, $B(r=R)=\frac{1}{(1-\frac{2GM}{R})}.$ (88) The results of this section are plotted in fig. 8 for the 4 sets of parameters. We compare the results between the MC and GC as well as with the EoS without kaons. The first observation is that the addition of kaons to the EoS agrees with the expectation, reducing the achievable maximum mass. This represents the so-called hyperon puzzle in which the appearance of new strange degrees of freedom softens the EoS such that it may not lead to sufficiently massive NS ($\sim 2M_{\odot}$). As can be seen, this is not the case in the generalized Skyrme model since we may obtain very high masses easily due to the contribution of the sextic term. Furthermore the radii of NS are also reduced, which benefits our concrete model since the radii for skyrmion crystals are in some cases too large. The main difference between the two different constructions is that the MC starts at a given density, hence it deviates from the $npe\mu$ EoS at a certain mass. On the other hand, since the GC changes the location of the minimum, it leads to different results also in the low mass region. However, both constructions practically merge in the high masses region, in which they follow the same $npe\mu\overline{K}$ EoS. As already explained, the thermodynamically stable region of the $E(L)$ curves and the corresponding EoS based on the Skyrme crystal is $L\leq L_{0}$ or, equivalently, $n_{B}\geq n_{0}$. As a consequence, NS based on the Skyrme crystal have $n_{B}=n_{0}$ at the NS surface or, in other words, Skyrme crystal NS have no crust. In the right panel of fig. 8, therefore, we show the result of adding a crust to the NS by joining the GC equations of state of the Skyrme crystal with a standard nuclear physics EoS for low densities (the corresponding EoS are shown in fig. 6). Concretely, we use the BCPM EoS Sharma _et al._ (2015) and joint the two EoS at the pressure $p=p_{*}$ where the two EoS coincide, i.e., $\rho_{\rm BCPM}(p_{*})=\rho_{\rm,crystal}(p_{*})$, exactly as we did in Adam _et al._ (2020). In terms of the baryon density, the joining occurs at $n_{B,*}\sim 1.1n_{0}$ for the parameter sets 1-3, and for $n_{B,*}\sim 1.2n_{0}$ for the set 4. Again as in Adam _et al._ (2020), we assume a smooth joining between the two EoS (concretely, described by a quadratic interpolation) in order to avoid an artificial phase transition at $p_{*}$. We also plot in fig. 8 the most likely mass-radius relations for the NS corresponding to GW170817 Abbott _et al._ (2017) and GW190425 Abbott _et al._ (2020) events (orange and blue regions). The green regions represent the estimations for the mass and radius values of PSR J0740+6620 (top) Miller _et al._ (2019) and J0030+0451 (bottom) Riley _et al._ (2021). The purple region constraints the mass-radius curves from the statistical analysis done in Altiparmak _et al._ (2022). We find that the NS resulting from the addition of a crust to the Skyrme crystal EoS with a nonzero kaon condensate agree very well with these recent constraints. Further, the softening of the EoS due to the presence of kaons and the resulting smaller NS radii are important for this agreement. Figure 8: Mass-Radius curves of NS with a kaon condensed core. The different sets of parameters that we consider are shown with different colors. Left panel: Solid lines represent $npe\mu$ matter, dashed-dotted lines are obtained with a MC and the dashed with the GC. Right panel: The effect of adding a standard nuclear physics crust to the Skyrme crystal EoS with kaon condensate obtained from the Gibbs construction (GC). ## VI Conclusions The quantization of the isospin degrees of freedom in Adam _et al._ (2022b) allowed us to find the dependence of the isospin contribution to the energy (hence the isospin chemical potential) on the lattice length of the skyrmion crystals, which in turn determines the proton (and electron) fraction in $\beta$-equilibrated skyrmion crystals. In this paper, we have used this information to precisely determine the critical density at which charged kaons will condense inside neutron stars described by the Skyrme crystal. Although the prediction of the condensation of charged mesons at high (isospin) chemical potential is common in nuclear matter literature, and there have been some partial results within the Skyrme model Westerberg (1995); Park _et al._ (2010), we have, to our knowledge, for the first time provided a framework which allows to precisely calculate the value of the density where kaon condensation sets in. This value turns out to be around twice nuclear saturation for physically relevant sets of parameters, so the prediction from the Skyrme model is that strangeness will be present at the core of neutron stars in the form of an (anti)kaon condensate. Further, we have computed the EoS for skyrmion matter in the condensate phase, which becomes softer due to the additional degrees of freedom. This has appreciable effects on the Mass- Radius curves for physically relevant parameters in the model, reducing the maximum radii by about 0.5-1 km. We found that a correct treatment of the resulting first-order phase transition between the phase with and without kaon condensate by a Gibbs construction, as originally advocated in Glendenning (1992), is important for these results. Despite the prediction of a very early onset of kaon condensation as compared with other nuclear EoS, the maximum mass limit for neutron stars doesn’t get significantly reduced, due to the sextic term being dominant at such densities. The fact that the presence of a kaon condensate does not pose a problem for reaching high masses in the Skyrme model, however, does not imply that the hyperon problem is completely solved in Skyrmion-based EoS. Indeed, in addition to kaons, one should take into account also hyperon degrees of freedom within the Skyrme model. Whilst hyperons can be succesfully described within this model Klebanov (1990), and even a proposal to study modifications of hyperon properties in dense matter has bee provided in Hong _et al._ (2019), it is not clear for us how to apply these ideas to Skyrmion crystals. A possibility would be to extend the isospin quantization scheme proposed in Adam _et al._ (2022b) to the three flavor case, and quantize the whole $SU(3)$ rotation group. This method, however, is much more technically involved, and relies on the assumption of an approximate $SU(3)$ symmetry of the Hamiltonian, which only holds for sufficiently large densities. ###### Acknowledgements. The authors acknowledge financial support from the Ministry of Education, Culture, and Sports, Spain (Grant No. PID2020-119632GB-I00), the Xunta de Galicia (Grant No. INCITE09.296.035PR and Centro singular de investigación de Galicia accreditation 2019-2022), the Spanish Consolider-Ingenio 2010 Programme CPAN (CSD2007-00042), and the European Union ERDF. AW is supported by the Polish National Science Centre, grant NCN 2020/39/B/ST2/01553. AGMC is grateful to the Spanish Ministry of Science, Innovation and Universities, and the European Social Fund for the funding of his predoctoral research activity (_Ayuda para contratos predoctorales para la formación de doctores_ 2019). MHG is also grateful to the Xunta de Galicia (Consellería de Cultura, Educación y Universidad) for the funding of his predoctoral activity through _Programa de ayudas a la etapa predoctoral_ 2021\. ## Appendix A Derivation of the WZW and sextic terms contribution to $V_{K}$ To work with the WZW and sextic terms, it is useful to employ the formalism of Lie algebra-valued differential forms, which we will extensively do in this appendix. Let us first review the basic properties of such objects and establish the notation that we will follow. A $\mathfrak{g}$-valued differential form $\alpha$ can be written in terms of the Lie algebra generators $T_{a}$, as $\alpha=\alpha^{a}\otimes T_{a}$. The exterior derivative is then simply obtained as $d\alpha=d\alpha^{a}\otimes T_{a}$. Furthermore, the wedge product on $\mathfrak{g}$-valued forms is defined as $\alpha\wedge\beta=\alpha^{a}\wedge\beta^{b}\otimes T_{a}T_{b}.$ (89) So that the following useful properties hold: $\displaystyle d\alpha\wedge\beta$ $\displaystyle=(d\alpha)\wedge\beta+(-1)^{\absolutevalue{\alpha}}\alpha\wedge(d\beta)$ (90) $\displaystyle\Tr{\alpha\wedge\beta}$ $\displaystyle=(-1)^{\absolutevalue{\alpha}\absolutevalue{\beta}}\Tr{\beta\wedge\alpha},$ (91) where $\absolutevalue{\alpha}$ denotes the degree of $\alpha$. Also, by linearity of the trace, both the trace and the exterior derivative commute, i.e. $d\Tr{\alpha}=\Tr{d\alpha}.$ (92) To alleviate the notation, in the following we will drop the wedge product symbol and denote the product (89) simply by $\alpha\beta$. Then, for instance, if $\alpha$ and $\beta$ denote two 1-forms, we have $\Tr{\alpha\beta}=-\Tr{\beta\alpha}$, $d(\alpha\beta)=d\alpha\beta-\alpha d\beta.$ Let us now perform the most general chiral transformation to the Skyrme field $U_{\pi}$, given by $U=g_{l}U_{\pi}g_{r}^{\dagger}$, with $(g_{l},g_{r})\in SU(3)_{L}\times SU(3)_{R}$ and define the following $\mathfrak{su}(3)$-valued differential forms, $\begin{split}V=U^{\dagger}&dU,\,\,L=U_{\pi}^{\dagger}dU_{\pi},\,\,\alpha=g_{l}^{\dagger}dg_{l},\,\beta=g_{r}dg_{r}^{\dagger}.\end{split}$ (93) By definition, we have the following relation between the forms above: $V=(U_{\pi}g_{r})^{\dagger}[\alpha+U_{\pi}(L-\beta)U^{\dagger}_{\pi}]U_{\pi}g_{r}.$ (94) On the other hand, the WZW action is then given by the pullback of a volume 5-form $\Omega_{5}$ by an extended Skyrme field $U:D^{5}\rightarrow SU(3)$ 333The result is of course independent of such extension, because $\pi_{4}(SU(3))$ vanishes. integrated over an auxiliar 5-dimensional disk $D$ whose boundary is the spacetime manifold $M$, $S_{\rm WZW}=-i\frac{N_{C}}{240\pi^{2}}\int_{D}U^{*}(\Omega_{5})$ (95) The form $U^{*}(\Omega_{5})$ can be expressed in terms of $L$ as $\displaystyle S_{\rm WZW}(L)$ $\displaystyle=-i\frac{N_{C}}{240\pi^{2}}\int_{D}\Tr\\{V^{5}\\}=$ $\displaystyle=$ $\displaystyle-\frac{iN_{C}}{240\pi^{2}}\int_{D}\Tr\\{[\alpha+U_{\pi}(L-\beta)U^{\dagger}_{\pi}]^{5}\\}.$ (96) Let us denote the 1-form $U_{\pi}(L-\beta)U^{\dagger}_{\pi}$ by $\omega$. The exterior derivative of this form is: $\displaystyle d\omega$ $\displaystyle=dU_{\pi}(L-\beta)U^{\dagger}_{\pi}+U_{\pi}(dL-d\beta)U^{\dagger}_{\pi}-U_{\pi}(L-\beta)dU^{\dagger}_{\pi}=$ $\displaystyle=-U^{\dagger}_{\pi}(dL+d\beta+L\beta+\beta L)U_{\pi},$ (97) where we have used the fact that both $\beta$ and $L$ satisfy the Maurer- Cartan equation $d\beta=-\beta^{2}$. Moreover, one can straightforwardly see that $\omega^{2}=U^{\dagger}_{\pi}(L-\beta)^{2}U_{\pi}=U^{\dagger}_{\pi}(-d\beta- dv-\beta v-v\beta)U_{\pi}=d\omega$ (98) Knowing this, we have $\begin{split}S_{\rm WZW}(L)=&-i\frac{N_{C}}{240\pi^{2}}\int_{\mathcal{M}}\Tr\\{[\alpha+\omega]^{5}\\}\\\ =&S_{\rm WZW}(\alpha)+S_{\rm WZW}(\omega)-\frac{iN_{C}}{48\pi^{2}}\int_{\mathcal{M}}\Tr\\{\alpha^{4}\omega+\omega^{4}\alpha+\alpha^{2}\omega^{3}+\omega^{2}\alpha^{3}+\alpha\omega\alpha\omega^{2}+\omega\alpha\omega\alpha^{2}\\}\\\ =&S_{\rm WZW}(\alpha)+S_{\rm WZW}(\omega)-\frac{iN_{C}}{48\pi^{2}}\int_{\partial\mathcal{M}}\Tr\\{\omega^{3}\alpha-\alpha^{3}\omega-\frac{1}{2}(\alpha\omega)^{2}\\},\end{split}$ (99) where we have used eqs. 90, 91 and 92, the relation (98) for $\omega$, the M-C equation for $\alpha$ and Stokes’ theorem in the last step. Repeating the same calculation for $S_{\rm WZW}(\omega)$ yields: $\displaystyle S_{\rm WZW}(\omega)=$ $\displaystyle-i\frac{N_{C}}{240\pi^{2}}\int_{\mathcal{M}}\Tr\\{[L-\beta]^{5}\\}=$ (100) $\displaystyle=$ $\displaystyle-S_{\rm WZW}(\beta)+S_{\rm WZW}(L)-$ $\displaystyle-\frac{iN_{C}}{48\pi^{2}}\int_{\partial\mathcal{M}}\\!\\!\Tr\\{L^{3}\beta-\beta^{3}L-\frac{1}{2}(L\beta)^{2}\\},$ so that $\displaystyle S_{\rm WZW}$ $\displaystyle(V)=S_{\rm WZW}(L)+S_{\rm WZW}(\alpha)-S_{\rm WZW}(\beta)-$ $\displaystyle-\frac{iN_{C}}{48\pi^{2}}$ $\displaystyle\int_{\partial\mathcal{M}}\Tr\\{L^{3}\beta-\beta^{3}L-\frac{1}{2}(L\beta)^{2}+\omega^{3}\alpha-\alpha^{3}\omega-\frac{1}{2}(\alpha\omega)^{2}\\}.$ (101) Eq.(101) shows that a chiral transformation of the $SU(3)$ Skyrme field induces an additional local term in the action due to the nontrivial transformation of the (nonlocal) WZW term. Furthermore, if we fix the chiral transformation fields to only depend on one spacetime coordinate, $g_{l/r}(x)\equiv g_{l/r}(t)$, any power of $\alpha$ and $\beta$ will vanish in the local, $4$-dimensional effective term. Taking this into account, we arrive to the final result: $S_{\rm WZW}(V)=S_{\rm WZW}(L)-\frac{iN_{C}}{48\pi^{2}}\int_{\partial\mathcal{M}}\Tr\\{L^{3}(\beta+U^{\dagger}_{\pi}\alpha U_{\pi})\\},$ (102) from where eq. 51 is readily obtained. Let us now turn to the sextic term. The coordinate free version of $\mathcal{L}_{6}$ is given by $\mathcal{L}_{6}=\lambda^{2}\pi^{4}\mathcal{B}\wedge\star\mathcal{B}$ (103) where $\star$ denotes the Hodge star operator, and $\mathcal{B}$ the 1-form in spacetime whose coordinates in a local chart coincide with the Baryon current, $\mathcal{B}=B_{\mu}dx^{\mu}$. We can construct such form as the Hodge dual of the baryon number density three-form, $\displaystyle b$ $\displaystyle=(24\pi^{2})^{-1}U^{*}(\Omega_{3})=\frac{1}{24\pi^{2}}\Tr\\{L_{\mu}L_{\nu}L_{\rho}\\}dx^{\mu}\wedge dx^{\nu}\wedge dx^{\rho}$ $\displaystyle=\frac{1}{3!}b_{\mu\nu\rho}dx^{\mu}\wedge dx^{\nu}\wedge dx^{\rho},$ (104) i.e. $\displaystyle\mathcal{B}=\star b=$ $\displaystyle\frac{1}{3!}b^{\nu\rho\sigma}\varepsilon_{\nu\rho\sigma\mu}dx^{\mu}=$ $\displaystyle=$ $\displaystyle\frac{1}{24\pi^{2}}\varepsilon_{\mu}^{\,\,\,\nu\rho\sigma}\Tr\\{L_{\nu}L_{\rho}L_{\sigma}\\}=B_{\mu}dx^{\mu}.$ (105) Thus $\mathcal{L}_{6}=\lambda^{2}\pi^{4}\star b\wedge b$. Expressing the sextic term in such form is most useful for calculating its contribution to the kaon potential employing the formalism of differential forms in the same way as for the WZW term. Indeed, we see that $b=\frac{1}{24\pi^{2}}\Tr\\{L^{3}\\}=\frac{1}{24\pi^{2}}\Tr\\{\omega^{3}+3w\omega^{2}\\},$ (106) where we used the fact that $w^{n}=0$ for $n\geq 2$ because the kaon field $\Sigma$ only depends on time. Hence, also $v^{n}=0$ for $n\geq 2$ and $vw=0$ hold, and, given that $\omega=U_{\pi}^{\dagger}(\beta-v)U_{\pi}$, we may write $\displaystyle b=$ $\displaystyle\frac{1}{24\pi^{2}}\Tr\\{\beta^{3}+3(wU_{\pi}^{\dagger}(\beta-v)^{2}U_{\pi}-\beta^{2}v)\\}=$ $\displaystyle=$ $\displaystyle\frac{1}{24\pi^{2}}\Tr\\{\beta^{3}+3\beta^{2}(U_{\pi}wU_{\pi}^{\dagger}-v)\\}.$ (107) Thus, the baryon current density in the kaon condensed phase will be modified by $B^{\mu}=B^{\mu}_{\pi}+C^{\mu}$, where $B^{\mu}_{\pi}$ is the baryon current due to the pionic background and $C^{\mu}=\frac{1}{8\pi^{2}}\varepsilon^{\mu\nu\rho\sigma}\Tr\\{R_{\nu}R_{\rho}(U_{\pi}\Sigma\partial_{\sigma}\Sigma^{\dagger}U_{\pi}^{\dagger}-\Sigma^{\dagger}\partial_{\sigma}\Sigma).\\}$ (108) For a time independent pion background and a homogeneous kaon condensate, we have $B_{\pi}^{\mu}C_{\mu}=0$, and hence the only constribution from the sextic term ($\propto B_{\mu}B^{\mu}$) to $V_{K}$ comes from the additional term: $\displaystyle C_{\mu}C^{\mu}$ $\displaystyle=\frac{1}{64\pi^{4}}\varepsilon^{\mu\nu\rho\sigma}\epsilon_{\mu\alpha\beta}\Tr\\{R_{\nu}RE_{\rho}\xi_{\sigma}\\}\Tr\\{R_{\alpha}R_{\beta}\xi_{\gamma}\\}=$ $\displaystyle=\frac{-1}{64\pi^{4}}\varepsilon_{ijk}\varepsilon_{ilm}\Tr\\{R_{j}R_{k}\xi_{0}\\}Tr\\{R_{l}R_{m}\xi_{0}\\}$ (109) where $\xi_{\mu}=U_{\pi}\Sigma\partial_{\mu}\Sigma^{\dagger}U_{\pi}^{\dagger}-\Sigma^{\dagger}\partial_{\mu}\Sigma$. ## Appendix B Explicit expressions We show here the explicit expressions of the (third component of the) inertia tensor and the kaon isospin current, divided into separated contributions coming from the quadratic, quartic, sextic and Wess-Zumino-Witten terms. A superindex labels the origin of each contribution. $\displaystyle\Lambda_{33}$ $\displaystyle\equiv\Lambda_{3}=\Lambda_{3}^{(2)}+\Lambda_{3}^{(4)}+\Lambda_{3}^{(6)}$ (110) $\displaystyle\Delta_{3}$ $\displaystyle\equiv\Delta=\Delta^{(2)}+\Delta^{(4)}+\Delta^{(6)}+\Delta^{(\rm WZW)}$ (111) with $\displaystyle\Lambda^{(2)}_{3}$ $\displaystyle=2a\Tr\\{T_{3}T_{3}\\}=$ (112) $\displaystyle=\frac{\pi_{1}^{2}+\pi_{2}^{2}}{2}(1+\cos^{2}\phi)^{2}+(1+\sigma)\sin^{2}(2\phi)/4,$ $\displaystyle\Lambda^{(4)}_{3}$ $\displaystyle=-4b\Tr\\{[T_{3},L_{k}][T_{3},L_{k}]\\}=2(1+\cos^{2}\phi)\times$ $\displaystyle\times\big{[}(1-\pi^{2}_{3})\partial_{i}\sigma^{2}+\sigma\pi_{3}\partial_{i}\sigma\partial_{i}\pi_{3})+(\sigma\leftrightarrow\pi_{3})\big{]}+$ $\displaystyle+\partial_{i}n^{2}(1+\sigma)/4\sin^{2}(2\phi),$ (113) $\displaystyle\Lambda^{(6)}_{3}$ $\displaystyle=-\frac{c}{32\pi^{4}}\varepsilon^{lmn}\Tr\\{T_{3}L_{m}L_{n}\\}\varepsilon_{lrs}\Tr\\{T_{3}L_{r}L_{s}\\}=$ $\displaystyle=$ $\displaystyle\frac{\lambda^{2}f^{2}_{\pi}e^{4}\mu_{K}^{2}}{2}(1+\cos^{2}(\tilde{\phi}))^{2}(\partial_{i}\pi_{3}\partial_{j}\sigma-\partial_{i}\sigma\partial_{j}\pi_{3})^{2}$ (114) $\displaystyle\Delta^{(2)}$ $\displaystyle=2a\Tr\\{T_{3}L_{0}\\}=$ (115) $\displaystyle=-i\mu_{K}[(\pi_{1}^{2}+\pi_{2}^{2})(\cos^{4}\phi-1)+(1+\sigma)\sin^{2}(2\phi)/2],$ $\displaystyle\Delta^{(4)}$ $\displaystyle=-4b\Tr\\{[T_{a},L_{k}][T_{b},L_{k}]\\}=-2i\mu_{K}\big{[}2(1-\cos^{4}\phi)\times$ $\displaystyle\times(\pi_{1}^{2}\partial_{i}\pi_{2}^{2}+\pi_{2}^{2}\partial_{i}\pi_{1}^{2}-2\pi_{1}\pi_{2}\partial_{i}\pi_{1}\partial_{i}\pi_{2}-\partial_{i}n^{2}(\pi_{1}^{2}+\pi^{2}_{2}))+$ $\displaystyle+\partial_{i}n^{2}/4(1+\sigma)\sin^{2}(2\phi)\big{]},$ (116) $\displaystyle\Delta^{(6)}$ $\displaystyle=-\frac{c}{32\pi^{4}}\varepsilon^{lmn}\Tr\\{T_{3}L_{m}L_{n}\\}\varepsilon_{lrs}\Tr\\{L_{0}L_{r}L_{s}\\}=$ $\displaystyle=i\mu_{K}\lambda^{2}f^{2}_{\pi}e^{4}(1+\cos^{2}\phi)\sin^{2}\phi(\partial_{i}\pi_{3}\partial_{j}\sigma-\partial_{i}\sigma\partial_{j}\pi_{3})^{2}$ (117) $\displaystyle\Delta$ ${}^{(\rm WZW)}=-\frac{N_{C}B_{\rm cell}}{2}\sin^{2}{\left(\phi\right)}$ (118) ## References * Li _et al._ (2018) J. J. Li, A. Sedrakian, and F. Weber, Phys. Lett. B 783, 234 (2018), arXiv:1803.03661 [nucl-th] . * Glendenning (1985) N. K. Glendenning, Astrophys. J. 293, 470 (1985). * Hartle _et al._ (1975) J. B. Hartle, R. F. Sawyer, and D. J. Scalapino, Astrophys. J. 199, 471 (1975). * Celenza _et al._ (1977) L. S. Celenza, W. Nutt, and C. Shakin, Physics Letters B 72, 23 (1977). * Kaplan and Nelson (1988) D. B. Kaplan and A. E. Nelson, Nucl. Phys. A 479, 273c (1988). * Glendenning and Schaffner-Bielich (1999) N. K. Glendenning and J. Schaffner-Bielich, Phys. Rev. C 60, 025803 (1999), arXiv:astro-ph/9810290 . * Pal _et al._ (2000) S. Pal, D. Bandyopadhyay, and W. Greiner, Nucl. Phys. A 674, 553 (2000), arXiv:astro-ph/0001039 . * Heiselberg _et al._ (1993) H. Heiselberg, C. J. Pethick, and E. F. Staubo, Phys. Rev. Lett. 70, 1355 (1993). * Benvenuto and Lugones (1999) O. G. Benvenuto and G. Lugones, Mon. Not. Roy. Astron. Soc. 304, L25 (1999). * Annala _et al._ (2020) E. Annala, T. Gorda, A. Kurkela, J. Nättilä, and A. Vuorinen, Nature Physics 16, 907 (2020). * McLerran and Pisarski (2007) L. McLerran and R. D. Pisarski, Nucl. Phys. A 796, 83 (2007), arXiv:0706.2191 [hep-ph] . * McLerran and Reddy (2019) L. McLerran and S. Reddy, Phys. Rev. Lett. 122, 122701 (2019), arXiv:1811.12503 [nucl-th] . * Tolos and Fabbietti (2020) L. Tolos and L. Fabbietti, Prog. Part. Nucl. Phys. 112, 103770 (2020), arXiv:2002.09223 [nucl-ex] . * Bombaci (2017) I. Bombaci, JPS Conf. Proc. 17, 101002 (2017), arXiv:1601.05339 [nucl-th] . * Vidaña (2016) I. Vidaña, J. Phys. Conf. Ser. 668, 012031 (2016), arXiv:1509.03587 [nucl-th] . * Skyrme (1962) T. H. R. Skyrme, Nuclear Physics 31, 556 (1962). * Adkins _et al._ (1983) G. S. Adkins, C. R. Nappi, and E. Witten, Nuclear Physics B 228, 552 (1983). * Ding and Yan (2007) G.-J. Ding and M.-L. Yan, Phys. Rev. C 75, 034004 (2007), arXiv:nucl-th/0702037 . * Halcrow and Harland (2020) C. Halcrow and D. Harland, Phys. Rev. Lett. 125, 042501 (2020). * Lau and Manton (2014) P. H. C. Lau and N. S. Manton, Phys. Rev. Lett. 113, 232503 (2014). * Halcrow _et al._ (2017) C. J. Halcrow, C. King, and N. S. Manton, Phys. Rev. C 95, 031303 (2017). * Adam _et al._ (2013) C. Adam, C. Naya, J. Sanchez-Guillen, and A. Wereszczynski, Phys. Rev. Lett. 111, 232501 (2013), arXiv:1312.2960 [nucl-th] . * Gillard _et al._ (2015) M. Gillard, D. Harland, and M. Speight, Nucl. Phys. B 895, 272 (2015). * Gudnason (2016) S. B. Gudnason, Phys. Rev. D 93, 065048 (2016). * Naya and Sutcliffe (2018) C. Naya and P. Sutcliffe, Phys. Rev. Lett. 121, 232002 (2018). * Adam _et al._ (2010) C. Adam, J. Sánchez-Guillén, and A. Wereszczyński, Physics Letters B 691, 105–110 (2010). * Naya (2019) C. Naya, International Journal of Modern Physics E 28, 1930006 (2019). * Adam _et al._ (2020) C. Adam, A. G. Martín-Caro, M. Huidobro, R. Vázquez, and A. Wereszczynski, Phys. Lett. B 811, 135928 (2020). * Adam _et al._ (2022a) C. Adam, A. G. Martin-Caro, M. Huidobro, R. Vazquez, and A. Wereszczynski, Phys. Rev. D 105, 074019 (2022a), arXiv:2109.13946 [hep-th] . * Adam _et al._ (2022b) C. Adam, A. G. Martín-Caro, M. Huidobro, R. Vázquez, and A. Wereszczynski, (2022b), arXiv:2202.00953 [nucl-th] . * Adam _et al._ (2015) C. Adam, C. Naya, J. Sanchez-Guillen, R. Vazquez, and A. Wereszczynski, Physics Letters B 742, 136–142 (2015). * Callan and Klebanov (1985) C. G. Callan and I. Klebanov, Nuclear Physics B 262, 365 (1985). * Klebanov (1990) I. Klebanov, “Strangeness in the skyrme model,” in _Hadrons and Hadronic Matter_, edited by D. Vautherin, F. Lenz, and J. W. Negele (Springer US, Boston, MA, 1990) pp. 223–262. * Fiorella Burgio and Fantina (2018) G. Fiorella Burgio and A. F. Fantina, Astrophys. Space Sci. Libr. 457, 255 (2018). * Essick _et al._ (2021) R. Essick, P. Landry, A. Schwenk, and I. Tews, Phys. Rev. C 104, 065804 (2021). * Tang _et al._ (2021) S.-P. Tang, J.-L. Jiang, M.-Z. Han, Y.-Z. Fan, and D.-M. Wei, Phys. Rev. D 104, 063032 (2021). * de Tovar _et al._ (2021) P. B. de Tovar, M. Ferreira, and C. Providência, Phys. Rev. D 104, 123036 (2021). * Gil _et al._ (2021) H. Gil, P. Papakonstantinou, and C. H. Hyun, (2021), arXiv:2110.09802 [nucl-th] . * Li _et al._ (2021) B.-A. Li, B.-J. Cai, W.-J. Xie, and N.-B. Zhang, Universe 7, 182 (2021). * Klebanov (1985) I. Klebanov, Nuclear Physics B 262, 133 (1985). * Blom _et al._ (1989) U. Blom, K. Dannbom, and D. Riska, Nuclear Physics 493, 384 (1989). * Nyman and Riska (1990) E. M. Nyman and D. Riska, Reports on Progress in Physics 53, 1137 (1990). * Schmitt (2010) A. Schmitt, _Dense matter in compact stars: A pedagogical introduction_, Vol. 811 (2010) arXiv:1001.3294 [astro-ph.SR] . * Note (1) Actually, that the charged (in particular, the negatively charged) kaons will condense first is true in our approach (whenever $\mu_{e}>0$), since the chemical potential associated to neutral kaons is zero, so that the onset of neutral kaon condensation is given by $m^{*}_{K}=0$. * Note (2) Remember that we are using the mostly minus convention for the metric signature. * Park _et al._ (2019) B.-Y. Park, W.-G. Paeng, and V. Vento, Nucl. Phys. A 989, 231 (2019), arXiv:1904.04483 [hep-ph] . * Canfora _et al._ (2020) F. Canfora, M. Lagos, and A. Vera, Eur. Phys. J. C 80, 697 (2020), arXiv:2007.11543 [hep-th] . * Glendenning (1992) N. K. Glendenning, Phys. Rev. D 46, 1274 (1992). * Bhattacharyya _et al._ (2010) A. Bhattacharyya, I. N. Mishustin, and W. Greiner, J. Phys. G 37, 025201 (2010), arXiv:0905.0352 [nucl-th] . * Sharma _et al._ (2015) B. K. Sharma, M. Centelles, X. Viñas, M. Baldo, and G. F. Burgio, Astron. Astrophys. 584, A103 (2015). * Abbott _et al._ (2017) B. P. Abbott _et al._ (LIGO , VIRGO), Phys. Rev. Lett. 119, 161101 (2017). * Abbott _et al._ (2020) B. Abbott _et al._ (LIGO Scientific, Virgo), Astrophys. J. Lett. 892, L3 (2020). * Miller _et al._ (2019) M. C. Miller _et al._ , Astrophys. J. Lett. 887, L24 (2019), arXiv:1912.05705 [astro-ph.HE] . * Riley _et al._ (2021) T. E. Riley _et al._ , Astrophys. J. Lett. 918, L27 (2021). * Altiparmak _et al._ (2022) S. Altiparmak, C. Ecker, and L. Rezzolla, (2022), arXiv:2203.14974 [astro-ph.HE] . * Westerberg (1995) K. M. Westerberg, Phys. Rev. D 51, 5030 (1995), arXiv:hep-ph/9411430 . * Park _et al._ (2010) B.-Y. Park, J.-I. Kim, and M. Rho, Phys. Rev. C 81, 035203 (2010), arXiv:0912.3213 [hep-ph] . * Hong _et al._ (2019) K.-H. Hong, U. Yakhshiev, and H.-C. Kim, Phys. Rev. C 99, 035212 (2019), arXiv:1806.06504 [nucl-th] . * Note (3) The result is of course independent of such extension, because $\pi_{4}(SU(3))$ vanishes.
# Manipulating non-reciprocity in a two-dimensional magnetic quantum walk Quan Lin Beijing Computational Science Research Center, Beijing 100084, China Wei Yi<EMAIL_ADDRESS>CAS Key Laboratory of Quantum Information, University of Science and Technology of China, Hefei 230026, China CAS Center For Excellence in Quantum Information and Quantum Physics, Hefei 230026, China Peng Xue<EMAIL_ADDRESS>Beijing Computational Science Research Center, Beijing 100084, China ###### Abstract Non-reciprocity is an important topic in fundamental physics and quantum- device design, as much effort has been devoted to its engineering and manipulation. Here we experimentally demonstrate non-reciprocal transport in a two-dimensional quantum walk of photons, where the directional propagation is highly tunable through dissipation and synthetic magnetic flux. The non- reciprocal dynamics hereof is a manifestation of the non-Hermitian skin effect, with its direction continuously adjustable through the photon-loss parameters. By contrast, the synthetic flux originates from an engineered geometric phase, which competes with the non-Hermitian skin effect through magnetic confinement. We further demonstrate how the non-reciprocity and synthetic flux impact the dynamics of the Floquet topological edge modes along an engineered boundary. Our results exemplify an intriguing strategy for achieving tunable non-reciprocal transport, highlighting the interplay of non- Hermiticity and gauge fields in quantum systems of higher dimensions. Open quantum systems are ubiquitous in nature, and exhibit rich and complex behaviors unknown to their closed counterparts openbook . The recent progresses in non-Hermitian physics offer fresh insights into open systems from a unique perspective, giving rise to exotic symmetries and new paradigms of topology benderreview ; review2 ; photonpt1 ; review3 ; nonHtopo1 ; nonHtopo2 ; Lee ; WZ1 . A much studied non-Hermitian phenomenon of late is the non-Hermitian skin effect (NHSE) WZ1 ; WZ2 ; murakami ; ThomalePRB ; stefano ; tianshu ; Budich ; mcdonald ; alvarez ; lli ; yzsgbz ; stefano2 ; lyc ; fangchenskin2 ; teskin ; photonskin ; metaskin ; scienceskin ; teskin2d ; dzou ; fangchenskin ; kawabataskin , whereby a macroscopic number of eigenstates become exponentially localized toward the boundaries. The NHSE has significant impact on the band topology WZ1 ; WZ2 ; murakami , the spectral symmetry nonblochpt1 ; nonblochpt2 ; XDW+21 , and dynamics quench1 ; quench2 ; coldatom . One of the most salient dynamic signatures of the NHSE is the directional bulk flow stefano ; coldatom ; ql ; ql2 , which is closely connected to the global topology of the spectrum on the complex plane fangchenskin . Such non- reciprocal dynamics can have potential applications in topological transport and quantum-device design, but the generation and control of this peculiar form of non-reciprocity, particularly in higher dimensions, remain experimentally unexplored. In this work, we experimentally demonstrate the tuning of non-reciprocal transport in photonic quantum walks on a synthetic two-dimensional square lattice. The non-reciprocal dynamics underlies the NHSE of the two-dimensional quantum walk—the unidirectional flow leads to the accumulation of eigenstates toward boundaries in the corresponding direction. By tuning the photon-loss parameters, we show how the direction of the flow (hence the direction of the NHSE) can be continuously adjusted. The much discussed corner skin effect in two dimensions is but a special case here, corresponding to a directional flow along the diagonal of the square lattice. By engineering the quantum-walk setup, we also introduce a synthetic flux to the lattice hof ; sai , which we observe to suppress the non-reciprocal dynamics. Such a suppression is the result of the competition between two localization mechanisms: magnetic confinement and NHSE chenwei . We quantitatively characterize the tunability of the non-reciprocity through loss and flux, and further demonstrate their impact on the dynamics of topological edge modes along the boundary. Our experiment is the first observation of the magnetic suppression of NHSE, and further illustrates the flexible control over the NHSE-induced non-reciprocity in higher dimensions. Figure 1: Two-dimensional non-Hermitian quantum walk with a synthetic gauge field. a Schematics for possible movements of a walker at spatial position $(x,y)$ during each time step. b A time-multiplexed implementation of the two- dimensional photonic quantum walk. The photons are initialized at position $(0,0)$ in the superposition of the polarizations $(\ket{H}+i\ket{V})/\sqrt{2}$. Once coupled into the setup through a low- reflectivity beam splitter (BS, reflectivity $3\%$), their polarization state is manipulated by a half-wave plate (HWP). The photonic wave packets are split by a polarizing beam splitter (PBS) and routed through a pair of single-mode fibers (SMF) of length $287.03$m and $270$m, respectively, implementing a temporal step in the $x$ direction. A temporal step in the $y$ direction is implemented by another two-PBS loop based on the same principle, but in the free space instead of fibers. At each step, photons are partially coupled out to a polarization resolving detection of the arrival time via avalanche photodiodes (APDs). ND: neutral density filter; AOM: optical switch acousto- optic modulator; EOM: electro-optic modulator. c Illustration of the operation sequence of the time-multiplexed quantum walk. Here $V_{\text{EOM}}$ is the control voltage applied to the EOMs. Figure 2: Tunable non-reciprocity and NHSE. The walker with the polarization $(\ket{H}+i\ket{V})/\sqrt{2}$ starts from the lattice site $(0,0)$ with $\alpha=0$. Probability distributions are measured after $16$ time steps. a Probability distribution for a Hermitian two-dimensional quantum walk with $\gamma_{x}=\gamma_{y}=0$. b Probability distribution following a non-Hermitian quantum walk with $\gamma_{x}=\gamma_{y}=0.125$. c Directional displacements after the final time step ($t=16$) for quantum walks with varying $\gamma_{x}$ and $\gamma_{y}$. (Left) Color contour of the azimuthal angle of the displacement $\bm{d}$ on the $x$–$y$ plane. (Right) Measured (blue arrows) and simulated (gray arrows) of the displacement along the red vertical line of the color contour (left panel). d Probability distribution following a non-Hermitian quantum walk in the presence of domain walls (marked by red dashed lines), with $\gamma_{x}=0.125$ for $x\geq-6$, $\gamma_{x}=-0.125$ for $x<-6$, $\gamma_{y}=0.125$ for $y\geq-6$, and $\gamma_{y}=-0.125$ for $y<-6$. Figure 3: Magnetic suppression of the non-reciprocal transport. The walker is initialized in the state $\frac{1}{\sqrt{2}}(\ket{H}+i\ket{V})\otimes\ket{x=0}\ket{y=0}$. Measured probability distributions of $16$-time-step quantum walks with the tuning parameter $\alpha=0.05$, and the loss parameter $\gamma_{x}=0$, $\gamma_{y}=0.1$ in a, with $\alpha=0.15$, $\gamma_{x}=0$, and $\gamma_{y}=0.1$ in b, with $\alpha=0.05$ and $\gamma_{x}=\gamma_{y}=0.1$ in c, and with $\alpha=0.15$ and $\gamma_{x}=\gamma_{y}=0.1$ in d. e Norm of the directional displacement $\bm{d}$ for $16$-time-step quantum walks under different flux $\alpha$. Symbols represent the experimental data, and curves are the corresponding numerical simulations. Error bars are due to the statistical uncertainty in photon-number-counting. f Measured (blue) and simulated (gray) $\bm{d}$. Blue arrows represent the experimental results, and gray ones indicate the corresponding numerical simulations. Results Time-multiplexed two-dimensional quantum walk. In discrete-time quantum walks, the walker state $|\psi(t)\rangle$ evolves according to $|\psi(t)\rangle=U^{t}|\psi(0)\rangle$, where $t$ indicates the discrete time steps, and $U$ is thus identified as the Floquet operator that periodically drives the system. We consider such a quantum walk on a two-dimensional square lattice, with the Floquet operator $U=M_{y}S_{y}PCM_{x}S_{x}C.$ (1) Here the shift operators are defined as $S_{j}=\sum_{\bm{r}}\ket{0}\bra{0}\otimes\ket{\bm{r}-\bm{e}_{j}}\bra{\bm{r}}+\ket{1}\bra{1}\otimes\ket{\bm{r}+\bm{e}_{j}}\bra{\bm{r}}$, with $\bm{r}=(x,y)\in\mathbb{Z}^{2}$ labeling the coordinates of the lattice sites, $j\in\\{x,y\\}$, and $\bm{e}_{x}=(1,0)$ and $\bm{e}_{y}=(0,1)$. The shift operators move the walker in the corresponding directions, depending on the walker’s internal degrees of freedom in the basis of $\\{|0\rangle,|1\rangle\\}$ (dubbed the coin states). These coin states are subject to rotations under the coin operator $C=\frac{1}{\sqrt{2}}\begin{pmatrix}1&1\\\ 1&-1\end{pmatrix}\otimes\mathds{1}_{\bm{r}}$, where $\mathds{1}_{\bm{r}}=\sum_{\bm{r}}\ket{\bm{r}}\bra{\bm{r}}$. The gain-loss operators are given by (here $j\in\\{x,y\\}$) $\displaystyle M_{j}(\gamma_{j})=\begin{pmatrix}e^{\gamma_{j}}&0\\\ 0&e^{-\gamma_{j}}\end{pmatrix}\otimes\mathds{1}_{\bm{r}},$ (2) which make the quantum walk non-unitary for finite $\gamma_{x}$ or $\gamma_{y}$. A key ingredient to our scheme is the phase-shift operator, defined as $P=\sum_{\bm{r}}\begin{pmatrix}e^{i2\pi\alpha x}&0\\\ 0&e^{-i2\pi\alpha x}\end{pmatrix}\otimes\ket{\bm{r}}\bra{\bm{r}},$ (3) which enforces a position-dependent geometric phase on the walker, so that the latter acquires a phase $2\pi\alpha$ when going around any single plaquette of the square lattice (see Fig. 1a). Similar to that of the Hofstadter model hof , the accumulated phase shift of the walker on the lattice is equal to the Aharonov-Bohm phase of a charged particle in a uniform magnetic field, with a magnetic flux $\alpha$ threaded through each plaquette. We therefore regard $\alpha$ as the synthetic flux, which takes value in the range $\left[0,1\right)$. We experimentally implement the two-dimensional quantum walk above using photons. As illustrated in Fig. 1, the overall architecture is that of a fiber network sch ; ql ; ql2 , through which attenuated single-photon pulses are sent, with each full cycle around the network representing a discrete time step. The coin states $\\{|0\rangle,|1\rangle\\}$ are encoded in the photon polarizations $\\{|H\rangle,|V\rangle\\}$. The spatial degrees of freedom of the square lattice are encoded in the time domain, following a time- multiplexed scheme. This is achieved by building path-dependent time delays into the four different paths (labeled $x\pm 1$ and $y\pm 1$ in Fig. 1a) within the network (see Methods for details). The superpositions of multiple well-resolved pulses within the same discrete time step thus represent those of multiple spatial positions at the given time step (see Fig. 1b). The shift and coin operators are implemented with beam splitters (BSs) and wave plates (WPs), and the phase operator with one of the electro-optical modulators (EOM1 in Fig. 1). We further implement polarization-dependent loss operators $M^{\prime}_{j}=e^{-\gamma_{j}}M_{j}$ in each path, using a combination of the WPs and the EOMs. The time-evolved state driven by $U$ is then related to that in the experiment by adding a factor $e^{(\gamma_{x}+\gamma_{y})t}$ to the latter. For all experiments, avalanche photo-diodes (APDs) with temporal and polarization resolutions are employed to record the probability distribution of the walker states. This enables us to construct the site-resolved population of the synthetic lattice, with $\displaystyle P_{\text{exp}}(x,y,t)=\frac{N(x,y,t)}{\sum_{x,y}N(x,y,t)},$ (4) where $N(x,y,t)$ is the total photon number on site $(x,y)$ at time $t$. Figure 4: Dynamic detection of topological edge modes in $16$-time-step two- dimensional magnetic quantum walks. The walker is initialized in the state $\frac{1}{\sqrt{2}}(\ket{H}+i\ket{V})\otimes\ket{x=0}\ket{y=0}$ under a domain-wall geometry. Specifically, in a, b, c, we set $\alpha=0.05$ for $x\leq 0$, and $\alpha=-0.05$ for $x>0$. In d, e, f, we set $\alpha=1/3$ for $x\leq 0$, and $\alpha=-1/3$ for $x>0$. The left, middle, and right columns show the measured probability distribution with the parameters $\gamma_{x}=\gamma_{y}=0$, $\gamma_{x}=0.1,\gamma_{y}=0$, and $\gamma_{x}=\gamma_{y}=0.1$, respectively. NHSE and tunable non-reciprocity. In the absence of flux, quantum walks driven by $U$ already show non-reciprocal transport under finite photon losses. In Fig. 2a, we show the measured populations of the synthetic lattice sites after $t=16$ time steps. Starting from a local initial state at $\bm{r}=(0,0)$, the propagation in the synthetic spatial dimensions is symmetric along the four lattice directions (Fig. 2a). However, under finite photon-loss parameters, the final-time photon distribution becomes asymmetric with a preferred direction. Such a directional flow is the signature of the non-reciprocal transport. For instance, when $\gamma_{x}=\gamma_{y}\neq 0$, as shown in Fig. 2b, the flow is diagonal to the square lattice. By tuning the ratio of $\gamma_{y}/\gamma_{x}$, we can continuously adjust the direction of the asymmetric pattern. This is explicitly shown in Fig. 2c, where we define the directional displacement $\displaystyle\bm{d}(t)=\sum_{x,y}\bm{r}P_{\text{exp}}(x,y,t).$ (5) As $\gamma_{y}/\gamma_{x}$ varies, the direction of the displacement at the final time step can be continuously tuned (see the left panel of Fig. 2c). In our experiment, we adjust $\gamma_{y}/\gamma_{x}$ in the range of $[-1,1]$ for $16$-time-step quantum walks. Correspondingly, the measured polar angle of $\bm{d}$ changes from $3\pi/4$ to $-3\pi/4$ (the right panel of Fig. 2c). Underlying this loss-induced directional flow is the NHSE in two dimensions. While it is straightforward to show that the direction of the non-reciprocal transport also indicates the direction of the eigenstate accumulation under the open boundary condition supp , from an experimental perspective, we observe the dynamic localization of the walker toward the boundary when a domain-wall boundary condition is imposed (see Fig. 2d). Combined with the theoretical spectral analysis that there are no topological edge states present under the parameters of Fig. 2d, it is clear that the localization is due to the NHSE. Magnetic suppression of the NHSE. In Figs. 3a-d, we show the final population distribution with the synthetic flux switched on, following $16$-time-step quantum walks. Compared to Fig. 2, the directional flow appears to be increasingly suppressed under larger $\alpha$, regardless of its direction. In Figs. 3e and f, we show the absolute values of the directional displacement $\bm{d}$ as functions of $\alpha$, for various loss parameters. The suppression is the largest when $\alpha$ is tuned in between $0$ and $0.5$. Such a suppression reflects the competition between the magnetic confinement and the NHSE, and can be used for the manipulation of the non-reciprocal transport. Impact on topological edge states. In the absence of loss, the Floquet operator $U$ describes an anomalous Floquet Chern insulator, characterized by the Floquet topological invariant sai ; asb ; rud , which can be calculated for each quasienergy gap and is fully responsible for the topological edge states. Here we experimentally investigate how the NHSE under loss and magnetic confinement impact the topological edge states. For this purpose, we engineer a domain-wall configuration by choosing different values of $\alpha$ on either side of $x=0$. As shown in Fig. 4a, for lossless quantum walks, a pair of topological edge modes emerge, moving in opposite directions along the boundary. This is consistent with the prediction of the Floquet topological invariant (see Methods and supp ). When only the loss parameter $\gamma_{x}$ is turned on, the NHSE induces a horizontal directional flow toward the region with $x<0$. From the measured population following a $16$-time-step quantum walk (see Fig. 4b), both the bulk flow and the topological edge states are clearly visible. Since the directional flow is perpendicular to the boundary, it has no direct impact on the motion of the topological edge states. This is no longer the case when both $\gamma_{x}$ and $\gamma_{y}$ become finite, as in Fig. 4c. Here, besides a diagonal bulk flow induced by the corner skin effect, the topological edge modes moving in the negative (positive) $y$ direction are enhanced (suppressed) by the NHSE. In Figs. 4d-f, we show the final probability distribution under a larger synthetic flux $\alpha$. Compared to Figs. 4a-c, the bulk propagation is significantly suppressed, whereas the topological edge modes are largely unaffected by flux. This suggests that magnetic confinement is helpful for the dynamics detection of topological edge modes in systems with the NHSE. Discussion. We have experimentally demonstrated how the interplay of synthetic flux and dissipation enables the full control over the non-reciprocal transport underlying the NHSE. Since the quantum walk simulates an anomalous Floquet Chern insulator, we further illustrate how the motion of topological edge modes on the boundary is affected by the tuning parameters. While the high tunability can be exploited for topological device design, our implementation of a dissipative anomalous Floquet Chern insulator further raises theoretical questions as to how the NHSE affects the bulk-boundary correspondence herein. Our experiment also paves the way for engineering more exotic forms of the non-Hermitian skin effect in higher dimensions fangchenskin2 using quantum-walk dynamics. ## References * (1) Breuer, H. P. & Petruccione, F. The Theory of Open Quantum Systems. (Oxford University Press, 2007). * (2) Bender, C. M. Making sense of non-Hermitian Hamiltonians. Rep. Prog. Phys. 70, 947 (2007). * (3) El-Ganainy, R. et al. Non-Hermitian physics and PT symmetry. Nat. Phys. 14, 11-19 (2018). * (4) Miri, M.-A. & Alù, A. Exceptional points in optics and photonics. Science 363, eaar7709 (2019). * (5) Ashida, Y., Gong, Z. & Ueda, M. Non-Hermitian physics. Advances in Physics 69, 3 (2020). * (6) Kawabata, K., Shiozaki, K., Ueda, M. & Sato, M. Symmetry and topology in non-Hermitian physics. Phys. Rev. X 9, 041015 (2019). * (7) Zhou, H. & Lee, J. Y. Periodic table for topological bands with non-Hermitian symmetries. Phys. Rev. B 99, 235112 (2019). * (8) Lee, T. E. Anomalous edge state in a non-Hermitian lattice. Phy. Rev. Lett. 16, 133903 (2016). * (9) Yao, S. & Wang, Z. Edge states and topological invariants of non-Hermitian systems. Phy. Rev. Lett. 121, 086803 (2018). * (10) Yao, S., Song, F. & Wang, Z. Non-Hermitian Chern bands. Phy. Rev. Lett. 121, 136802 (2018). * (11) Yokomizo, K. & Murakami, S. Non-Bloch band theory of non-Hermitian systems. Phy. Rev. Lett. 123, 066404 (2019). * (12) Lee, C. H. & Thomale, R. Anatomy of skin modes and topology in non-Hermitian systems. Phys. Rev. B 99, 201103 (2019). * (13) Longhi, S. Probing non-Hermitian skin effect and non-Bloch phase transitions. Phys. Rev. Research 1, 023013 (2019). * (14) Deng, T.-S. & Yi, W. Non-Bloch topological invariants in a non-Hermitian domain wall system. Phys. Rev. B 100, 035102 (2019). * (15) Kunst, F. K., Edvardsson, E., Budich, J. C. & Bergholtz, E. J. Biorthogonal bulk-boundary correspondence in non-Hermitian systems. Phy. Rev. Lett. 121, 026808 (2018). * (16) McDonald, A., Pereg-Barnea, T. & Clerk, A. Phase-dependent chiral transport and effective non-Hermitian dynamics in a bosonic Kitaev-Majorana chain. Phys. Rev. X 8, 041031 (2018). * (17) Alvarez, V. M., Vargas, J. B. & Torres, L. F. Non-Hermitian robust edge states in one dimension: Anomalous localization and eigenspace condensation at exceptional points. Phys. Rev. B 97, 121401 (2018). * (18) Li, L., Lee, C. H., Mu, S. & Gong, J. Critical non-Hermitian skin effect. Nat. Commun. 11, 5491 (2020). * (19) Yang, Z., Zhang, K., Fang, C. & Hu, J. Non-Hermitian bulk-boundary correspondence and auxiliary generalized Brillouin zone theory. Phys. Rev. Lett. 125, 226402 (2020). * (20) Longhi, S. Self-healing of non-Hermitian topological skin modes. Phy. Rev. Lett. 128, 157601 (2022). * (21) Li, Y., Liang, C., Wang, C., Lu, C. & Liu, Y.-C. Gain-loss-induced hybrid skin-topological effect. Phy. Rev. Lett. 128, 223903 (2022). * (22) Zhang, K., Yang, Z. & Fang, C. Universal non-Hermitian skin effect in two and higher dimensions. Nat. Commun. 13, 2496 (2022). * (23) Helbig, T. et al. Generalized bulk–boundary correspondence in non-Hermitian topolectrical circuits. Nat. Phys. 16, 747 (2020). * (24) Xiao, L. et al. Non-Hermitian bulk–boundary correspondence in quantum dynamics. Nat. Phys. 16, 761 (2020). * (25) Ghatak, A., Brandenbourger, M., Van Wezel, J. & Coulais, C. Observation of non-Hermitian topology and its bulk–edge correspondence in an active mechanical metamaterial. Proc. Natl. Acad. Sci. 117, 29561 (2020). * (26) Weidemann, S. et al. Topological funneling of light. Science 368, 311-314 (2020). * (27) Hofmann, T. et al. Reciprocal skin effect and its realization in a topolectrical circuit. Phys. Rev. Research 2, 023265 (2020). * (28) D. Zou. et al. Observation of hybrid higher-order skin-topological effect in non-Hermitian topolectrical circuits. Nat. Commun. 12, 7201 (2021). * (29) Zhang, K., Yang, Z. & Fang, C. Correspondence between winding numbers and skin modes in non-Hermitian systems. Phys. Rev. Lett. 125, 126402 (2020). * (30) Okuma, N., Kawabata, K., Shiozaki, K. & Sato, M. Topological origin of non-Hermitian skin effects. Phys. Rev. Lett. 124, 086801 (2020). * (31) Longhi, S. Non-Bloch PT symmetry breaking in non-Hermitian photonic quantum walks. Opt. Lett. 44, 5804 (2019). * (32) Hu, Y.-M., Wang, H.-Y., Wang, Z. & Song, F. Geometric Origin of non-Bloch PT symmetry breaking. arXiv:2210.13491. * (33) Xiao, L. et al. Observation of non-Bloch parity-time symmetry and exceptional points. Phys. Rev. Lett. 126, 230402 (2021). * (34) Li, T., Sun, J.-Z., Zhang, Y.-S & Yi, W. Non-Bloch quench dynamics. Phys. Rev. Research 3, 023022 (2021). * (35) Wang, K. et al. Detecting non-Bloch topological invariants in quantum dynamics. Phys. Rev. Lett. 127, 270602 (2021). * (36) Liang, Q. et al. Dynamic signatures of non-Hermitian skin effect and topology in ultracold atoms. Phys. Rev. Lett. 129, 070401 (2022). * (37) Lin, Q. et al. Observation of non-Hermitian topological Anderson insulator in quantum dynamics. Nat. Commun. 13, 3229 (2022). * (38) Lin, Q. et al. Topological phase transitions and mobility edges in non-Hermitian quasicrystals. Phys. Rev. Lett. 129, 113601 (2022). * (39) Hofstadter, D. R. Energy levels and wave functions of Bloch electrons in rational and irrational magnetic fields. Phys. Rev. B 14, 2239 (1976). * (40) Sajid, M., Asbóth, J. K., Meschede, D., Werner, R. F. & Alberti, A. Creating anomalous Floquet Chern insulators with magnetic quantum walks. Phys. Rev. B 99, 214303 (2019). * (41) Shao, K. et al. Cyclotron quantization and mirror-time transition on nonreciprocal lattices. Phys. Rev. B 106, L081402 (2022). * (42) Schreiber, A. et al. A 2D quantum walk simulation of two-particle dynamics. Science 336, 55-58 (2012). * (43) see Supplemental Material. * (44) Rudner, M. S., Lindner, N. H., Berg, E. & Levin, M. Anomalous edge states and the bulk-edge correspondence for periodically driven two-dimensional systems. Phys. Rev. X 3, 031005 (2013). * (45) Asbóth, J. K. & Alberti, A. Spectral flow and global topology of the Hofstadter butterfly. Phys. Rev. Lett. 118, 216801 (2017). Methods Experimental setup. We adopt a time-multiplexed scheme for the experimental realization of photonic quantum walks sch ; ql ; ql2 . As illustrated in Fig. 1, the photon source is provided by a pulsed laser with a central wavelength of $808$nm, a pulse width of $88$ps, and a repetition rate of $15.625$kHz. The pulses are attenuated by a neutral density filter, such that an average photon number per pulse is less than $2.4\times 10^{-4}$, which ensures a negligible probability of multi-photon events. The photons are coupled in and out of a time-multiplexed setup through a BS with a reflectivity of $3\%$, corresponding to a low coupling rate of photons into the network. Such a low- reflectivity BS also enables the out-coupling of photons for measurement. A HWP with the setting angle $\pi/8$ is used to implement the coin operator $C$. Four different paths in a fiber network correspond to the four different directions a walker can take in one step on a two-dimensional lattice. Two-PBS loops are used to realize polarization-dependent optical delays. The shift operator $S_{x}$ is implemented by separating photons corresponding to their two polarization components and routing them through the fibre loops, respectively. Polarization-dependent time delay is then introduced. Since the lengths of the two fiber loops are $287.03$m and $270.00$m, respectively, the time difference of photons traveling through two fiber loops is $80$ns. The shift operator $S_{y}$ is implemented by another two-PBS loop based on the same principle, where the vertical component of photons is delayed relative to the horizontal component by a $1.61$m free space path difference. The corresponding time difference in the $y$ direction is then $4.83$ns. The position-dependent phase operator $P$ is implemented using the first electro-optical modulator (EOM1). The rise/fall times of EOM ($4$ns) are much shorter than the time difference between adjacent positions ($80$ns and $4.83$ns for $x$ and $y$ directions, respectively), which enable us to control the parameter $\phi$ precisely. To realize a polarization-dependent loss operation $M^{\prime}_{x}(\gamma_{x})=e^{-\gamma_{x}}M_{x}(\gamma_{x})$, two HWPs and an EOM are introduced into each fiber loop. Here HWPs are used to keep the polarizations of photons unchanged before and after they pass through the fiber loops. For $\gamma_{x}>0$, for the short loop, the voltage of EOM3 is tuned to $0$. Thus, after passing through the first PBS, horizontally polarized photons are all transmitted by the second PBS and are subject to further time evolution. Whereas for the long loop, by controlling the voltage of EOM2 to satisfy $\cos\theta/2=e^{-2\gamma_{x}}$, we flip part of the photons $(1-e^{4\gamma_{x}})$ with vertical polarization into horizontal ones. They are subsequently transmitted by the second PBS, and leak out of the setup. Otherwise, for $\gamma_{x}<0$, horizontally polarized photons are all transmitted by the second PBS and are subject to further time evolution in the long loop. By contrast, for the short loop, part of the photons $(1-e^{4\gamma_{x}})$ with vertical polarization are flipped by EOM3, transmitted by the second PBS and subsequently leak out of the setup. We use the same method to realize $M^{\prime}_{y}(\gamma_{y})$. We compare the ideal theoretical distribution with the measured distribution via the similarity, $S(t)=\sum_{x,y}\sqrt{P_{\text{th}}(x,y,t)P_{\text{exp}}(x,y,t)},$ (6) which quantifies the equality of two probability distributions. Here $S=0$ stands for completely orthogonal distributions, and $S=1$ for identical distributions. We observe $S\geq 0.914$ in Fig. 2, $S\geq 0.922$ in Fig. 3, and $S\geq 0.930$ in Fig. 4, respectively. Here the theoretical value $P_{\text{th}}(x,y,t)$ is given by $\displaystyle P_{\text{th}}(x,y,t)=\sum_{m=H,V}\frac{|(\langle\bm{r}|\otimes\langle m|)e^{-(\gamma_{x}+\gamma_{y})t}|\psi(t)\rangle|^{2}}{\sum_{\bm{r},m}|(\langle\bm{r}|\otimes\langle m|)e^{-(\gamma_{x}+\gamma_{y})t}|\psi(t)\rangle|^{2}},$ (7) where $|\psi(t)\rangle$ is the time-evolved walker state under $U$. In our experiment, photon loss is caused by the loss of photons through an optical element. Our round-trip single-loop efficiency is about $0.66$ even for a unitary quantum walk. This is calculated by multiplying the transmission rates of each optical component used in the round trip, including the transmission rates of the BS ($\sim 0.97$), the collection efficiency from free space to fiber ($\sim 0.78$), the EOM ($\sim 0.96$), and all other optical components ($\sim 0.91$). We therefore estimate the single-loop efficiency as $0.78\times 0.97\times 0.96\times 0.91\simeq 0.66$. Floquet topological invariant. The walker state evolves according to $\displaystyle|\psi(t)\rangle=U^{t}|\psi(0)\rangle=e^{-iH_{\rm{eff}}t}|\psi(0)\rangle,$ (8) where $H_{\rm{eff}}=i\ln U$ is defined as the effective Hamiltonian. While the quantum walk is identified as the periodically-driven Floquet dynamics, the eigenenergies of $H_{\rm{eff}}$ constitute the quasienergy spectrum of the Floquet system. We fix the branch cut of the logarithm such that the quasienergy spectrum lies within the range $[-\pi,\pi)$. To calculate the Floquet topological invariant, we follow Refs. sai ; asb ; rud , and define $U^{\prime}=e^{i\tilde{E}}M_{y}S_{y}P^{\prime}CM_{x}S_{x}C,$ (9) where $e^{i\tilde{E}}$ shifts the quasienergy spectrum by $-\tilde{E}$. The modified phase-shift operator is $P^{\prime}(\beta,\alpha)=\sum_{\bm{r}}\exp\left[i\sigma_{z}(\beta\lfloor x/q\rfloor+2\pi\alpha x)\right]\otimes\ket{\bm{r}}\bra{\bm{r}},$ (10) where $\alpha=p/q$, $\beta=2\pi/s$, and $\lfloor x/q\rfloor$ is the greatest integer less than or equal to $x/q$. Here $p$ and $q$ are coprime integers, and $s$ is a sufficiently large integer (in our case, for $\alpha=1/3$, $s=15$ is sufficient). We denote the eigenvalues of $U^{\prime}$ as $e^{-iE_{j}}$, and the topological invariant for the quasienergy gap (corresponding to $U$) comprising $\tilde{E}$ can be calculated through $R=\frac{1}{2\pi}\left(\sum_{j=1}^{2sq}E_{j}(1/s,\alpha,\tilde{E})-\sum_{j=1}^{2sq}E_{j}(0,\alpha,\tilde{E})\right).$ (11) Under an open boundary condition, the value of $R$ indicates the number of anomalous Floquet edge states emerging within the quasienergy gap. In Fig. 2d, all quasienergy gaps are closed, hence there are no topological edge states along the boundaries, and the gap topological invariants are ill- defined. In Fig. 4, for the $x\leq 0$ ($x>0$) region of the domain-wall configurations, we have $p=1,q=20$ ($p=-1,q=20$) in Figs. 4a, b, c, and $p=1,q=3$ ($p=-1,q=3$) in Figs. 4d, e, f, respectively. While there are now a host of quasienergy gaps, for any given gap, the topological invariants $R$ of the two regions are always finite and differ by a sign supp . As a consequence, Floquet topological edge modes emerge along the domain-wall boundary. We find that the Floquet topological invariant $R$ is capable of predicting the anomalous topological edge states under all our experimental parameters, despite the presence of the NHSE. Whether the NHSE can have significant impact on $R$ beyond our experimental parameters (particularly when the photon loss is further increased) is an interesting theoretical question that we leave to future studies. Acknowledgments We thank Chen Fang for helpful discussions. This work has been supported by the National Natural Science Foundation of China (Grant Nos. 92265209, 12025401 and 11974331). W. Y. acknowledges support from the National Key Research and Development Program of China (Grant Nos. 2016YFA0301700 and 2017YFA0304100). ## A Supplemental Material for “Manipulating non-reciprocity in a two- dimensional magnetic quantum walk” In this Supplemental Material, we provide numerical results characterizing the non-Hermitian skin effect (NHSE) and the Floquet topological invariants. ### A.1 Quasienergy spectra The presence of NHSE can be confirmed by examining the quasienergy spectra of the effective Hamiltonian under different boundary conditions, and the spatial distribution of eigenstates under the open boundary condition. As discussed in the main text (Methods section), we define the effective Hamiltonian $H_{\rm{eff}}=i\ln U$, where $U$ is the Floquet operator in Eq. (1) of the main text, and the branch cut of the logarithm is taken to be the negative real axis. The quasienergy spectrum of $H_{\rm{eff}}$ thus lies within the range $[-\pi,\pi)$. In Figs. S1a and c, we show the quasienergy spectra under both the periodic (grey) and open (red) boundary conditions. The collapse of the spectra, when the boundary condition is changed from periodic to open, strongly suggests the presence of the NHSE. This is explicitly confirmed in Figs. S1b and d, where we show the spatial distribution of the eigenstates under the open boundary condition. They accumulate to one edge or one corner, depending on the loss parameters. Note that $H_{\rm{eff}}$ and $U$ share the same set of eigenstates. Figure S1: a Numerically calculated quasienergy spectra of $H_{\rm{eff}}$ on the complex plane, with $\gamma_{x}=\gamma_{y}=0.125$, under either the periodic boundary condition (grey) or the open boundary condition (red). b Spatial distribution of the eigenstates under the open boundary condition, with the same parameters used in a. c Numerically calculated quasienergy spectra of $H_{\rm{eff}}$ with $\gamma_{x}=0.1$, $\gamma_{y}=0$, under the periodic boundary condition (grey) and the open boundary condition (red), respectively. d Spatial distribution of the eigenstates under the open boundary condition, with the same parameters used in c. For numerical calculations, a lattice size of $45\times 45$ is taken as an example, with $\alpha=0$. ### A.2 Floquet topological invariants and edge states In Figs. S2a and b, we show the real components of the quasienergy spectra under the domain-wall geometry, respectively under the parameters of Fig. 4d and Fig. 4f of the main text. While the Floquet topological edge states are visible within each quasienergy gap, the quasienergy spectra are close to each other, because of the smallness of the loss parameters in Fig. S2b. The Floquet topological edge states can be characterized by the gap invariant $R$ defined in the Methods section of the main text. In Fig. S2c, we show the calculated gap invariants $R$ for the left (red) and right (black) regions of the domain-wall configuration in Fig. S2b. The topological invariants of the two regions are always finite but differ by their signs. Importantly, within each quasienergy gap, the difference in the gap topological invariants between the two regions is $2$. This corresponds to the number of Floquet topological edge states within each quasienergy gap, as illustrated in Fig. S2b. Figure S2: a, b The real component of the quasienergy spectra under the domain-wall geometry of Figs. 4d and f in the main text, with $\gamma_{x}=\gamma_{y}=0$ in a, and $\gamma_{x}=\gamma_{y}=0.1$ in b. Floquet topological edge states (red) are seen to emerge within each quasienergy gap. c Floquet topological invariants $R$ as functions of the real component of the quasienergy. Within each quasienergy gap, the red (black) line indicates $R$ of the left (right) region. The shaded areas are the density of states normalized to the total number of states, indicating the regions of quasienergy bands. For numerical calculations, we take a lattice with $45\times 45$ sites, and fix $\alpha=1/3$ and $\alpha=-1/3$ for the left and right regions, respectively.
# Kick-motion Training with DQN in AI Soccer Environment ††thanks: This work was supported by the Institute for Information communications Technology Promotion (IITP) grant funded by the Korean government (MSIT) (No.2020-0-00440, Development of Artificial Intelligence Technology that continuously improves itself as the situation changes in the real world). Bumgeun Park Cho Chun Shik Graduate School of Mobility Korea Advanced Institute of Science and Technology (KAIST) Daejeon, South Korea <EMAIL_ADDRESS>Jihui Lee Division of Future Vehicle Korea Advanced Institute of Science and Technology (KAIST) Daejeon, South Korea <EMAIL_ADDRESS>Taeyoung Kim Cho Chun Shik Graduate School of Mobility Korea Advanced Institute of Science and Technology (KAIST) Daejeon, South Korea <EMAIL_ADDRESS>Dongsoo Har Cho Chun Shik Graduate School of Mobility Korea Advanced Institute of Science and Technology (KAIST) Daejeon, South Korea <EMAIL_ADDRESS> ###### Abstract This paper presents a technique to train a robot to perform kick-motion in AI soccer by using reinforcement learning (RL). In RL, an agent interacts with an environment and learns to choose an action in a state at each step. When training RL algorithms, a problem called the curse of dimensionality (COD) can occur if the dimension of the state is high and the number of training data is low. The COD often causes degraded performance of RL models. In the situation of the robot kicking the ball, as the ball approaches the robot, the robot chooses the action based on the information obtained from the soccer field. In order not to suffer COD, the training data, which are experiences in the case of RL, should be collected evenly from all areas of the soccer field over (theoretically infinite) time. In this paper, we attempt to use the relative coordinate system (RCS) as the state for training kick-motion of robot agent, instead of using the absolute coordinate system (ACS). Using the RCS eliminates the necessity for the agent to know all the (state) information of entire soccer field and reduces the dimension of the state that the agent needs to know to perform kick-motion, and consequently alleviates COD. The training based on the RCS is performed with the widely used Deep Q-network (DQN) and tested in the AI Soccer environment implemented with Webots simulation software. ###### Index Terms: Reinforcement learning (RL), Deep Q-Network (DQN), AI soccer, Curse of dimensionality (COD), Coordinate transformation matrix (CTM) ## I Introduction Reinforcement Learning (RL), which is one of the widely used machine learning techniques, makes an agent maximize the accumulated rewards in given environment [1]. To do so requires the use of experiences, e.g., experience replay, during training. The agent selects an action based on the current state and receives the reward from the environment. Through this interaction, the agent can learn which action is more useful to achieve the goal. Recently, a deep neural network (DNN) for nonlinear approximation enables the RL to handle more complicated tasks such as playing a range of Atari games [2], playing the game of GO [3, 4], controlling the robot arm [5, 6, 7], and planning path for mobile robots [8, 9]. AI soccer environment, which is one of the environments for training multi- agent RL algorithms, is introduced in the previous work [10]. The main target of AI Soccer is winning the soccer game by controlling each robot. Each robot’s movement consists of many small motions such as ”going to a target spot”, ”blocking robot or ball”, ”positioning”, ”heading”, and ”kicking”. Therefore, the entire AI soccer training can be divided into various tasks of training motions. The state, action, and reward, which are three basic elements for RL, should be designed depending on the motion to be trained. The curse of dimensionality(COD) occurs when the dimension of the state increases and the amount of training data is inappropriately small. The inappropriately small data for training causes the RL models to be overfitted and consequently leads to deteriorated performance. In this paper, we consider training kick-motion with a deep Q-network (DQN). To train kick-motion successfully, the state should include various information about the robot and ball on the soccer field, which is described by their current position, speed, and the direction of the movement of ball or robot. However, designing the state to include large information can cause the COD, unless experiences are gained evenly over all areas of the soccer field. To eliminate the need for the agent to know all the game information over entire soccer field and to reduce the dimension of the state, we attempt to design the state for efficient kick-motion training by using a coordinate transformation matrix (CTM) used to transform coordinates in robotics or computer vision [11]. The main contributions of this paper are as follows: 1. 1. The attempted designing of the state avoids deterioration arising from the COD and improve the performance of the RL algorithm. 2. 2. The RCS taken for AI Soccer eliminates the need for excessive exploration and reduces the dimension of the state. ## II Background ### II-A AI soccer environment In this paper, experiments are conducted on the AI soccer environment employed for AI World Cup event [12]. The environment consists of a ball, two teams of robots and a soccer field. Each team consists of five robots each driven by two wheels. Like the human soccer game, each team tries to beat the other team by controlling their robots to take the best action for each game state. The robots are controlled at each timestep by setting wheel speeds. The two separately controllable wheels extend the range of actions and increase the degree of freedom. Value of the best action can be measured by corresponding reward. Each robot, as an agent, takes an action based on the situation of the field to get a higher accumulated reward. This is a very challenging task, because various strategies are needed for the agents to adapt to the vast state space which is observed in dynamic game situation. ### II-B Curse of dimensionality (COD) Traditional RL algorithms suffer from the COD when the dimension of the space in the environment is excessively high [13]. As the dimension of the state increases, the number of experiences needed to train an agent grows exponentially. The lack of experiences makes the RL model overfitted. Although using function approximation such as DNN alleviates this problem, the performance of the algorithm is still deteriorated. In the AI soccer environment, large information describing the (game) state and two-wheeled actions with a high degree of freedom increase the dimension of the state space and action space, which readily causes the COD. ### II-C Deep Q-network (DQN) The DQN algorithm, which is a widely used RL algorithm, is a deep RL method that approximates an action-value function, called Q-function by using a DNN. The Q-function with parameters $\theta$ is defined as follows: $Q^{\theta}(s,a)=\mathbb{E}_{s,a,\theta}[\sum_{t=1}^{\infty}\gamma^{t-1}r_{t}]$ (1) where a, s, r, and $\gamma$ are action, state, reward, and discount factor respectively. DQN is trained to minimize the loss function, which can be presented as follows: $L(\theta)=\mathbb{E}[(y_{t}-Q^{\theta}(s_{t},a_{t}))^{2}]$ (2) The target Q function $y_{t}$ is calculated by reward and next state $s^{\prime}$ as follows: $y_{t}=r(s_{t},a_{t})+\gamma max_{a^{\prime}}Q^{\theta^{-}}(s^{\prime},a^{\prime})$ (3) where $\theta^{-}$ represents the parameters of a fixed and separate target network. Figure 1: Illustration of the absolute coordinate system (ACS). The origin of the ACS is placed in the center of the field. The orange dotted line presents the direction of ball movement. The yellow arrow presents the direction the robot movement. Figure 2: Illustration of the relative coordinate system (RCS). The origin of the RCS is placed in the center of the robot, and the direction the robot is aligned with the $X^{\prime}$-axis. ## III Relative coordinate system for description of game state In the task of training the robot to perform kick-motion, the robot chooses at each timestep the best action out of a set of actions, taking into account the information about the ball and the robot such as the current coordinates, the velocity of the ball and the robot, and the approaching angle of the ball to the robot. To avoid the COD and train the robot to kick the ball at any location, it is required to use evenly distributed experiences over entire field. For this purpose, we adopt the relative coordinate system (RCS) to represent the state, instead of the absolute coordinate system (ACS). The RCS can alleviate the COD and make training process straightforward. Most important benefit obtained from the use of the RCS is that the robot can achieve the task at any location after training at only one location. In the AI soccer environment, all raw information given to the robot from the environment is presented in the ACS format. The origin of the ACS is defined as the center of the field with $X$-axis and $Y$-axis as shown in Fig. 1. In this case, the state given to the robot is defined as follows: $S_{ACS}=[R_{x},R_{y},R_{\theta},B_{x},B_{y},V_{R_{x}},V_{R_{y}},V_{B_{x}},V_{B_{y}},\theta]$ (4) Each element in (4) respectively represents $X$ and $Y$ coordinates of the robot, the heading angle of the robot, $X$ and $Y$ coordinates of the ball, $X$ and $Y$ direction velocity of the robot, $X$ and $Y$ direction velocity of the ball, and the angle of the direction in which the ball moves. The high dimensional state space formulated in (4) can deteriorate the performance of RL algorithms. With the RCS, the dimensional space is reduced and training becomes efficient. In the RCS, the origin is defined at the center of the robot with $X^{\prime}$-axis and $Y^{\prime}$-axis and the direction the robot movement is the same as the direction of $X^{\prime}$-axis shown in Fig. 2. With the RCS, the values of $R_{x}$, $R_{y}$, and $R_{\theta}$ are always zero since the origin of the RCS is the center of the robot. The value of $V_{R}{}_{y}$ is also always zero because the robot can take only one of forward, backward, or stop, during the task of the kick-motion. As a result, the state can be defined as follows: $S_{RCS}=[B_{x}^{RCS},B_{y}^{RCS},V_{R_{x}}^{RCS},V_{B_{x}}^{RCS},V_{B_{y}}^{RCS},\theta^{RCS}]$ (5) where $(\cdot)^{RCS}$ means $(\cdot)$ in the RCS. The relationship between the RCS and the ACS can be presented using the CTM as shown in (6). $M=\begin{bmatrix}\cos(R_{\theta})&-\sin(R_{\theta})&R_{x}\\\ \sin(R_{\theta})&\cos(R_{\theta})&R_{y}\\\ 0&0&1\end{bmatrix}$ (6) The position, velocity, and angle in the RCS can be computed with the CTM and those in the ACS, and can be shown as (7), (8), and (9). $\begin{bmatrix}B_{x}^{RCS}&V_{B_{x}}^{RCS}\\\ B_{y}^{RCS}&V_{B_{y}}^{RCS}\\\ 1&1\end{bmatrix}=M^{-1}\begin{bmatrix}B_{x}&V_{B_{x}}\\\ B_{y}&V_{B_{y}}\\\ 1&1\end{bmatrix}$ (7) $V_{R_{x}}^{RCS}={V_{R_{x}}}^{2}+{V_{R_{y}}}^{2}$ (8) $\theta^{RCS}=\theta-R_{\theta}$ (9) The change of the coordinate system from the ACS to the RCS reduces the dimension of the state by 60%, which means significantly easier training process. Figure 3: Illustration of a situation in which the ball and the robot are in contact. The blue line connects from the center of the robot to the center of the ball and the yellow arrow presents the direction toward the target point. ## IV Experiments This section describes AI Soccer environment and presents the result of experiments conducted to show the effect of the COD. ### IV-A Experimental Environment Experiments are conducted for the kick-motion task in the AI soccer environment developed with Webot physics simulator [14]. The goal of the kick- motion task is that robot kicks the ball to a target point. The robot is facing the target point and the ball is located at a random point with a distance of 2m from the robot. With the start of an episode of training, the ball starts to move at a random speed to a random point near the robot. At each timestep, the robot chooses an action based on the state. Each episode consist of 40 experiences over 40 timesteps. The RL framework for this task can be described as follows: * States: State consists of 6 values shown as (5) in Section III. * Rewards: Reward is given as a sum of 3 types of reward, $r_{contact}+r_{theta}+r_{velocity}$. The $r_{contact}$ is a reward of contact, which is -0.1 before the robot contact the ball and 0 after the contact. The $r_{Theta}$ is a reward according to distance of the contact point from the center of the robot, which is represented as $3\cos{\theta_{contact}}$, where $\theta_{contact}$ is the angle between the blue and the yellow lines in Fig. 3. The $r_{velocity}$ is a reward according to the speed of the contact with the ball, which is represented as $3.92V_{contact}$, where $V_{contact}$ is the velocity of the robot at the contact. * Actions: List of available actions: going straight forward with steady-state speeds of 2.55 (m/s), stop, going straight backward with steady-state speeds of -2.55 (m/s). ### IV-B Experimental Results The performances with the kick-motion task are evaluated in terms of the sum of the reward for each training episode. To reduce granularity of the reward variation, the moving average of the sum of the reward across 250 episodes is taken. Figure 4: Total reward according to the number of dummy variables. It shows that increasing the dimension of state space slows down the training speed and as a result degrades performance. The blue line represents the model with the RCS and the others represent comparative models. Especially, the purple line is equivalent to the model with the ACS. To present the intuitive understanding of the COD, models used to compare with the proposed method are trained with states including ’dummy variables’, which are some meaningless random variables. The states for the comparison models are the state in the RCS with ’dummy variables’. In Fig. 4, the name of ’$+n$’ means that $n$ dummy variables are added to the state. Figure 4 shows the results of the experiments. The (total accumulated) reward of the action for RCS based state exceeds -4 from 1500-th episode, which is about 500 episodes faster than by ACS based method. The RCS based method achieves the highest reward value of 6.8, although the reward value of the RCS based method is similar to the others in the latter part of the training, e.g., 3600-4700 episodes. ## V Conclusion Efficient training of the RL algorithms with high dimensional state space is hard to achieve, due to the COD. In this paper, a method to reduce the dimension of the state space for training the kick-motion task in the AI soccer environment is attempted. The attempted method uses the CTM to convert the state in the ACS to the state in the RCS. The dimension of the state space required for the agent to achieve the kick-motion is reduced from ten to six since 4 values in the state are always zero by the RCS based method. The experimental result shows that the kick-motion can be achieved with the reduced state by the RCS based method. The result obtained along with the comparative models with different number of variables, which are trained with the states including dummy variables, shows that the lower dimension of the state space can avoid deterioration due to the COD. ## Acknowledgment This work was supported by the Institute for Information communications Technology Promotion (IITP) grant funded by the Korean government (MSIT) (No.2020-0-00440, Development of Artificial Intelligence Technology that continuously improves itself as the situation changes in the real world). ## References * [1] Richard S Sutton and Andrew G Barto. Reinforcement learning: An introduction. MIT press, 2018. * [2] Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller. Playing atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602, 2013. * [3] David Silver, Aja Huang, Chris J Maddison, Arthur Guez, Laurent Sifre, George Van Den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Veda Panneershelvam, Marc Lanctot, et al. Mastering the game of go with deep neural networks and tree search. nature, 529(7587):484–489, 2016. * [4] David Silver, Julian Schrittwieser, Karen Simonyan, Ioannis Antonoglou, Aja Huang, Arthur Guez, Thomas Hubert, Lucas Baker, Matthew Lai, Adrian Bolton, et al. Mastering the game of go without human knowledge. nature, 550(7676):354–359, 2017. * [5] Minah Seo, Luiz Felipe Vecchietti, Sangkeum Lee, and Dongsoo Har. Rewards prediction-based credit assignment for reinforcement learning with sparse binary rewards. IEEE Access, 7:118776–118791, 2019. * [6] Luiz Felipe Vecchietti, Taeyoung Kim, Kyujin Choi, Junhee Hong, and Dongsoo Har. Batch prioritization in multigoal reinforcement learning. IEEE Access, 8:137449–137461, 2020. * [7] Luiz Felipe Vecchietti, Minah Seo, and Dongsoo Har. Sampling rate decay in hindsight experience replay for robot control. IEEE Transactions on Cybernetics, 2020. * [8] Inhwan Kim, Sarvar Hussain Nengroo, and Dongsoo Har. Reinforcement learning for navigation of mobile robot with lidar. In 2021 5th International Conference on Electronics, Communication and Aerospace Technology (ICECA), pages 148–154. IEEE, 2021. * [9] Woohyeon Moon, Bumgeun Park, Sarvar Hussain Nengroo, Taeyoung Kim, and Dongsoo Har. Path planning of cleaning robot with reinforcement learning. arXiv preprint arXiv:2208.08211, 2022. * [10] Taeyoung Kim, Luiz Felipe Vecchietti, Kyujin Choi, Sanem Sariel, and Dongsoo Har. Two-stage training algorithm for ai robot soccer. PeerJ Computer Science, 7:e718, 2021. * [11] Richard Hartley and Andrew Zisserman. Multiple view geometry in computer vision. Cambridge university press, 2003. * [12] Chansol Hong, Inbae Jeong, Luiz Felipe Vecchietti, Dongsoo Har, and Jong-Hwan Kim. Ai world cup: Robot-soccer-based competitions. IEEE Transactions on Games, 13(4):330–341, 2021. * [13] Abhijit Gosavi. Reinforcement learning: A tutorial survey and recent advances. INFORMS Journal on Computing, 21(2):178–192, 2009. * [14] Olivier Michel. Cyberbotics ltd. webots™: professional mobile robot simulation. International Journal of Advanced Robotic Systems, 1(1):5, 2004\.
# Statistical Chronometry of Meteorites. I. A Test of ${}^{26}{\rm Al}$ homogeneity and the Pb-Pb age of the Solar System’s $t\\!\\!=\\!\\!0$. Steven J. Desch Daniel R. Dunlap Emilie T. Dunham Curtis D. Williams Prajkta Mane ###### Abstract We use rapidly cooled achondrites to test the assumption of ${}^{26}{\rm Al}$ homogeneity in the solar nebula, by checking if there is a single value of $t_{\rm SS}$, the absolute “Pb-Pb” age of the Solar System’s $t\\!\\!=\\!\\!0$, that makes concordant their ages from the Al-Mg and Pb-Pb systems. We find that values $t_{\rm SS}=4568.42\pm 0.24$ Myr do make these ages concordant, and therefore the hypothesis of homogeneous ${}^{26}{\rm Al}$ is not falsified. This age, defined to be when the solar nebula had $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})=5.23\times 10^{-5}$, is significantly older than the $\approx$ 4567.3 Myr inferred from direct measurements of Pb-Pb ages in CAIs. Discrepancies between the Al-Mg and Pb-Pb chronometers in chondrules and CAIs have previously been interpreted as arising from heterogeneities in ${}^{26}{\rm Al}$, under the presumption that the Al-Mg and Pb-Pb systems in CAIs closed simultaneously. We examine this assumption and show that resetting is to be expected in CAIs. In particular, we quantitatively demonstrate that it is plausible that Pb-Pb ages of CAIs were reset at late times, without resetting the earlier Al-Mg ages, if they were transiently heated in the same manner as chondrules. We critically examine Pb-Pb isochrons, refining data and suggesting best practices for their calculation and reporting. We advocate reporting chronometry as times of formation after $t\\!\\!=\\!\\!0$ rather than absolute ages, as only the former is useful for astrophysical models of the solar nebula. We advocate averaging of multiple samples, rather than anchoring to individual meteorites, to improve precision. ###### keywords: Solar System formation 1530 , Planet formation 1241 , Meteorites 1038 , Achondrites 15 , Chondrites 228 ††journal: Icarus [inst1]organization=School of Earth and Space Exploration, Arizona State University,addressline=PO Box 871404, city=Tempe, postcode=85287-1404, state=Arizona, country=USA [inst2]organization=Oak Ridge National Laboratory, addressline=1 Bethel Valley Rd, city=Oak Ridge, postcode=37830, state=Tennessee, country=USA [inst6]organization=Department of Earth, Planetary and Space Sciences, University of California, Los Angeles, addressline=PO Box 951567, city=Los Angeles, postcode=90095-1567, state=California, country=USA [inst3]organization=Earth and Planetary Sciences Department, University of California, Davis,addressline=One Shields Ave., city=Davis, postcode=95616, state=California, country=USA [inst4]organization=Lunar and Planetary Institute, USRA, addressline=3600 Bay Area Blvd., city=Houston, postcode=77058, state=Texas, country=USA We re-evaluate and average data from the Al-Mg and Pb-Pb radiometric dating systems, for achondrites and chondrules, to attain better accuracy and precision in the time of formation of meteorites and meteoritic inclusions. We find remarkable concordancy between the systems, provided the Pb-Pb age of the Solar System is $4568.4\pm 0.2$ Myr, older than previous estimates; this suggests late resetting of the Pb-Pb system in Ca-rich, Al-rich inclusions (CAIs), and supports homogeneity of ${}^{26}{\rm Al}$ in the solar nebula. ## 1 Introduction ### 1.1 The Al-Mg and Pb-Pb Chronometers To learn about the birth of planets and the story of our Solar System’s first few million years, we study meteorites that bear witness to this era. It is especially crucial to constrain the times at which their constituent components formed, the times at which their parent planetesimals accreted and melted, when collisions occurred, and to constrain the relative order of these events in the context of the solar nebula. The need of astrophysical models is to know the time $\Delta t$ after $t\\!\\!=\\!\\!0$ that an event occurred, where $t\\!\\!=\\!\\!0$ is a defined event or time in Solar System history. To obtain these times $\Delta t$, radiometric dating systems are employed. The Al-Mg system is most commonly used. Calcium-rich, aluminum-rich inclusions (CAIs), incorporated live ${}^{26}{\rm Al}$, a short-lived radionuclide (SLR) that decays to ${}^{26}{\rm Mg}$ with a half-life of 0.717 Myr, or mean-life $\tau_{26}=1.034$ Myr (Auer et al., 2009; Kondev, 2021). CAIs are among the first meteoritic components to have formed in the Solar System, apparently during a short time interval that can be associated with $t\\!\\!=\\!\\!0$. Most CAIs incorporated Al with an isotopic ratio $\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$}\approx(5.23\pm 0.13)\times 10^{-5}$ (Jacobsen et al., 2008). A useful working hypothesis is that this was the widespread and spatially homogeneous abundance of ${}^{26}{\rm Al}$ in the solar nebula when CAIs formed. We define $t\\!\\!=\\!\\!0$ to be the time in the solar nebula when ${}^{26}{\rm Al}/{}^{27}{\rm Al}$ equalled $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{\rm SS}\equiv 5.23\times 10^{-5}$ exactly. Regardless of the “true” initial value, most CAIs appear to have formed in a short time interval around this time (Liu et al., 2019), so this is a convenient definition. Assuming homogeneity of ${}^{26}{\rm Al}$, if a different object formed with a lower initial ratio $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$, it must have formed later, at a time $\Delta t_{26}=\tau_{26}\,\ln\left(\frac{(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{\rm SS}}{(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}}\right).$ (1) This measures a time of formation relative to $t\\!\\!=\\!\\!0$. In practice, the Al-Mg chronometer often has uncertainty of less than $\pm 0.1$ Myr (§2). None of the original ${}^{26}{\rm Al}$ in CAIs exists today, and so its one- time existence must be inferred. Its initial abundance relative to stable Al is determined from a correlation between the isotopic ratio $({}^{26}{\rm Mg}/{}^{24}{\rm Mg})$ and the ratio $({}^{27}{\rm Al}/{}^{24}{\rm Mg})$, as measured in different minerals that condensed or crystallized from the same isotopic reservoir within the inclusion (or igneous achondrite). If the minerals have not been disturbed, e.g., not heated or altered in a way that would lead to mobilization of Mg isotopes, then the correlation must be linear and form an “internal isochron” in a plot of $({}^{26}{\rm Mg}/{}^{24}{\rm Mg})$ versus $({}^{27}{\rm Al}/{}^{24}{\rm Mg})$ ratios of the inclusion’s minerals. The slope of the isochron is the inferred initial $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ ratio at the time the inclusion achieved isotopic closure. Because ${}^{26}{\rm Al}$ must be inferred from an isochron, there are caveats associated with its determination. First, determination of $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ is only possible if the data form a linear isochron. This requires a statistical test: the mean squares weighted deviation (MSWD)] of the linear correlation must not exceed a maximum value $\approx 1+2(2/(N-2))^{1/2}$, where $N$ is the number of data points (Wendt and Carl, 1991). A non-linear fit suggests disturbance or an additional process; without a model of this process, $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ cannot be inferred. Second, it must be remembered that the initial ratio $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ records not the formation of the CAI, but the time at which the Al and Mg isotopes no longer moved through the inclusion (i.e., when it reached isotopic closure). On the parent body, aqueous alteration is a likely cause of disturbance. On the parent body or in the solar nebula, a common cause of disturbance is heat- related diffusion of isotopes. Often $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ records the time at which the CAI remained below a critical temperature, the “closure temperature,” such that diffusion of Mg isotopes is slower than continued cooling. This is discussed further in (§ 2.2). The Pb-Pb system provides a second way to measure times of formation. In this system, ${}^{235}{\rm U}$ essentially decays to ${}^{207}{\rm Pb}$ with half- life $t_{1/2}=703.81\pm 0.96(1\sigma)$ Myr, and ${}^{238}{\rm U}$ essentially decays to ${}^{206}{\rm Pb}$ with half-life $t_{1/2}=4468.3\pm 4.8(1\sigma)$ Myr (Jaffey et al., 1971; Villa et al., 2016), but no natural isotopes decay to ${}^{204}{\rm Pb}$. The ratio of radiogenic ${}^{207}{\rm Pb}$ to radiogenic ${}^{206}{\rm Pb}$ is therefore $\left(\frac{{}^{207}{\rm Pb}}{{}^{206}{\rm Pb}}\right)_{\rm r}=\left(\frac{{}^{235}{\rm U}}{{}^{238}{\rm U}}\right)_{\rm t}\,\frac{\exp(+t/\tau_{235})-1}{\exp(+t/\tau_{238})-1},$ (2) where $({}^{235}{\rm U}/{}^{238}{\rm U})_{\rm t}$ is the isotope ratio measured in the sample today, and $t$ is the time that has elapsed since the inclusion formed. The radiogenic Pb isotopic ratio is found by creating different washes and leachates and residues by dissolution of a sample by different acids, and then measuring Pb isotopes in each. A regression of $y={}^{207}{\rm Pb}/{}^{206}{\rm Pb}$ vs. $x={}^{204}{\rm Pb}/{}^{206}{\rm Pb}$ in each fraction yields a line (sometimes called the “inverse isochron”) with $y$-intercept (limit of zero non-radiogenic component) equal to the left- hand side of Equation 2. Different washes or leachates may incorporate variable amounts of Pb from terrestrial contamination or initial non- radiogenic Pb, either of which may cause a particular data point to fall off an otherwise reasonable regression. As with Al-Mg isochrons, adherence to a line can be tested using the MSWD of the linear regression. Before about 2010, it was assumed that all samples were characterized by $\mbox{${}^{238}{\rm U}/{}^{235}{\rm U}$}=137.88$; but it has become recognized that CAIs, achondrites, etc., vary significantly in this ratio. This is attributable to variable amounts of radiogenic ${}^{235}{\rm U}$ from decay of ${}^{247}{\rm Cm}$ (Brennecka et al., 2010; Tissot et al., 2016), but also evaporation and other processes (Tissot and Dauphas, 2015; Tissot et al., 2017, 2019). A fractional change of $10^{-3}$ in the U isotope ratio corresponds to a shift in the age of 1.45 Myr, and typical variations in CAIs can lead to inaccuracies in their ages $>1$ Myr. Therefore it is now standard to report “U-corrected” Pb-Pb ages, in which ${}^{238}{\rm U}/{}^{235}{\rm U}$ is measured in the sample. An additional issue is that not all uranium isotopic ratios are reported using the same assumed values for standards. This may lead to variations in ${}^{238}{\rm U}/{}^{235}{\rm U}$ on the order of $0.05\mbox{\text{\textperthousand}}$, leading to variations in inferred Pb-Pb ages of 0.07 Myr (F. Tissot, personal communication). Future renormalization of the data appears warranted. The uncertainty in Pb-Pb absolute ages is often misunderstood to be $<1$ Myr. (i.e., $\pm y$ in the notation of Tissot et al. (2017)), but ignores other sources of uncertainty, the largest of which is the half-lives of U. Although both are known at the 0.1% level, the uncertainty in absolute Pb-Pb ages is $\pm 9$ Myr, due to uncertainties in the ${}^{238}{\rm U}$ and especially ${}^{235}{\rm U}$ half-lives (Ludwig, 2000; Tissot et al., 2017). However, when two Pb-Pb ages are computed using the same half-lives, the systematic uncertainties cancel, and the ($2\sigma$) precision can approach $0.3-0.5$ Myr (Tissot et al., 2017). It is essential that the half-lives be identical. Other half-lives are sometimes used—e.g., the ${}^{235}{\rm U}$ half-life of 703.20 Myr (Mattinson, 2010) used by Palk et al. (2018) in their Pb-Pb dating of enstatite chondrites—but fortunately almost all absolute ages in the literature are computed using the half-lives quoted above from Jaffey et al. (1971) and Villa et al. (2016). The Pb-Pb system is the most precise and versatile absolute chronometer, but its greater utility and precision comes from using it as a relative chronometer to determine the sequence of events in the solar nebula. To be effective as a relative chronometer, absolute ages measured by the Pb-Pb system, $t_{\rm Pb}$, must be converted into times after $t\\!\\!=\\!\\!0$, $\Delta t_{\rm Pb}$, by subtracting them from the absolute age of $t\\!\\!=\\!\\!0$, which we denote $t_{\rm SS}$: $\Delta t_{\rm Pb}=t_{\rm SS}-t_{\rm Pb}.$ (3) Here $t_{\rm SS}$ should be interpreted as the Pb-Pb age that would be obtained by measuring the age of a sample that achieved isotopic closure at $t\\!\\!=\\!\\!0$, using the same uranium half-lives as assumed for other samples. Determination of $t_{\rm SS}$, the “age of the Solar System,” is one focus of this paper. ### 1.2 Absolute Age of CAIs and the Solar System To date, most determinations of $t_{\rm SS}$ and the time elapsed since $t\\!\\!=\\!\\!0$ have been made by radiometrically dating CAIs. As described above, an inverse isochron is made by measuring Pb isotopes in different washes, leachates and residues (“fractions”) obtained by reacting a sample with various acids, and for each fraction obtaining isotopic fractionation- corrected $y={}^{207}{\rm Pb}/{}^{206}{\rm Pb}$ and $x={}^{204}{\rm Pb}/{}^{206}{\rm Pb}$ ratios. Assuming a linear correlation, the $y$-intercept of the line is the purely radiogenic end-member $({}^{207}{\rm Pb}/{}^{206}{\rm Pb})$, and this plus a measurement of ${}^{235}{\rm U}/{}^{238}{\rm U}$ allows one to solve for $t$ in Equation 2. The uncertainty in the age comes from adding in quadrature the uncertainties in the ${}^{235}{\rm U}/{}^{238}{\rm U}$ ratio and the uncertainties in the $y$-intercept from the linear regression. The latter derives from the measurement uncertainties in the $x$ and $y$ data points. Perhaps because of the difficulty of the measurement (wet chemistry of small samples while avoiding terrestrial contamination) and the newness of the uranium correction, there are only four CAIs with peer-reviewed, U-corrected Pb-Pb ages. Amelin et al. (2010) dated Allende CAI SJ101 at $4567.18\pm 0.50$ Myr. MacPherson et al. (2017) reported an internal isochron with $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(5.20\pm 0.53)\times 10^{-5}$ for this CAI. Connelly et al. (2012) found ages for Efremovka CAIs 22E, 31E and 32E of $4567.35\pm 0.28$ Myr, $4567.23\pm 0.29$ Myr, and $4567.38\pm 0.21$ Myr. Larsen et al. (2011) found excesses of ${}^{26}{\rm Mg}$ in 22E and 31E that are consistent with formation with $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}\approx 5.23\times 10^{-5}$. Assuming that all of these CAIs experienced isotopic closure in Pb-Pb at identical times, that time is the weighted average of the Pb-Pb ages, given by Connelly et al. (2012) as $4567.30\pm 0.16$ Myr. Further assuming that they all closed in Al-Mg at this same time, this time is taken to be $t\\!\\!=\\!\\!0$. These four CAIs are the only ones with peer-reviewed, U-corrected Pb-Pb ages, and all appear consistent with this average age, so this value has been widely adopted as the time elapsed since $t\\!\\!=\\!\\!0$. There are strong hints of older ages of CAIs, however. Bouvier et al. (2011a) reported a U-corrected Pb-Pb age for NWA 6991 CAI B4 of $4567.94\pm 0.31$ Myr. This CAI also has canonical $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ (Wadhwa et al., 2014). In addition, Bouvier and Wadhwa (2010) inferred an age $4568.22\pm 0.18$ Myr in NWA 2364 CAI B1. These ages have not been widely accepted, perhaps because the former has not been published in the refereed literature, and the latter was not U-corrected using a direct measurement of U isotopes, but rather a correlation between Th/U and ${}^{238}{\rm U}/{}^{235}{\rm U}$ found by Brennecka et al. (2010); this relation is not universal (Tissot et al., 2016), and so might not yield the correct age adjustment for this CAI. The two latter ages do not agree with $4567.30\pm 0.16$ Myr to within the uncertainties. If accepted or confirmed, CAIs B1 and SJ101 would be seen to have formed at times that differ from $4567.3$ Myr by over 1.0 Myr (at the $4\sigma$ level). More Pb-Pb dates of CAIs would be enormously helpful, to determine whether CAIs reached isotopic closure in the Pb-Pb system at the same time or across a range of times. Because the four CAIs averaged by Connelly et al. (2012) seem to have formed at a single time, that time is taken to be $t\\!\\!=\\!\\!0$. However, as is well-known and as is discussed below, if CAIs actually formed with canonical $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}\approx 5.23\times 10^{-5}$ at 4567.30 Myr ago, then almost all other chronometers would be left discordant. This has led to the proposal that ${}^{26}{\rm Al}$ must have been spatially or temporally heterogeneous in the solar nebula, with the CAI-forming region holding nearly four times as much ${}^{26}{\rm Al}$ as the rest of the solar nebula (Larsen et al., 2011; Bollard et al., 2019). In this paper, we explore other reasons why Al-Mg and Pb-Pb ages might not match. One important factor we identify is that the practice of building isochrons from a subset of fractions means that some Pb-Pb ages are inaccurate and their uncertainties have been considerably underestimated. Another pertinent fact is that CAIs were likely transiently heated at random times up to 3 Myr after $t\\!\\!=\\!\\!0$; we demonstrate that the Pb-Pb system could have been isotopically reset, without disturbing the Al-Mg system. As these are plausible scenarios, it is a distinct possibility that the reported Pb-Pb ages of CAIs might not actually record the time of $t\\!\\!=\\!\\!0$. An alternative to using radiometric dating of CAIs to determine $t_{\rm SS}$ is to mathematically find values that reduce the discrepancies between the Pb- Pb and other chronometers to levels consistent with measurement errors. Many other groups have sought to correlate Al-Mg and other chronometers. For example, Lugmair and Shukolyukov (1998) correlated Mn-Cr ages against Pb-Pb ages in many achondrites, finding a range of ages 4568 to 4571 Myr would minimize the discrepancies between the chronometers. These ages overlapped with the then-accepted age of CAIs, $4566\pm 2$ Myr (Göpel et al., 1991), so it was possible to conclude the chronometers were concordant. Of course, none of these ages was U-corrected. Nyquist et al. (2009) correlated Al-Mg and Mn- Cr ages to show that their ages were linearly correlated and thereby constrain the half-life of ${}^{53}{\rm Mn}$ and the Solar System initial value $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{\rm SS}$. They then used the initial $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{0}$ inferred for the achondrite LEW 86010 to determine the time of its formation after $t\\!\\!=\\!\\!0$. They added this to the Pb-Pb age of LEW 86010 to derive $t_{\rm SS}=4568.2$ Myr. This Pb-Pb age was not U-corrected, either. Sanborn et al. (2019) directly correlated U-corrected Pb-Pb ages against Al-Mg formation times for several achondrites and CAIs. Although not explicitly stated, their Figure 6 makes clear that they would infer a value of $t_{\rm SS}\approx 4567.8$ Myr, with an uncertainty that we estimate at $\sim 0.5$ Myr. Recently, Piralla et al. (2023) also directly correlated U-corrected Pb- Pb ages against Al-Mg formation times of several achondrites and concluded $t_{\rm SS}=4568.7$ Myr. We further discuss these treatments below. One key aspect missing from these treatments is that they did not check whether the values of $t_{\rm SS}$ they derived led to a statistically valid fit between the Al-Mg and other chronometers, i.e., whether the discrepancies between their inferred $\Delta t_{26}$ and other formation times are attributable solely to measurement uncertainties. Without this assessment of concordancy, it is impossible to test the underlying assumption of ${}^{26}{\rm Al}$ homogeneity. ### 1.3 Outline In this work we test the hypothesis that ${}^{26}{\rm Al}$ was distributed homogeneously in the solar nebula. We do this by comparing Al-Mg and Pb-Pb ages in the samples above, as in the work by Sanborn et al. (2019). A prediction of the hypothesis is that in samples in which the systems must have closed simultaneously, a single value $t_{\rm SS}$ must reconcile the chronometers, in a statistical sense. If a single value of $t_{\rm SS}$ does not reconcile the Al-Mg and Pb-Pb ages, then the hypothesis of ${}^{26}{\rm Al}$ homogeneity is falsified. If the chronometers are reconciled, then that supports the homogeneous hypothesis, and futher suggests that the range of values of $t_{\rm SS}$ that reconciles them is the Pb-Pb age of the Solar System at $t\\!\\!=\\!\\!0$. In §2 we compile and discuss the meteoritic data, which comprises five achondrites from the non-carbonaceous chondrite (NC) isotopic reservoir, two achondrites from the carbonaceous chondrite (CC) isotopic reservoir, four chondrules, and one CAI. We review closure temperatures and why there can be no discrepancies between Al-Mg and Pb-Pb ages for the rapidly cooled achondrites, and therefore why these can falsify the homogeneous hypothesis. We discuss why the other samples, especially the CAI, do not necessarily test the hypothesis. We re-analyze the Pb-Pb isochrons from all samples, to determine the Pb-Pb age and the uncertainty for each. In §3 we take weighted means of the data to derive the value of $t_{\rm SS}$ that minimizes the discrepancies between the Al-Mg and Pb-Pb chronometers in the appropriate samples. We calculate the goodness-of-fit metric $\chi_{\nu}^{2}$ (also known as MSWD) to test whether the samples are adequately fit by a single value of $t_{\rm SS}$. We find the the Pb-Pb and Al-Mg ages are indeed reconciled in a statistical sense for the seven achondrites, and even the 4 chondrules as well, if $t_{\rm SS}=4568.42\pm 0.24$ Myr. Had they not been, this would have falsified the hypothesis of homogeneity, but instead this finding supports it. We find that the same $t_{\rm SS}$ that minimizes the discrepancies between the Al-Mg and Pb-Pb systems among the NC achondrites also makes the chondrules concordant, but cannot reconcile the CAI SJ101, which appears reset. In §4 we compare our approach to previous similar approaches. We discuss how the data support homogeneity of ${}^{26}{\rm Al}$ and discuss why astrophysical models strongly support such an interpretation. One of our primary conclusions is that the Pb-Pb age of CAI SJ101 and other CAIs may have been reset 1.2 Myr after it formed and recorded an Al-Mg age, which was not reset. We strongly support reporting times of formation after $t\\!\\!=\\!\\!0$ rather than absolute ages. Astrophysical models are unable to use absolute ages, only relative formation times, and relative times of formation are far more precise anyway. In §5 we draw conclusions. It does not appear that ${}^{26}{\rm Al}$ was heterogeneous in the solar nebula. We argue that statistical approaches, like those of Nyquist et al. (2009), Sanborn et al. (2019), Piralla et al. (2023), and presented here, are more reliable determinants of $t_{\rm SS}$ than direct measurements of CAIs. We advocate for reporting dates as times after $t\\!\\!=\\!\\!0$, not absolute ages, and we advocate against use of individual anchors. ## 2 Meteoritic Data ### 2.1 Which Samples test Homogeneity? The hypothesis of homogeneous ${}^{26}{\rm Al}$ makes the testable prediction that times of formation of appropriate samples found using the Al-Mg chronometer should match times of formation found using the Pb-Pb chronometer. Several conditions must be met for a sample to be an appropriate test. The most important is that the $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ and Pb-Pb age actually can be found in an object; the former requires a valid Al-Mg isochron, and the latter a valid Pb-Pb isochron and measurement of the ${}^{238}{\rm U}/{}^{235}{\rm U}$ ratio in the sample. It also must be certain that both the Al-Mg and Pb-Pb systems date the same event. The Al-Mg ages can be defined as a time of formation after $t\\!\\!=\\!\\!0$, $\Delta t_{26}$, using Equation 1. The Pb-Pb absolute ages must be converted to a time of formation after $t\\!\\!=\\!\\!0$, $\Delta t_{\rm Pb}$, using Equation 3. These requirements drive our choice of samples. To our knowledge, the objects in which both Al-Mg and U-corrected Pb-Pb ages have been determined includes seven bulk achondrites, four chondrules, and one CAI. The achondrites include five with the signatures of having formed in the NC isotopic reservoir: the quenched (volcanic) angrites D’Orbigny, SAH 99555, and NWA 1670; the eucrite- like achondrite Asuka 881394; and the ungrouped achondrite NWA 7325 (paired with NWA 8486). Two of the achondrites have signatures of having formed in the CC isotopic reservoir: the ungrouped achondrites NWA 2976 (paired with NWA 011) and NWA 6704 (paired with NWA 6693). In some samples, we are confident that the Al-Mg and Pb-Pb systems probably date the same event; in others we have no evidence to support that. There are many reasons why the two isotopic systems might not date the same event. As discussed below, bulk achondrites generally record the time at which the temperature dropped below a “closure temperature”. This can be hundreds of K different for different minerals and isotopic systems, so Al-Mg and Pb-Pb can record different times. If the cooling rate is slow, only $\sim 10^{2}\,{\rm K}\,{\rm Myr}^{-1}$, then those times can be $\sim 10^{6}$ years apart. Alternatively, chondrules notoriously were exposed to transient heating events with peak temperatures $>1800$ K, followed by cooling rates $\sim 10-10^{3}\,{\rm K}\,{\rm hr}^{-1}$ (Desch et al., 2012). Being located in the nebula at the same time as chondrules before parent-body accretion, CAIs also could have been exposed to these same heating events. Below we explore how such transient heating events could have reset the Pb-Pb isochron at a late time, and even do so without resetting the Al-Mg chronometer, which would record a much earlier event. ### 2.2 Closure Temperatues of the Al-Mg and Pb-Pb Systems To quantify the largest possible difference between the times recorded by the Al-Mg and Pb-Pb systems, we calculate their closure temperatures, the temperatures below which isotopic redistribution within the sample (here, by thermal diffusion) is slower than further cooling. Closure temperatures are found by comparing the cooling rate of the sample against the diffusion rate of key elements in specific minerals, and depend on the grain size of the minerals (more precisely, the lengthscales over which isotopic gradients must be preserved). It is assumed the diffusion coefficient of the element obeys an Arrhenius relationship: $D(T)=D_{0}\,\exp\left(-E/RT\right),$ (4) where $D_{0}$ is the pre-factor and $E$ the activation energy, both specific to the element and mineral, and $R$ the gas constant. In that case the closure temperature is given by $T_{\rm c}=\frac{E}{R}\,\left[\ln\left(\frac{A\,T_{\rm c}^{2}\,D_{0}}{(E/R)\,(dT/dt)\,a^{2}}\right)\right]^{-1}$ (5) (Dodson, 1973), where $A=55$ (for a sphere; $A=27$ for a cylinder, $A=8.7$ for a plane), $a$ is the grain radius, $dT/dt$ is the cooling rate, and it is assumed that the cooling is from a peak temperature above $T_{\rm c}$. The closure temperature must be estimated, substituted into the right-hand side, and the solution iterated to convergence. The majority of Pb-Pb dates are based on data from pyroxene grains. We have calculated the closure temperature for Pb in clinopyroxene using $D_{0}=44\,{\rm m}^{2}\,{\rm s}^{-1}$, and $E/R=62,400$ K (Cherniak, 1998). Al-Mg measurements have been taken in a variety of minerals, including anorthite, melilite, pyroxene and spinel. For Mg diffusion in anorthite, we assume $D_{0}=1.2\times 10^{-6}\,{\rm m}^{2}\,{\rm s}^{-1}$ and $E/R=33440$ K (LaTourrette and Wasserburg, 1998). For Mg diffusion in melilite, we assume $D_{0}=3.02\times 10^{-9}\,{\rm m}^{2}\,{\rm s}^{-1}$ and $E/R=28990$ K (Ito and Ganguly, 2009). For Mg diffusion in pyroxene, we assume $D_{0}=1.9\times 10^{-9}\,{\rm m}^{2}\,{\rm s}^{-1}$ and $E/R=24300$ K (Müller et al., 2013). For Mg diffusion in spinel, we assume $D_{0}=2.77\times 10^{-7}\,{\rm m}^{2}\,{\rm s}^{-1}$ and $E/R=38560$ K (Liermann and Ganguly, 2002). In general, during transient heating events, Pb-Pb is easier to reset than Al- Mg. In Figure 1 we plot the calculated closure temperatures $T_{\rm c}$ of Pb and Mg in clinopyroxene, as functions of the cooling rate for grain sizes of 20 and $200\,\mu{\rm m}$. At slow cooling rates characteristic of secular cooling of asteroids, $\sim 100\,{\rm K}\,{\rm Myr}^{-1}$ ($3\times 10^{-12}\,{\rm K}\,{\rm s}^{-1}$), $T_{\rm c}\approx 1040$ K for Pb, but lower ($\approx 670$ K) for Mg. The order of isotopic closure tends to reverse for small inclusions in transient heating events with cooling rates $\sim 1-10^{3}\,{\rm K}\,{\rm hr}^{-1}$ ($3\times 10^{-4}-0.3\,{\rm K}\,{\rm s}^{-1}$). Figure 1: Closure temperatures of Pb in clinopyroxene (blue curves) and Mg in clinopyroxene (red curves), as a function of cooling rate, assuming grain sizes (diameters) of $200\,\mu{\rm m}$ (top), and $20\,\mu{\rm m}$ (bottom). For slow cooling rates typical of secular cooling on asteroids (e.g., 100 K/Myr), Al-Mg has a lower closure temperature than the Pb-Pb system and is more easily reset. For a cooling rate 500 K/hr, associated with transient heating events experienced by CAIs or chondrules, the closure temperatures are reversed. At a hypothetical peak temperature of $\approx 1730$ K and cooling rate 500 K/hr, the Al-Mg system would remain closed in clinopyroxene grains $100\,\mu{\rm m}$ in size (closure temperature 1764 K, red diamond), but the Pb-Pb system in pyroxene grains $100\,\mu{\rm m}$ in size would be disturbed (closure temperature 1704 K, blue diamond). The quenched or “volcanic” angrites (D’Orbigny, SAH 99555, NWA 1670) are widely accepted, based on petrologic features suggesting a lava erupted onto an asteroid surface, to have cooled very rapidly, at rates $\approx 300\,{\rm K}\,{\rm hr}^{-1}$ in the 1273 - 1573 K range (Keil, 2012). Likewise, NWA 7325 shares these petrologic features, and is inferred from Mg diffusion in plagioclase to have cooled at rates $\sim 500-650\,{\rm K}\,{\rm hr}^{-1}$, at high temperatures (Yang et al., 2019). The achondrite NWA 6704 was inferred by Hibiya et al. (2019) from textures to have cooled at rates $1-10^{2}\,{\rm K}\,{\rm hr}^{-1}$ at high temperatures, and from olivine-spinel geospeedometry to have cooled at rates $<10^{-2}\,{\rm K}\,{\rm hr}^{-1}$ at 1123 - 1243 K. The cooling rate of Asuka 881394 has not been constrained petrologically; but given its early time of formation in common with the achondrites above, we assume it also cooled rapidly. It is very likely that these achondrites formed from complete melts at high temperatures and continued to cool below the closure temperatures of Al-Mg and Pb-Pb in a matter of years, so that these two systems closed effectively simultaneously. Al-Mg and (properly determined) Pb-Pb ages must be concordant in these achondrites. In contrast to the achondrites, the Al-Mg and Pb-Pb systems will not necessarily have closed at the same time in chondrules and CAIs. This is not because they were not rapidly cooled, but because their last heating event did not reset both chronometers. Chondrules are silicate melts that crystallized following a transient heating event in the solar nebula, typified by peak temperatures anywhere from subsolidus or barely above the solidus (Wasson and Rubin, 2003; Ruzicka et al., 2008), to $>1800$ K or more, and cooling rates $\sim 10-10^{3}\,{\rm K}\,{\rm hr}^{-1}$ (Desch et al., 2012). Many chondrules appear to have experienced multiple such events. A representative thermal history might be heating to a peak temperature $\approx 1700-1800$ K, followed by cooling at $500\,{\rm K}\,{\rm hr}^{-1}$ (a cooling rate consistent with all chondrule textures). A majority of CAIs also were transiently heated in the solar nebula. Compact type A CAIs were at least partially melted or annealed, and type B and C CAIs have igneous textures showing they were partially or completely melted at least once. During their last heating event, type B CAIs appear from zoning in melilite to have cooled at rates of $0.5-50\,{\rm K}\,{\rm hr}^{-1}$ (Stolper and Paque, 1986; Jones et al., 2000), or (from the need to limit diffusion of oxygen isotopes) $>6\times 10^{4}\,{\rm K}\,{\rm hr}^{-1}$ (Kawasaki et al., 2021). These hours-long cooling times are of the same order as the cooling times experienced by chondrules. Both chondrules and CAIs resided in the same regions of the solar nebula, for the same $\sim 2-4$ Myr it took before chondrites formed (Desch et al., 2018). Given that a large fraction of the material in chondrites is chondrules, all of which were transiently heated by some event, it would be remarkable if CAIs avoided being transiently heated as well. It seems very likely that CAIs must have been transiently heated, some to the melting point and some to only subsolidus temperatures, over the same period of time as chondrules (1-3 Myr after $t\\!\\!=\\!\\!0$; Villeneuve et al. 2009). In fact, evidence exists of this late heating of CAIs (Mane et al., 2022). These transient heating events could have reset isotopic systems in chondrules and CAIs, depending on the peak temperatures and the cooling rates. If a CAI were to be heated to peak temperatures of about 1750 K, and cool at $500\,{\rm K}\,{\rm hr}^{-1}$, pre-existing minerals would survive. Pyroxene (enstatite) grains survive to temperatures of 1830 K (Greenwood and Hess, 1996), although the more Ti-rich pyroxene more typically found in CAIs may partially melt at lower temperatures $\approx 1500$ K (A. Davis, personal communication). Spinel grains survive to similar temperatures $>1800$ K (Whattam et al., 2022). For $20\,\mu{\rm m}$ grains ($a=10\mu{\rm m}$) and cooling rates of 500 K/hr, we calculate closure temperatures of 2071 K in spinel, 2022 K in melilite, 1764 K in pyroxene, and 1687 K in anorthite. With the possible exception of anorthite, which would be slightly disturbed, a heating event with peak temperatures just above 1700 K would not reset the Al-Mg chronometer. In contrast, the closure temperature of the Pb-Pb system in $20\,\mu{\rm m}$ pyroxene grains would be no higher than 1703 K, so Pb-Pb ages would be reset, especially if the more Ti-rich pyroxene grains were to partially melt. The types of transient heating events experienced by chondrules and assuredly also CAIs are capable of resetting the Pb-Pb chronometer without necessarily resetting the Al-Mg chronometer, at least if it Al-Mg ages are determined using isochrons based on pyroxene and spinel. Chondrules and CAIs may have formed at $t\\!\\!=\\!\\!0$, or any time over the next 2-3 Myr, before a transient heating event reset their Pb-Pb ages without necessarily resetting their Al-Mg ages. Therefore chondrules and CAIs are not samples that reliably or necessarily test the homogeneity hypothesis. It has been noted before that the discrepancies between the Al-Mg and Pb-Pb systems imply a late-stage resetting of the Pb-Pb system. Bouvier and Wadhwa (2010) suggested that diffusion of Pb in pyroxene might be faster than diffusion of Mg in melilite or anorthite, although they did not outline a scenario in which this would occur, identify a nebular or parent-body setting, or quantify the thermal histories that would be required. With these caveats, we examine the ages recorded by each sample. ### 2.3 Achondrites #### 2.3.1 D’Orbigny (quenched angrite) D’Orbigny is a quenched angrite with an unshocked and unbrecciated texture consisting of large laths of anorthite intergrown with Al-Ti-bearing diopside- hedenbergite, Ca-rich olivine and kirschsteinite as well as abundant glasses and round vugs (Keil, 2012). The glass is not of shock origin or externally introduced at a later time (Glavin et al., 2004). The glass and vugs speak to a rapid cooling rate $\sim 300\,{\rm K}\,{\rm hr}^{-1}$ (Keil, 2012). D’Orbigny has long been considered an anchor in which the different isotopic systems likely closed simultaneously. The initial $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ ratio in D’Orbigny has been determined by multiple groups: Spivak-Birndorf et al. (2009) found $(5.06\pm 0.92)\times 10^{-7}$, MSWD $=2.5$; Schiller et al. (2010) found $(3.88\pm 0.27)\times 10^{-7}$, MSWD $=1.9$; and Schiller et al. (2015) found $(3.98\pm 0.15)\times 10^{-7}$, MSWD $=1.9$. Combining these data and rejecting one outlier, Sanborn et al. (2019) found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(3.93\pm 0.39)\times 10^{-7}$, the value we adopt. It has also been reported to have been $(3.97\pm 0.21)\times 10^{-7}$, but from an isochron with MSWD above the threshold for acceptance (Kleine and Wadhwa, 2017). For our adopted initial ratio of $5.23\times 10^{-5}$, we find a time of formation $\Delta t_{26}=5.06\pm 0.10\,{\rm Myr}$ (Equation 1). Amelin (2008b) measured Pb isotopic ratios using whole-rock and pyroxene fractions in D’Orbigny. Excluding two outliers with excess initial Pb, and one with low Pb concentration, and using 9 of the 13 fractions, their regression yields an intercept corresponding to an age $4564.42\pm 0.12$ Myr, and MSWD $=1.18$ (assuming ${}^{238}{\rm U}/{}^{235}{\rm U}=137.88$), We concur with their selection of points and confirm their regression, finding $4564.40\pm 0.13$ Myr, with MSWD $=1.14$. The ${}^{238}{\rm U}/{}^{235}{\rm U}$ ratio in whole-rock samples of D’Orbigny has been measured by Brennecka and Wadhwa (2012), who found $137.780\pm 0.021$, and Tissot et al. (2017), who found $137.793\pm 0.025$. We take the weighted mean of these, or $137.785\pm 0.016$, and use this to update the age and its uncertainty according to equations 3 and 4. We find $4563.42\pm 0.21$ Myr. An important caveat is that the ${}^{238}{\rm U}/{}^{235}{\rm U}$ ratio was measured in whole-rock samples, whereas the Pb-Pb ages were determined using data primarily from Pb in (clino)pyroxene (especially the points with the most radiogenic Pb). This matters because the U in pyroxene appears to be isotopically light, most probably because of isotopic fractionation during magmatic differentiation; Tissot et al. (2017) model this process and conclude that pyroxene grains have $\delta^{238}{\rm U}$ roughly $0.25\mbox{\text{\textperthousand}}$ below the value for $\delta^{238}{\rm U}$ in the silicate melt in both D’Orbigny and Angra dos Reis. Whether the ${}^{238}{\rm U}/{}^{235}{\rm U}$ ratio in the pyroxene grains that figure into the isochron is well represented by the ratio in the bulk samples depends on $f_{\rm cpx}$, the fraction of all U residing in clinopyroxene, with $\delta^{238}{\rm U}_{\rm cpx}=\delta^{238}{\rm U}_{\rm bulk}-(1-f_{\rm cpx})\,(0.25\mbox{\text{\textperthousand}}).$ (6) Tissot et al. (2017) estimate $f_{\rm cpx}\approx 50\%$ in most angrites and similar achondrites, so that using whole-rock U isotopic ratios overestimates the ages of these achondrites, by about $0.19$ Myr, but possibly by up to $0.38$ Myr ($f_{\rm cpx}=0$), or not at all ($f_{\rm cpx}=1$). We therefore favor $4563.24\pm 0.21\,{\rm Myr}$. #### 2.3.2 SAH 99555 (quenched angrite) SAH 99555 is a quenched angrite with an unshocked, fine-grained texture composed of anorthite, Al-Ti-bearing hedenbergite, olivine and mm-sized vesicles (Keil, 2012). We take the value $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(3.64\pm 0.18)\times 10^{-6}$ determined by Schiller et al. (2015). This yields a time of formation $\Delta t_{26}=5.14\pm 0.05\,{\rm Myr}$. The Pb isotopic ratios of SAH 99555 were determined by both Amelin (2008a) and Connelly et al. (2008). Amelin (2008a) built an isochron using 8 out of 10 whole-rock fractions, rejecting two clear outliers. Assuming ${}^{238}{\rm U}/{}^{235}{\rm U}=137.88$, they found an age $4564.86\pm 0.38$ Myr, MSWD $=1.5$. We essentially reproduce this regression, finding $4564.88\pm 0.10$ Myr, MSWD $=1.5$. Connelly et al. (2008) built an isochron using 8 of 11 whole-rock leachates, plus the residue from the pyroxene samples. Assuming ${}^{238}{\rm U}/{}^{235}{\rm U}=137.88$, they found an age $4564.58\pm 0.14$ Myr, MSWD $=0.99$. We again essentially reproduce this regression, finding $4564.58\pm 0.07$ Myr, MSWD $=1.17$. The weighted mean of the ages is $4564.614\pm 0.131$ Myr. For the U isotopic rate, we use the weighted mean of the values measured by Tissot et al. (2017), ${}^{238}{\rm U}/{}^{235}{\rm U}=137.805\pm 0.029$, and by Connelly et al. (2012), ${}^{238}{\rm U}/{}^{235}{\rm U}=137.784\pm 0.024$ (after renormalizing to the same U standard as Tissot et al. 2017), and adopt ${}^{238}{\rm U}/{}^{235}{\rm U}=137.793\pm 0.019$. With this correction, we find an age $4563.70\pm 0.24\,{\rm Myr}$. Again we note the caveat that the ${}^{238}{\rm U}/{}^{235}{\rm U}$ ratio is measured in the bulk sample, not the pyroxene grains, meaning the value of $t_{\rm Pb}$ should be reduced by about 0.19 Myr. We therefore favor $4563.51\pm 0.24\,{\rm Myr}$. #### 2.3.3 NWA 1670 (quenched angrite) NWA 1670 is a quenched angrite with a porphyritic texture including large olivine megacrysts in a fine-grained matrix of olivine, pyroxene, kirsch- steinite and anorthite as well as other accessory minerals (Keil, 2012). These indicate rapid cooling at $\sim 300\,{\rm K}\,{\rm hr}^{-1}$ (Mikouchi et al., 2003). It was analyzed by Schiller et al. (2015), who found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(5.92\pm 0.59)\times 10^{-7}$, which yields a time of formation $\Delta t_{26}=4.63\pm 0.10\,{\rm Myr}$. Schiller et al. (2015) also measured Pb isotopic ratios to determine a Pb-Pb age of $4565.39\pm 0.24$ Myr, assuming ${}^{238}{\rm U}/{}^{235}{\rm U}=137.786\pm 0.013$. We note that this value appears to have been taken from whole-rock measurements of other angrites (Connelly et al., 2012). Besides this, we were not able to confirm this age, because it is unclear which points to include in the regression. This isochron was based on using only six of the 13 available fractions, with no clear criteria for including or excluding points from the regression. As a result, the uncertainty in the age has been considerably underestimated, as we now demonstrate. As described above, the absolute age of a sample can be found by performing a linear regression on the data $y_{i}={}^{207}{\rm Pb}/{}^{206}{\rm Pb}$ vs. $x_{i}={}^{204}{\rm Pb}/{}^{206}{\rm Pb}$, where each point $i$ represents an isotopic measurement in a different wash, leachate, or residue of a sample after it has been dissolved in various acids. Provided there is only one non- radiogenic source of lead—either primordial lead or terrestrial contamination—the data should array along a line, and the $y$-intercept of that line yields the absolute age according to Equation 2. Primordial lead is a likely component of meteorites and inclusions, and terrestrial contamination is pervasive (Gopel et al., 1994; Palk et al., 2018), so it is likely that one of the fractions may contain an anomalous amount of one source of Pb or the other, and fall off the isochron. It is almost always necessary to exclude one or more points from the Pb-Pb isochron regression, and this is indeed the case for all the other Pb-Pb isochrons discussed in this paper. For example, in their analyses of D’Orbigny and SAH 99555, Amelin (2008b) and Amelin (2008a), excluded from their regressions data points with elevated common Pb, or low total Pb. In other analyses of Asuka 881394 (Wadhwa et al., 2009; Wimpenny et al., 2019), NWA 2976 (Bouvier et al., 2011b), and NWA 6704 (Amelin et al., 2019), points were excluded essentially only if the ${}^{206}{\rm Pb}/{}^{204}{\rm Pb}$ was below a threshold value (indicating too little radiogenic Pb). In contrast, Connelly et al. (2008) (SAH 99555), Schiller et al. (2015) (NWA 1670), and Bollard et al. (2017) (NWA 5697 chondrules) did not include or exclude points based on stated criteria such as dissolution order or threshold ${}^{206}{\rm Pb}/{}^{204}{\rm Pb}$. We believe they instead excluded up to half their data based on whether the data fell off a pre-determined isochron. This procedure is prone to confirmation bias, as we illustrate using the example of NWA 1670. Schiller et al. (2015) obtained Pb isotopic data for NWA 1670 from the following fractions: two acid washes of whole-rock samples (W1, W2), eight leachates (L1-L8) and a residue (R) of whole-rock samples, and two leachates (C-L3, C-L4) of clinopyroxene grains. In Figure 2 we plot these 13 data (except for W1 and W2, with very high ${}^{204}{\rm Pb}/{}^{206}{\rm Pb}\approx 0.01$). When we perform a York regression (York et al., 2004) on the same 6 data points selected by Schiller et al. (2015) (L1, L2, L4, L5, L7, C-L3), we infer a slope $4.093\pm 0.009$ and intercept $0.624205\pm 0.000038$, with MSWD $=0.30$. With ${}^{238}{\rm U}/{}^{235}{\rm U}=137.786\pm 0.013$, we infer a Pb-Pb age of $4564.40\pm 0.22$ Myr. This is almost identical to the result obtained by Schiller et al. (2015), who found an age $4564.39\pm 0.24$ Myr, with MSWD $=0.31$. Except perhaps for W1 and W2, the reasons for excluding data points from the regression were not based on wash order or ${}^{204}{\rm Pb}/{}^{206}{\rm Pb}$ ratio, as this would not explain why L4 and L7 were included, but L6 and L8 were not, or why C-L3 was included but C-L4 was not. Schiller et al. (2015) state that the reasons for excluding data points was the need to have components that only sampled radiogenic and terrestrial Pb, not initial Pb. They point out that the regression line is consistent with mixing between radiogenic and terrestrial Pb; we also find that this regression line lies only 0.0038 above the terrestrial Pb point (${}^{204}{\rm Pb}/{}^{206}{\rm Pb}=0.0542$, ${}^{207}{\rm Pb}/{}^{206}{\rm Pb}=0.84228$; Stacey and Kramers 1975). Schiller et al. (2015) interpret points above the line (e.g., L3) to contain minor amounts of initial Pb, and points below the line (e.g., L6, L8, R, C-L4) to have lost radiogenic Pb. Notably, these interpretations are not based on any physical characteristics of the samples, but based on where they lie relative to the regression line. Thus, this interpretation is circular logic. Points were excluded from the regression; those points that were not included in the regression fell off the regression line; for that reason alone, their exclusion was justified. The fits obtained by such analyses have low MSWD, but significantly, an MSWD of only 0.30 is “too good to be true”. The probability of six data points with random measurement errors exhibiting a fit with such a low MSWD is $\approx 5\%$ (Wendt and Carl, 1991), which is usually the threshold for acceptability. This entire approach is prone to confirmation bias. Figure 2: Pb-Pb isochron of NWA 1670, based on the six data points regressed by Schiller et al. (2015), in black, with excluded points in red. The intercept of this isochron, and its uncertainty, yield an age $4564.40\pm 0.21$ Myr. Indeed, we have found a range of combinations of points that yield more probable isochrons (in terms of MSWD) and a wide variety of ages. In Figure 3, we show a regression using a different set of five points (W1, L1, L2, L6, C-L3). We infer a slope $4.151\pm 0.006$ and intercept $0.623937\pm 0.000036$, with MSWD $=1.71$. This line passes only 0.006 above terrestrial Pb, and a regression of five points with MSWD $=1.71$ is actually just as likely as one with 0.30. The points that lie below the line (L8, R, C-L4) can be equally interpreted as having lost radiogenic Pb, and the points above the line (L3, L4, L5, L7) as having incorporated initial Pb. Thus, this is an equally valid isochron, but the age derived from it is $4563.77\pm 0.21$ Myr, 0.63 Myr younger than the age derived in Figure 1. Figure 3: Pb-Pb isochron of NWA 1670, based a different subset of data points. The intercept of this isochron, and its uncertainty, yield an age $4563.77\pm 0.21$ Myr, 0.63 Myr younger than the age from Figure 1. Likewise, in Figure 4, we show a regression using a different set of six points (L4, L5, L6, L7, R, C-L3). We infer a slope $4.000\pm 0.018$ and intercept $0.624309\pm 0.000051$, with MSWD $=1.12$. This line passes only 0.001 below terrestrial Pb, and a regression of six points with MSWD $=1.12$ is very probable. The points that lie below the line (L8, C-L4) can be equally interpreted as having lost radiogenic Pb, and the points above the line (L1, L2, L3) as having incorporated initial Pb. This isochron is as valid as the others, perhaps even more so because it passes closest to terrestrial Pb, has the most probable MSWD, and there is some order to which fractions are included in the regression. The age derived from this regression is $4564.64\pm 0.23$ Myr, 0.24 Myr older than the isochron derived in Figure 1. Figure 4: Pb-Pb isochron of NWA 1670, based a different subset of data points. The intercept of this isochron, and its uncertainty, yield an age $4564.64\pm 0.23$ Myr, 0.24 Myr older than the age from Figure 1. Through these examples, we have established that two isochrons equally valid to the one derived by Schiller et al. (2015) yield a range of ages, from 0.63 Myr younger to 0.24 Myr older. These isochrons are equally valid, because they use equal numbers of apparently equally valid points. Without objective criteria like ${}^{204}{\rm Pb}/{}^{206}{\rm Pb}$ ratio or Pb concentration to select points, points included in our regressions can equally well be argued to be valid because they lie on the regression line; points excluded from the regression can equally well be argued to exhibit loss of radiogenic Pb or excess of primordial Pb. Without further information, we consider the 95% confidence interval of possible Pb-Pb ages of NWA 1670 to extend from 4563.55 to 4564.87 Myr, and we take its Pb-Pb age to be $4564.21\pm 0.66\,{\rm Myr}$. This encompasses the age $4564.39\pm 0.24$ Myr reported by Schiller et al. (2015), but acknowledges a larger uncertainty. This age uncertainty may be magnified, depending on whether the isochron draws mostly from pyroxenes or whole rock washes and leachates, because of isotopic fractionation of U isotopes. Again we note the caveat that the ${}^{238}{\rm U}/{}^{235}{\rm U}$ ratio is measured in the bulk sample, not the pyroxene grains, meaning the value of $t_{\rm Pb}$ should be reduced by about 0.19 Myr. We therefore favor $4564.02\pm 0.66\,{\rm Myr}$. #### 2.3.4 Asuka 881394 (eucrite-like achondrite) Asuka 881394 is a eucrite-like achondrite with a coarse-grained igneous texture with near equal amounts of anorthite and pyroxene. The granoblastic texture suggests post-formation, low-grade metamorphism could have affected Asuka 881394 (Wimpenny et al., 2019). While originally classified as a cumulate eucrite (Takeda, 1997), Asuka 881394 has geochemcial and isotopic qualities that preclude classification as such. As discusssed by Wimpenny et al. (2019), the major element chemistry of the primary phases do not resemble cumulate eucrites (e.g., its plagioclase is too calcic, and the Mg-rich pyroxenes do not show evidence of inversion textures). Additionally, the $\Delta^{17}{\rm O}$ oxygen isotope composition of Asuka 881394 is 15${\sigma}$ above the mean value for HEDs (Scott et al., 2009). The Al-Mg systematics of Asuka 881394 have been measured by three groups: Nyquist et al. (2003), who found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(1.18\pm 0.14)\times 10^{-6}$, reanalyzed by Wimpenny et al. (2019) as $(1.18\pm 0.31)\times 10^{-6}$; Wadhwa et al. (2009), who found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(1.28\pm 0.07)\times 10^{-6}$, eliminating one outlier from the regression; and Wimpenny et al. (2019), who found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(1.48\pm 0.12)\times 10^{-6}$. We take the weighted average of these, $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(1.31\pm 0.06)\times 10^{-6}$. This yields a time of formation $\Delta t_{26}=3.82\pm 0.04\,{\rm Myr}$. Pb-Pb dating of Asuka 881394 was done by Wimpenny et al. (2019), who built an isochron using their pyroxene and whole-rock residue data, plus one whole-rock wash point, plus the most radiogenic residues (${}^{206}{\rm Pb}/{}^{204}{\rm Pb}>400$) from the analysis by Wadhwa et al. (2009). They also determined ${}^{238}{\rm U}/{}^{235}{\rm U}=137.786\pm 0.038$ in the bulk rock, and found a Pb-Pb age of $4564.95\pm 0.53\,{\rm Myr}$, based on an isochron with MSWD $=1.4$. Again we note the caveat that the ${}^{238}{\rm U}/{}^{235}{\rm U}$ ratio is measured in the bulk sample, not the pyroxene grains, meaning the value of $t_{\rm Pb}$ should be reduced by about 0.19 Myr. We therefore favor $4564.76\pm 0.53\,{\rm Myr}$. ### 2.4 NWA 7325 (ungrouped achondrite) NWA 7325 is an ungrouped achondrite with a medium-grained cumulate texture consisting of Mg-rich olivine, Cr-bearing diopside and Ca-rich plagioclase (Goodrich et al., 2017). The positive Eu anomaly in NWA 7325 is a geochemical sign of being a cumulate rock, but also can be interpreted as the result of melting or of a basaltic or gabbroic lithology (Barrat et al., 2015). The Al-Mg systematics of NWA 7325 were measured by Koefoed et al. (2016), who found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(3.03\pm 0.14)\times 10^{-7}$, which yields a time of formation $\Delta t_{26}=5.33\pm 0.05\,{\rm Myr}$. Koefoed et al. (2016) also analyzed its Pb-Pb systematics, restricting their regression to pyroxene residues, and building an isochron that excluded those points with ${}^{206}{\rm Pb}/{}^{204}{\rm Pb}<50$, as they included obvious terrestrial contamination. They did not measure the U isotopic ratio, instead adopting a value ${}^{238}{\rm U}/{}^{235}{\rm U}=137.794$ as representative of materials from the inner Solar System (Goldmann et al., 2015). Koefoed et al. (2016) found a Pb-Pb age for NWA 7325 of $4563.4\pm 2.6\,{\rm Myr}$. More recently, Cartwright et al. (2016) analyzed NWA 8486, which is paired with NWA 7325. Combining the data for the two achondrites, and apparently assuming the same U isotopic ratio, they found an age of $4563.9\pm 1.7$ Myr. As it is consistent with the Koefoed et al. (2016) result but provides a more extreme test, we adopt this value with its smaller uncertainties. Again we note the caveat that the ${}^{238}{\rm U}/{}^{235}{\rm U}$ ratio is measured in the bulk sample, not the pyroxene grains, meaning the value of $t_{\rm Pb}$ should be reduced by about 0.19 Myr. We therefore favor $4563.7\pm 1.7\,{\rm Myr}$. #### 2.4.1 NWA 2976 (ungrouped carbonaceous achondrite) NWA 2976 (paired with NWA 011) is an unshocked, unbrecciated, ungrouped achondrite with coarse-grained pigeonite surrounded by fine-grained plagioclase with prevalent 120∘ triple junctions. (Yamaguchi et al., 2002). The 54Cr, 50Ti and ${\Delta}^{17}$O isotope systematics of NWA 011 have strong affinities to carbonaceous chondrites, and to the CR clan of chondrites in particular (Floss et al., 2005; Sanborn et al., 2019). The major element data presented by Floss et al. (2005) point to oxidizing conditions during formation of NWA 011/2976. The Al-Mg systematics of NWA 2976 were measured by Bouvier et al. (2011b), who found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(3.94\pm 0.16)\times 10^{-7}$. Schiller et al. (2010) also reported $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(4.91\pm 0.46)\times 10^{-7}$ in NWA 2976. Sugiura and Yamaguchi (2007) reported $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(6.93\pm 2.12)\times 10^{-7}$ in the paired achondrite NWA 011. We adopt the weighted mean of the values from Bouvier et al. (2011b) and Schiller et al. (2010), $(4.05\pm 0.15)\times 10^{-7}$ and find a time of formation $\Delta t_{26}=5.03\pm 0.04\,{\rm Myr}$. For NWA 2976, Bouvier et al. (2011b) regressed the five whole-rock and pyroxene samples obtained from the last dissolution step, with ${}^{206}{\rm Pb}/{}^{204}{\rm Pb}>1010$. They measured ${}^{238}{\rm U}/{}^{235}{\rm U}=137.751\pm 0.018$, and found a Pb-Pb age for NWA 2976 of $4262.89\pm 0.59$ Myr from an isochron with MSWD $=0.02$. We agree with their choice of data to regress and reproduce their regression. Connelly et al. (2012) measured an isotopic ratio ${}^{238}{\rm U}/{}^{235}{\rm U}=137.787\pm 0.011$. We combine these, finding ${}^{238}{\rm U}/{}^{235}{\rm U}=137.777\pm 0.009$, which then yields an age $4563.16\pm 0.57\,{\rm Myr}$, As before, it is important to note that the U isotopic ratio was measured in whole-rock samples, and may not represent the isotopic ratio in the pyroxenes. Here we argue, though, that for the two ‘carbonaceous’ achondrites NWA 2976 and NWA 6704, likely $f_{\rm cpx}\approx 1$, most of the U probably resides in the pyroxene grains, and the whole-rock isotopic ratio probably is appropriate. Both of these achondrites are more oxidized. As members of the CC isotopic reservoir, they would be expected to contain more water and other volatiles. Both Warren et al. (2013) and Hibiya et al. (2019) estimated that the CC achondrite NWA 6693, paired with NWA 6704, formed under oxygen fugacity conditions 2 log units above the iron-wüstite buffer ($\Delta{\rm IW}=+2$), as opposed to angrites, which formed at $\Delta{\rm IW}=+1$ (Brett et al., 1977; Jurewicz et al., 1993; McKay et al., 1994; Tissot et al., 2022). More importantly, Izawa et al. (2022) concluded that the paired achondrite NWA 011, paired with NWA 2976, experienced hydrous magmatism and a water-bearing melt. Under these redox conditions, U is expected to be mostly in the U(iv) $4^{+}$ valence state (as insoluble UO2) with some in the U(vi) $6^{+}$ state as the soluble ${\rm UO}_{2}^{2+}$ uranyl ion. (Chevreux et al., 2021). At IW+2, compared to IW+1, slightly more U speciates as U(vi) instead of U(iv), but the more important factor is the presence of H2O in the magma. This could drastically change the partitioning of U due to the ability of water to depolymerize the aluminosilicate network and produce Non-Bridging Oxygen (NBO) sites. These are the coordination sites by which U (of either valence) can enter a growing crystal. We expect that in a hydrous magma, U will readily partition into the major crystallizing phases, as opposed to remaining in the melt to partition into late-forming phosphates as in an anhydrous magma. As a result, the pyroxene grains will take up most the U and share the same U isotopic ratio as the whole rock. Assuming $f_{\rm cpx}\approx 1$, we do not apply an age correction to NWA 2976 or NWA 6704. By our reasoning, it should not be applied to any achondrite formed from hydrous magmatism. #### 2.4.2 NWA 6704 (ungrouped carbonaceous achondrite) NWA 6704 is an unshocked, ungrouped achondrite with a medium-grained, cumulate texture comprised of low-Ca pyroxene along with Ni-rich olivine and sodic plagioclase (Hibiya et al., 2019). Warren et al. (2013) argue the coarse texture and bulk subchondritic MgO/SiO2 indicate origin as an igneous cumulate. However, Hibiya et al. (2019) argue that these same textural features are best explained by a rapid initial crystallization, which agrees with the apparent absence of geochemical evidence for significant magmatic differentiation. The Ni-rich phases along with the V/(Al+Cr) ratio in the spinels suggest formation at relatively oxidizing oxygen fugacity $\Delta{\rm IW}=+2$, 2 log units above the Iron-Wüstite) buffer (Warren et al., 2013). Both Hibiya et al. (2019) and Warren et al. (2013) demonstrate the volatile- depleted nature of NWA 6704/6693. Sanborn et al. (2019) showed the ${\Delta}^{17}$O, ${\epsilon}^{50}$Ti and ${\epsilon}^{54}$Cr isotope composition of NWA 6704/6693 agree with the CR clan of chondrites. The Al-Mg systematics of NWA 6704 were measured by Sanborn et al. (2019), who found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(3.15\pm 0.38)\times 10^{-7}$, the value we adopt, which yields a time of formation $\Delta t_{26}=5.29\pm 0.13\,{\rm Myr}$. The Pb-Pb age of NWA 6704 was determined by Amelin et al. (2019), who regressed 13 pyroxene fraction data with ${}^{206}{\rm Pb}/{}^{204}{\rm Pb}>700$, excluding one outlier. Based on a measured ${}^{238}{\rm U}/{}^{235}{\rm U}=137.7784\pm 0.0097$, their regression yielded an age $4562.76^{+0.22}_{-0.30}\,{\rm Myr}$. As with NWA 2976, we assume the U isotopic ratio in the pyroxenes is well represented by the whole rock ratio, and do not apply a correction. ### 2.5 NWA 5697 (L3) Chondrules The Al-Mg and Pb-Pb systematics were simultaneously measured by Bollard et al. (2017) and Bollard et al. (2019) for eight chondrules from the carbonaceous chondrite Allende (CV3) and the ordinary chondrite NWA 5697 (L3). However, although Bollard et al. (2017) reported Pb-Pb dates for eight chondrules analyzed for Al-Mg, only four of these (NWA 5697: 2-C1, 3-C5, 5-C2, 11-C1) were U-corrected using a ${}^{238}{\rm U}/{}^{235}{\rm U}$ measured in the chondrule; the rest (NWA 5697: 5-C10, C1, C3; and Allende: C30) assumed a bulk chondrite value ${}^{238}{\rm U}/{}^{235}{\rm U}=137.786\pm 0.013$. Because some chondrules are resolvably different from this assumed value (e.g., NWA 5697 5-C2 has ${}^{238}{\rm U}/{}^{235}{\rm U}=137.756\pm 0.029$), we do not assume all chondrules should have this value, and we restrict our analysis only to those four measured ordinary chondrite chondrules. As well, Bollard et al. (2017) constructed their isochrons using a subset of the fractions. As with the case of NWA 1670, we have found it necessary to re-analyze the Pb-Pb isochrons of these four chondrules. Each of these recalculated Pb-Pb ages of the chondrules encompasses the values reported by Bollard et al. (2017), but acknowledges a much greater uncertainty. Chondrule 2-C1: For this chondrule, Bollard et al. (2017) regressed 11 out of 20 fractions (L3, L4, L6, L8, L9, R, W11-2, L3-2, L4-2, L7-2, L8-2) and reported an age $4567.57\pm 0.56$ Myr and MSWD $=1.20$. Performing the same regression, we find a similar $4567.45\pm 0.48$ Myr and MSWD $=1.12$. Regressing a different subset of 11 fractions (L3, L4, L5, L6, L8, L9, R, L4-2, L7-2, L8-2, L9-2) yields an age $4567.33\pm 0.48$ Myr and MSWD $=1.88$. And regressing yet another subset of 11 fractions (L1, L2, L7, L9, R, L2-2, L3-2, L4-2, L7-2, L8-2, L9-2) yields an age $4567.85\pm 0.50$ Myr and MSWD $=0.44$. We take the age of chondrule 2-C1 to be $4567.60\pm 0.75\,{\rm Myr}$. Bollard et al. (2019) found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(7.56\pm 0.13)\times 10^{-6}$ (MSWD=1.3), which yields a time of formation $\Delta t_{26}=2.00\pm 0.21\,{\rm Myr}$. Chondrule 3-C5: For this chondrule, Bollard et al. (2017) regressed 10 out of 14 fractions (W11, L1, L2, L3, L4, L5, L6, L7, L8, L9) and reported an age $4566.20\pm 0.63$ Myr and MSWD $=1.27$. Performing the same regression, we find a similar $4566.13\pm 0.51$ Myr and MSWD $=1.29$. Simply removing point L8 from the regression, we find an age $4565.54\pm 0.54$ Myr and MSWD $=0.46$. Regressing instead a different set of nine points (W11, L2, L4, L5, L6, L7, L8, L11, l12), we find instead an age $4566.57\pm 0.56$ Myr and MSWD $=0.63$. We take the age of chondrule 3-C5 to be $4566.07\pm 1.07\,{\rm Myr}$. Bollard et al. (2019) found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(7.04\pm 1.51)\times 10^{-6}$ (MSWD$=0.9$, which yields a time of formation $\Delta t_{26}=1.84\pm 0.21\,{\rm Myr}$. Chondrule 5-C2: For this chondrule, Bollard et al. (2017) regressed 8 out of 15 fractions (W11, L3, L4, L6, L7, L8, L9, L10) and reported an age $4567.54\pm 0.52$ and MSWD $=0.66$, a result we reproduce exactly. Regressing a subset of 9 fractions (L1, L2, L3, L4, L6, L7, L9, L10, L11) yields an age $4566.84\pm 0.56$ Myr and a more probable MSWD $=0.94$. It is not clear why the other points should be rejected. Regressing a subset of eight data points (W10, W11, L2, L3, L4, L6, L8, L10, L11) yields an age $4567.70\pm 0.49$ Myr and MSWD $=0.72$. We take the age of chondrule 5-C2 to be $4567.24\pm 0.95\,{\rm Myr}$. Bollard et al. (2019) found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(8.85\pm 1.83)\times 10^{-6}$ (MSWD$=1.8$), which yields a time of formation after $t\\!\\!=\\!\\!0$ of $\Delta t_{26}=2.07\pm 0.22\,{\rm Myr}$. Chondrule 11-C1: Finally, for this chondrule, Bollard et al. (2017) regressed 8 out of 15 fractions (W11, L4, L5, L6, L7, L8, L9, L10) and reported an age $4565.84\pm 0.72$ and MSWD $=1.16$. We find similar but different $4565.64\pm 0.55$ Myr, MSWD $=1.29$. Regressing a subset of eight data points (W11, L2, L4, L5, L7, L9, L10, L12) yields an age $4565.36\pm 0.59$ Myr and MSWD $=0.85$. Regressing a different subset of 10 fractions (W11, L2, L4, L5, L6, L7, L8, L9, L10, L12) yields an age $4565.69\pm 0.56$ Myr and MSWD $=1.77$. We take the age of chondrule 11-C1 to be $4565.51\pm 0.74\,{\rm Myr}$. Bollard et al. (2019) found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(5.55\pm 1.84)\times 10^{-6}$ (MSWD$=0.7$), which yields a time of formation $\Delta t_{26}=2.32\pm 0.34\,{\rm Myr}$. ### 2.6 CAIs Although the age of $t\\!\\!=\\!\\!0$ is commonly taken to be the Pb-Pb age of CAIs, the assumption that the Pb-Pb chronometer closed at the same time that the Al-Mg did (at $t\\!\\!=\\!\\!0$) is untested. Given that CAIs almost certainly spent millions of years in the same region of the solar nebula where chondrules formed, it is highly probable they could have been subject to transient heating events like those that formed chondrules. As discussed above (§2.2), this could have reset the Pb-Pb chronometer without resetting the Al- Mg chronometer. With this in mind, we ask whether any of the CAIs from which Connelly et al. (2012) determined a Pb-Pb age (22E, 31E, 33E, SJ101), actually show evidence that the Pb-Pb chronometer closed at $t\\!\\!=\\!\\!0$, when $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}\approx 5\times 10^{-5}$. To our knowledge, Al-Mg systematics have not been measured in CAI 33E. Larsen et al. (2011) measured bulk isotopic ratios in CAIs 22E and 31E, but not an internal isochron; if a late resetting of the Al-Mg system in these CAIs occurred, this would not have been detected by bulk measurements. This leaves only CAI SJ101. Amelin et al. (2010) measured ${}^{238}{\rm U}/{}^{235}{\rm U}=137.876\pm 0.043$ and an isochron with intercept $0.625000\pm 0.000092$ (MSWD $=1.07$) and derived a Pb-Pb age of $4567.18\pm 0.50$ Myr for this CAI. Its Al-Mg systematics were measured by MacPherson et al. (2017), who found $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(5.2\pm 0.5)\times 10^{-5}$. It is widely assumed that the event recorded by the Pb-Pb chronometer is the same as the one recorded by the Al-Mg chronometer, i.e., the CAI’s formation. This in turn rests on the assumption that any thermal event resetting the Pb-Pb chronometer would cause diffusion of Mg and reset the Al-Mg chronometer, as well. As discussed in §2.2, however, a typical chondrule-forming transient heating event would have reset the Pb-Pb chronometer, and also the Al-Mg system in anorthite, but not in melilite, pyroxene, or spinel. It is highly significant that the isochron built by MacPherson et al. (2017) is built from these three minerals, and that it clearly shows that anorthite is slightly disturbed, with points in anorthite falling below the isochron. The CAI SJ101 data admit the possibility that it formed and its Al-Mg chronometer was set at $t\\!\\!=\\!\\!0$, but that it was later transiently heated to peak temperatures $\approx 1750$ K and cooling rates $\approx 500\,{\rm K}\,{\rm hr}^{-1}$ like chondrules. It cannot be known that the Pb-Pb system closed at the same time as the Al-Mg system. For completeness, we also examine the Pb-Pb isochrons of other CAIs. For the CAI 22E, Connelly et al. (2012) reported a Pb-Pb age of $4567.35\pm 0.28$ Myr, based on a regression with MSWD $=0.91$ and measured $\mbox{${}^{238}{\rm U}/{}^{235}{\rm U}$}=137.627\pm 0.022$. They generated 20 different leachates/washes/residues in their analysis, but based the intercept on a regression involving only 11 of these points (L1, L3, L6, L7, L8, L9, L10, L11, W7, W8, W9). Performing a York regression on these data points, and propagating the uncertainties in the ${}^{238}{\rm U}/{}^{235}{\rm U}$ ratio and intercept in quadrature, we find a similar age $4567.32\pm 0.27$ Myr, on an isochron with MSWD $=1.08$. As was the case for NWA 1670, we were able to find many combinations of 11 data points that yielded acceptable isochrons (MSWD $<1.94$, the maximum for 11 points; Wendt and Carl 1991). For one set of 11 points (W1, W5, W6, W7, W8, W9, L3, L6, L9, L10, L11), we found MSWD $=0.29$, and an age $4567.21\pm 0.29$ Myr. For a different set of 11 points (W1, W5, W6, W7, W9, L1, L3, L4, L6, L7, L8), we found MSWD $=0.77$ and an age $4567.54\pm 0.31$ Myr. As with the case of NWA 1670, no physical criteria were laid out for the exclusion of data points. Connelly et al. (2008) argued that the points they included in the regression eliminated are free of contamination by terrestrial Pb, on the basis that the regression line extends to primordial Pb, and they eliminated points below this line (i.e., toward terrestrial Pb). However, their regression line falls below primordial Pb, suggesting (by their criteria) that actually all their data have some terrestrial contamination. Moreover, some included data (e.g., W7, L7) in fact plot above the regression line. In the absence of independent criteria for judging the isochrons, we consider all the isochrons described above to be equally valid, and conclude that the 95% confidence interval for the Pb-Pb age of CAI 22E should extend from 4566.92 to 4567.85 Myr, i.e., $4567.39\pm 0.47$ Myr. We find basically the same Pb-Pb age for 22E as Connelly et al. (2012), but find the uncertainty has been underestimated by about a factor of almost 2. We find similar outcomes for CAIs 31E and 32E. We find no issue with the reported Pb-Pb ages of the other CAIs SJ101, B1, or B4. It is worth noting that CAI B1 records $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(5.03\pm 0.26)\times 10^{-5}$ (Bouvier and Wadhwa, 2010). We conclude that CAIs actually do record Pb-Pb ages over a range of times from about 4568.2 Myr to 4567.2 Myr, and that their Pb-Pb ages may have been reset about 1 Myr after they formed. ### 2.7 Summary In Table 1 we convert the $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ initial ratios of each sample into a time of formation after $t\\!\\!=\\!\\!0$ assuming an abundance $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{\rm SS}=5.23\times 10^{-5}$ at $t\\!\\!=\\!\\!0$ and mean-life 1.034 Myr, and we compile our adopted Pb-Pb ages, $t_{\rm Pb}$, for each achondrite and chondrule, as well as the implied value of $t_{\rm SS}=t_{\rm Pb}+\Delta t_{26}$. For the first five achondrites we have applied the 0.19 Myr correction of Tissot et al. (2017), making the Pb-Pb ages younger. We have not applied this to the achondrites NWA 2976 or NWA 6704, because we assume their whole-rock U isotopic ratios reflect the composition of the pyroxene grains in which the Pb-Pb ages were determined. We have not applied the correction to chondrules because ${}^{238}{\rm U}/{}^{235}{\rm U}$ was measured in them directly. The uncertainties in $t_{\rm SS}$ are found by adding the uncertainties in $t_{\rm Pb}$ and $\Delta t_{26}$ in quadrature. Table 1: Adopted $\Delta t_{26}$ and Pb-Pb ages ages of seven bulk achondrites and four chondrules, plus the implied value of $t_{\rm SS}=t_{\rm Pb}+\Delta t_{26}$. Sample | $\Delta t_{26}$ (Myr) | $t_{\rm Pb}$ (Myr) | $t_{\rm SS}$ (Myr) ---|---|---|--- D’Orbigny | $5.06\pm 0.10$ | $4563.24\pm 0.21$ | $4568.30\pm 0.23$ SAH 99555 | $5.14\pm 0.05$ | $4563.51\pm 0.24$ | $4568.65\pm 0.25$ NWA 1670 | $4.64\pm 0.10$ | $4564.02\pm 0.66$ | $4568.66\pm 0.67$ Asuka 881394 | $3.82\pm 0.04$ | $4564.76\pm 0.53$ | $4568.58\pm 0.53$ NWA 7325 | $5.33\pm 0.05$ | $4563.71\pm 1.7$ | $4569.04\pm 1.7$ NWA 2796 | $5.03\pm 0.04$ | $4563.17\pm 0.57$ | $4568.20\pm 0.57$ NWA 6704 | $5.29\pm 0.13$ | $4562.76\pm 0.26$ | $4568.05\pm 0.29$ NWA 5697 2-C1 | $2.00\pm 0.21$ | $4567.60\pm 0.75$ | $4569.60\pm 0.78$ NWA 5697 3-C5 | $1.84\pm 0.21$ | $4566.07\pm 1.07$ | $4567.91\pm 1.09$ NWA 5976 5-C2 | $2.07\pm 0.22$ | $4567.24\pm 0.95$ | $4569.31\pm 0.97$ NWA 5697 11-C1 | $2.32\pm 0.34$ | $4565.51\pm 0.74$ | $4567.83\pm 0.81$ These samples all imply values of $t_{\rm SS}$ ranging from 4567.8 and 4569.6 Myr, centered around 4568.7 Myr. The average uncertainty in the Pb-Pb model ages $t_{\rm SS}$ is 0.8 Myr, due almost entirely to the uncertainties in the Pb-Pb ages of each sample. ## 3 Determination of $t_{\rm SS}$ ### 3.1 Statistical analysis As seen from Table 1, the majority of achondrites and chondrules imply that $t\\!\\!=\\!\\!0$ of the Solar System occurred more than 4568 Myr ago. The best estimate of the Pb-Pb age of $t\\!\\!=\\!\\!0$, $t_{\rm SS}^{*}$, is the weighted average of the estimates of $t_{{\rm SS},i}$ from each sample (indexed by $i$): $t_{\rm SS}^{*}=\left[\sum_{i=1}^{N}w_{i}\right]^{-1}\,\times\,\sum_{i=1}^{N}\,w_{i}\left[t_{{\rm Pb},i}+\tau_{26}\,\ln\left(R_{26,{\rm SS}}/R_{26,i}\right)\right],$ (7) where $R_{26,i}$ and $t_{{\rm Pb},i}$ are the initial ratio $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ and Pb-Pb age in sample $i$, $w_{i}=1/\sigma_{i}^{2}$, $\sigma_{t26,i}^{2}=\tau_{26}^{2}(\sigma_{R26,i}/R_{26,i})^{2}$, and $\sigma_{i}^{2}=\sigma_{t26,i}^{2}+\sigma_{t{\rm Pb},i}^{2}$ accounts for the uncertainties in the ages. The uncertainty in the weighted mean is $\left[\sum_{i=1}^{N}w_{i}\right]^{-1/2}$. Although the uncertainties in each sample’s model age are typically 0.8 Myr, by averaging several estimates of $t_{\rm SS}$ together, the uncertainty in $t_{\rm SS}$ can be reduced. This treatment gives the best estimate (and the uncertainty in our estimate) of $t_{\rm SS}^{*}$. Whether the ages actually are concordant with that estimate of $t_{\rm SS}{{}^{*}}$ must be determined by calculating the $z$ scores of each age. We first determine for each sample $i$ our best estimate of when that achondrite or chondrule formed, $\Delta t_{i}$, assuming both the Al-Mg and Pb-Pb systems closed simultaneously. We calculate $\Delta t_{{\rm Pb},i}=t_{\rm SS}^{*}-t_{{\rm Pb},i}$ and $\Delta t_{i}=\left(\frac{1}{\sigma_{t26,i}^{2}}+\frac{1}{\sigma_{t{\rm Pb},i}^{2}}\right)^{-1}\,\left(\frac{\Delta t_{26,i}}{\sigma_{t26,i}^{2}}+\frac{\Delta t_{{\rm Pb},i}}{\sigma_{t{\rm Pb},i}^{2}}\right).$ (8) The $z$ scores measure the departure of the measured time of formation of a sample from this modeled time of formation, in units of $1\sigma$: $z_{26,i}=2\frac{\Delta t_{26,i}-\Delta t_{i}}{\sigma_{t26,i}}$ (9) and $z_{{\rm Pb},i}=2\frac{\Delta t_{{\rm Pb},i}-\Delta t_{i}}{\sigma_{t{\rm Pb},i}},$ (10) where the factor of 2 recognizes that the $\sigma$ are reported as $2\sigma$ errors. If $\left|z_{26,i}\right|>2$ or $\left|z_{{\rm Pb},i}\right|>2$, then it is improbable ($<5\%$) that that single time of formation $\Delta t_{26,i}$ or $\Delta t_{{\rm Pb},i}$ is being adequately fit by the data. Of course, in a dataset with $>20$ ages being fit, it would be improbable not to have at least one age with $\left|z\right|>2$. This is accounted for by calculating the $\chi_{\nu}^{2}$ or MSWD quantity for the dataset: $\chi_{\nu}^{2}=\frac{1}{2A-1}\,\sum_{i=1}^{A}\left(z_{26,i}^{2}+z_{{\rm Pb},i}^{2}\right),$ (11) where $A$ is the number of samples (achondrites, chondrules, etc.) being considered, and it is recognized that the number of data being fit is $N=2A$ (two ages for each sample), and the number of degrees of freedom in the dataset is 1 (only $t_{\rm SS}^{*}$ is being fit). This quantity must not exceed a critical value $\chi_{\nu,{\rm max}}^{2}\approx 1+2(2/(N-1))^{1/2}$ or else there is a $<5\%$ probability that the dataset as a whole is being adequately fit by the single value of $t_{\rm SS}^{*}$ (Wendt and Carl, 1991). ### 3.2 Achondrites We first consider just the seven achondrites as a group, since we are relatively certain that each achondrite cooled rapidly enough that the Al-Mg and Pb-Pb systems closed simultaneously and were not reset. We find $t_{\rm SS}^{*}=4568.39$ Myr. Among this set of 14 times of formation to be fit, we find that two (the Pb-Pb ages of SAH 99555 and NWA 6704) are barely fit, at the $2.0\sigma$ and $2.1\sigma$ levels; then again, in a sample of 14 formation times, we would expect 5%, or about 1, to be $>2.0\sigma$ discrepant. Overall, the data are fit very well, with $\chi_{\nu}^{2}=0.98<\chi_{\nu,{\rm max}}^{2}=1.72$ (48% probability). These data are concordant. This is our preferred case. The above analysis assumes a correction of 0.19 Myr applied to Pb-Pb ages of the five achondrites that did not involve hydrous silicate melts, as suggested by Tissot et al. (2017) for angrites, assuming $f_{\rm cpx}=0.5$ in NC achondrites and $f_{\rm cpx}=1$ in CC achondrites. Had we assumed $f_{\rm cpx}=0$ in the NC achondrites and applied a correction of -0.38 Myr, the fit would be improved: $t_{\rm SS}^{*}=4568.24$ Myr, with all 14 $z$ scores $<2$, and $\mbox{$\chi_{\nu}^{2}$}=0.57$. Had we assumed $f_{\rm cpx}=1$ in the NC achondrites and applied no correction, the fit would be worsened: $t_{\rm SS}^{*}=4568.52$ Myr, with the Pb-Pb ages of SAH 99555 and NWA 6704 discordant at the $>2.5\sigma$ and $>2.9\sigma$ levels, and $\mbox{$\chi_{\nu}^{2}$}=1.63$ (7% probability). If $f_{\rm cpx}<1$ for the CC achondrites, this would significantly worsen the fits. We do not treat them as free parameters, but we consider it likely that $f_{\rm cpx}\approx 0.5$ at most in the NC achondrites and $f_{\rm cpx}\approx 1$ in the CC achondrites, due to the hydrous nature of the melt. We adopt these values of $f_{\rm cpx}$ and age corrections going forward. ### 3.3 Chondrules We next consider just the four chondrules. We fit these eight formation times with $t_{\rm SS}^{*}=4568.76$ Myr, but barely. Two of the Pb-Pb ages have $z>2.0$, and $\mbox{$\chi_{\nu}^{2}$}=1.94$, barely below the critical value 2.07, indicating a probability $\approx 6\%$ that the chondrule data are concordant with each other. However, the two values of $t_{\rm SS}^{*}$ implied by the achondrites (4568.39 Myr) and chondrules (4568.76 Myr) are only 0.34 Myr apart, suggesting the achondrites and chondrules might be concordant with each other. If we adopt the value of $t_{\rm SS}^{*}=4568.39$ Myr from the achondrites fit (or 4568.36 Myr from Paper II), the four chondrules are fit well, except for one anomalous $z$ score, the Pb-Pb formation time of 2-C1 being discordant at the $3.0\sigma$ level. We find $\mbox{$\chi_{\nu}^{2}$}=1.35<\mbox{$\chi_{\nu}^{2}$}_{\rm max}=1.56$ (13% probability). If we average the seven achondrites and four chondrules together, we find $t_{\rm SS}^{*}=4568.42$ Myr. Among the 14 achondrite formation times, only one (the Pb-Pb age of NWA 6704) is off, at the $2.3\sigma$ level. Among the eight chondrule formation times, only the Pb-Pb age of 2-C1 is off, at the $2.9\sigma$ level. That is, two out of 22 formation times have $z$ scores $>2$, which is close the value (1.1) expected statistically. For this ensemble, $\mbox{$\chi_{\nu}^{2}$}=1.35$, which does not exceed $\mbox{$\chi_{\nu}^{2}$}_{\rm max}=1.56$ for these 22 data points, and is significant (probability 12%). This fit is slightly better for the chondrules, not much worse for the achondrites, and can be considered concordant, although the Pb-Pb age of 2-C1 is a concern. We can vary $t_{\rm SS}^{*}$ about this value and ask what range of $t_{\rm SS}^{*}$ yields $\mbox{$\chi_{\nu}^{2}$}\leq 1.62$. We find the range is 4568.31 to 4568.54 Myr, implying the Pb-Pb age of $t\\!\\!=\\!\\!0$ is roughly $4568.42\pm 0.12\,{\rm Myr}$. ### 3.4 CAIs For none of the values of $t_{\rm SS}^{*}$ above is the formation times of CAI SJ101 concordant with the others. This would require $t_{\rm SS}^{*}=4567.17\pm 0.51$ Myr, 1.3 Myr younger than our preferred value. For our preferred $t_{\rm SS}^{*}=4568.42$ Myr, the Pb-Pb $z$ score of SJ101 is discordant at the $5.2\sigma$ level. ### 3.5 Effects of varying half-lives These calculations assume a half-life of ${}^{26}{\rm Al}$ of 0.717 Myr, but the value is not known this precisely. As reviewed by Nishiizumi (2003), estimates had converged by the 1970s at $\approx 0.72$ Myr, but with some range. Rightmire et al. (1958) reported $0.738\pm 0.29(1\sigma)$ Myr based on gamma ray measurements; but this was superseded by the work of Samworth et al. (1972), who revised the branching ratios and reported $0.716\pm 0.032(1\sigma)$ Myr. Norris et al. (1983) undertook independent gamma ray spectrometry and found $0.705\pm 0.024(1\sigma)$ Myr. Middleton et al. (1983) used activity accelerator mass spectrometry to find 0.699 and 0.705 Myr using two different standards; they recommended $0.702\pm 0.056(1\sigma)$ Myr. Finally, Thomas et al. (1984) found $0.78\pm 0.05(1\sigma)$ Myr based on a technique involving the gamma ray cross section. Auer et al. (2009) took the weighted mean of these values to derive $0.717\pm 0.017(1\sigma)$ Myr, and this is commonly cited. The value listed in the chart of the nuclides (Walker et al., 1989), 0.73 Myr, also is commonly cited, but the source of this value is not clear. We adopt the Auer et al. (2009) value. This is also the value listed in the compilation of Kondev (2021). We have repeated the calculations above, varying the half-life of ${}^{26}{\rm Al}$ across its allowed range $0.717\pm 0.034(2\sigma)$ Myr. For a half-life of 0.683 Myr, we find $t_{\rm SS}^{*}=4568.17$ Myr,for the $A=7$ samples. For a half-life of 0.751 Myr, we find $t_{\rm SS}^{*}=4568.63$ Myr. The $2\sigma$ uncertainty of $\pm 0.049$ Myr in the ${}^{26}{\rm Al}$ mean-life translates into a difference of $\pm 0.22$ Myr in $t_{\rm SS}^{*}$. Of course the uncertainties in the ${}^{238}{\rm U}$ and ${}^{235}{\rm U}$ half-lives lead to $40\times$ larger uncertainties in the Pb-Pb times of formation, $\pm 9$ Myr (Tissot et al., 2017). It is understood that the reported values of $t_{\rm SS}^{*}$ refer to the Pb-Pb age of a sample that closed at $t\\!\\!=\\!\\!0$, assuming standard half-lives (§1.1). Figure 5: Times of formation after $t\\!\\!=\\!\\!0$ implied by the Al-Mg system, $\Delta t_{26}$, and the age determined by the Pb-Pb system, $t_{\rm Pb}$, for the five NC achondrites (blue), two CC achondrites (purple), four chondrules (red), and one CAI (orange) in which both are measured. The black line denotes the locus of points yielding $t_{\rm Pb}+\Delta t_{26}=4568.42$ Myr, the apparent value of $t_{\rm SS}$. All 14 formation times of the 7 achondrites are consistent with that value, except the Pb-Pb age of NWA 6704 is inconsistent at the $2.3\sigma$ level. The achondrite Erg Chech 002 has a Pb-Pb age (Reger et al., 2023) but two very different reported Al-Mg formation times (Barrat et al., 2021; Reger et al., 2023); because of this discrepancy will be the subject of a forthcoming paper. The 8 chondrule formations times are marginally consistent with that value, except the Pb-Pb age of 2-C1 is inconsistent at the $2.9\sigma$ level. Still, the overall goodness-of-fit parameter for the combined 22 ages is $\mbox{$\chi_{\nu}^{2}$}=1.36$, which is statistically significant (12% probability). The ages are concordant. The light black line is shifted to the right by 0.19 Myr, corresponding to $t_{\rm SS}=4568.61$ Myr if the five NC achondrites’ Pb-Pb ages had not been made 0.19 Myr younger to account for the isotopic fractionation of U in pyroxene grains relative to whole-rock measurements, as suggested by Tissot et al. (2017). In that scenario, the ages would be barely concordant. CAI SJ101 is not concordant with the other samples, which we attribute to a late resetting of the Pb-Pb chronometer without resetting the Al-Mg chronometer (§2.2). ## 4 Discussion ### 4.1 Comparison to other treatments Our approach is similar to, but distinct from, a few other attempts to determine the age of the Solar System’s $t\\!\\!=\\!\\!0$ by statistically correlating Pb-Pb ages against the Al-Mg or other systems. Lugmair and Shukolyukov (1998) correlated Mn-Cr ages against Pb-Pb ages of several achondrites to determine $t_{\rm SS}$. They found a range of values 4568 to 4571 Myr, compared to the then-accepted age of CAIs, $4566\pm 2$ Myr (Göpel et al., 1991), although again none of these Pb-Pb ages was U-corrected. According to this data, an age 4568 Myr would be consistent with both CAI formation and $t_{\rm SS}$. These Pb-Pb ages were not as precise as more modern methods allow, and were not in any event U-corrected. A different approach was taken by Nyquist et al. (2009) to estimate $t_{\rm SS}$ (their “$T_{\rm SS}$”). They compiled $\log_{10}(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ and $\log_{10}(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{0}$ ratios, as well as Pb-Pb ages, for several samples: the angrites D’Orbigny and Sahara (SAH) 99555; the eucrite-like Asuka 881394; ureilites DaG 165 and Dag 319; an estimate for the howardite-eucrite-diogenite (HED) parent body; and Semarkona chondrules. It is difficult to convert $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{0}$ ratios directly into $\Delta t_{53}$ ages because the initial abundance $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{\rm SS}$ and even the half-life of ${}^{53}{\rm Mn}$ are poorly known. Nyquist et al. (2009) did not correlate $\Delta t_{26}$ ages against Pb-Pb ages $t_{\rm Pb}$, but instead used the Al-Mg system to calibrate the Mn-Cr system, then found the Mn-Cr time of formation $\Delta t_{53}$ for one achondrite, Lewis Cliff (LEW) 86010, and used its measured Pb- Pb age $t_{\rm Pb}$ as an anchor, to derive $t_{\rm SS}=t_{\rm Pb}+\Delta t_{53}$. Specifically, Nyquist et al. (2009) regressed the Al-Mg and Mn-Cr data to show that the quantities were linearly correlated as $\log_{10}(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{0}=(-5.485\pm 0.028)$ $+(0.23\pm 0.04)\times\left[\log_{10}(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}-(-6.252\pm 0.074)\right]$ (12) The slope in this linear correlation, $0.23\pm 0.04(2\sigma)$, which should be the ratio of half-lives, Nyquist et al. (2009) estimated from measurements should be $0.20\pm 0.02(1\sigma)$. This suggests a ${}^{53}{\rm Mn}$ half-life $3.12\pm 0.65(2\sigma)$ Myr. This linear correlation was extended to the value $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=5.1\times 10^{-5}$, the value they assigned to the solar nebula at $t\\!\\!=\\!\\!0$, to find the ratio in the solar nebula value then, $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{\rm SS}=9.14\times 10^{-6}$. We note that this implicitly assumes that ${}^{26}{\rm Al}$ and ${}^{53}{\rm Mn}$ were homogeneous (or identically heterogeneous) between the regions of CAI and achondrite formation. From there, using the inferred value $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{0}=1.35\times 10^{-6}$ in the achondrite LEW 86010, they calculated $\Delta t_{53}=10.23$ Myr based on a ${}^{53}{\rm Mn}$ half-life of 3.7 Myr, with 10% uncertainty ($1\sigma$). Adopting a Pb-Pb age of $4558.55\pm 0.15$ Myr for LEW 86010, Nyquist et al. (2009) then calculated $t_{\rm SS}=4568.8\pm 1.0$ Myr. Nyquist et al. (2009) repeated this calculation using a slope $0.20\pm 0.02$ in the correlation above, finding $t_{\rm SS}=4568.1\pm 0.6$ Myr. They then took the weighted mean of these two estimates to get their best estimate of $t_{\rm SS}=4568.2\pm 0.5$ Myr. It is notable that this value of $t_{\rm SS}$ is much older than the accepted Pb-Pb ages of CAIs measured today ($4567.30\pm 0.16$ Myr), and even the majority of Pb-Pb ages of CAIs used in the analysis of Nyquist et al. (2009). However, there are many caveats that may have prevented adoption of this value as the age of the Solar System. First, of course, none of the Pb-Pb ages in their analysis was U-corrected, including those of LEW 86010 or the CAIs. Thus the value of $t_{\rm SS}$ potentially could be shifted by about $>1$ Myr (Brennecka and Wadhwa, 2012; Tissot et al., 2017). Second, the calculation of $t_{\rm SS}$ was done quite indirectly. Since ${}^{26}{\rm Al}$ was implicitly assumed to be homogeneous anyway, it might have been simpler and more precise to correlate $\Delta t_{26}$ against Pb-Pb ages directly, as Nyquist et al. (2009) did in their Figure 3; this, incidentally, would have yielded a value $t_{\rm SS}\approx 4569.3$ Myr. Third, although some statistical approaches were employed, ultimately the value of $t_{\rm SS}$ relied on a single anchor, LEW 86010. This increases the uncertainty in $t_{\rm SS}$ due to the combined uncertainties in $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{\rm SS}$, $\tau_{53}$, $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{0}$ in LEW 86010, and the Pb-Pb age of LEW 86010. In fact, the $1\sigma$ uncertainty in the half-life of ${}^{53}{\rm Mn}$ is $\pm 10\%$ (Honda and Imamura, 1971), and this dominates the uncertainty in $t_{\rm SS}$ in their treatment, so Nyquist et al. (2009) should have reported $t_{\rm SS}=4568.2\pm 1.0$ Myr. In similar fashion, Sanborn et al. (2019), correlated times of formation derived from Al-Mg systematics, and Pb-Pb ages, for several achondrites and the CAIs listed above. They were focused on the slope (which yields a ${}^{26}{\rm Al}$ half-life of $0.69\pm 0.12$ Myr), and did not report the intercept, necessary for finding the Pb-Pb age of $t\\!\\!=\\!\\!0$ when $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})=5.23\times 10^{-5}$. Nevertheless, it is clear by inspection of their Figure 6 that they would have inferred an age $4567.8$ Myr, with an uncertainty $\sim 0.5$ Myr, regressing the achondrite and CAI data. Their regression would appear to be discordant or just barely concordant with the CAI Pb-Pb ages of Connelly et al. (2012). More recently, Piralla et al. (2023) have attempted to determine $t_{\rm SS}$ using three approaches. In the first, they averaged the model ages of eight volcanic achondrites, ($t_{\rm Pb}+\Delta t_{26}$ or $t_{\rm Pb}+\Delta t_{182}$), like those reported in our Table 1. Through a Monte Carlo approach, they determined $t_{\rm SS}=4568.50_{-0.99}^{+0.91}\,{\rm Myr}$ using the Al- Mg ages. Their second approach was to compare Pb-Pb ages and Al-Mg formation times of their measured spinels and, essentially, anchor to their oldest chondrule. They found $t_{\rm SS}=4568.36\pm 0.59$ Myr. Their third approach was to combine a probabilistic approach to their use of the oldest chondrule, to derive $t_{\rm SS}=4568.67_{-0.59}^{+0.79}\,{\rm Myr}$. Our approach is similar in some ways to these methods, but is marked by many innovations. First, we have selected data based on the need to have measured, U-corrected Pb-Pb ages as well as Al-Mg ages from internal isochrons, for all samples; we abandon CAIs in favor of achondrites. We have quantitatively demonstrated the possibility that the Al-Mg and Pb-Pb systems did not close simultaneously in CAIs that were transiently heated. We have carefully compiled and vetted the Pb-Pb isochrons to make sure they are being analyzed in the same fashion, and that they provide robust estimates of both the ages and the uncertainties. This involved applying examining how points were included or excluded from the isochron, properly averaging and propagating the errors when combining multiple analyses, etc. We have applied the likely corrections of $\delta^{238}{\rm U}\approx-0.19$ Myr to the angrite-like achondrites, as advocated by Tissot et al. (2017), where geochemically appropriate. We have directly correlated Pb-Pb ages against Al-Mg ages, as others have, but most importantly, we have applied statistical techniques to assess the goodness of fit and concordancy. ### 4.2 Heterogeneity of ${}^{26}{\rm Al}$? We have calculated an age of $t_{\rm SS}=4568.42\pm 0.24$ Myr for the time at which the solar nebula was characterized by the canonical ratio $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})=5.23\times 10^{-5}$ (assuming homogeneity of ${}^{26}{\rm Al}$). In contrast, the ages of CAIs E22, E31, SJ101, and B4, which are generally considered to record near-canonical $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ ratios (Larsen et al., 2011; MacPherson et al., 2017; Bouvier and Wadhwa, 2010), are $4567.3\pm 0.2$ Myr (Connelly et al., 2012) to $4568.2\pm 0.2$ Myr (Bouvier et al., 2011a). Because a homogeneous amount of ${}^{26}{\rm Al}$ at 4568.42 Myr would have decayed by a factor of 3.3 by 4567.18 Myr, it has been inferred that ${}^{26}{\rm Al}$ was not homogeneously distributed in the solar system (e.g., Larsen et al., 2011). For example, Bollard et al. (2019) concluded that the reservoir from which achondrites formed was depleted in ${}^{26}{\rm Al}$ by a factor of 3.8 compared to the CAI-forming region. They reached this conclusion by obtaining the $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ in eight chondrules, taking the Pb-Pb ages $t_{\rm Pb}$ inferred for the same chondrules by Bollard et al. (2017), then extrapolating back to the time of $t\\!\\!=\\!\\!0$ at $t_{\rm SS}$, to estimate the ${}^{26}{\rm Al}/{}^{27}{\rm Al}$ of the material comprising each chondrule: $\left(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$}\right)_{t=0}=(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}\,\exp\left(+(t_{\rm SS}-t_{\rm Pb})/\tau_{26}\right).$ (13) Taking $t_{\rm SS}=4567.3$ Myr, they inferred values of $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{t=0}$ ranging from $0.4\times 10^{-5}$ to $2.7\times 10^{-5}$. The average was $1.36\times 10^{-5}$, with hints of a bimodal distribution: the four lowest values averaged to $5\times 10^{-6}$, and the four highest values to $1.8\times 10^{-5}$. They concluded that ${}^{26}{\rm Al}$ was heterogeneous. However, if one instead assumes $t_{\rm SS}=4568.42$ Myr, then the CAIs must have taken 1.24 Myr longer to form than previously thought, and their $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{t=0}$ values should be multiplied by $\exp(+1.24/1.034)=3.3$. In that case, the average value of $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{t=0}$ implied by the eight chondrules would be $4.5\times 10^{-5}$, very close to the canonical value. Conclusions of ${}^{26}{\rm Al}$ heterogeneity based on discordancy between the Al-Mg and Pb-Pb chronometers ultimately derive from the assumption that the Pb-Pb system closed at $t\\!\\!=\\!\\!0$, i.e., at the same time as the Al-Mg system. As we have demonstrated, the 22 Al-Mg and Pb-Pb ages of seven achondrites and four chondrules are made concordant in a statistical sense with a single value of $t_{\rm SS}$. It could have been the case that no single value of $t_{\rm SS}$ could make the samples, especially the achondrites, concordant; this would have falsified the hypothesis of ${}^{26}{\rm Al}$ homogeneity. Instead, the concordance of the data strongly suggest that ${}^{26}{\rm Al}$ was homogeneous and that CAI SJ101 and the other CAIs are the anomaly. Given that transient heating events like those experienced by chondrules can reset the Pb-Pb chronometer without resetting the Al-Mg system, we consider CAIs to be unreliable testers of the homogeneity hypothesis. The consistency of the data with homogeneity of ${}^{26}{\rm Al}$ is exactly what is expected from astrophysical models, which do not generally predict conditions in which the CAI-forming region would have $3-4$ times the level of ${}^{26}{\rm Al}$ as other reservoirs. Such large variations in ${}^{26}{\rm Al}/{}^{27}{\rm Al}$ would not be expected in the molecular cloud, as molecular cloud cores take several Myr to collapse, during which time turbulent diffusion will readily mix materials across several parsecs (Pan et al., 2012), an expectation borne out by the surprising chemical similarity of stars born in the same cluster (Feng and Krumholz, 2014; Armillotta et al., 2018). On the scale of a molecular cloud core, $<1$ pc across, there should be no differences in the composition of the accreting material. Any spatial variations that did exist across a molecular cloud core would be further mixed during collapse (Kuffmeier et al., 2017). Models have been proposed in which the infalling material varied over time; for example, to explain stable isotope anomalies, Nanne et al. (2019) proposed that early-accreted material might be richer in supernova-derived material than later-accreted material. But in such models, heterogeneity in the molecular cloud is merely assumed or asserted, not demonstrated. Moreover, as in the model of (Nanne et al., 2019), increased ${}^{26}{\rm Al}$ in the inner disk would require the cloud core interior to contain stellar ejecta, but not its exterior, an unlikely scenario. Because heterogeneities of ${}^{26}{\rm Al}$ cannot be inherited from the molecular cloud, they would have to result from late injections of ${}^{26}{\rm Al}$-bearing material from supernovas or Wolf-Rayet stars, into the protoplanetary disk. Late injection has long been invoked as an explanation for why some hibonite-dominated grains apparently lacked ${}^{26}{\rm Al}$: presumably they formed before ${}^{26}{\rm Al}$ was injected (Sahijpal and Goswami, 1998). However, the fact that only certain rare corundum- or hibonite-dominated inclusions exhibit a lack of ${}^{26}{\rm Al}$ (e.g., Krot et al. 2012) more strongly suggests a chemical heterogeneity, as suggested by (Larsen et al., 2020), and not spatial or temporal variations. As demonstrated by Ouellette et al. (2007), it is possible to inject dust grains from stellar ejecta into a protoplanetary disk; but it is highly improbable for a disk to form close enough to a stellar source to receive relevant amounts of ejecta (Ouellette et al., 2010). Although it has not been quantitatively explored whether CAIs could form with $4\times$ more ${}^{26}{\rm Al}$ than the rest of the disk, a simple argument suggests CAIs should form with less ${}^{26}{\rm Al}$, not more: over the scale of the disk, the stellar ejecta injects a uniform mass of ${}^{26}{\rm Al}$ per area; but that injected material is more diluted by greater surface densities in the inner disk where CAIs form. Approaching the problem from a different angle, the abundances of about a dozen short-lived radionuclides in the solar nebula are very successfully explained as being inherited from the molecular cloud (contaminated by supernovae and Wolf-Rayet winds), and not through late injection or irradiation in the solar nebula (Young, 2020). This demands these isotopes, including ${}^{26}{\rm Al}$, would have been spatially well mixed in the solar nebula from before $t\\!\\!=\\!\\!0$. Ascribing ${}^{26}{\rm Al}$ heterogeneities to something to do with star formation is too simplistic. Astrophysical models overall suggest ${}^{26}{\rm Al}$ should have been homogeneous in the Sun’s protoplanetary disk, and that the discrepancy between the Al-Mg and Pb-Pb chronometers has to do with the Pb-Pb chronometer itself. ### 4.3 Relative vs. Absolute Ages The determination of $t_{\rm SS}=4568.42\pm 0.24$ Myr allows us to use the Pb- Pb data available for a number of achondrites, and unlock the potential of the Pb-Pb system as a relative chronometer. Relative ages are much more precise than absolute ages: absolute ages are uncertain to within $\pm 9$ Myr ($2\sigma$) due to the uncertainties in the half-lives of ${}^{235}{\rm U}$ and ${}^{238}{\rm U}$ (Amelin, 2006; Tissot et al., 2017). These systematic uncertainties cancel when using the Pb-Pb system to calculate relative ages, and the precision can approach $\pm 0.3$ Myr, a factor of 30 more precise. Even ignoring the half-life uncertainties and comparing to the measurement uncertainties, the uncertainties in times of formation after $t\\!\\!=\\!\\!0$ are typically 0.13 Myr, 5 times better than the uncertainties in the model ages, which are typically 0.75 Myr. Relative ages are also much more relevant to models of the protoplanetary disk. All astrophysical models of planet formation desperately need quantification of the order of events in the solar nebula, relative to a commonly accepted event at a defined $t\\!\\!=\\!\\!0$. These could then be compared to astronomical constraints on the ages of young stellar objects, for which the timing of evolutionary stages is typically precise to within $<1$ Myr (Haisch et al., 2001; Mamajek, 2009). In contrast, there is almost no model of stellar or planetary evolution, that uses information of how long ago the protoplanetary disk stage was, that would be affected if it turned out the Solar System were 4560 Myr or 4580 Myr old instead of 4569 Myr old. For example, the tightest astrophysical constraints from helioseismology on the Sun’s age are $4600\pm 40(1\sigma)$ Myr (Houdek and Gough, 2011), and $4587\pm 7$ Myr or $4569\pm 6(1\sigma)$ Myr, based on different databases of physical quantities (Bonanno and Fröhlich, 2015). For these allied fields of study, “4.57 Gyr ago” is precise enough. In addition, the need to choose anchors to report absolute ages introduces confusion and imprecision, especially to those not working in the meteoritics community, as a recent example from the literature illustrates. The two achondrites NWA 11119 and Erg Chech 002 are each among the most ancient crustal rocks in the meteoritic record. Which formed first? Srinivasan et al. (2018) reported $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(1.69\pm 0.09)\times 10^{-6}$ in NWA 11119, while Barrat et al. (2021) reported $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}=(5.72\pm 0.07)\times 10^{-6}$ for Erg Chech 002. Assuming they both formed from material sampling the canonical $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})$ ratio, then NWA 11119 formed at $\Delta t_{26}=3.55\pm 0.06$ Myr, and Erg Chech 002 at $\Delta t_{26}=2.29\pm 0.01$ Myr. If these values are correct and if these two achondrites formed from a common reservoir, then clearly NWA 11119 formed a time $\Delta t=\tau_{26}\,\ln\left[(5.72\pm 0.07)/(1.69\pm 0.09)\right]$ $=1.26\pm 0.08$ Myr after Erg Chech 002. However, the only dating information reported by Srinivasan et al. (2018) in the abstract of their paper is that NWA 11119 has an absolute age of $4564.8\pm 0.3$ Myr. Likewise, the only dating information Barrat et al. (2021) reported in the abstract of their paper was that Erg Chech 002 has an absolute age $4565.0$ (presumably $\pm 0.3$) Myr. Based on these headline ages, the age difference is $0.2\pm 0.4$ Myr, which is clearly incompatible with an age difference of $1.26\pm 0.08$ Myr. Only a very careful reading of the papers will clear up the discrepancy. The age $4564.8\pm 0.3$ Myr for NWA 11119 was derived by Srinivasan et al. (2018) by anchoring to Al-Mg and Pb-Pb ages of D’Orbigny, although at no point did they explicitly state the $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ and Pb-Pb age of D’Orbigny they were assuming. In contrast, the $4565.0$ Myr for Erg Chech 002 was derived by Barrat et al. (2021) by anchoring to CAIs, implicitly assuming their Pb-Pb age is 4567.3 Myr, although at no point did they state the assumed Pb-Pb age of CAIs, recognize the collateral assumption that ${}^{26}{\rm Al}$ would have to be heterogeneous, or even cite the source (Connelly et al., 2012) of these assumptions. So many assumptions are introduced when using anchors that the entire meaning of a sample age becomes opaque. Chronometry papers are replete with phrases like “if anchored to D’Orbigny, then… but if anchored to CAIs, then…”. This is confusing and not even necessary, as the quantity of greatest interest is the time of formation after $t\\!\\!=\\!\\!0$. Moreover, the use of absolute ages necessarily introduces unneeded uncertainty, because of the uncertainties in even Pb-Pb relative ages, $\pm 0.3$ Myr, or $\pm 0.7$ Myr in the example of NWA 6704 above. Had the formation times of NWA 11119 and Erg Chech 002 been reported as times after $t\\!\\!=\\!\\!0$, the precision in the difference in their formation times would have been $\pm 0.08$, a factor of 4 more precise than could be obtained using the ages reported for them. For all these reasons, we strongly advocate reporting all ages, even Pb-Pb ages, as times of formation relative to $t\\!\\!=\\!\\!0$, avoiding the use of anchors, and stating all assumptions clearly. This means reporting the assumed value of $t_{\rm SS}$ for Pb-Pb dating, or the value of $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{\rm SS}$ or even $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{\rm SS}$ that was assumed. We strongly advocate avoiding use of anchors to report absolute ages. ### 4.4 Improved Chronometry Fixing $t_{\rm SS}=4568.42\pm 0.24$ Myr allows us to use the Pb-Pb system as a relative chronometer, in concert with the Al-Mg chronometer, improving the precision of dating. In Table 2 we present for various meteoritic samples: our calculated time of formation after $t\\!\\!=\\!\\!0$ based on the Al-Mg system, $\Delta t_{26}$; our calculated time time of formation based on the Pb-Pb system and $t_{\rm SS}=4568.42$ Myr, $\Delta t_{\rm Pb}$; and the weighted mean of these, $\Delta t$. To connect to the broader literature, we also list the implied absolute age, based on $t_{\rm SS}=4568.42$ Myr (and the assumption of commonly used U half-lives), but we emphasize again that the relative ages are what matter for chronometry. The uncertainties reflect only the uncertainty in the Al-Mg ages, as is appropriate when taking the differences in model ages. We also note that in a companion paper (Desch et al. 2023, hereafter Paper II) we update these estimates using information from the Mn-Cr and Hf-W systems, and many more achondrites; these shift the inferred times of formation $\Delta t$ and model ages, by about 0.07 Myr. Table 2: Adopted times of formation $\Delta t_{26}$ and $\Delta t_{\rm Pb}$ (assuming $t_{\rm SS}=4568.42$ Myr), and weighted mean, $\Delta t$, plus implied Pb-Pb age, of seven bulk achondrites and four chondrules. Sample | $\Delta t_{26}$ (Myr) | $\Delta t_{\rm Pb}$ (Myr) | $\Delta t$ (Myr) | Model age ---|---|---|---|--- D’Orbigny | $5.06\pm 0.10$ | $4.99\pm 0.21$ | $5.05\pm 0.09$ | $4563.37\pm 0.09$ SAH 99555 | $5.14\pm 0.05$ | $4.72\pm 0.24$ | $5.12\pm 0.05$ | $4563.30\pm 0.05$ NWA 1670 | $4.64\pm 0.10$ | $4.21\pm 0.66$ | $4.63\pm 0.10$ | $4563.79\pm 0.10$ Asuka 881394 | $3.82\pm 0.04$ | $3.47\pm 0.53$ | $3.82\pm 0.04$ | $4564.60\pm 0.04$ NWA 7325 | $5.33\pm 0.05$ | $4.52\pm 1.70$ | $5.33\pm 0.05$ | $4563.09\pm 0.05$ NWA 2796 | $5.03\pm 0.04$ | $5.26\pm 0.57$ | $5.03\pm 0.04$ | $4563.39\pm 0.04$ NWA 6704 | $5.29\pm 0.13$ | $5.66\pm 0.26$ | $5.36\pm 0.11$ | $4563.13\pm 0.13$ NWA 5697 2-C1 | $2.00\pm 0.21$ | $0.82\pm 0.75$ | $1.91\pm 0.20$ | $4566.51\pm 0.20$ NWA 5697 3-C5 | $1.84\pm 0.21$ | $2.35\pm 1.07$ | $1.86\pm 0.21$ | $4566.56\pm 0.21$ NWA 5976 5-C2 | $2.07\pm 0.22$ | $1.18\pm 0.95$ | $2.03\pm 0.21$ | $4566.39\pm 0.21$ NWA 5697 11-C1 | $2.32\pm 0.34$ | $2.91\pm 0.74$ | $2.42\pm 0.31$ | $4566.00\pm 0.31$ Having a precise and accurate value for $t_{\rm SS}$ obtained by averaging of multiple samples allows improved chronometry across the board, not just for the samples above with Al-Mg dates, but also for samples with Mn-Cr measurements. For example, if a sample has a Pb-Pb age $t_{\rm Pb}$ and a measurement of $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{0}$, one could extrapolate backward to find the implied value of $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{\rm SS}$ at $t\\!\\!=\\!\\!0$ in the solar nebula: $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{\rm SS}=(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{0}\,\exp\left(+\Delta t_{Pb}/\tau_{53}\right).$ (14) Even samples formed too late to have Al-Mg ages could be used to estimate $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{\rm SS}$, as Nyquist et al. (2009) did using LEW 86010. However, we advocate using averages of many such samples to derive $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{\rm SS}$. Once it is derived, a single measurement of $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{0}$ in a sample, even without a Al-Mg or Pb-Pb date, would suffice to date the formation time of the sample. Deriving such quantities is one goal of the companion Paper II (Desch et al., 2023). ## 5 Conclusions We have correlated Pb-Pb ages and Al-Mg times of formation of seven achondrites and four chondrules for which both $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}$ and U-corrected Pb-Pb ages $t_{\rm Pb}$ exist. We define $t\\!\\!=\\!\\!0$ to be the time in the solar nebula when $(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})=5.23\times 10^{-5}$ (assuming homogeneity of ${}^{26}{\rm Al}$). The time of formation after $t\\!\\!=\\!\\!0$ is then $\Delta t_{26}=\tau_{26}\,\ln\left[(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{\rm SS}/(\mbox{${}^{26}{\rm Al}/{}^{27}{\rm Al}$})_{0}\right]$, where $\tau_{26}=1.034$ Myr is the mean-life of ${}^{26}{\rm Al}$. The time of formation according to the Pb-Pb system is $\Delta t_{\rm Pb}=t_{\rm SS}-t_{\rm Pb}$, where $t_{\rm SS}$ is the Pb-Pb age of samples that achieved isotopic closure at $t\\!\\!=\\!\\!0$, using standard half-lives for ${}^{235}{\rm U}$ and ${}^{238}{\rm U}$. The achondrites rapidly cooled and should have achieved simultaneous isotopic closure in both systems. In these samples in particular, the times of formation $\Delta t_{26}$ and $\Delta t_{\rm Pb}$ should match if ${}^{26}{\rm Al}$ was homogeneous. These samples therefore test the hypothesis of homogeneity. Specifically, if a single value of $t_{\rm SS}$ does not reconcile (in a statistical sense) the $\Delta t_{26}$ and $\Delta t_{\rm Pb}$ formation times for the achondrites, this would falsify the homogeneous hypothesis. We find that the hypothesis of ${}^{26}{\rm Al}$ homogeneity is not falsified. The seven achondrites’ ages are reconciled by $t_{\rm SS}=4568.42\pm 0.24$ Myr. Thirteen of their 14 formation times are concordant at the $<2\sigma$ level, and one (the Pb-Pb age of NWA 6704) is discrepant at the $2.4\sigma$ level, consistent with the scatter being due only to measurement errors. Likewise, the goodness-of-fit parameter for the ensemble is a statistically significant $\mbox{$\chi_{\nu}^{2}$}=0.98$ (47% probability). This supports the assumption that ${}^{26}{\rm Al}$ was homogeneous. Because transient heating events might have reset one chronometer in chondrules but not the other, chondrules do not necessarily test homogeneity. Nevertheless, they also appear concordant. Combining them with the seven achondrites, we find that the 22 ages are reconciled for values of $t_{\rm SS}=4568.42\pm 0.24$ Myr. Only two of the 22 ages are discordant: the Pb-Pb ages of NWA 6704 (at the $2.3\sigma$ level) and chondrule 2-C1 (at the $2.9\sigma$ level). Still, $\mbox{$\chi_{\nu}^{2}$}=1.36$ for this ensemble, which is statistically significant (12% probability). Despite the spread in the chondrule ages, they also appear concordant, further supporting the assumption of homogeneity. There is only a single CAI, SJ101, for which a U-corrected Pb-Pb age and an internal Al-Mg isochron both exist. These ages are quite discrepant. Although this has been interpreted in the past as heterogeneity of ${}^{26}{\rm Al}$, we have demonstrated that the underlying assumption—that the Al-Mg and Pb-Pb systems closed at the same time—is not justified. If this CAI were subjected to a transient heating event in the solar nebula like those that chondrules experienced, then the Pb-Pb system would be reset, the Al-Mg system in anorthite would be marginally reset, and the Al-Mg system in pyroxene, melilite and spinel would not be modified. The Al-Mg system could record the time of the CAI’s formation at $t\approx 0$, while the Pb-Pb system would record a time 1-3 Myr later. The isochron of SJ101 is exactly consistent with this behavior, making it untenable to argue that the Al-Mg and Pb-Pb systems had to close simultaneously, or that CAIs’ Pb-Pb chronometers record $t\\!\\!=\\!\\!0$. As part of our analysis, we reevaluated Pb-Pb ages of all the samples. We performed the regressions as a check, and averaged the intercepts of the Pb-Pb isochrons and averaged the ${}^{238}{\rm U}/{}^{235}{\rm U}$ ratios to reevaluate the Pb-Pb ages, where appropriate. We discovered that the isochrons built by Connelly et al. (2008) [SAH 99555], Connelly et al. (2012) [CAIs], Schiller et al. (2015) [NWA 1670], and Bollard et al. (2017) [NWA 5697 chondrules] employed a methodology for selecting points to regress that is prone to confirmation bias, leading to ages that are overly precise. We have reevaluated the uncertainties in these ages. We have also applied the isotopic correction factor $\delta^{238}{\rm U}\approx-0.3\mbox{\text{\textperthousand}}$ advocated by Tissot et al. (2017), to account for the fact that pyroxene grains (from which Pb-Pb isochrons are typically built) are isotopically lighter than uranium in the usually-measured whole rock (in which uranium isotopes are usually measured), leading to Pb-Pb ages being overestimated by about 0.19 Myr. We argue that this correction should not be applied to the achondrites NWA 2976 and NWA 6704 from the CC isotopic reservoir, due to them forming from hydrous magmas. The difficulty of measurements of Pb-Pb ages in individual CAIs, and the controversy over the results, makes clear that statistical approaches like that of Nyquist et al. (2009), Sanborn et al. (2019), Piralla et al. (2023), and the approach taken here are preferred. Averages of many samples are more reliable and lead to greater precision, than use of single anchors. While we find $t_{\rm SS}=4568.42\pm 0.24$ Myr from the samples above, we encourage further measurements of Al-Mg and Pb-Pb dates for many more systems for which the Al-Mg and Pb-Pb systems are likely to have achieved isotopic closure simultaneously, especially volcanic achondrites. This will severely test the prediction of homogeneity, but if ${}^{26}{\rm Al}$ was homogeneous then we expect $t_{\rm SS}$ to be determined even more precisely as more measurements are made. We advocate a move away from reporting formation times as absolute ages determined by anchoring to samples. Use of anchors introduces a number of assumptions about the ages, which are rarely stated explicitly in papers. At any rate, absolute ages are not needed for models of protoplanetary disks and planet formation, or for most purposes; formation times $\Delta t$ relative to $t\\!\\!=\\!\\!0$, or time differences between two events, are demanded. Relative ages are much more precise than Pb-Pb absolute ages, which have uncertainties of $\pm 9$ Myr due to uncertainties in the ${}^{235}{\rm U}$ half-life. The real utility of the Pb-Pb system is as a relative dating system, with precision $\pm 0.3$ Myr. This requires fixing $t_{\rm SS}$ to define $\Delta t_{\rm Pb}=t_{\rm SS}-t_{\rm Pb}$. We hope that our finding of $t_{\rm SS}=4568.42\pm 0.24$ Myr, will allow precise relative dating of samples using the Pb-Pb system. This increased precision should open the door to more precise chronometry using other isotopic systems. In the companion Paper II (Desch et al., 2023), we refine $t_{\rm SS}$ and show how knowledge of $t_{\rm SS}$ allows us to use statistical approaches to better determine $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{\rm SS}$ and $(\mbox{${}^{182}{\rm Hf}/{}^{180}{\rm Hf}$})_{\rm SS}$ at $t\\!\\!=\\!\\!0$ in the solar nebula, despite the difficulties of measuring these quantities in CAIs directly. This allows a measurement of $(\mbox{${}^{53}{\rm Mn}/{}^{55}{\rm Mn}$})_{0}$ or $(\mbox{${}^{182}{\rm Hf}/{}^{180}{\rm Hf}$})_{0}$ in a sample to be converted into a time of formation directly, without the need for anchors or absolute ages. Acknowledgments: The authors would like to acknowledge the efforts of cosmochemists from multiple laboratories around the world whose work makes possible the data cited in Table 1 and throughout this paper. Statistical chronometry necessarily distills very difficult and painstaking analytical work into mere numbers to be crunched, but the efforts to obtain those numbers are appreciated. We thank Zachary Torrano for useful discussions. We thank Francois Tissot and two other anonymous reviewers, whose suggestions greatly improved the quality of our work. The work herein benefitted from collaborations and/or information exchange within NASA’s Nexus for Exoplanetary System Science research coordination network sponsored by NASA’s Space Mission Directorate (grant NNX15AD53G, PI Steve Desch). Emilie Dunham gratefully acknolwedges support from a 51 Pegasi b Fellowship, grant #2020-1829. The data in Table 1 and the calculations by which we derived our results are included as an Excel spreadsheet as Research Data. ## References * Amelin (2006) Amelin, Y., 2006. The prospect of high-precision Pb isotopic dating of meteorites. Meteoritics and Planetary Science 41, 7–17. doi:10.1111/j.1945-5100.2006.tb00189.x. * Amelin (2008a) Amelin, Y., 2008a. The U Pb systematics of angrite Sahara 99555. Geochimica et Cosmochimica Acta 72, 4874–4885. doi:10.1016/j.gca.2008.07.008. * Amelin (2008b) Amelin, Y., 2008b. U Pb ages of angrites. Geochimica et Cosmochimica Acta 72, 221–232. doi:10.1016/j.gca.2007.09.034. * Amelin et al. (2010) Amelin, Y., Kaltenbach, A., Iizuka, T., Stirling, C.H., Ireland, T.R., Petaev, M., Jacobsen, S.B., 2010. U-Pb chronology of the Solar System’s oldest solids with variable 238U/ 235U. Earth and Planetary Science Letters 300, 343–350. doi:10.1016/j.epsl.2010.10.015. * Amelin et al. (2019) Amelin, Y., Koefoed, P., Iizuka, T., Fernandes, V.A., Huyskens, M.H., Yin, Q.Z., Irving, A.J., 2019. U-Pb, Rb-Sr and Ar-Ar systematics of the ungrouped achondrites Northwest Africa 6704 and Northwest Africa 6693. Geochimica et Cosmochimica Acta 245, 628–642. doi:10.1016/j.gca.2018.09.021. * Armillotta et al. (2018) Armillotta, L., Krumholz, M.R., Fujimoto, Y., 2018. Mixing of metals during star cluster formation: statistics and implications for chemical tagging. Monthly Notices of the Royal Astronomical Society 481, 5000–5013. doi:10.1093/mnras/sty2625, arXiv:1807.01712. * Auer et al. (2009) Auer, M., Wagenbach, D., Wild, E.M., Wallner, A., Priller, A., Miller, H., Schlosser, C., Kutschera, W., 2009\. Cosmogenic 26Al in the atmosphere and the prospect of a 26Al/ 10Be chronometer to date old ice. Earth and Planetary Science Letters 287, 453–462. doi:10.1016/j.epsl.2009.08.030. * Barrat et al. (2021) Barrat, J.A., Chaussidon, M., Yamaguchi, A., Beck, P., Villeneuve, J., Byrne, D.J., Broadley, M.W., Marty, B., 2021\. A 4,565-My-old andesite from an extinct chondritic protoplanet. Proceedings of the National Academy of Science 118, 2026129118. doi:10.1073/pnas.2026129118, arXiv:2105.01911. * Barrat et al. (2015) Barrat, J.A., Greenwood, R.C., Verchovsky, A.B., Gillet, P., Bollinger, C., Langlade, J.A., Liorzou, C., Franchi, I.A., 2015\. Crustal differentiation in the early solar system: Clues from the unique achondrite Northwest Africa 7325 (NWA 7325). Geochimica et Cosmochimica Acta 168, 280–292. doi:10.1016/j.gca.2015.07.020. * Bollard et al. (2017) Bollard, J., Connelly, J.N., Whitehouse, M.J., Pringle, E.A., Bonal, L., Jørgensen, J.K., Nordlund, Å., Moynier, F., Bizzarro, M., 2017. Early formation of planetary building blocks inferred from Pb isotopic ages of chondrules. Science Advances 3, e1700407. doi:10.1126/sciadv.1700407, arXiv:1708.02631. * Bollard et al. (2019) Bollard, J., Kawasaki, N., Sakamoto, N., Olsen, M., Itoh, S., Larsen, K., Wielandt, D., Schiller, M., Connelly, J.N., Yurimoto, H., Bizzarro, M., 2019. Combined U-corrected Pb-Pb dating and 26Al-26Mg systematics of individual chondrules - Evidence for a reduced initial abundance of 26Al amongst inner Solar System chondrules. Geochimica et Cosmochimica Acta 260, 62–83. doi:10.1016/j.gca.2019.06.025. * Bonanno and Fröhlich (2015) Bonanno, A., Fröhlich, H.E., 2015\. A Bayesian estimation of the helioseismic solar age. Astronomy and Astrophysics 580, A130. doi:10.1051/0004-6361/201526419, arXiv:1507.05847. * Bouvier et al. (2011a) Bouvier, A., Brennecka, G.A., Wadhwa, M., 2011a. Absolute Chronology of the First Solids in the Solar System, in: Workshop on Formation of the First Solids in the Solar System, p. 9054. * Bouvier et al. (2011b) Bouvier, A., Spivak-Birndorf, L.J., Brennecka, G.A., Wadhwa, M., 2011b. New constraints on early Solar System chronology from Al-Mg and U-Pb isotope systematics in the unique basaltic achondrite Northwest Africa 2976. Geochimica et Cosmochimica Acta 75, 5310–5323. doi:10.1016/j.gca.2011.06.033. * Bouvier and Wadhwa (2010) Bouvier, A., Wadhwa, M., 2010\. The age of the Solar System redefined by the oldest Pb-Pb age of a meteoritic inclusion. Nature Geoscience 3, 637–641. doi:10.1038/ngeo941. * Brennecka and Wadhwa (2012) Brennecka, G.A., Wadhwa, M., 2012\. Uranium isotope compositions of the basaltic angrite meteorites and the chronological implications for the early Solar System. Proceedings of the National Academy of Science 109, 9299–9303. doi:10.1073/pnas.1114043109. * Brennecka et al. (2010) Brennecka, G.A., Weyer, S., Wadhwa, M., Janney, P.E., Zipfel, J., Anbar, A.D., 2010\. 238U/235U Variations in Meteorites: Extant 247Cm and Implications for Pb-Pb Dating. Science 327, 449\. doi:10.1126/science.1180871. * Brett et al. (1977) Brett, R., Huebner, J.S., Sato, M., 1977. Measured oxygen fugacities of the Angra dos Reis achondrite as a function of temperature. Earth and Planetary Science Letters 35, 363–368. doi:10.1016/0012-821X(77)90139-X. * Cartwright et al. (2016) Cartwright, J.A., Amelin, Y., Koefoed, P., Wadhwa, M., 2016\. U-Pb Age of the Ungrouped Achondrite NWA 8486, in: LPI Editorial Board (Ed.), 79th Annual Meeting of the Meteoritical Society, p. 6231. * Cherniak (1998) Cherniak, D.J., 1998. Pb diffusion in clinopyroxene. Chemical Geology 150, 105–117. doi:10.1016/S0009-2541(98)00056-4. * Chevreux et al. (2021) Chevreux, P., Tissandier, L., Laplace, A., Vitova, T., Bahl, S., Guyadec, F.L., Deloule, E., 2021. Uranium solubility and speciation in reductive soda-lime aluminosilicate glass melts. Journal of Nuclear Materials 544, 152666. doi:10.1016/j.jnucmat.2020.152666. * Connelly et al. (2012) Connelly, J.N., Bizzarro, M., Krot, A.N., Nordlund, Å., Wielandt, D., Ivanova, M.A., 2012\. The Absolute Chronology and Thermal Processing of Solids in the Solar Protoplanetary Disk. Science 338, 651\. doi:10.1126/science.1226919. * Connelly et al. (2008) Connelly, J.N., Bizzarro, M., Thrane, K., Baker, J.A., 2008\. The Pb Pb age of Angrite SAH99555 revisited. Geochimica et Cosmochimica Acta 72, 4813–4824. doi:10.1016/j.gca.2008.06.007. * Desch et al. (2023) Desch, S.J., Dunlap, Daniel, R., Dunham, E.T., Williams, C.D., Mane, P., 2023. Statistical Chronometry of Meteorites: II Abundances and Homogeneity of Short-Lived Radionuclides. Geochimica and Cosmochimica Acta submitted, 00. doi:10.3847/1538-4365/aad95f, arXiv:1710.03809. * Desch et al. (2018) Desch, S.J., Kalyaan, A., O’D. Alexander, C.M., 2018. The Effect of Jupiter’s Formation on the Distribution of Refractory Elements and Inclusions in Meteorites. The Astrophysical Journal Supplement Series 238, 11. doi:10.3847/1538-4365/aad95f, arXiv:1710.03809. * Desch et al. (2012) Desch, S.J., Morris, M.A., Connolly, H.C., Boss, A.P., 2012\. The importance of experiments: Constraints on chondrule formation models. Meteoritics and Planetary Science 47, 1139–1156. doi:10.1111/j.1945-5100.2012.01357.x. * Dodson (1973) Dodson, M.H., 1973. Closure temperature in cooling geochronological and petrological systems. Contributions to Mineralogy and Petrology 40, 259–274. doi:10.1007/BF00373790. * Feng and Krumholz (2014) Feng, Y., Krumholz, M.R., 2014\. Early turbulent mixing as the origin of chemical homogeneity in open star clusters. Nature 513, 523–525. doi:10.1038/nature13662, arXiv:1408.6543. * Floss et al. (2005) Floss, C., Taylor, L.A., Promprated, P., Rumble, D., I., 2005\. Northwest Africa 011: A “eucritic” basalt from a non-eucrite parent body. Meteoritics and Planetary Science 40, 343. doi:10.1111/j.1945-5100.2005.tb00387.x. * Glavin et al. (2004) Glavin, D.P., Kubny, A., Jagoutz, E., Lugmair, G.W., 2004\. Mn-Cr isotope systematics of the D’Orbigny angrite. Meteoritics and Planetary Science 39, 693–700. doi:10.1111/j.1945-5100.2004.tb00112.x. * Goldmann et al. (2015) Goldmann, A., Brennecka, G., Noordmann, J., Weyer, S., Wadhwa, M., 2015. The uranium isotopic composition of the Earth and the Solar System. Geochimica et Cosmochimica Acta 148, 145–158. doi:10.1016/j.gca.2014.09.008. * Goodrich et al. (2017) Goodrich, C.A., Kita, N.T., Yin, Q.Z., Sanborn, M.E., Williams, C.D., Nakashima, D., Lane, M.D., Boyle, S., 2017\. Petrogenesis and provenance of ungrouped achondrite Northwest Africa 7325 from petrology, trace elements, oxygen, chromium and titanium isotopes, and mid-IR spectroscopy. Geochimica et Cosmochimica Acta 203, 381–403. doi:10.1016/j.gca.2016.12.021. * Göpel et al. (1991) Göpel, C., Manhes, G., Allegre, C.J., 1991. Constraints on the time of accretion and thermal evolution of chondrite parent bodies by precise U-Pb dating of phosphates. Meteoritics 26, 338\. * Gopel et al. (1994) Gopel, C., Manhes, G., Allegre, C.J., 1994. U-Pb systematics of phosphates from equilibrated ordinary chondrites. Earth and Planetary Science Letters 121, 153–171. doi:10.1016/0012-821X(94)90038-8. * Greenwood and Hess (1996) Greenwood, J.P., Hess, P.C., 1996\. Congruent melting kinetics: constraints on chondrule formation., in: Chondrules and the Protoplanetary Disk, pp. 205–211. * Haisch et al. (2001) Haisch, Karl E., J., Lada, E.A., Lada, C.J., 2001. Disk Frequencies and Lifetimes in Young Clusters. The Astrophysical Journal 553, L153–L156. doi:10.1086/320685, arXiv:astro-ph/0104347. * Hibiya et al. (2019) Hibiya, Y., Archer, G.J., Tanaka, R., Sanborn, M.E., Sato, Y., Iizuka, T., Ozawa, K., Walker, R.J., Yamaguchi, A., Yin, Q.Z., Nakamura, T., Irving, A.J., 2019\. The origin of the unique achondrite Northwest Africa 6704: Constraints from petrology, chemistry and Re-Os, O and Ti isotope systematics. Geochimica et Cosmochimica Acta 245, 597–627. doi:10.1016/j.gca.2018.04.031. * Honda and Imamura (1971) Honda, M., Imamura, M., 1971\. Half-Life of Mn53. Physical Review C 4, 1182–1188. doi:10.1103/PhysRevC.4.1182. * Houdek and Gough (2011) Houdek, G., Gough, D.O., 2011\. On the seismic age and heavy-element abundance of the Sun. Monthly Notices of the Royal Astronomical Society 418, 1217–1230. doi:10.1111/j.1365-2966.2011.19572.x, arXiv:1108.0802. * Ito and Ganguly (2009) Ito, M., Ganguly, J., 2009\. Mg Diffusion in Minerals in CAIs: New Experimental Data for Melilites and Implications for the Al-Mg Chronometer and Thermal History of CAIs, in: 40th Annual Lunar and Planetary Science Conference, p. 1753. * Izawa et al. (2022) Izawa, M.R.M., Yokoyama, S., Yamashita, S., Okuchi, T., Jephcoat, A.P., Cloutis, E.A., Applin, D.M., Hall, B.J., 2022\. Evidence for Hydrous Magmatism on the Northwest Africa 011 Parent Asteroid, in: 53rd Lunar and Planetary Science Conference, p. 2369. * Jacobsen et al. (2008) Jacobsen, B., Yin, Q.z., Moynier, F., Amelin, Y., Krot, A.N., Nagashima, K., Hutcheon, I.D., Palme, H., 2008\. 26Al- 26Mg and 207Pb- 206Pb systematics of Allende CAIs: Canonical solar initial 26Al/ 27Al ratio reinstated. Earth and Planetary Science Letters 272, 353–364. doi:10.1016/j.epsl.2008.05.003. * Jaffey et al. (1971) Jaffey, A.H., Flynn, K.F., Glendenin, L.E., Bentley, W.C., Essling, A.M., 1971. Precision Measurement of Half-Lives and Specific Activities of 235U and 238U. Physical Review C 4, 1889–1906. doi:10.1103/PhysRevC.4.1889. * Jones et al. (2000) Jones, R.H., Lee, T., Connolly, H. C., J., Love, S.G., Shang, H., 2000. Formation of Chondrules and CAIs: Theory VS. Observation, in: Mannings, V., Boss, A.P., Russell, S.S. (Eds.), Protostars and Planets IV, p. 927. * Jurewicz et al. (1993) Jurewicz, A.J.G., Mittlefehldt, D.W., Jones, J.H., 1993. Experimental partial melting of the Allende (CV) and Murchison (CM) chondrites and the origin of asteroidal basalts. Geochimica et Cosmochimica Acta 57, 2123–2139. doi:10.1016/0016-7037(93)90098-H. * Kawasaki et al. (2021) Kawasaki, N., Itoh, S., Sakamoto, N., Simon, S.B., Yamamoto, D., Yurimoto, H., Marrocchi, Y., 2021. Oxygen and Al-Mg isotopic constraints on cooling rate and age of partial melting of an Allende Type B CAI, Golfball. Meteoritics and Planetary Science 56, 1224–1239. doi:10.1111/maps.13701. * Keil (2012) Keil, K., 2012. Angrites, a small but diverse suite of ancient, silica-undersaturated volcanic-plutonic mafic meteorites, and the history of their parent asteroid. Chemie der Erde / Geochemistry 72, 191–218. doi:10.1016/j.chemer.2012.06.002. * Kleine and Wadhwa (2017) Kleine, T., Wadhwa, M., 2017\. Chronology of Planetesimal Differentiation. pp. 224–245. doi:10.1017/9781316339794.011. * Koefoed et al. (2016) Koefoed, P., Amelin, Y., Yin, Q.Z., Wimpenny, J., Sanborn, M.E., Iizuka, T., Irving, A.J., 2016. U-Pb and Al-Mg systematics of the ungrouped achondrite Northwest Africa 7325. Geochimica et Cosmochimica Acta 183, 31–45. doi:10.1016/j.gca.2016.03.028. * Kondev (2021) Kondev, F.G., 2021. Nuclear Data Sheets for A=201. arXiv e-prints , arXiv:2107.07613arXiv:2107.07613. * Krot et al. (2012) Krot, A.N., Makide, K., Nagashima, K., Huss, G.R., Ogliore, R.C., Ciesla, F.J., Yang, L., Hellebrand, E., Gaidos, E., 2012. Heterogeneous distribution of 26Al at the birth of the solar system: Evidence from refractory grains and inclusions. Meteoritics and Planetary Science 47, 1948–1979. doi:10.1111/maps.12008. * Kuffmeier et al. (2017) Kuffmeier, M., Haugbølle, T., Nordlund, Å., 2017. Zoom-in Simulations of Protoplanetary Disks Starting from GMC Scales. The Astrophysical Journal 846, 7\. doi:10.3847/1538-4357/aa7c64, arXiv:1611.10360. * Larsen et al. (2011) Larsen, K.K., Trinquier, A., Paton, C., Schiller, M., Wielandt, D., Ivanova, M.A., Connelly, J.N., Nordlund, Å., Krot, A.N., Bizzarro, M., 2011\. Evidence for Magnesium Isotope Heterogeneity in the Solar Protoplanetary Disk. The Astrophysical Journal 735, L37. doi:10.1088/2041-8205/735/2/L37. * Larsen et al. (2020) Larsen, K.K., Wielandt, D., Schiller, M., Krot, A.N., Bizzarro, M., 2020. Episodic formation of refractory inclusions in the Solar System and their presolar heritage. Earth and Planetary Science Letters 535, 116088. doi:10.1016/j.epsl.2020.116088. * LaTourrette and Wasserburg (1998) LaTourrette, T., Wasserburg, G.J., 1998\. Mg diffusion in anorthite: implications for the formation of early solar system planetesimals. Earth and Planetary Science Letters 158, 91–108. doi:10.1016/S0012-821X(98)00048-X. * Liermann and Ganguly (2002) Liermann, H.P., Ganguly, J., 2002\. Diffusion kinetics of Fe 2+ and Mg in aluminous spinel . experimental determination and applications. Geochimica et Cosmochimica Acta 66, 2903–2913. doi:10.1016/S0016-7037(02)00875-X. * Liu et al. (2019) Liu, M.C., Han, J., Brearley, A.J., Hertwig, A.T., 2019\. Aluminum-26 chronology of dust coagulation and early solar system evolution. Science Advances 5, eaaw3350. doi:10.1126/sciadv.aaw3350. * Ludwig (2000) Ludwig, K.R., 2000. Decay constant errors in U-Pb concordia-intercept ages. Chemical Geology 166, 315–318. doi:10.1016/S0009-2541(99)00219-3. * Lugmair and Shukolyukov (1998) Lugmair, G.W., Shukolyukov, A., 1998\. Early solar system timescales according to 53Mn- 53Cr systematics. Geochimica et Cosmochimica Acta 62, 2863–2886. doi:10.1016/S0016-7037(98)00189-6. * MacPherson et al. (2017) MacPherson, G.J., Bullock, E.S., Tenner, T.J., Nakashima, D., Kita, N.T., Ivanova, M.A., Krot, A.N., Petaev, M.I., Jacobsen, S.B., 2017. High precision Al-Mg systematics of forsterite-bearing Type B CAIs from CV3 chondrites. Geochimica et Cosmochimica Acta 201, 65–82. doi:10.1016/j.gca.2016.12.006. * Mamajek (2009) Mamajek, E.E., 2009. Initial Conditions of Planet Formation: Lifetimes of Primordial Disks, in: Usuda, T., Tamura, M., Ishii, M. (Eds.), Exoplanets and Disks: Their Formation and Diversity, pp. 3–10. doi:10.1063/1.3215910, arXiv:0906.5011. * Mane et al. (2022) Mane, P., Wallace, S., Bose, M., Wallace, P., Wadhwa, M., Weber, J., Zega, T.J., 2022. Earliest evidence of nebular shock waves recorded in a calcium-aluminum-rich Inclusion. Geochimica et Cosmochimica Acta 332, 369–388. doi:10.1016/j.gca.2022.06.006. * Mattinson (2010) Mattinson, J.M., 2010. Analysis of the relative decay constants of 235U and 238U by multi-step CA-TIMS measurements of closed-system natural zircon samples. Chemical Geology 275, 186–198. doi:10.1016/j.chemgeo.2010.05.007. * McKay et al. (1994) McKay, G., Le, L., Wagstaff, J., Crozaz, G., 1994\. Experimental partitioning of rare earth elements and strontium: Constraints on petrogenesis and redox conditions during crystallization of Antarctic angrite Lewis Cliff 86010. Geochimica et Cosmochmica Acta 58, 2911–2919. doi:10.1016/0016-7037(94)90124-4. * Middleton et al. (1983) Middleton, R., Klein, J., Raisbeck, G.M., Yiou, F., 1983\. Accelerator mass spectrometry with 26Al. Nuclear Instruments and Methods in Physics Research 218, 430–438. doi:10.1016/0167-5087(83)91017-7. * Mikouchi et al. (2003) Mikouchi, T., McKay, G., Koizumi, E., Monkawa, A., Miyamoto, M., 2003. Northwest Africa 1670: A New Quenched Angrite. Meteoritics and Planetary Science Supplement 38, 5218. * Müller et al. (2013) Müller, T., Dohmen, R., Becker, H.W., ter Heege, J.H., Chakraborty, S., 2013. Fe-Mg interdiffusion rates in clinopyroxene: experimental data and implications for Fe-Mg exchange geothermometers. Contributions to Mineralogy and Petrology 166, 1563–1576. doi:10.1007/s00410-013-0941-y. * Nanne et al. (2019) Nanne, J.A.M., Nimmo, F., Cuzzi, J.N., Kleine, T., 2019\. Origin of the non-carbonaceous-carbonaceous meteorite dichotomy. Earth and Planetary Science Letters 511, 44–54. doi:10.1016/j.epsl.2019.01.027. * Nishiizumi (2003) Nishiizumi, K., 2003. Cosmogenic nuclides studies from meteorites to the Earth and beyond. Geochimica et Cosmochimica Acta Supplement 67, 336. * Norris et al. (1983) Norris, T.L., Gancarz, A.J., Rokop, D.J., Thomas, K.W., 1983\. Half-life of Al-26. Lunar and Planetary Science Conference Proceedings 88, B331–B333. doi:10.1029/JB088iS01p0B331. * Nyquist et al. (2009) Nyquist, L.E., Kleine, T., Shih, C.Y., Reese, Y.D., 2009\. The distribution of short-lived radioisotopes in the early solar system and the chronology of asteroid accretion, differentiation, and secondary mineralization. Geochimica et Cosmochimica Acta 73, 5115–5136. doi:10.1016/j.gca.2008.12.031. * Nyquist et al. (2003) Nyquist, L.E., Reese, Y., Wiesmann, H., Shih, C.Y., Takeda, H., 2003. Fossil 26Al and 53Mn in the Asuka 881394 eucrite: evidence of the earliest crust on asteroid 4 Vesta. Earth and Planetary Science Letters 214, 11–25. doi:10.1016/S0012-821X(03)00371-6. * Ouellette et al. (2007) Ouellette, N., Desch, S.J., Hester, J.J., 2007. Interaction of Supernova Ejecta with Nearby Protoplanetary Disks. The Astrophysical Journal 662, 1268–1281. doi:10.1086/518102, arXiv:0704.1652. * Ouellette et al. (2010) Ouellette, N., Desch, S.J., Hester, J.J., 2010. Injection of Supernova Dust in Nearby Protoplanetary Disks. The Astrophysical Journal 711, 597–612. doi:10.1088/0004-637X/711/2/597. * Palk et al. (2018) Palk, C., Andreasen, R., Rehkämper, M., Stunt, A., Kreissig, K., Coles, B., Schönbächler, M., Smith, C., 2018\. Variable Tl, Pb, and Cd concentrations and isotope compositions of enstatite and ordinary chondrites—Evidence for volatile element mobilization and decay of extinct 205Pb. Meteoritics and Planetary Science 53, 167–186. doi:10.1111/maps.12989. * Pan et al. (2012) Pan, L., Desch, S.J., Scannapieco, E., Timmes, F.X., 2012\. Mixing of Clumpy Supernova Ejecta into Molecular Clouds. The Astrophysical Journal 756, 102\. doi:10.1088/0004-637X/756/1/102, arXiv:1206.6516. * Piralla et al. (2023) Piralla, M., Villeneuve, J., Schnuriger, N., Bekaert, D., Marrocchi, Y., 2023. A unified chronology of dust formation in the early solar system. Icarus 10, 230–237. doi:10.1016/j.icarus.2023.115427. * Reger et al. (2023) Reger, P.M., Roebbert, Y., Neumann, W., Gannoun, A., Regelous, M., Schwarz, W.H., Ludwig, T., Trieloff, M., Weyer, S., Bouvier, A., 2023\. Al-Mg and U-Pb chronological records of Erg Chech 002 ungrouped achondrite meteorite. Geochimica et Cosmochimica Acta 343, 33–48. doi:10.1016/j.gca.2022.12.025. * Rightmire et al. (1958) Rightmire, R.A., Kohman, T.P., Hinten-Berger, H., 1958. Über die Halbwertszeit des langlebigen 26Al. Zeitschrift Naturforschung Teil A 13, 847–853. doi:10.1515/zna-1958-1010. * Ruzicka et al. (2008) Ruzicka, A., Floss, C., Hutson, M., 2008. Relict olivine grains, chondrule recycling, and implications for the chemical, thermal, and mechanical processing of nebular materials. Geochimica et Cosmochimica Acta 72, 5530–5557. doi:10.1016/j.gca.2008.08.017. * Sahijpal and Goswami (1998) Sahijpal, S., Goswami, J.N., 1998\. Refractory Phases in Primitive Meteorites Devoid of 26Al and 41Ca:Representative Samples of First Solar System Solids? The Astrophysical Journal 509, L137–L140. doi:10.1086/311778. * Samworth et al. (1972) Samworth, E.A., Warburton, E.K., Engelbertink, G.A., 1972. Beta Decay of the 26Al Ground State. Physical Review C 5, 138–142. doi:10.1103/PhysRevC.5.138. * Sanborn et al. (2019) Sanborn, M.E., Wimpenny, J., Williams, C.D., Yamakawa, A., Amelin, Y., Irving, A.J., Yin, Q.Z., 2019. Carbonaceous achondrites Northwest Africa 6704/6693: Milestones for early Solar System chronology and genealogy. Geochimica et Cosmochimica Acta 245, 577–596. doi:10.1016/j.gca.2018.10.004. * Schiller et al. (2010) Schiller, M., Baker, J.A., Bizzarro, M., 2010. 26Al- 26Mg dating of asteroidal magmatism in the young Solar System. Geochimica et Cosmochimica Acta 74, 4844–4864. doi:10.1016/j.gca.2010.05.011. * Schiller et al. (2015) Schiller, M., Connelly, J.N., Glad, A.C., Mikouchi, T., Bizzarro, M., 2015. Early accretion of protoplanets inferred from a reduced inner solar system 26Al inventory. Earth and Planetary Science Letters 420, 45–54. doi:10.1016/j.epsl.2015.03.028. * Scott et al. (2009) Scott, E.R.D., Greenwood, R.C., Franchi, I.A., Sanders, I.S., 2009\. Oxygen isotopic constraints on the origin and parent bodies of eucrites, diogenites, and howardites. Geochimica et Cosmochimica Acta 73, 5835–5853. doi:10.1016/j.gca.2009.06.024. * Spivak-Birndorf et al. (2009) Spivak-Birndorf, L., Wadhwa, M., Janney, P., 2009. 26Al-26Mg systematics in D’Orbigny and Sahara 99555 angrites: Implications for high-resolution chronology using extinct chronometers. Geochimica et Cosmochimica Acta 73, 5202–5211. doi:10.1016/j.gca.2009.02.038. * Srinivasan et al. (2018) Srinivasan, P., Dunlap, D.R., Agee, C.B., Wadhwa, M., Coleff, D., Ziegler, K., Zeigler, R., McCubbin, F.M., 2018\. Silica-rich volcanism in the early solar system dated at 4.565 Ga. Nature Communications 9, 3036\. doi:10.1038/s41467-018-05501-0. * Stacey and Kramers (1975) Stacey, J.S., Kramers, J.D., 1975\. Approximation of terrestrial lead isotope evolution by a two-stage model. Earth and Planetary Science Letters 26, 207–221. doi:10.1016/0012-821X(75)90088-6. * Stolper and Paque (1986) Stolper, E., Paque, J.M., 1986\. Crystallization sequences of Ca-Al-rich inclusions from Allende: The effects of cooling rate and maximum temperature. Geochimica et Cosmochimica Acta 50, 1785–1806. doi:10.1016/0016-7037(86)90139-0. * Sugiura and Yamaguchi (2007) Sugiura, N., Yamaguchi, A., 2007\. Al-Mg and Mn-Cr Ages of Northwest Africa 011 Achondrite, in: 38th Annual Lunar and Planetary Science Conference, p. 1431. * Takeda (1997) Takeda, H., 1997. Mineralogical records of early planetary processes on the HED parent body with reference to Vesta. Meteoritics and Planetary Science 32, 841–853. doi:10.1111/j.1945-5100.1997.tb01574.x. * Thomas et al. (1984) Thomas, J.H., Rau, R.L., Skelton, R.T., Kavanagh, R.W., 1984\. Half-life of 26Al. Physical Review C 30, 385–387. doi:10.1103/PhysRevC.30.385. * Tissot et al. (2022) Tissot, F.L.H., Collinet, M., Namur, O., Grove, T.L., 2022\. The case for the angrite parent body as the archetypal first-generation planetesimal: Large, reduced and Mg-enriched. Geochimica et Cosmochimica Acta 338, 278–301. doi:10.1016/j.gca.2022.09.031. * Tissot and Dauphas (2015) Tissot, F.L.H., Dauphas, N., 2015\. Uranium isotopic compositions of the crust and ocean: Age corrections, U budget and global extent of modern anoxia. Geochimica et Cosmochimica Acta 167, 113–143. doi:10.1016/j.gca.2015.06.034. * Tissot et al. (2016) Tissot, F.L.H., Dauphas, N., Grossman, L., 2016. Origin of uranium isotope variations in early solar nebula condensates. Science Advances 2, e1501400–e1501400. doi:10.1126/sciadv.1501400, arXiv:1603.01780. * Tissot et al. (2017) Tissot, F.L.H., Dauphas, N., Grove, T.L., 2017. Distinct 238U/235U ratios and REE patterns in plutonic and volcanic angrites: Geochronologic implications and evidence for U isotope fractionation during magmatic processes. Geochimica et Cosmochimica Acta 213, 593–617. doi:10.1016/j.gca.2017.06.045. * Tissot et al. (2019) Tissot, F.L.H., Ibanez-Mejia, M., Boehnke, P., Dauphas, N., McGee, D., Grove, T.L., Harrison, T.M., 2019. 238U/235U measurement in single-zircon crystals: implications fo the Hadean environment, magmatic differentiation and geochronology. Journal of Analytical Atomic Spectrometry 34, 2035–2052. doi:10.1039/C9JA00205G. * Villa et al. (2016) Villa, I.M., Bonardi, M.L., De Bièvre, P., Holden, N.E., Renne, P.R., 2016. IUPAC-IUGS status report on the half-lives of 238U, 235U and 234U. Geochimica et Cosmochimica Acta 172, 387–392. doi:10.1016/j.gca.2015.10.011. * Villeneuve et al. (2009) Villeneuve, J., Chaussidon, M., Libourel, G., 2009. Homogeneous Distribution of 26Al in the Solar System from the Mg Isotopic Composition of Chondrules. Science 325, 985\. doi:10.1126/science.1173907. * Wadhwa et al. (2009) Wadhwa, M., Amelin, Y., Bogdanovski, O., Shukolyukov, A., Lugmair, G.W., Janney, P., 2009\. Ancient relative and absolute ages for a basaltic meteorite: Implications for timescales of planetesimal accretion and differentiation. Geochimica et Cosmochimica Acta 73, 5189–5201. doi:10.1016/j.gca.2009.04.043. * Wadhwa et al. (2014) Wadhwa, M., Kita, N.T., Nakashima, D., Bullock, E.S., MacPherson, G.J., Bouvier, A., 2014\. High Precision 26Al-26Mg Systematics for an Almost Pristine Refractory Inclusion: Implications for the Absolute Age of the Solar System, in: Lunar and Planetary Science Conference, p. 2698\. * Walker et al. (1989) Walker, F.W., Parrington, J.R., Feiner, F., 1989. Nuclides and Isotopes. General Electric Co. * Warren et al. (2013) Warren, P.H., Rubin, A.E., Isa, J., Brittenham, S., Ahn, I., Choi, B.G., 2013\. Northwest Africa 6693: A new type of FeO-rich, low-$\Delta$17O, poikilitic cumulate achondrite. Geochimica et Cosmochimica Acta 107, 135–154. doi:10.1016/j.gca.2012.12.025. * Wasson and Rubin (2003) Wasson, J.T., Rubin, A.E., 2003\. Ubiquitous low-FeO relict grains in type II chondrules and limited overgrowths on phenocrysts following the final melting event. Geochimica et Cosmochimica Acta 67, 2239–2250. doi:10.1016/S0016-7037(03)00023-1. * Wendt and Carl (1991) Wendt, I., Carl, C., 1991\. The statistical distribution of the mean squared weighted deviation. Chemical Geology (Isotope Geoscience Section) 86, 275–285. doi:10.1016/0168-9622(91)90010-TGet. * Whattam et al. (2022) Whattam, S.A., Hewins, R.H., Seo, J., Devouard, B., 2022\. Refractory inclusions as Type IA chondrule precursors: Constraints from melting experiments. Geochimica et Cosmochimica Acta 319, 30–55. doi:10.1016/j.gca.2021.12.022. * Wimpenny et al. (2019) Wimpenny, J., Sanborn, M.E., Koefoed, P., Cooke, I.R., Stirling, C., Amelin, Y., Yin, Q.Z., 2019. Reassessing the origin and chronology of the unique achondrite Asuka 881394: Implications for distribution of 26Al in the early Solar System. Geochimica et Cosmochimica Acta 244, 478–501. doi:10.1016/j.gca.2018.10.006. * Yamaguchi et al. (2002) Yamaguchi, A., Clayton, R., Mayeda, T., Ebihara, M., Oura, Y., Miura, Y., Haramura, H., Misawa, K., Kojima, H., Nagao, K., 2002\. A New Source of Basaltic Meteorites Inferred from Northwest Africa 011. Science 296, 334–336. doi:10.1126/science.1069408. * Yang et al. (2019) Yang, J., Zhang, C., Miyahara, M., Tang, X., Gu, L., Lin, Y., 2019\. Evidence for early impact on a hot differentiated planetesimal from Al-rich micro-inclusions in ungrouped achondrite Northwest Africa 7325. Geochimica et Cosmochimica Acta 258, 310–335. doi:10.1016/j.gca.2019.03.010. * York et al. (2004) York, D., Evensen, N.M., Martínez, M.L., De Basabe Delgado, J., 2004. Unified equations for the slope, intercept, and standard errors of the best straight line. American Journal of Physics 72, 367–375. doi:10.1119/1.1632486. * Young (2020) Young, E.D., 2020. The birth environment of the solar system constrained by the relative abundances of the solar radionuclides, in: Elmegreen, B.G., Tóth, L.V., Güdel, M. (Eds.), Origins: From the Protosun to the First Steps of Life, pp. 70–77. doi:10.1017/S1743921319001777, arXiv:1909.06361.
# The canonical trace of determinantal rings Antonino Ficarra, Jürgen Herzog, Dumitru I. Stamate and Vijaylaxmi Trivedi Antonino Ficarra, Department of mathematics and computer sciences, physics and earth sciences, University of Messina, Viale Ferdinando Stagno d’Alcontres 31, 98166 Messina, Italy<EMAIL_ADDRESS>Jürgen Herzog, Fakultät für Mathematik, Universität Duisburg-Essen, 45117 Essen, Germany <EMAIL_ADDRESS>Dumitru I. Stamate, Faculty of Mathematics and Computer Science, University of Bucharest, Str. Academiei 14, Bucharest – 010014, Romania<EMAIL_ADDRESS>Vijaylaxmi Trivedi, School of Mathematics, Tata Institute of Fundamental Research, Homi Bhabha Road, Mumbai-400005, India<EMAIL_ADDRESS> ###### Abstract. We compute the canonical trace of generic determinantal rings and provide a sufficient condition for the trace to specialize. As an application we determine the canonical trace $\operatorname{tr}(\omega_{R})$ of a Cohen–Macaulay ring $R$ of codimension two, which is generically Gorenstein. It is shown that if the defining ideal $I$ of $R$ is generated by $n$ elements, then $\operatorname{tr}(\omega_{R})$ is generated by the $(n-2)$-minors of the Hilbert-Burch matrix of $I$. ###### Key words and phrases: canonical traces, determinantal rings, Teter numbers, perfect codimension 2 ideals ###### 2010 Mathematics Subject Classification: Primary 13H10; Secondary 13M05. This paper was written while the first, the third and the fourth author visited the Faculty of Mathematics of Essen. D.I. Stamate was partly supported by a grant of the Ministry of Research, Innovation and Digitization, CNCS - UEFISCDI, project number PN-III-P1-1.1-TE-2021-1633, within PNCDI III. ## Introduction Let $(R,{\mathfrak{m}})$ be a Cohen–Macaulay local ring which admits a canonical module or a finitely generated graded Cohen–Macaulay $K$-algebra with graded maximal ideal ${\mathfrak{m}}$. In both cases we denote by $\omega_{R}$ the canonical module of $R$. The trace of an $R$-module $M$ is defined to be the ideal $\operatorname{tr}_{R}(M)=\sum_{\varphi\in\operatorname{Hom}_{R}(M,R)}\varphi(M)\subseteq R.$ We omit the index $R$ in $\operatorname{tr}_{R}(M)$, if it is clear from the context in which ring $R$ we are computing the trace. The trace of $\omega_{R}$ is called the canonical trace of $R$. There is a particular interest in the canonical trace because it determines the non-Gorenstein locus of $R$. It also allows to define nearly Gorenstein rings, which are the rings for which ${\mathfrak{m}}\subseteq\operatorname{tr}_{R}(\omega_{R})$. This class of rings have first been considered in [11]. The name “nearly Gorenstein” was introduced in [9]. In 1974 William Teter [13] studied $0$-dimensional local rings which can be represented as Gorenstein rings modulo their socle. Such rings are nowadays called Teter rings. It has been shown (see [13],[11] and [5]) that $R$ is a Teter ring if and only if there exists an epimorphism $\varphi\colon\omega_{R}\rightarrow{\mathfrak{m}}$. This result shows that a Teter ring is nearly Gorenstein. Inspired by this result, the authors of [6] define the Teter number $\textup{teter}(R)$ of $R$ as the smallest integer $t$ for which there exist $R$-module homomorphisms $\varphi_{1},\ldots,\varphi_{t}\colon\omega_{R}\rightarrow\operatorname{tr}(\omega_{R})$ with $\operatorname{tr}(\omega_{R})=\sum_{i=1}^{t}\varphi_{i}(\omega_{R})$. For further studies of the canonical trace we refer the reader to [4]. In this paper we study the canonical trace of determinantal rings. Let $K$ be a field, and let $X$ be an $m\times n$ matrix of indeterminates with $m<n$. Furthermore, assume that $1<r+1\leq m$ and let $R=K[X]/I_{r+1}(X)$, where $I_{r+1}(X)$ is the ideal of $K[X]$ generated by the $(r+1)$-minors of $X$. Section 1 of this paper is devoted to prove that $\operatorname{tr}(\omega_{R})=I_{r}(X)^{n-m}R$, see Theorem 1.1. As a consequence, the non-Gorenstein locus and the singular locus of $R$ coincide. In Section 2 we determine the Teter number of $R$. It can be easily seen that in general the Teter number of a ring is bounded above by the minimal number of generators of the anti-canonical module $\omega_{R}^{-1}$. We show in Theorem 2.3 that for our determinantal rings this upper bound is reached, and hence by [3, Proposition 4.1] it is equal to the determinant of $\big{[}\binom{2n-m-j}{n-i}\big{]}_{1\leq i,j\leq r}$. It is natural to ask whether the canonical trace specializes. It is known and easily seen that if ${\mathbf{x}}$ is an $R$-sequence and $\overline{R}=R/{\mathbf{x}}R$, then $\operatorname{tr}(\omega_{R})\overline{R}\subseteq\operatorname{tr}(\omega_{\overline{R}})$. If equality holds, we say that canonical trace specializes. In Section 3 we give an explicit example which shows that the trace does not always specialize. On the other hand, we prove in Theorem 3.1 that the canonical trace specializes if $R$ and $\overline{R}$ are generically Gorenstein and $\omega_{R}^{2}$ is a Cohen–Macaulay $R$-module. For our determinantal rings this condition on $\omega_{R}^{2}$ is satisfied if and only if $n\leq 2m-r$, see [3, Theorem 4.3]. We should stress the fact that Theorem 3.1 only gives a sufficient condition for the canonical trace to specialize. Indeed, we don’t have any example of a determinantal ring whose canonical trace does not specialize. Let $I$ be graded perfect ideal of height $2$ in the polynomial ring $S=K[x_{1},\ldots,x_{s}]$. Assume that $I$ is minimally generated by $n$ homogeneous elements and that $R=S/I$ is generically Gorenstein. Theorem 3.1 is used in Corollary 3.5 to show that $\operatorname{tr}(\omega_{R})=I_{n-2}(A)R$, where $A$ is a Hilbert-Burch matrix of $I$. This result is applied to rings defined by generic perfect monomial ideals of grade $2$ and also to numerical semigroup rings generated by $3$ elements, to recover a result in [10]. ## 1\. The canonical trace of determinantal rings The aim of this section is to prove ###### Theorem 1.1. Let $K$ be a field, and let $X=(x_{ij})$ be an $m\times n$ matrix of indeterminates with $m\leq n$. Let $1\leq r+1\leq m$ and $R=K[X]/I_{r+1}(X)$. Then $\operatorname{tr}(\omega_{R})=I_{r}(X)^{n-m}R.$ The proof of this theorem needs some preparations. Let $X=(x_{ij})$ be an $m\times n$ matrix of indeterminates with $m\leq n$. Let $A\subseteq[m],B\subseteq[n]$ with $|A|=|B|=s$, and assume that $A=\\{a_{1}<\ldots<a_{s}\\}$ and $B=\\{b_{1}<\ldots<b_{s}\\}$. We denote by $[A|B]=[a_{1},\dots,a_{s}|b_{1},\dots,b_{s}]$ the $r$-minor of $X$, whose $i$-th row is the $a_{i}$-th row of $X$ and whose $j$-th column is the $b_{j}$-th column of $X$. Let $[C|D]=[c_{1},\dots,c_{t}|d_{1},\dots,d_{t}]$ be any other minors of $X$. We set $[A|B]\leq[C|D]\ \Longleftrightarrow\ s\geq t\ \text{and}\ a_{i}\leq c_{i},\ b_{i}\leq d_{i},\ i=1,\dots,t.$ Then $\leq$ is a partial order on the set of all minors of $X$. Let $r+1\leq m$. It is well-known that $K[x_{ij}]/I_{r+1}(X)$ is an algebra with straightening laws. In particular, if $u,v$ are incomparable $r$-minors, we have a straightening relation (1) $u\cdot v=\sum_{i}\lambda_{i}u_{i}v_{i},$ where $\lambda_{i}\in K$, $\lambda_{i}\neq 0$, and for all $i$, $u_{i},v_{i}$ are $r$-minors with $u_{i}\leq v_{i}$, $u_{i}<u$ and $u_{i}<v$. We denote again by $X$ the image of the matrix $X$ in $R$. Let $P$ be the ideal of $R$ generated by the $r$-minors of the first $r$ rows of $X$. Likewise, let $Q$ be the ideal of $R$ generated by the $r$-minors of the first $r$ columns of $X$. Furthermore, we denote by $\delta$ the minor $[1,\dots,r|1,\dots,r]$. If $I$ is a graded ideal in $R$, $\mu(I)$ denotes the minimal number of generators. ###### Lemma 1.2. With the notation introduced, we have for all $\ell\geq 1$, $\mu((PQ)^{\ell})=\mu(P^{\ell})\mu(Q^{\ell}).$ ###### Proof. Let $Y=(y_{ij})$ be an $m\times r$ matrix and $Z=(z_{ij})$ be an $r\times n$ matrix of indeterminates. For an integer $k$, we set $[k]=\\{1,\dots,k\\}$. Furthermore, if $I\subseteq[m]$ with $|I|=r$, we denote by $Y_{I}$ the $r\times r$ submatrix of $Y$ whose rows are indexed by $I$. Similarly, if $J\subseteq[n]$ with $|J|=r$, we denote by $Z_{J}$ the $r\times r$ submatrix of $Z$ whose columns are indexed by $J$. We set $\delta_{Y}=\det(Y_{[r]})$ and $\delta_{Z}=\det(Z_{[r]})$. For the proof we use the isomorphism $\displaystyle\varphi:R\rightarrow K[Y\cdot Z],\ \ x_{ij}\mapsto(Y\cdot Z)_{ij}\ \ \textup{for all}\ \ i\in[m],\ j\in[n].$ Let $u=[1,\dots,r|J]\in P$ be an $r$-minor, with $J\subseteq[n]$, $|J|=r$. Then $\varphi(u)=\det((Y\cdot Z)_{i\in[r],j\in J})=\det(Y_{[r]}\cdot Z_{J})=\delta_{Y}\det(Z_{J}).$ Thus $\varphi(P)=\delta_{Y}I_{r}(Z)$. Similarly, we have $\varphi(Q)=\delta_{Z}I_{r}(Y)$. This implies that $\varphi(P^{\ell}Q^{\ell})=(\delta_{Y}\delta_{Z})^{\ell}I_{r}(Y)^{\ell}I_{r}(Z)^{\ell}.$ Therefore, $\mu((PQ)^{\ell})=\mu(I_{r}(Y)^{\ell}I_{r}(Z)^{\ell})=\mu(I_{r}(Y)^{\ell})\mu(I_{r}(Z)^{\ell})=\mu(P^{\ell})\mu(Q^{\ell}).$ The second equality follows from the fact that the generators of $I_{r}(Y)$ and $I_{r}(Z)$ are polynomials in different sets of variables. ∎ ###### Corollary 1.3. With the notation introduced, we have $PQ=\delta I_{r}(X).$ ###### Proof. Let $u\in P$ and $v\in Q$. If $u=\delta$ or $v=\delta$, then $uv\in\delta I_{r}(X)$. Suppose both of them are different from $\delta$. Then $u$ and $v$ are incomparable $r$-minors. The only $r$-minor which is less than $u$ and $v$ is $\delta$. Therefore, equation (1) implies that for each $i$, $u_{i}=\delta$. This shows that $PQ\subseteq\delta I_{r}(X)$. It is clear that $\mu(\delta I_{r}(X))=\mu(I_{r}(X))=\binom{m}{r}\binom{n}{r}$. Hence, if we show that $\mu(PQ)=\binom{m}{r}\binom{n}{r}$, then equality follows. This follows from the previous lemma with $\ell=1$, because $\mu(P)=\binom{n}{r}$ and $\mu(Q)=\binom{m}{r}$. ∎ Theorem 1.1 is now a consequence of the following more general result, observing that $Q^{n-m}$ is the canonical ideal $\omega_{R}$ of $R$ [1, Theorem 7.3.6]. ###### Theorem 1.4. We have $\operatorname{tr}(Q^{\ell})=I_{r}(X)^{\ell}R$, for all $\ell\geq 1$. ###### Proof. By Corollary 1.3 we have $Q^{\ell}(\delta^{-1}P)^{\ell}=I_{r}(X)^{\ell}$. On [1, Page 315] it is noted that $\operatorname{height}(I_{r}(X)R)>1$. Therefore, for any prime ideal $L$ of $R$ of height one, we have $Q^{\ell}R_{L}(\delta^{-1}P)^{\ell}R_{L}=R_{L}$. This implies that $(Q^{\ell})^{-1}R_{L}=(\delta^{-1}P)^{\ell}R_{L}$. Since $Q^{\ell}$ and $(\delta^{-1}P)^{\ell}$ are divisorial ideals and the previous equation holds for all height one prime ideals $L$ of $R$, it follows that $(Q^{\ell})^{-1}=(\delta^{-1}P)^{\ell}$. Here we use the fact that if $I$ is a divisorial ideal, then $I=\bigcap_{\mathfrak{p}}IR_{\mathfrak{p}}$, where the intersection is taken over all height one prime ideals $\mathfrak{p}$ of $R$. Since $Q^{\ell}$ has positive grade, by [9, Lemma 1.1] $\operatorname{tr}(Q^{\ell})=Q^{\ell}(Q^{\ell})^{-1}$, and hence the desired result follows. ∎ ###### Corollary 1.5. Suppose the determinantal ring $R$ is not Gorenstein. Then the non-Gorenstein locus and the singular locus of $R$ coincide. ###### Proof. It is noted in [9, Lemma 2.1] that the canonical trace determines the non- Gorenstein locus. Our determinantal ring $R$ is Gorenstein if and only if $m=n$. Hence, if $R$ is not Gorenstein, Theorem 1.1 implies that $\sqrt{\operatorname{tr}(\omega_{R})}=I_{r}(X)R$. By [1, Proposition 7.3.4] $I_{r}(X)R$ determines the singular locus. This implies the assertion. ∎ ## 2\. The Teter number of determinantal rings In [6] the notion of Teter ring was introduced. Let $R$ be a Cohen–Macaulay ring which is either a local ring with canonical module or a graded $K$-algebra over a field $K$. Since by definition $\operatorname{tr}(\omega_{R})=\sum_{\varphi\in\textup{Hom}_{R}(\omega_{R},R)}\varphi(\omega_{R})$, it is natural to determine the smallest number $t$ of maps $\varphi_{1},\dots,\varphi_{t}\in\textup{Hom}_{R}(\omega_{R},R)$ such that $\operatorname{tr}(\omega_{R})=\sum_{i=1}^{t}\varphi_{i}(\omega_{R})$. We call such a number, the Teter number of $R$, and denote it by $\textup{teter}(R)$. If $\textup{teter}(R)=1$, $R$ is called a ring of Teter type. The purpose of this section is to determine the Teter number of a determinantal ring. For this aim, we adopt the following strategy. Recall that a Cohen–Macaulay $R$ is _generically Gorenstein_ if $R_{\mathfrak{p}}$ is Gorenstein for all minimal prime ideals $\mathfrak{p}$ or $R$. If this is the case, then $\omega_{R}$ can be identified with an ideal of height one in $R$. Hence $\operatorname{tr}(\omega_{R})=\omega_{R}\cdot\omega_{R}^{-1}$, by [9, Lemma 1.1]. Since $\textup{Hom}_{R}(\omega_{R},R)\cong\omega_{R}^{-1}$, each map $\varphi:\omega_{R}\rightarrow R$ is multiplication by some element $x\in\omega_{R}^{-1}$. Hence, the Teter number is the smallest number $t$ such that there exist $x_{1},\dots,x_{t}\in\omega_{R}^{-1}$ with $x_{1}\omega_{R}+\dots+x_{t}\omega_{R}=\operatorname{tr}(\omega_{R}).$ Set $J=(x_{1},\dots,x_{t})$. Therefore, $J\omega_{R}=\operatorname{tr}(\omega_{R})$. Let $\mu(\omega_{R}^{-1})$ be the minimal number of generators of $\omega_{R}^{-1}$. Since $\omega_{R}^{-1}\omega_{R}=\operatorname{tr}(\omega_{R})$, our discussion yields the following upper bound: ###### Proposition 2.1. Let $R$ be a ring as introduced above. Then $\textup{teter}(R)\leq\mu(\omega_{R}^{-1})$. Below we discuss a special case where this upper bound is reached. ###### Lemma 2.2. Let $R$ be a ring as introduced above. Assume that $\mu(\operatorname{tr}(\omega_{R}))=\mu(\omega_{R})\mu(\omega_{R}^{-1})$. Then $\textup{teter}(R)=\mu(\omega_{R}^{-1})$. ###### Proof. Suppose by contradiction that $\textup{teter}(R)<\mu(\omega_{R}^{-1})$. Then, we can find $J\subseteq\omega_{R}^{-1}$ with $\mu(J)<\mu(\omega_{R}^{-1})$ and $J\omega_{R}=\operatorname{tr}(\omega_{R})$. Therefore, $\mu(\operatorname{tr}(\omega_{R}))\leq\mu(J)\mu(\omega_{R})<\mu(\omega_{R}^{-1})\mu(\omega_{R})=\mu(\operatorname{tr}(\omega_{R})).$ A contradiction. Since we also have $\textup{teter}(R)\leq\mu(\omega_{R}^{-1})$, the conclusion follows. ∎ ###### Theorem 2.3. Let $K$ be a field, and let $X=(x_{ij})$ be an $m\times n$ matrix of indeterminates with $m<n$. Let $1<r+1\leq m$ and $R=K[X]/I_{r+1}(X)$. Then $\textup{teter}(R)=\det\bigg{[}\binom{2n-m-j}{n-i}\bigg{]}_{1\leq i,j\leq r}.$ ###### Proof. By using the notation introduced before Corollary 1.3, we know that $\omega_{R}=Q^{n-m}$, $\omega_{R}^{-1}=(\delta^{-1}P)^{n-m}$ and $\operatorname{tr}(\omega_{R})=I_{r}(X)^{n-m}$. Obviously, $\mu(\omega_{R}^{-1})=\mu(P^{n-m})$ and $\mu(\operatorname{tr}(\omega_{R}))=\mu(\delta^{n-m}I_{r}(X)^{n-m})$. By Lemma 1.2 $\mu(P^{n-m})\mu(Q^{n-m})=\mu((PQ)^{n-m})$. By Lemma 2.2 it follows that $\textup{teter}(R)=\mu(P^{n-m})$ and using [3, Proposition 4.1] we get the formula in the statement. ∎ We conclude this section with the following remark. ###### Remark 2.4. Suppose $R$ is generically Gorenstein and that $\omega_{R}$ is a divisorial ideal. Let $J\subseteq\omega_{R}^{-1}$ such that $J\omega_{R}=\operatorname{tr}(\omega_{R})$. Then $J\omega_{R}=\omega_{R}^{-1}\omega_{R}$. Hence, $\omega_{R}:(\omega_{R}J)=\omega_{R}:(\omega_{R}\omega_{R}^{-1}).$ On the other hand $\omega_{R}:(\omega_{R}J)=(\omega_{R}:\omega_{R}):J=R:J=J^{-1}$ and $\omega_{R}:(\omega_{R}\omega_{R}^{-1})=(\omega_{R}:\omega_{R}):\omega_{R}^{-1}=R:\omega_{R}^{-1}=(\omega_{R}^{-1})^{-1}=\omega_{R}$, because we assumed that $\omega_{R}$ is divisorial. Therefore, $J^{-1}=\omega_{R}$. Because $J$ may not be divisorial, we only obtain the following inclusion, $J\subseteq(J^{-1})^{-1}=\omega_{R}^{-1}.$ Hence, if $J\omega_{R}=\operatorname{tr}(\omega_{R})$, the divisorial closure $(J^{-1})^{-1}$ of $J$ has to be equal to the anti-canonical module $\omega_{R}^{-1}$, but $J$ itself may be a proper sub-ideal of $\omega_{R}^{-1}$. ## 3\. Reduction of traces Let $R$ be a local ring or a positively graded graded $K$-algebra which is Cohen–Macaulay and admits a canonical module $\omega_{R}$. Let ${\mathbf{x}}=x_{1},\ldots,x_{m}$ be a (graded) sequence of elements in $R$, and set $\overline{M}=M/({\mathbf{x}})M$. It is easy to see that $\operatorname{tr}(\omega_{R})\overline{R}\subseteq\operatorname{tr}(\omega_{\overline{R}})$, if ${\bf x}$ is an $R$-sequence. The question is whether this inclusion may be strict. When $R$ is generically Gorenstein, then $\omega_{R}$ may be identified with an ideal, and we may consider the powers of $\omega_{R}$. ###### Theorem 3.1. Let $R$ be a ring as introduced above, and assume that $R$ is generically Gorenstein. Then $\operatorname{tr}(\omega_{R})\overline{R}=\operatorname{tr}(\omega_{\overline{R}}),$ if $\omega_{R}^{2}$ is Cohen–Macaulay and $\overline{R}$ is generically Gorenstein as well. The proof of the theorem is based on the following ###### Lemma 3.2. Let $R$ be a Cohen–Macaulay ring as before and assume that $R$ is generically Gorenstein. Then $\operatorname{Hom}_{R}(\omega_{R}^{2},\omega_{R})\cong\operatorname{Hom}_{R}(\omega_{R},R).$ ###### Proof. Let $U$ be the kernel of the canonical map $\omega_{R}\otimes_{R}\omega_{R}\rightarrow\omega_{R}^{2}$. Then we obtain the exact sequence $0\rightarrow\operatorname{Hom}_{R}(\omega_{R}^{2},\omega_{R})\rightarrow\operatorname{Hom}_{R}(\omega_{R}\otimes_{R}\omega_{R},\omega_{R})\rightarrow\operatorname{Hom}_{R}(U,\omega_{R})$ Since $R$ is generically Gorenstein it follows that $\omega_{R_{P}}\cong R_{P}$ for all minimal prime ideals $P$ of $R$. This implies that $\omega_{R}\otimes_{R}\omega_{R}\rightarrow\omega_{R}^{2}$ becomes an isomorphism after localization at a minimal prime ideal of $R$. Hence, $U_{P}=0$ for all minimal prime ideal $P$ of $R$. Since $\operatorname{Ass}\operatorname{Hom}_{R}(U,\omega_{R})=\operatorname{Supp}U\cap\operatorname{Ass}\omega_{R}$ and since $\operatorname{Ass}\omega_{R}=\operatorname{Ass}R$, we conclude that $\operatorname{Ass}\operatorname{Hom}_{R}(U,\omega_{R})=\emptyset$. This implies that $\operatorname{Hom}_{R}(U,\omega_{R})=0$. Therefore, $\operatorname{Hom}_{R}(\omega_{R}^{2},\omega_{R})\cong\operatorname{Hom}_{R}(\omega_{R}\otimes_{R}\omega_{R},\omega_{R}).$ By using that $\operatorname{Hom}_{R}(\omega_{R},\omega_{R})\cong R$, we see that $\operatorname{Hom}_{R}(\omega_{R}\otimes_{R}\omega_{R},\omega_{R})\cong\operatorname{Hom}_{R}(\omega_{R},\operatorname{Hom}_{R}(\omega_{R},\omega_{R}))\cong\operatorname{Hom}_{R}(\omega_{R},R).$ Thus, the desired conclusion follows. ∎ ###### Proof of Theorem 3.1. Consider the commutative diagram ${\operatorname{Hom}_{R}(\omega_{R},R)\times\omega_{R}}$${R}$${\overline{\operatorname{Hom}_{R}(\omega_{R},R)}\times\overline{\omega_{R}}}$${\overline{R}}$$\scriptstyle{\varphi}$$\scriptstyle{\alpha}$$\scriptstyle{\beta}$$\scriptstyle{\psi}$ with its natural maps. Then $\operatorname{Im}(\varphi)=\operatorname{tr}(\omega_{R})$, and hence $\operatorname{Im}(\beta\circ\varphi)=\operatorname{tr}(\omega_{R})\overline{R}$. It follows that $\operatorname{Im}(\psi)=\operatorname{tr}(\omega_{R})\overline{R}$. Lemma 3.2, implies that $\operatorname{Hom}_{R}(\omega_{R},R)\cong\operatorname{Hom}_{R}(\omega_{R}^{2},\omega_{R})$, since we assume that $R$ is generically Gorenstein. By assumption, $\omega_{R}^{2}$ is Cohen–Macaulay, and since $\dim\omega_{R}^{2}=\dim R$, it is a maximal Cohen–Macaulay module. Thus, [1, Theorem 3.3.10(c)(ii)] implies that $\operatorname{Ext}^{i}_{R}(\omega_{R}^{2},\omega_{R})=0$ for $i>0$, and hence it follows from [1, Proposition 3.3.3] that $\overline{\operatorname{Hom}_{R}(\omega_{R}^{2},\omega_{R})}\cong\operatorname{Hom}_{\overline{R}}(\overline{\omega_{R}^{2}},\overline{\omega_{R}})\cong\operatorname{Hom}_{\overline{R}}(\omega^{2}_{\overline{R}},\omega_{\overline{R}}).$ Now we use that $\overline{R}$ is generically Gorenstein, and apply again Lemma 3.2 to obtain that $\operatorname{Hom}_{\overline{R}}(\omega^{2}_{\overline{R}},\omega_{\overline{R}})\cong\operatorname{Hom}_{\overline{R}}(\omega_{\overline{R}},\overline{R})$. This shows that $\operatorname{Im}\psi=\operatorname{tr}(\omega_{\overline{R}}),$ and completes the proof of the theorem. ∎ In the following example, using Macaulay2 [7] we verified that the canonical trace does not specialize in general. ###### Example 3.3. Consider the monomial ideal $I=(x_{1}y_{1},x_{2}y_{2},x_{3}y_{3},x_{1}x_{2},x_{2}y_{3},x_{1}x_{3})\subset S=K[x_{1},x_{2},x_{3},y_{1},y_{2},y_{3}]$. Then $R=S/I$ is a Cohen–Macaulay $K$-algebra and the canonical trace of $R$ is $\operatorname{tr}(\omega_{R})=(x_{1},x_{2},x_{3}^{2},x_{3}y_{1},x_{3}y_{2},y_{1}y_{2},y_{1}y_{3},y_{2}y_{3},y_{3}^{2})R.$ The element $x_{1}-y_{1}$ is regular on $R$. We set $\overline{R}=R/(x_{1}-y_{1})R$. Then, $\operatorname{tr}(\omega_{\overline{R}})=(x_{2},x_{3}^{2},x_{3}y_{2},y_{1},y_{3})\overline{R}.$ Hence, the canonical trace does not specialize in such a case. In this example $R$ and $\overline{R}$ are both generically Gorenstein, but $\omega_{R}^{2}$ is not Cohen–Macaulay, because $\operatorname{depth}\omega_{R}^{2}=2$, while $\operatorname{depth}R=3$. Let $K$ be a field, and let $S=K[x_{1},\ldots,x_{s}]$ be the graded polynomial ring with $\deg x_{i}=d_{i}>0$ for $i=1,\ldots,s$. Furthermore, let $A=(f_{ij})$ be an $m\times n$-matrix of homogeneous polynomials of $S$ such that $\deg f_{ij}=a_{i}-b_{j}$, where $a_{1},\ldots,a_{m}$ and $b_{1},\ldots,b_{n}$ positive integers and where $f_{ij}=0$ if $a_{i}-b_{j}\leq 0$. Under these assumptions all minors of $A$ are homogeneous polynomials. Hence $R=S/I_{r+1}(A)$ is a graded $K$-algebra. We have $\operatorname{height}I_{r+1}(A)\leq\operatorname{height}I_{r+1}(X)=(n-r)(m-r)$. If equality holds, then $R$ is Cohen–Macaulay. ###### Corollary 3.4. With the notation introduced, assume that $\operatorname{height}I_{r+1}(A)=(n-r)(m-r)$, that $R$ is generically Gorenstein and that $n\leq 2m-r$. Then $\operatorname{tr}(\omega_{R})=I_{r}(A)^{n-m}R.$ ###### Proof. Let $X$ be the $m\times n$ matrix whose entries are the indeterminates $x_{ij}$, $i=1,\ldots,m$ and $j=1,\ldots,n$, and let $S^{\prime}=S[X]$ and $R^{\prime}=S^{\prime}/I^{\prime}$, where $I^{\prime}=I_{r+1}(X)S^{\prime}$. We set $\deg x_{ij}=b_{j}-a_{i}$ for all $i$ and $j$. Then $I^{\prime}$ is a graded ideal in the non-standard graded polynomial ring $S^{\prime}$, and $R$ is a specialization of $R^{\prime}$. In other words, the sequence ${\mathbf{g}}=x_{11}-f_{11},\ldots,x_{n-1,n}-f_{n-1,n}$ is a homogeneous regular sequence on $S^{\prime}$ and $R^{\prime}$ with $S^{\prime}/({\mathbf{g}})=S$ and $R^{\prime}/({\mathbf{g}})=R$. The ring $R^{\prime}$ is generically Gorenstein, since it is a Cohen–Macaulay domain, and $R$ is generically Gorenstein, by assumption. Therefore, by Theorem 3.1, $\operatorname{tr}(\omega_{R})=\operatorname{tr}(\omega_{R^{\prime}})R=I_{r}(A)^{n-m}R$, if $\omega_{R^{\prime}}^{2}$ is Cohen–Macaulay. This condition for $\omega_{R^{\prime}}$ is obviously satisfied if and only if the square of the canonical module $Q^{n-m}$ of $K[X]/I_{r+1}(X)$ is Cohen–Macaulay. By [3, Theorem 4.3] this is the case if and only if $2(n-m)\leq n-r$, equivalently if $n\leq 2m-r$. This completes the proof of the corollary. ∎ As an application, we consider a graded perfect ideal $I$ of height $2$ in the polynomial ring $S=K[x_{1},\ldots,x_{s}]$ minimally generated by $n$ homogeneous elements. Then $R=S/I$ has a graded resolution $0\rightarrow\bigoplus_{j=1}^{n-1}S(-b_{j})\stackrel{{\scriptstyle A}}{{\longrightarrow}}\bigoplus_{i=1}^{n}S(-a_{i})\rightarrow R\rightarrow 0$ with $A$ an $n\times(n+1)$-matrix. The Hilbert-Burch theorem [1, Theorem 1.4.17] tells us that $I$ is generated by the maximal minors of $A$, which is the same as the maximal minors of $A^{T}$ \- the transpose of $A$. In particular, $I$ is obtained as a specialization of the ideal of maximal minors of the generic $(n-1)\times n$ matrix $X$. The matrix $A$ is called the Hilbert-Burch matrix of $I$ (with respect to the given resolution). ###### Corollary 3.5. Let $I$ be a graded perfect ideal of height $2$ in the polynomial ring $S=K[x_{1},\ldots,x_{s}]$ minimally generated by $n$ homogeneous elements and with Hilbert-Burch matrix $A$. Let $R=S/I$, and assume that $R$ is generically Gorenstein. Then $\operatorname{tr}(\omega_{R})=I_{n-2}(A)R.$ ###### Proof. Since $R$ is the specialization of $K[X]/I_{n-1}(X)$ with $X$ an $(n-1)\times n$-matrix of indeterminates and since $r=n-2$, Corollary 3.4 yields the desired conclusion. ∎ As a very special case of Theorem 3.1 we recover a result in [10]. Let $H$ be a numerical semigroup generated by 3 elements, and let $K$ be any field. The numerical semigroup ring $K[H]$ is a $1$-dimensional (non-standard) graded domain. This implies that the defining ideal $I_{H}$ of $K[H]$ is a graded perfect height 2 prime ideal in the polynomial ring in 3 variables. By [8], the ideal $I_{H}$ is generated by $3$ elements, so that its Hilbert-Burch matrix $A_{H}$ is a $2\times 3$ matrix. Thus, Corollary 3.5 implies that the canonical trace of $K[H]$ is generated by the entries of $A_{H}$ modulo $I_{H}$, see [10, Proposition 3.1] for more details. In [2, Remark 6.3] it is observed that any perfect monomial ideal of height $2$ arises from a generic perfect monomial ideal of height 2 by specialization. The family of generic perfect monomial ideals of height 2 are parameterized by trees. This has been worked out in detail in [12], whose presentation we follow here. To each tree $\Gamma$ on the vertex set $[n]=\\{1,\ldots,n\\}$ one assigns an $(n-1)\times n$-matrix $A(\Gamma)=(a_{ij})$, whose entries are either variables or $0$. One chooses a total order of the edges of $\Gamma$, and assigns to the $k$th edge $\\{i,j\\}$ of $\Gamma$, $i<j$, the $k$th row of $A(\Gamma)$ as follows: $\displaystyle a_{k\ell}=\begin{cases}-x_{ij},&\text{if $\ell=i$}\\\ x_{ji},&\text{if $\ell=j$}\\\ 0,&\text{otherwise}.\end{cases}$ Let $I(\Gamma)=I_{n-1}(A(\Gamma))$. Then $I(\Gamma)$ is a perfect squarefree monomial ideal of height $2$, whose Hilbert-Burch matrix is the transpose of $A(\Gamma)$. The generators of $I(\Gamma)$ are determined as follows: let $i,j$ be two distinct vertices of $\Gamma$. Then there exists a unique path from $i$ to $j$, that is, a sequence of pairwise distinct numbers $i=i_{0},i_{1},i_{2},\ldots,i_{k-1},i_{k}=j$, where $\\{i_{\ell},i_{\ell+1}\\}$ is an edge for $\ell=1,\ldots,k-1$. We set $b(i,j)=i_{1}$. Let $v_{j}$ be the minor of $A(\Gamma)$ which is obtained by omitting the $j$th column of $A(\Gamma)$. Then $v_{j}=\pm\prod_{i=1\atop i\neq j}^{n}x_{ib(i,j)}$ for $j=1,2,\ldots,n$, and $I(\Gamma)=(v_{1},\ldots,v_{n})$. The ideal $I(\Gamma)$ is a squarefree monomial ideal, and therefore generically Gorenstein. It follows from Corollary 3.5 that the $(n-2)$-minors of $A(\Gamma)$ modulo $I(\Gamma)$ generate the canonical trace of $R=S/I(\Gamma)$, where $S$ is the polynomial ring over $K$ generated by the entries of $A(\Gamma)$. Expanding the $(n-1)$-minor $v_{j}$ along the $i$th row, we see that the $n-2$ minors of $A(\Gamma)$ are the monomials $v_{j}/x_{ib(i,j)}$ with $j=1,\ldots,n$, $i=1,\ldots n$ and $i\neq j$. In general, these monomials do not form a minimal set of generators of the canonical trace. For a monomial ideal $I$ we denote by $I(x)$ the monomial ideal which is obtained from $I$ by substituting $x$ by $1$ in each monomial generator of $I$. This operation is a special case of monomial localization. The above discussions give us the following result. ###### Corollary 3.6. Let $\Gamma$ be a tree on the vertex set $[n]$, and let $R=S/I(\Gamma)$, where $S$ is the polynomial ring over $K$ generated by the entries of $A(\Gamma)$. Then $\operatorname{tr}(\omega_{R})=\big{(}\sum I(\Gamma)(x_{ij})\big{)}/I(\Gamma),$ where the sum is taken over all $i$ and $j$ for which $x_{ij}$ is a variable appearing in $A(\Gamma)$. ###### Example 3.7. Let $\Gamma$ be the tree with the edges $\\{1,2\\},\\{2,3\\},\\{3,4\\},\\{3,5\\}$. Then $\displaystyle A(\Gamma)=\begin{pmatrix}-x_{12}&x_{21}&0&0&0\\\ 0&-x_{23}&x_{32}&0&0\\\ 0&0&-x_{34}&x_{43}&0\\\ 0&0&-x_{35}&0&x_{53}\end{pmatrix}$ and $\displaystyle I(\Gamma)$ $\displaystyle=$ $\displaystyle(x_{21}x_{32}x_{43}x_{53},x_{12}x_{32}x_{43}x_{53},x_{12}x_{23}x_{43}x_{53},x_{12}x_{23}x_{34}x_{53},x_{12}x_{23}x_{43}x_{35}).$ We write $A(\Gamma)$ modulo $I(\Gamma)$ as $\displaystyle\begin{pmatrix}-a&b&0&0&0\\\ 0&-c&d&0&0\\\ 0&0&-e&f&0\\\ 0&0&-g&0&h\end{pmatrix}.$ Then the canonical trace of the ring $R$ defined by $I(\Gamma)$ is the ideal of $3$-minors of this matrix, and we obtain $\operatorname{tr}(\omega_{R})=(a\,c\,e,\,a\,c\,g,\,a\,c\,f,\,a\,d\,f,\,a\,f\,g,\,b\,d\,f,\,b\,f\,g,\,c\,f\,g,\,a\,c\,h,\,\\\ a\,d\,h,\,a\,e\,h,\,b\,d\,h,\,b\,e\,h,\,c\,e\,h,\,a\,f\,h,\,b\,f\,h,\,c\,f\,h,\,d\,f\,h).$ ## References * [1] W. Bruns and J. Herzog, Cohen–Macaulay Rings, revised ed., Cambridge Stud. Adv. Math., 39, Cambridge University Press, Cambridge, 1998 * [2] W. Bruns and J. Herzog, On multigraded resolutions, Math. Proc. Camb. Phil. Soc. 118(1995), 245–257. * [3] W. Bruns, T. Römer, A. Wiebe, Initial algebras of determinantal rings, Cohen–Macaulay and Ulrich ideals. Michigan Mathematical Journal, 53 (2005), 71–81. * [4] H. Dao, T. Kobayashi, R. Takahashi, Trace ideals of canonical modules, annihilators of Ext modules, and classes of rings close to being Gorenstein. Journal of Pure and Applied Algebra, 225(2021), 106655. * [5] J. Elias, MS, Takatuji. On Teter rings. Proceedings of the Royal Society of Edinburgh Section A: Mathematics. 147 (2017), 125–39. * [6] O. Gasanova, J. Herzog, T. Hibi, S. Moradi. Rings of Teter type. Nagoya Mathematical Journal, 248 (2022), 1005–1033. * [7] D. R. Grayson, M. E. Stillman. Macaulay2, a software system for research in algebraic geometry. Available at http://www.math.uiuc.edu/Macaulay2. * [8] J. Herzog, Generators and relations of abelian semigroups and semigroup rings. manuscripta math. 3(1970), 175–193. * [9] J. Herzog, T. Hibi and D.I. Stamate, The trace of the canonical module, Israel Journal of Mathematics 233 (2019), 133–165. * [10] J. Herzog, J., T. Hibi, D.I. Stamate, Canonical trace ideal and residue for numerical semigroup rings. Semigroup Forum 103(2021), 550–566. * [11] C. Huneke, A. Vraciu, Rings that are almost Gorenstein, Pacific J. Math. 225 (2006), 85–102. * [12] M. Naeem, Cohen–Macaulay monomial ideals of codimension $2$, manuscripta math. 127(2008), 533–545. * [13] W. Teter, Rings which Are a Factor of a Gorenstein Ring by Its Socle, Inventiones math. 23 (1974), 153–162.
# Dirac’s Theorem and Multigraded syzygies Antonino Ficarra, Jürgen Herzog Antonino Ficarra, Department of mathematics and computer sciences, physics and earth sciences, University of Messina, Viale Ferdinando Stagno d’Alcontres 31, 98166 Messina, Italy <EMAIL_ADDRESS>Jürgen Herzog, Fakultät für Mathematik, Universität Duisburg-Essen, 45117 Essen, Germany<EMAIL_ADDRESS> ###### Abstract. Let $G$ be a simple finite graph. A famous theorem of Dirac says that $G$ is chordal if and only if $G$ admits a perfect elimination order. It is known by Fröberg that the edge ideal $I(G)$ of $G$ has a linear resolution if and only if the complementary graph $G^{c}$ of $G$ is chordal. In this article, we discuss some algebraic consequences of Dirac’s theorem in the theory of homological shift ideals of edge ideals. Recall that if $I$ is a monomial ideal, $\textup{HS}_{k}(I)$ is the monomial ideal generated by the $k$th multigraded shifts of $I$. We prove that $\textup{HS}_{1}(I)$ has linear quotients, for any monomial ideal $I$ with linear quotients generated in a single degree. For and edge ideal $I(G)$ with linear quotients, it is not true that $\textup{HS}_{k}(I(G))$ has linear quotients for all $k\geq 0$. On the other hand, if $G^{c}$ is a proper interval graph or a forest, we prove that this is the case. Finally, we discuss a conjecture of Bandari, Bayati and Herzog that predicts that if $I$ is polymatroidal, $\textup{HS}_{k}(I)$ is polymatroidal too, for all $k\geq 0$. We are able to prove that this conjecture holds for all polymatroidal ideals generated in degree two. ###### Key words and phrases: monomial ideal, multigraded shifts, homological shift ideals, edge ideals, polymatroidal ideals ###### 2020 Mathematics Subject Classification: Primary 13F20; Secondary 13H10 . ## Introduction Let $S=K[x_{1},\ldots,x_{n}]$ be the standard graded polynomial ring with coefficients in a field $K$ and $G$ be a simple graph on the vertex set $V(G)=\\{1,\dots,n\\}$ and with edge set $E(G)$. The edge ideal of $G$ is the ideal $I(G)$ in $S$ generated by the monomials $x_{i}x_{j}$, such that $\\{i,j\\}\in E(G)$. The classification of all Cohen–Macaulay edge ideals and the classification of all edge ideals with linear resolution are fundamental problems. While the first problem is widely open and considered to be intractable in general, for the second problem we have a complete answer. The complementary graph $G^{c}$ of $G$ is the graph with vertex set $V(G^{c})=V(G)$ and where $\\{i,j\\}$ is an edge of $G^{c}$ if and only if $\\{i,j\\}\notin E(G)$. Ralph Fröberg in [9] proved that $I(G)$ has a linear resolution if and only $G^{c}$ is chordal, that is, it has no induced cycles of length bigger than three. In turn, the classical and fundamental Dirac’s theorem on chordal graphs says that a graph $G$ is chordal if and only if $G$ admits a perfect elimination order [5]. Recently, a new research trend in the theory of monomial ideals has been initiated by the second author, Moradi, Rahimbeigi and Zhu in [13], see, also, [2, 3, 4, 6, 7, 14]. For ${\bf a}=(a_{1},\dots,a_{n})\in{\mathbb{Z}}_{\geq 0}^{n}$, we denote $x_{1}^{a_{1}}\cdots x_{n}^{a_{n}}$ by ${\bf x^{a}}$. Let $I\subset S$ be a monomial ideal and let ${\mathbb{F}}$ be its minimal multigraded free $S$-resolution. Then, the $k$th free $S$-module in ${\mathbb{F}}$ is $F_{k}=\bigoplus_{j=1}^{\beta_{k}(I)}S(-{\bf a}_{kj})$, where ${\bf a}_{kj}\in{\mathbb{Z}}_{\geq 0}^{n}$ are the $k$th multigraded shifts of $I$. The _$k$ th homological shift ideal_ of $I$ is the monomial ideal generated by the monomials ${\bf x}^{{\bf a}_{kj}}$ for $j=1,\ldots,\beta_{k}(I)$. Note that $\textup{HS}_{0}(I)=I$. It is natural to ask what combinatorial and homological properties are satisfied by all $\textup{HS}_{k}(I)$, $k=0,\ldots,\operatorname{pd}(I)$. Any such property is called an homological shift property of $I$. If all $\textup{HS}_{k}(I)$ have linear quotients, or linear resolution, we say that $I$ has homological linear quotients or homological linear resolution, respectively. In this article we discuss the algebraic consequences of Dirac’s theorem on chordal graphs related to the theory of homological shift ideals of edge ideals. The article is structured as follows. In Section 1, we investigate arbitrary monomial ideals with linear quotients generated in one degree. Our main theorem states that for such an ideal $I$, $\textup{HS}_{1}(I)$ always has linear quotients. The proof relies upon the fact that certain colon ideals are generated by linear forms (Lemma 1.1). In particular, $\textup{HS}_{1}(I)$ has a linear resolution. At present we are not able to generalize this result for all monomial ideals with linear resolution. In this case, one could expect even that $\textup{HS}_{1}(I)$ also has linear quotients, if $I$ has a linear resolution. On the other hand, if $I$ is generated in more than one degree, in Example 1.4 we show that Theorem 1.3 is no longer valid. Sections 2 and 3 are devoted to homological shifts of edge ideals with linear resolution. Let $G$ be a graph and $I(G)$ be its edge ideal. For unexplained terminology look at Section 2. Unfortunately, even if $I(G)$ has linear resolution, it may not have homological linear resolution in general, (Example 2.3). At present we do not have a complete classification of all edge ideals with homological linear quotients or homological linear resolution. Thus, we determine many classes of cochordal graphs whose edge ideals have homological linear resolution. In particular, for proper interval graphs and forests, we prove that the edge ideals of their complementary graphs have homological linear quotients, (Theorems 2.4 and 3.1). For the proof of the first result we introduce the class of reversible chordal graphs, and show that any proper interval graph is a reversible graph, (Lemma 2.5). For the second result, we consider two operations on chordal graphs that preserve the homological linear quotients property. Namely, adding whiskers to a chordal graph and taking unions of disjoint chordal graphs, (Propositions 3.2 and 3.4). Using these results, it is easy to see that $I(G)$ has homological linear quotients, if $G$ is a forest. Indeed, any forest is the union of pairwise disjoint trees, and any tree can be constructed by iteratively adding whiskers to a previously constructed tree on a smaller vertex set. In the last section, we consider polymatroidal ideals. An equigenerated monomial ideal $I$ is called polymatroidal if its minimal set of monomial generators $G(I)$ corresponds to the set of bases of a discrete polymatroid, see [12, Chapter 12]. Polymatroidal ideals are characterized by the fact that they have linear quotients with respect to the lexicographic order induced by any ordering of the variables. Such characterization is due to Bandari and Rahmati-Asghar [3]. It was conjectured by Bandari, Bayati and Herzog that all homological shift ideals of a polymatroidal ideal are polymatroidal. At present this conjecture is widely open. On the other hand, Bayati proved that the conjecture holds for any squarefree polymatroidal ideal [2]. The second author of this paper, Moradi, Rahimbeigi and Zhu proved that it holds for polymatroidal ideals that satisfy the strong exchange property [13, Corollary 3.6]; whereas the first author of this paper proved that $\textup{HS}_{1}(I)$ is again polymatroidal if $I$ is such [6], pointing towards the validity of the conjecture in general. We prove in Theorem 4.5 that for any polymatroidal ideal $I$ generated in degree two, all homological shift ideals are polymatroidal. In the squarefree case, $I$ may be seen as the edge ideal of a cochordal graph and we apply our criterion on reversibility of perfect elimination orders. Unfortunately our methods are very special and they can not be applied to prove that homological shifts of polymatroidal ideals, generated in higher degree than two, are polymatroidal. ## 1\. The first Homological shift of ideals with Linear quotients Let $S=K[x_{1},\dots,x_{n}]$ be the standard graded polynomial ring, with $K$ a field. A monomial ideal $I\subset S$ has linear quotients if for some ordering $u_{1},\dots,u_{m}$ of its minimal set of monomial generators $G(I)$, all colon ideals $(u_{1},\dots,u_{i-1}):u_{i}$, $i=1,\dots,m$, are generated by variables. We call $u_{1},\dots,u_{m}$ an admissible order of $I$. Such order is called non-increasing if $\deg(u_{1})\leq\deg(u_{2})\leq\dots\leq\deg(u_{m})$. By [16, Lemma 2.1], an ideal with linear quotients always has a non-increasing admissible order. So, from now, we consider only non-increasing admissible orders. Let $u_{1},\dots,u_{m}$ be an admissible order of an ideal $I\subset S$ having linear quotients. For $i\in\\{1,\dots,m\\}$, we let $\textup{set}(u_{i})=\\{j:x_{j}\in(u_{1},\dots,u_{i-1}):u_{i}\\}.$ Given a non-empty subset $A$ of $\\{1,\dots,n\\}$, we set ${\bf x}_{A}=\prod_{i\in A}x_{i}$ and ${\bf x}_{\emptyset}=1$. The multigraded version of [15, Lemma 1.5] implies that (1) $\textup{HS}_{k}(I)\ =\ (u_{i}{\bf x}_{A}\ :\ i=1,\dots,m,\ A\subseteq\textup{set}(u_{i}),\ |A|=k).$ The ideal $(u_{1},\dots,u_{i-1}):u_{i}$ is generated by the monomials $u_{j}:u_{i}=\textup{lcm}(u_{j},u_{i})/u_{i}$. Hence, $I$ has linear quotients if and only if for all $i=1,\dots,m$ and all $j<i$ there exists $\ell<i$ such that $u_{\ell}:u_{i}=x_{p}$ for some $p$, and $x_{p}$ divides $u_{j}:u_{i}$. Hereafter, we denote the set $\\{1,\dots,n\\}$ by $[n]$. For a monomial $u\in S$ and $i\in[n]$, the $x_{i}$-degree of $u$ is the integer $\deg_{x_{i}}(u)=\max\\{j\geq 0:x_{i}^{j}\ \textup{divides}\ u\\}$. For the proof of our main result we need Corollary 1.2 of the following lemma. ###### Lemma 1.1. Let $I$ be an equigenerated graded ideal with linear relations. Let $f_{1},\dots,f_{m}$ be a minimal set of generators of $I$. Then, for any $1\leq i\leq m$, $(f_{1},\dots,f_{i-1},f_{i+1},\dots,f_{m}):f_{i}$ is generated by linear forms. ###### Proof. To simplify the notation, we may assume that $i=m$, and we set $J=(f_{1},\dots,f_{m-1}):f_{m}$. Since the $f_{i}$ are homogeneous elements, $J$ is a graded ideal. Let $r_{m}\in J$ be an homogeneous element. Then, there exist $r_{1},\ldots,r_{m-1}$ such that $r_{m}f_{m}=-\sum_{i=1}^{m-1}r_{i}f_{i}$ with $\deg(r_{i})=\deg(r_{m})$ for $i=1,\ldots,m-1$. Therefore, $r=(r_{1},\ldots,r_{m})$ is a homogeneous relation of $I$. By assumption, the relation module of $I$ is generated by linear relations, say $\ell_{i}=(\ell_{i1},\ldots,\ell_{im})$ for $i=1,\ldots,t$. Therefore, there exist homogeneous elements $s_{i}\in S$ such that $r=\sum_{i=1}^{t}s_{i}\ell_{i}$. This implies that $r_{m}=\sum_{i=1}^{t}s_{i}\ell_{i,m}$. Since $\ell_{i,m}\in J$, the desired conclusion follows. ###### Corollary 1.2. Let $I$ be an equigenerated monomial ideal with linear quotients and let $u_{1},\dots,u_{m}$ be its minimal monomial generators. Then, for any $1\leq i\leq m$, $(u_{1},\dots,u_{i-1},u_{i+1},\dots,u_{m}):u_{i}$ is generated by variables. ###### Theorem 1.3. Let $I\subset S$ be an equigenerated monomial ideal having linear quotients. Then $\textup{HS}_{1}(I)$ has linear quotients. ###### Proof. We proceed by induction on $m\geq 1$. For $m=1$ or $m=2$ there is nothing to prove. Let $m>2$ and set $J=(u_{1},\dots,u_{m-1})$. Let $L=(x_{i}:i\in\textup{set}(u_{m}),x_{i}u_{m}\notin\textup{HS}_{1}(J))$. Then, by equation (1), $\textup{HS}_{1}(I)=\textup{HS}_{1}(J)+u_{m}L.$ By inductive hypothesis, $\textup{HS}_{1}(J)$ has linear quotients. Let $v_{1},\dots,v_{r}$ be an admissible order of $\textup{HS}_{1}(J)$. If $L=(x_{j_{1}},\dots,x_{j_{s}})$, we claim that $v_{1},\dots,v_{r},x_{j_{1}}u_{m},\dots,x_{j_{s}}u_{m}$ is an admissible order of $\textup{HS}_{1}(I)$. We only need to show that (2) $(v_{1},\dots,v_{r},x_{j_{1}}u_{m},\dots,x_{j_{t-1}}u_{m}):x_{j_{t}}u_{m}$ is generated by variables, for all $t=1,\dots,s$. Note that each generator $x_{j_{\ell}}u_{m}:x_{j_{t}}u_{m}=x_{j_{\ell}}$, with $\ell<t$ is already a variable. Consider now a generator $v_{\ell}:x_{j_{t}}u_{m}$ for some $\ell=1,\dots,r$. Then $v_{\ell}=x_{h}u_{j}$ for some $j<m$ and $h\in\text{set}(u_{j})$. Moreover, we can write $x_{j_{t}}u_{m}=x_{p}u_{k}$ for some $k<m$. If $j=k$, then $v_{\ell}:x_{j_{t}}u_{m}=x_{h}u_{k}:x_{p}u_{k}=x_{h}$ is a variable and there is nothing to prove. Suppose now $j\neq k$. Since $u_{1},\dots,u_{m-1}$ is an admissible order, by Corollary 1.2 $Q=(u_{1},\dots,u_{k-1},u_{k+1},\dots,u_{m-1}):u_{k}$ is generated by variables. Since $j\neq k$ and $j<m$, $u_{j}:u_{k}$ belongs to $Q$. Hence, we can find $b<m$, $b\neq k$ such that $u_{b}:u_{k}=x_{q}$ and $x_{q}$ divides $u_{j}:u_{k}$. Thus $x_{q}u_{k}\in\textup{HS}_{1}(J)$. Note that $x_{q}$ divides also $x_{h}u_{j}:x_{p}u_{k}$. Indeed $x_{q}$ divides $u_{j}:u_{k}$. If $x_{q}$ does not divide $x_{h}u_{j}:x_{p}u_{k}$, then necessarily $p=q$. But this would imply that $x_{j_{t}}u_{m}=x_{q}u_{k}\in\textup{HS}_{1}(J)$, against the fact that $x_{j_{t}}\in L$. Hence $x_{q}$ divides $x_{h}u_{j}:x_{p}u_{k}$. But $x_{q}u_{k}:x_{j_{t}}u_{m}=x_{q}u_{k}:x_{p}u_{k}=x_{q}$ belongs to the ideal (2). Hence $x_{h}u_{j}:x_{p}u_{k}$ is divided by a variable belonging to the ideal (2). This concludes our proof. It is natural to ask the following question. Let $I\subset S$ be a monomial ideal having a linear resolution. Is it true that $\textup{HS}_{1}(I)$ has a linear resolution, too? Theorem 1.3 is no longer valid for monomial ideals with linear quotients generated in more than one degree, as next example of Bayati et all shows [3]. ###### Example 1.4. ([3, Example 3.3]). Let $I=\left(x_{1}^{2},\,x_{1}x_{2},\,x_{2}^{4},\,x_{1}x_{3}^{4},\,x_{1}x_{3}^{3}x_{4},\,x_{1}x_{3}^{2}x_{4}^{2}\right)$ be an ideal of $S=K[x_{1},x_{2},x_{3},x_{4}]$. $I$ is a (strongly) stable ideal whose Borel generators are $x_{1}x_{2},x_{2}^{4},x_{1}x_{3}^{2}x_{4}^{2}$. It is well-known that stable ideals have linear quotients. Thus $I$ has linear quotients. Using Macaulay2 [10] the package [7], we verified that $\displaystyle\textup{HS}_{1}(I)\ =\ \big{(}$ $\displaystyle x_{1}^{2}x_{2},\,x_{1}x_{2}^{4},\,x_{1}x_{3}^{3}x_{4}^{2},\,x_{1}x_{2}x_{3}^{2}x_{4}^{2},\,x_{1}^{2}x_{3}^{2}x_{4}^{2},\,x_{1}x_{3}^{4}x_{4},$ $\displaystyle x_{1}x_{2}x_{3}^{3}x_{4},\,x_{1}^{2}x_{3}^{3}x_{4},\,x_{1}x_{2}x_{3}^{4},\,x_{1}^{2}x_{3}^{4}\big{)}$ has the following Betti table $\begin{array}[]{c|cccc}&0&1&2&3\\\ \hline\cr 3&1&.&.&.\\\ 4&.&.&.&.\\\ 5&1&1&.&.\\\ 6&8&15&8&1\\\ 7&.&.&.&.\\\ 8&.&3&5&2\end{array}$ We show that $\textup{HS}_{1}(I)$ does not have linear quotients. Suppose by contradiction that $\textup{HS}_{1}(I)$ has linear quotients. Then, since the Betti numbers of an ideal with linear quotients do not depend upon the characteristic of the underlying field $K$, we may assume that $K$ has characteristic zero. Hence $\textup{HS}_{1}(I)$ would be componentwise linear, see [12, Corollary 8.2.21]. However, this cannot be the case by virtue of [12, Theorems 8.2.22. and 8.2.23(a)]. Indeed $\beta_{1,1+8}(\textup{HS}_{1}(I))\neq 0$, while $\beta_{0,8}(\textup{HS}_{1}(I))=0$. ## 2\. Homological shifts of proper interval graphs Let $G$ be a finite simple graph with vertex set $V(G)=[n]$ and edge set $E(G)$. Let $K$ be a field. The edge ideal of $G$ is the squarefree monomial ideal $I(G)$ of $S=K[x_{1},\dots,x_{n}]$ generated by the monomials $x_{i}x_{j}$ such that $\\{i,j\\}\in E(G)$. A graph $G$ is complete if every $\\{i,j\\}$ with $i,j\in[n]$, $i\neq j$, is an edge of $G$. The open neighbourhood of $i\in V(G)$ is the set $N_{G}(i)=\big{\\{}j\in V(G):\\{i,j\\}\in E(G)\big{\\}}.$ A graph $G$ is called chordal if it has no induced cycles of length bigger than three. Recall that a perfect elimination order of $G$ is an ordering $v_{1},\dots,v_{n}$ of its vertex set $V(G)$ such that $N_{G_{i}}(v_{i})$ induces a complete subgraph on $G_{i}$, where $G_{i}$ is the induced subgraph of $G$ on the vertex set $\\{i,i+1,\dots,n\\}$. Hereafter, if $1,2,\dots,n$ is a perfect elimination order of $G$, we denote it by $x_{1}>x_{2}>\dots>x_{n}$. ###### Theorem 2.1. (Dirac). A simple finite graph $G$ is chordal if and only if $G$ admits a perfect elimination order. The complementary graph $G^{c}$ of $G$ is the graph with vertex set $V(G^{c})=V(G)$ and where $\\{i,j\\}$ is an edge of $G^{c}$ if and only if $\\{i,j\\}\notin E(G)$. A graph $G$ is called cochordal if and only if $G^{c}$ is chordal. ###### Theorem 2.2. (Fröberg). Let $G$ be a simple finite graph. Then, $I(G)$ has a linear resolution if and only if $G$ is cochordal. It is known by [12, Theorem 10.2.6] that $I(G)$ has linear resolution if and only if it has linear quotients. The theorems of Dirac and Fröberg classify all edge ideals with linear quotients. Furthermore if $x_{1}>x_{2}>\dots>x_{n}$ is a perfect elimination order of $G^{c}$, then $I(G)$ has linear quotients with respect to the lexicographic order $>_{\textup{lex}}$ induced by $x_{1}>x_{2}>\dots>x_{n}$. Now we turn to the homological shifts of edge ideals with linear quotients. Unfortunately, in general an edge ideal with linear quotients does not even has homological linear resolution as next example shows. ###### Example 2.3. Let $G$ be the following cochordal graph on six vertices. 123456 Let $I=I(G)\subset S=K[x_{1},\dots,x_{6}]$. Using the package [7] we verified that $\textup{HS}_{0}(I)$ and $\textup{HS}_{1}(I)$ have linear quotients. However the last homological shift ideal $\textup{HS}_{2}(I)=(x_{1}x_{2}x_{3}x_{4},\,x_{1}x_{4}x_{5}x_{6})$ has the following non-linear resolution $0\rightarrow S(-6)\rightarrow S(-4)^{2}\rightarrow(x_{1}x_{2}x_{3}x_{4},x_{1}x_{4}x_{5}x_{6})\rightarrow 0.$ In graph theory, one distinguished class of chordal graphs is the family of proper interval graphs. A graph $G$ is called an interval graph, if one can label its vertices with some intervals on the real line so that two vertices are adjacent in $G$, when the intersection of their corresponding intervals is non-empty. A proper interval graph is an interval graph such that no interval properly contains another. Now we are ready to state our main result in the section. ###### Theorem 2.4. Let $G$ be a cochordal graph on $[n]$ whose complementary graph $G^{c}$ is a proper interval graph. Then, $I(G)$ has homological linear quotients. In order to prove the theorem we introduce a more general class of graphs. We call a perfect elimination order $x_{1}>x_{2}>\dots>x_{n}$ of a chordal graph $G$ reversible if $x_{n}>x_{n-1}>\dots>x_{1}$ is also a perfect elimination order of $G$. We call a chordal graph $G$ reversible if $G$ admits a reversible perfect elimination order. Moreover, a cochordal graph $G$ is called reversible if and only if $G^{c}$ is reversible. ###### Lemma 2.5. Let $G$ be a proper interval graph. Then $G$ is reversible. ###### Proof. By [17, Theorem 1 and Lemma 1], up to a relabeling of the vertex set of $G$, the following property is satisfied: 1. $(*)$ for all $i<j$, $\\{i,j\\}\in E(G)$ implies that the induced subgraph of $G$ on $\\{i,i+1\dots,j\\}$ is a clique, _i.e._ , a complete subgraph. With such a labeling, both $x_{1}>x_{2}>\dots>x_{n}$ and $x_{n}>x_{n-1}>\dots>x_{1}$ are perfect elimination orders of $G$. By symmetry, it is enough to show that $x_{1}>x_{2}>\dots>x_{n}$ is a perfect elimination order. Let $i\in[n]$, $j,k\in N_{G}(i)$ with $j,k>i$. We prove that $\\{j,k\\}\in E(G)$. Suppose $j>k$. By $(*)$, the induced subgraph of $G$ on $\\{i,i+1\dots,j\\}$ is a clique. Since $j>k>i$, we obtain that $\\{j,k\\}\in E(G)$, as wanted. With this lemma at hand, Theorem 2.4 follows from the following more general result. ###### Theorem 2.6. Let $G$ be a cochordal graph on $[n]$, and let $x_{1}>\dots>x_{n}$ be a reversible perfect elimination order of $G^{c}$. Then, $\textup{HS}_{k}(I(G))$ has linear quotients with respect to the lexicographic order $>_{\textup{lex}}$ induced by $x_{1}>\dots>x_{n}$, for all $k\geq 0$. For the proof of this theorem, we need a description of the homological shift ideals. ###### Lemma 2.7. Let $G$ be a cochordal graph on $[n]$, and let $x_{1}>x_{2}>\dots>x_{n}$ be a perfect elimination order of $G^{c}$. Then, for all $\\{i,j\\}\in E(G)$, with $i<j$, (3) $\textup{set}(x_{i}x_{j})=\\{1,\dots,i-1\\}\cup(\\{i+1,\dots,j-1\\}\cap N_{G}(i)).$ In particular, $\displaystyle\textup{HS}_{k}(I(G))=\big{(}{\bf x}_{A}{\bf x}_{B}\ :$ $\displaystyle\ A,B\subseteq[n],A,B\neq\emptyset,\max(A)<\min(B),|A\cup B|=k+2,$ $\displaystyle\ \\{\max(A),b\\}\in E(G),\ \text{for all}\ b\in B\big{)}.$ ###### Proof. As remarked before $I(G)$ has linear quotients with respect to the lexicographic order $>_{\textup{lex}}$ induced by $x_{1}>x_{2}>\dots>x_{n}$. Let $\\{i,j\\}\in E(G)$ with $i<j$. Let us determine $\textup{set}(x_{i}x_{j})$. If $k\in\textup{set}(x_{i}x_{j})$, then $x_{k}(x_{i}x_{j})/x_{\ell}\in I(G)$ and $x_{k}(x_{i}x_{j})/x_{\ell}>_{\textup{lex}}x_{i}x_{j}$ for some $\ell\in\\{i,j\\}$. Note that $k<j$, indeed for $k>j$, both $x_{i}x_{k},x_{j}x_{k}$ are smaller than $x_{i}x_{j}$ in the lexicographic order. Thus either $k<i$ or $i<k<j$. We distinguish the two possible cases. Case 1. Suppose $k<i$. Assume that none of $x_{k}x_{i},x_{k}x_{j}$ is in $I(G)$. Then $\\{k,i\\},\\{k,j\\}\in E(G^{c})$. Since $x_{1}>x_{2}>\dots>x_{n}$ is a perfect elimination order, the induced graph of $G_{i}^{c}$ on the vertex set $N_{G_{k}^{c}}(k)$ is complete. But $i,j>k$ and $i,j\in N_{G_{k}^{c}}(k)$. Thus we would have $\\{i,j\\}\in E(G^{c})$, that is, $x_{i}x_{j}\notin I(G)$, absurd. Case 2. Suppose $i<k<j$. Since $k>i$, $x_{k}x_{j}<_{\textup{lex}}x_{i}x_{j}$. Thus $k\in\textup{set}(x_{i}x_{j})$ if and only if $x_{i}x_{k}\in E(G)$, that is $k\in N_{G}(i)$. The two cases above show that equation (3) holds. The formula for $\textup{HS}_{k}(I(G))$ follows immediately by applying equations (1) and (3). For the proof of the theorem we recall the concept of Betti splitting [8]. Let $I$, $I_{1}$, $I_{2}$ be monomial ideals of $S$ such that $G(I)$ is the disjoint union of $G(I_{1})$ and $G(I_{2})$. We say that $I=I_{1}+I_{2}$ is a Betti splitting if $\beta_{i,j}(I)=\beta_{i,j}(I_{1})+\beta_{i,j}(I_{2})+\beta_{i-1,j}(I_{1}\cap I_{2})\ \ \ \textup{for all}\ i,j.$ ###### Proof of Theorem 2.6.. We proceed by induction on $n\geq 1$. Let $G^{\prime}$ be the induced subgraph of $G$ on the vertex set $\\{2,3,\dots,n\\}$. Then $x_{2}>x_{3}>\dots>x_{n}$ is again a reversible perfect elimination order of $(G^{\prime})^{c}$ and $G^{\prime}$ is a reversible cochordal graph. Let $J=(x_{i}:x_{1}x_{i}\in I(G))$. Then, $I(G)=x_{1}J+I(G^{\prime})$ is a Betti splitting, because $G(I(G))$ is the disjoint union of $G(x_{1}J)$ and $G(I(G^{\prime}))$, and $x_{1}J$, $I(G^{\prime})$ have linear resolutions, see [8, Corollary 2.4]. Since $I(G^{\prime})\cap x_{1}J=x_{1}I(G^{\prime})$, [4, Proposition 1.7] gives $\textup{HS}_{k}(I(G))=x_{1}\big{(}\textup{HS}_{k-1}(I(G^{\prime}))+\textup{HS}_{k}(J)\big{)}+\textup{HS}_{k}(I(G^{\prime})).$ We claim that $\textup{HS}_{k}(I(G))$ has linear quotients with respect to the lexicographic order $>_{\textup{lex}}$ induced by $x_{1}>x_{2}>\dots>x_{n}$. For $k=0$ this is true. Let $k>0$. Let $u=x_{i_{1}}x_{j_{1}}{\bf x}_{F_{1}},v=x_{i_{2}}x_{j_{2}}{\bf x}_{F_{2}}\in G(\textup{HS}_{k}(I(G)))$, with $u>_{\textup{lex}}v$, $i_{1}<j_{1},i_{2}<j_{2}$, $x_{i_{1}}x_{j_{1}},\ x_{i_{2}}x_{j_{2}}\in I(G)$, $F_{1}\subseteq\textup{set}(u)$, $F_{2}\subseteq\textup{set}(v)$. We are going to prove that there exists $w\in G(\textup{HS}_{k}(I(G)))$ such that $w>_{\textup{lex}}v$, $w:v=x_{p}$ and $x_{p}$ divides $u:v$. We can write $u=x_{p_{1}}x_{p_{2}}\cdots x_{p_{k+2}},\ \ \ \ v=x_{q_{1}}x_{q_{2}}\cdots x_{q_{k+2}},$ with $p_{1}<p_{2}<\dots<p_{k+2}$, $q_{1}<q_{2}<\dots<q_{k+2}$. Since $u>_{\textup{lex}}v$ then $p_{1}=q_{1}$, $p_{2}=q_{2}$, $\dots$, $p_{s-1}=q_{s-1}$, $p_{s}<q_{s}$ for some $s\in\\{1,\dots,k+2\\}$. If $s=k+2$, then $u:v=x_{p_{k+2}}=x_{j_{1}}$ and there is nothing to prove. Therefore, we may assume $s<k+2$. Thus $p_{s}<q_{s}<q_{k+2}=j_{2}$. Set $p=p_{s}$ and $q=q_{s}$, then $x_{p}$ divides $u:v$. Suppose for the moment that $x_{1}$ divides $v$. Then by definition of $>_{\textup{lex}}$, $p_{1}=q_{1}=1$ and $x_{1}$ divides $u$, too. There are four cases to consider. Case 1. Suppose $i_{1}=i_{2}=1$. Setting $u^{\prime}=u/x_{1}$ and $v^{\prime}=v/x_{1}$, we have $u^{\prime},v^{\prime}\in G(\textup{HS}_{k}(J))$ and $u^{\prime}>_{\textup{lex}}v^{\prime}$. Since $J$ is an ideal generated by variables, it has homological linear quotients with respect to $>_{\textup{lex}}$. Hence, there exists $w^{\prime}\in G(\textup{HS}_{k}(J))$ with $w^{\prime}>_{\textup{lex}}v^{\prime}$ such that $w^{\prime}:v^{\prime}=x_{\ell}$ and $x_{\ell}$ divides $u^{\prime}:v^{\prime}$. Setting $w=x_{1}w^{\prime}$, we have that $w>_{\textup{lex}}v$ and $w\in G(x_{1}\textup{HS}_{k}(J))\subseteq G(\textup{HS}_{k}(I(G)))$. Hence $w:v=w^{\prime}:v^{\prime}=x_{\ell}$ and $x_{\ell}$ divides $u:v=u^{\prime}:v^{\prime}$. Case 2. Suppose $i_{1}>1$ and $i_{2}>1$. Setting $u^{\prime}=u/x_{1}$ and $v^{\prime}=v/x_{1}$, we have $u^{\prime},v^{\prime}\in G(\textup{HS}_{k-1}(I(G^{\prime})))$ and $u^{\prime}>_{\textup{lex}}v^{\prime}$. By inductive hypothesis, $I(G^{\prime})$ has homological linear quotients with respect to $>_{\textup{lex}}^{\prime}$ induced by $x_{2}>x_{3}>\dots>x_{n}$. Hence, there exists $w^{\prime}\in G(\textup{HS}_{k-1}(I(G^{\prime})))$ with $w^{\prime}>_{\textup{lex}}^{\prime}v^{\prime}$ such that $w^{\prime}:v^{\prime}=x_{\ell}$ and $x_{\ell}$ divides $u^{\prime}:v^{\prime}$. Setting $w=x_{1}w^{\prime}$, we have that $w>_{\textup{lex}}v$ and $w\in G(x_{1}\textup{HS}_{k-1}(I(G^{\prime})))\subseteq G(\textup{HS}_{k}(I(G)))$. Hence $w:v=w^{\prime}:v^{\prime}=x_{\ell}$ and $x_{\ell}$ divides $u:v=u^{\prime}:v^{\prime}$. Case 3. Suppose $i_{1}>1$ and $i_{2}=1$. Then $1=i_{2}<p<j_{2}$. Subcase 3.1. Assume $x_{1}x_{p}\in I(G)$, then $p\in\textup{set}(x_{i_{2}}x_{j_{2}})$. Setting $w=x_{p}(v/x_{q})$, by equation (1) $w\in G(\textup{HS}_{k}(I(G)))$, and $w>_{\textup{lex}}v$, because $p<q$. Moreover $w:v=x_{p}$ and $x_{p}$ divides $u:v$. Subcase 3.2. Assume that $x_{1}x_{p}\notin I(G)$. By hypothesis, $x_{n}>x_{n-1}>\dots>x_{1}$ is also a perfect elimination order of $G^{c}$. Thus, by Lemma 2.7, we can write $u={\bf x}_{A}{\bf x}_{B}$ with $A=\\{p_{k+2},p_{k+1},\dots,p_{r}\\}$, $B=\\{p_{r-1},\dots,p_{2},p_{1}\\}$ for some $r>1$ and with $\\{p_{r},p_{\ell}\\}\in E(G)$ for all $\ell=r-1,\dots,2,1$. Since $\\{1,p\\}=\\{p_{1},p_{s}\\}\notin E(G)$, by the above presentation of $u$, $s>r$. Using again Lemma 2.7, but considering the reversed perfect elimination order $x_{n}>x_{n-1}>\dots>x_{1}$, we see that $\displaystyle w$ $\displaystyle=x_{q_{s+1}}x_{q_{s+2}}\cdots x_{q_{k+2}}u/(x_{p_{s+1}}x_{p_{s+2}}\cdots x_{p_{k+2}})$ $\displaystyle={\bf x}_{(A\setminus\\{p_{s+1},p_{s+2},\dots,p_{k+2}\\})\cup\\{q_{s+1},q_{s+2},\dots,q_{k+2}\\}}{\bf x}_{B}\in G(\textup{HS}_{k}(I(G))).$ Moreover, $w>_{\textup{lex}}v$, $w:v=x_{p}$ and $x_{p}$ divides $u:v$, as desired. Case 4. Suppose $i_{1}=1$ and $i_{2}>1$. Recall that $p<j_{2}$. Moreover $p\neq i_{2}$, because $x_{p}$ divides $u:v$ but $x_{i_{2}}$ divides $v$. Thus there are two cases to consider. Subcase 4.1. Assume $p<i_{2}$. By Lemma 2.7, $p\in\textup{set}(x_{i_{2}}x_{j_{2}})$. If $q\neq i_{2}$, then $q<j_{2}$ and by equation (1) $w=x_{p}(v/x_{q})$ is a minimal generator of $\textup{HS}_{k}(I(G))$. Moreover $w>_{\textup{lex}}v$ and $w:v=x_{p}$ divides $u:v$, as wanted. Suppose now that $q=i_{2}$. If there exists $\ell$ such that $x_{\ell}$ divides $v$ and $i_{2}<\ell<j_{2}$, then $\ell>p$ and $w=x_{p}(v/x_{\ell})$ is a minimal generator of $\textup{HS}_{k}(I(G))$ such that $w>_{\textup{lex}}v$ and with $w:v=x_{p}$ dividing $u:v$, as wanted. Otherwise, suppose no such integer $\ell$ exists. Then, $s=k+1$, $q_{k+1}=i_{2}$ and $q_{k+2}=j_{2}$. Since $p\in\textup{set}(x_{i_{2}}x_{j_{2}})$, then $x_{p}x_{\ell}\in I(G)$, where $\ell\in\\{i_{2},j_{2}\\}$. Then $p<\ell$ and by Lemma 2.7 we see that $w=x_{p}(v/x_{\ell})$ is a minimal generator of $\textup{HS}_{k}(I(G))$ such that $w>_{\textup{lex}}v$ and with $w:v=x_{p}$ dividing $u:v$. Subcase 4.2. Assume now $i_{2}<p<j_{2}$. If $x_{i_{2}}x_{p}\in I(G)$, by Lemma 2.7, $p\in\textup{set}(x_{i_{2}}x_{j_{2}})$. Setting $w=x_{p}(v/x_{q})$, we have $w\in G(\textup{HS}_{k}(I(G)))$, $w>_{\textup{lex}}v$ and $w:v=x_{p}$ divides $u:v$. Suppose now that $x_{i_{2}}x_{p}\notin I(G)$. By hypothesis, $x_{n}>x_{n-1}>\dots>x_{1}$ is also a perfect elimination order of $G^{c}$. Thus, by Lemma 2.7, we can write $u={\bf x}_{A}{\bf x}_{B}$ with $A=\\{p_{k+2},p_{k+1},\dots,p_{r}\\}$, $B=\\{p_{r-1},\dots,p_{2},p_{1}\\}$ for some $r>1$ and with $\\{p_{r},p_{\ell}\\}\in E(G)$ for all $\ell=r-1,\dots,2,1$. Note that $i_{2}<p$, so $x_{i_{2}}$ divides $u$. Since $\\{i_{2},p\\}=\\{i_{2},p_{s}\\}\notin E(G)$, by the above presentation of $u$, $s>r$. Using again Lemma 2.7, but considering the reversed perfect elimination order $x_{n}>x_{n-1}>\dots>x_{1}$, we see that $\displaystyle w$ $\displaystyle=x_{q_{s+1}}x_{q_{s+2}}\cdots x_{q_{k+2}}u/(x_{p_{s+1}}x_{p_{s+2}}\cdots x_{p_{k+2}})$ $\displaystyle={\bf x}_{A}{\bf x}_{(B\setminus\\{p_{s+1},p_{s+2},\dots,p_{k+2}\\})\cup\\{q_{s+1},q_{s+2},\dots,q_{k+2}\\}}\in G(\textup{HS}_{k}(I(G))).$ Moreover, $w>_{\textup{lex}}v$, $w:v=x_{p}$ and $x_{p}$ divides $u:v$, as desired. Suppose now that $x_{1}$ does not divide $v$. Then $v\in G(\textup{HS}_{k}(I(G^{\prime})))$. If $x_{1}$ does not divide $u$, then $u\in G(\textup{HS}_{k}(I(G^{\prime})))$, too. Let $>_{\textup{lex}}^{\prime}$ be the lexicographic order induced by $x_{2}>x_{3}>\dots>x_{n}$. Since by induction $I(G^{\prime})$ has homological linear quotients with respect to $>_{\textup{lex}}^{\prime}$ and also $u>_{\textup{lex}}^{\prime}v$, there exists $w\in G(\textup{HS}_{k}(I(G^{\prime})))$, with $w>_{\textup{lex}}^{\prime}v$, $w:v=x_{\ell}$ and $x_{\ell}$ divides $u:v$. But also we have $w\in G(\textup{HS}_{k}(I(G)))$ and $w>_{\textup{lex}}v$. Otherwise if $x_{1}$ divides $u$, then $x_{1}$ divides $u:v$. Since $\textup{HS}_{k}(I(G^{\prime}))\subseteq\textup{HS}_{k-1}(I(G^{\prime}))$ and $k>0$, we can write $v=x_{t}w^{\prime}$ with $w^{\prime}\in G(\textup{HS}_{k-1}(I(G^{\prime})))$. Let $w=x_{1}w^{\prime}$. Then $w>_{\textup{lex}}v$ and $w:v=x_{1}$ divides $u:v$. Hence, the inductive proof is complete and the theorem is proved. ###### Remark 2.8. Let $x_{1}>x_{2}>\dots>x_{n}$ be a reversible perfect elimination order of $G^{c}$. By symmetry, Theorem 2.6 shows also that $\textup{HS}_{k}(I(G))$ has linear quotients with respect to the lexicographic order induced by $x_{n}>x_{n-1}>\dots>x_{1}$. ###### Examples 2.9. Let $n,m$ be two positive integers. 1. (a) Let $G=K_{n,m}$ be the complete bipartite graph. That is, $V(G)=[n+m]$ and $E(G)=\big{\\{}\\{i,j\\}:i\in[n],j\in\\{n+1,\dots,n+m\\}\big{\\}}$. For example, for $n=3$ and $m=4$ 1234567 It is easy to see that $G^{c}$ is the disjoint union of two complete graphs $\Gamma_{1}$ and $\Gamma_{2}$ on vertex sets $[n]$ and $\\{n+1,\dots,n+m\\}$ respectively. Furthermore, any ordering of the vertices is a perfect elimination order of $G^{c}$. Applying the previous theorem, $I(G)=(x_{1},\dots,x_{n})(x_{n+1},\dots,x_{m})$ has homological linear quotients with respect to the lexicographic order induced by any ordering of the variables. 2. (b) Let $G$ be the graph with vertex set $V(G)=[n+m]$ and edge set $E(G)=\big{\\{}\\{i,j\\}:i\in[n+m],n+1\leq j\leq n+m,i<j\big{\\}}.$ We claim that $G$ is a reversible cochordal graph. Indeed $G^{c}$ is the disjoint union of the complete graph $K_{n}$ on the vertex set $[n]$ together with the set of isolated vertices $\\{n+1,\dots,n+m\\}$. It easily seen that any ordering of the vertices is a perfect elimination order of $G^{c}$. Applying Theorem 2.6 $I(G)=(x_{1},\dots,x_{n})(x_{n+1},\dots,x_{m})+(x_{i}x_{j}:n+1\leq i<j\leq n+m)$ has homological linear quotients with respect to the lexicographic order induced by any ordering of the variables. ## 3\. Homological shifts of Trees In this section we construct several classes of edge ideals with homological linear quotients, by considering various operations on cochordal graphs that preserve the homological linear quotients property. As a main application of all these results we will prove the following theorem. ###### Theorem 3.1. Let $G$ be a graph such that $G^{c}$ is a forest. Then $I(G)$ has homological linear quotients. The squarefree Veronese ideal $I_{n,d}$ of degree $d$ in $S=K[x_{1},\dots,x_{n}]$ is the ideal of $S$ generated by all squarefree monomials of degree $d$ in $S$. It is well-known that $I_{n,d}$ has homological linear quotients, (see for instance [13, Corollary 3.2]). The first operation we consider consists in adding whiskers. Let $\Gamma^{\prime}$ be a graph on vertex set $[n-1]$. Let $i\in[n-1]$ and let $\Gamma$ be the graph with vertex set $[n]$ and edge set $V(\Gamma)=V(\Gamma^{\prime})\cup\\{\\{i,n\\}\\}$. $\Gamma$ is called the whisker graph of $\Gamma^{\prime}$ obtained by adding the whisker $\\{i,n\\}$ to $\Gamma^{\prime}$. ###### Proposition 3.2. Let $\Gamma^{\prime}$ be a graph on vertex set $[n-1]$ and $\Gamma$ be the graph on vertex set $[n]$ and edge set $V(\Gamma)=V(\Gamma^{\prime})\cup\\{\\{i,n\\}\\}$ for some $i\in[n-1]$. Set $G=\Gamma^{c}$. Suppose $I((\Gamma^{\prime})^{c})$ has homological linear quotients. Then $I(G)$ has homological linear quotients, too. ###### Proof. Since $\Gamma^{\prime}$ is chordal, obviously $\Gamma$ is chordal, too. Set $J=I((\Gamma^{\prime})^{c})$, $I=I(G)$ and $L=(x_{j}:j\in[n-1]\setminus\\{i\\})$. Since $N_{G^{c}}(n)=\\{i\\}$, we have the Betti splitting: (4) $I=x_{n}L+J.$ Since $G$ is cochordal, $\textup{HS}_{0}(I)$ and $\textup{HS}_{1}(I)$ have linear quotients. So we only have to show that $\textup{HS}_{k}(I)$ has linear quotients for $k\geq 2$. By equation (4), for all $k\geq 2$, $\textup{HS}_{k}(I)=x_{n}\textup{HS}_{k}(L)+x_{n}\textup{HS}_{k-1}(J)+\textup{HS}_{k}(J).$ Note that $\textup{HS}_{k}(L)$ is the squarefree Veronese ideal of degree $k+1$ in the polynomial ring $K[x_{j}:j\in[n-1]\setminus\\{i\\}]$. Thus $\textup{HS}_{k}(L)$ has linear quotients with admissible order, say, $u_{1},\dots,u_{m}$. Let $v_{1},\dots,v_{r}$ and $w_{1},\dots,w_{s}$ be admissible orders of $\textup{HS}_{k-1}(J)$ and $\textup{HS}_{k}(J)$, respectively. Let $v_{j_{1}},\dots,v_{j_{p}}$, with $j_{1}<j_{2}<\dots<j_{p}$, the monomials in $G(\textup{HS}_{k-1}(J))\setminus G(\textup{HS}_{k}(L))$. We claim that (5) $x_{n}u_{1},\dots,x_{n}u_{m},\ x_{n}v_{j_{1}},\dots,x_{n}v_{j_{p}},\ w_{1},\dots,w_{s}$ is an admissible order of $\textup{HS}_{k}(J)$. Let $\ell\in\\{1,\dots,m\\}$. Then $(x_{n}u_{1},\dots,x_{n}u_{\ell-1}):x_{n}u_{\ell}=(u_{1},\dots,u_{\ell-1}):u_{\ell}$ is generated by variables. Let $\ell\in\\{1,\dots,p\\}$. We show that $\displaystyle Q$ $\displaystyle=(x_{n}u_{1},\dots,x_{n}u_{m},x_{n}v_{j_{1}},\dots,x_{n}v_{j_{\ell-1}}):x_{n}v_{j_{\ell}}$ $\displaystyle=(u_{1},\dots,u_{m},v_{j_{1}},\dots,v_{j_{\ell-1}}):v_{j_{\ell}}$ is generated by variables. Consider $v_{j_{q}}:v_{j_{\ell}}$, then we can find $d<j_{\ell}$ such that $v_{d}:v_{j_{\ell}}$ is a variable that divides $v_{j_{q}}:v_{j_{\ell}}$. Either $d=j_{b}$, for some $b<\ell$, or $v_{d}\in\textup{HS}_{k}(L)$. In any case, $x_{n}v_{d}\in(u_{1},\dots,u_{m},v_{j_{1}},\dots,v_{j_{\ell-1}})$ and $v_{d}:v_{j_{\ell}}\in Q$ divides $v_{j_{q}}:v_{j_{\ell}}$. Consider now $u_{q}:v_{j_{\ell}}$, $1\leq q\leq m$. Hence $x_{i}$ divides $v_{j_{\ell}}$, lest $v_{j_{\ell}}\in G(\textup{HS}_{k}(L))$. But then $v_{j_{\ell}}/x_{i}\in\textup{HS}_{k-1}(L)$. Let $x_{t}$ dividing $u_{q}:v_{j_{\ell}}$. Then $u=x_{t}v_{j_{\ell}}/x_{i}\in\textup{HS}_{k}(L)$ and $u:v_{j_{\ell}}=x_{t}\in Q$ divides $u_{q}:v_{j_{\ell}}$. Finally, let $\ell\in\\{1,\dots,s\\}$. We show that $\displaystyle Q$ $\displaystyle=(x_{n}u_{1},\dots,x_{n}u_{m},x_{n}v_{j_{1}},\dots,x_{n}v_{j_{p}},w_{1},\dots,w_{\ell-1}):w_{\ell}$ $\displaystyle=(x_{n}\textup{HS}_{k}(L)+x_{n}\textup{HS}_{k-1}(J)):w_{\ell}+(w_{1},\dots,w_{\ell-1}):w_{\ell}$ is generated by variables. Since $w_{1},\dots,w_{s}$ is an admissible order, $(w_{1},\dots,w_{\ell-1}):w_{\ell}$ is generated by variables. Consider now a generator $x_{n}z:w_{\ell}$ with $z\in\textup{HS}_{k}(L)$ or $z\in\textup{HS}_{k-1}(J)$. Then $x_{n}$ divides $x_{n}z:w_{\ell}$. On the other hand $w_{\ell}/x_{t}\in\textup{HS}_{k-1}(J)$ for some $t$. But then $x_{n}w_{\ell}/x_{t}:w_{\ell}=x_{n}\in Q$ divides our generator. The three cases above show that (5) is an admissible order, as desired. Since any tree can be constructed iteratively by adding a whisker to a tree on a smaller vertex set at each step, the previous proposition implies immediately ###### Corollary 3.3. Let $G$ be a graph such that $G^{c}$ is a tree. Then $I(G)$ has homological linear quotients. The second operation we consider consists in joining disjoint graphs. Two graphs $\Gamma_{1}$ and $\Gamma_{2}$ are called disjoint if $V(\Gamma_{1})\cap V(\Gamma_{2})=\emptyset$. The join of $\Gamma_{1}$ and $\Gamma_{2}$ is the graph $\Gamma$ with vertex set $V(\Gamma)=V(\Gamma_{1})\cup V(\Gamma_{2})$ and edge set $E(\Gamma)=E(\Gamma_{1})\cup E(\Gamma_{2})$. ###### Proposition 3.4. Let $\Gamma_{1}$ and $\Gamma_{2}$ be disjoint chordal graphs such that $I(\Gamma_{1}^{c}),I(\Gamma_{2}^{c})$ have homological linear quotients. Let $\Gamma$ be the join of $\Gamma_{1}$ and $\Gamma_{2}$ and set $G=\Gamma^{c}$. Then $I(G)$ has homological linear quotients, too. ###### Proof. Obviously $\Gamma$ is chordal, too. Let $G_{1}=\Gamma_{1}^{c}$, $G_{2}=\Gamma_{2}^{c}$, $V(G_{1})=[n]$ and $V(G_{2})=\\{n+1,\dots,n+m\\}$. Set $L=(x_{1},\dots,x_{n})(x_{n+1},\dots,x_{m})$. Then, $I(G)=I(G_{1})+I(G_{2})+L$ Suppose $x_{1}>\dots>x_{n}$ and $x_{n+1}>\dots>x_{n+m}$ are perfect elimination orders of $\Gamma_{1}$ and $\Gamma_{2}$. Then $G=\Gamma^{c}$ is cochordal. Indeed, $x_{1}>\dots>x_{n}>x_{n+1}>\dots>x_{n+m}$ is a perfect elimination order of $\Gamma$. Let $>_{\textup{lex}}$ be the lexicographic order induced by such an ordering of the variables. Set, $I=I(G)$, $I_{1}=I(G_{1})$ and $I_{2}=I(G_{2})$. Then, $I,I_{1},I_{2}$ and $J$ have linear quotients with respect to $>_{\textup{lex}}$. Let $k\geq 0$ and $u\in G(\textup{HS}_{k}(I))$ such that $x_{i}x_{j}$ divides $u$ for some integers $i\in[n]$, $n+1\\!\leq\\!j\leq\\!n+m$. We claim that $u\in G(\textup{HS}_{k}(L))$. Let $i_{0}=\max\\{i\in[n]:x_{i}\ \textup{divides}\ u\\}$ and $j_{0}=\max\\{j\in\\{n+1,\dots,n+m\\}:x_{j}\ \textup{divides}\ u\\}$. Let $u/(x_{i_{0}}x_{j_{0}})={\bf x}_{F}$. Then $F\subseteq\\{1,\dots,i_{0}-1\\}\cup\\{n+1,\dots,j_{0}-1\\}=\textup{set}_{I}(x_{i_{0}}x_{j_{0}})$ and $x_{i_{0}}x_{j_{0}}\in L$. Thus, by equation (1), $u=x_{i_{0}}x_{j_{0}}{\bf x}_{F}\in\textup{HS}_{k}(L)$, as desired. This argument shows that any squarefree monomial $w\in K[x_{1},\dots,x_{n+m}]$ of degree $k+2$, containing as a factor any monomial $x_{i}x_{j}$ with $i\in[n]$ and $n+1\leq j\leq n+m$, is a generator of $\textup{HS}_{k}(L)$. From this remark, for all $k\geq 0$, it follows that $\textup{HS}_{k}(I)=\textup{HS}_{k}(L)+\textup{HS}_{k}(I_{1})+\textup{HS}_{k}(I_{2}).$ Note that $L$ is the edge ideal of a complete bipartite graph. By Examples 2.9(a), $L$ has homological linear quotients. Let $u_{1},\dots,u_{m}$ be an admissible order of $\textup{HS}_{k}(L)$. Moreover, let $v_{1},\dots,v_{r}$ and $w_{1},\dots,w_{s}$ be admissible orders of $\textup{HS}_{k}(I_{1})$ and $\textup{HS}_{k}(I_{2})$, respectively. Note that the monomials $u_{i},v_{j},w_{t}$ are all different, because all monomials $u_{i}$ contain a factor $x_{i_{0}}x_{j_{0}}$ with $i_{0}\in[n]$ and $j_{0}\in\\{n+1,\dots,n+m\\}$. Whereas, the $v_{j}$ are monomials in $K[x_{1},\dots,x_{n}]$ and the $w_{t}$ are monomials in $K[x_{n+1},\dots,x_{n+m}]$. We claim that (6) $u_{1},\dots,u_{m},\ v_{1},\dots,v_{r},\ w_{1},\dots,w_{s}$ is an admissible order of $\textup{HS}_{k}(I)$. Let $\ell\in\\{1,\dots,m\\}$. Then $(u_{1},\dots,u_{\ell-1}):u_{\ell}$ is generated by variables. Let $\ell\in\\{1,\dots,r\\}$. We show that $Q=(u_{1},\dots,u_{m},v_{1},\dots,v_{\ell-1}):v_{\ell}$ is generated by variables. Clearly $(v_{1},\dots,v_{\ell-1}):v_{\ell}$ is generated by variables. Consider now $u_{q}:v_{\ell}$, $1\leq q\leq m$. Recall that $v_{\ell}$ is a monomial in $K[x_{1},\dots,x_{n}]$. Thus $x_{j}$ divides $u_{q}:v_{\ell}$ for some $j\in\\{n+1,\dots,n+m\\}$. Consider $v_{\ell}/x_{t}$ for some $t$. Then $u=x_{j}(v_{\ell}/x_{t})\in\textup{HS}_{k}(L)$ and $u:v_{\ell}=x_{j}\in Q$, as desired. Finally, let $\ell\in\\{1,\dots,s\\}$. We show that $Q=(u_{1},\dots,u_{m},v_{1},\dots,v_{r},w_{1},\dots,w_{\ell-1}):w_{\ell}$ is generated by variables. Since $w_{1},\dots,w_{s}$ is an admissible order, $(w_{1},\dots,w_{\ell-1}):w_{\ell}$ is generated by variables. Consider now a generator $z:w_{\ell}$ with $z=u_{q}$ or $z=v_{q}$, for some $q$. Since $w_{\ell}$ is a monomial in $K[x_{n+1},\dots,x_{n+m}]$, $z:w_{\ell}$ is divided by a variable $x_{i}$, where $i\in[n]$. Consider $w_{\ell}/x_{t}$ for some $t$. Then $u=x_{i}(w_{\ell}/x_{t})\in\textup{HS}_{k}(L)$ and $u:w_{\ell}=x_{i}\in Q$, as desired. The three cases above show that (6) is an admissible order, as desired. ###### Proof of Theorem 3.1. Let $\Gamma=G^{c}$ be a forest and let $c$ be the number of connected components of $\Gamma$. If $c=1$, then $\Gamma$ is a tree, and by Corollary 3.3, $I(G)$ has homological linear quotients. Suppose $c>1$ and write $\Gamma=\Gamma_{1}\cup\Gamma_{2}$, where $\Gamma_{1}$ and $\Gamma_{2}$ are disjoint forests. The the numbers of connected components of $\Gamma_{1}$ and $\Gamma_{2}$ are smaller than $c$. Thus, by induction $I(\Gamma_{1}^{c})$ and $I(\Gamma_{2}^{c})$ have homological linear quotients. Applying Proposition 3.4, it follows that $I(G)$ has homological linear quotients, too. Let $G$ be a complete multipartite graph, then $G^{c}$ is the disjoint union of some complete graphs. Repeated applications of Proposition 3.4 yield ###### Corollary 3.5. Let $G$ be a complete multipartite graph. Then $I(G)$ has homological linear quotients. ## 4\. Polymatroidal ideals generated in degree two A polymatroidal ideal $I\subset S=K[x_{1},\dots,x_{n}]$ is a monomial ideal $I$ generated in a single degree verifying the following exchange property: for all $u,v\in G(I)$ with $u\neq v$ and all $i$ such that $\deg_{x_{i}}(u)>\deg_{x_{i}}(v)$, there exists $j$ such that $\deg_{x_{j}}(u)<\deg_{x_{j}}(v)$ and $x_{j}(u/x_{i})\in G(I)$. The name polymatroidal ideal is justified by the fact that their minimal generating set corresponds to the set of bases of a discrete polymatroid. A squarefree polymatroidal ideal is called matroidal. Any polymatroidal ideal also satisfy a dual version of the exchange property. ###### Lemma 4.1. ([11, Lemma 2.1]). Let $I\subset S$ be a polymatroidal ideal. Then, for all $u,v\in G(I)$ and all $i$ such that $\deg_{x_{i}}(u)>\deg_{x_{i}}(v)$, there exists $j$ such that $\deg_{x_{j}}(u)<\deg_{x_{j}}(v)$ and $x_{i}(v/x_{j})\in G(I)$. There are many useful characterization of polymatroidal ideals. The following one is due Bandari and Rahmati-Asghar. ###### Theorem 4.2. ([1, Theorem 2.4]). Let $I\subset S$ be a monomial ideal generated in a single degree. Then, $I$ is polymatroidal if and only if $I$ has linear quotients with respect to the lexicographic order induced by any ordering of the variables. It is expected by Bandari, Bayati and Herzog that the homological shift ideals $\textup{HS}_{k}(I)$ of a polymatroidal ideal $I$ are all polymatroidal, see [2, 13]. In this section, we provide an affirmative answer to this conjecture for all polymatroidal ideals generated in degree two. Firstly, we deal with the squarefree case. ###### Lemma 4.3. Let $I\subset S$ be a matroidal ideal generated in degree two, and let $G$ be the simple graph on $[n]$ such that $I=I(G)$. Then, any ordering of the variables is a perfect elimination order of $G^{c}$. ###### Proof. Up to relabeling, we can consider the ordering $x_{1}>x_{2}>\dots>x_{n}$. Let $j,k\in N_{G^{c}}(i)$ with $j,k>i$. We must prove that $\\{j,k\\}\in E(G^{c})$. By our assumption, $\\{i,j\\},\\{i,k\\}\notin E(G)$, that is $x_{i}x_{j},x_{i}x_{k}\notin I(G)=I$. Suppose by contradiction that $\\{j,k\\}\notin E(G^{c})$, then $\\{j,k\\}\in E(G)$, that is, $x_{j}x_{k}\in I(G)$. Pick any monomial $x_{i}x_{s}\in I(G)$. Then $\deg_{x_{i}}(x_{i}x_{s})>\deg_{x_{i}}(x_{j}x_{k})$. By Lemma 4.1, we can find $\ell$ with $\deg_{x_{\ell}}(x_{i}x_{s})<\deg_{x_{\ell}}(x_{j}x_{k})$ and $x_{i}(x_{j}x_{k})/x_{\ell}\in I(G)$. Thus, either $x_{i}x_{j}\in I(G)$ or $x_{i}x_{k}\in I(G)$. This is a contradiction. Hence $\\{j,k\\}\in E(G^{c})$, as desired. ###### Corollary 4.4. Let $I\subset S$ be a matroidal ideal generated in degree two. Then $\textup{HS}_{k}(I)$ is a matroidal ideal, for all $k\geq 0$. ###### Proof. Let $G$ be the simple graph on $[n]$ such that $I=I(G)$. By Lemma 4.3 and Theorem 2.2, $G^{c}$ is a reversible chordal graph and any ordering of the variables is a reversible perfect elimination order of $G^{c}$. By Theorem 2.6, for all $k\geq 0$, $\textup{HS}_{k}(I)$ has linear quotients with respect to the lexicographic order induced by any ordering of the variables. Thus, by Theorem 4.2, $\textup{HS}_{k}(I)$ is matroidal, for all $k\geq 0$. Now, we turn to the non-squarefree case. ###### Theorem 4.5. Let $I\subset S$ be a polymatroidal ideal generated in degree two. Then, $\textup{HS}_{k}(I)$ is a polymatroidal ideal, for all $k\geq 0$. ###### Proof. If $I$ is squarefree, the thesis follows from Corollary 4.4. Suppose $I$ is not squarefree. Up to a suitable relabeling, we can write $I=(J,x_{1}^{2},x_{2}^{2},\dots,x_{t}^{2})$, where $J$ is the squarefree part of $I$, _i.e._ , $G(J)=\\{u\in G(I):u\ \textup{is squarefree}\\}$ and $1\leq t\leq n$. Then $J$ is a matroidal ideal. Let $G$ be the simple graph on $[n]$ with $J=I(G)$, then $G^{c}$ is cochordal. Let $u_{1},\dots,u_{m}$ be an admissible order of $J$. We claim that $u_{1},\dots,u_{m},x_{1}^{2},x_{2}^{2},\dots,x_{t}^{2}$ is an admissible order of $I$. We only need to prove that $Q=(u_{1},\dots,u_{m},x_{1}^{2},\dots,x_{\ell-1}^{2}):x_{\ell}^{2}=(J,x_{1}^{2},\dots,x_{\ell-1}^{2}):x_{\ell}^{2}$ is generated by variables. Indeed, let $x_{i}x_{j}:x_{\ell}^{2}\in Q$ be a generator with $i\leq j$. If $x_{i}x_{j}:x_{\ell}^{2}$ is a variable, there is nothing to prove. Otherwise $x_{i}x_{j}:x_{\ell}^{2}=x_{i}x_{j}$, and $\ell\neq i,j$. Since $\deg_{x_{\ell}}(x_{\ell}^{2})>\deg_{x_{\ell}}(x_{i}x_{j})$, by the exchange property, $w=x_{k}(x_{\ell}^{2})/x_{\ell}=x_{k}x_{\ell}\in I$, with $k=i$ or $k=j$. Then $k\neq\ell$, $w=x_{k}x_{\ell}\in J$ and $w:x_{\ell}^{2}=x_{k}\in Q$ is a variable that divides $x_{i}x_{j}:x_{\ell}^{2}$, as desired. We claim that $\textup{set}(x_{\ell}^{2})=[n]\setminus\\{\ell\\}$, for all $\ell=1,\dots,t$. Let $i\in[n]\setminus\\{\ell\\}$. Then $x_{i}x_{j}\in G(I)$ for some $j$. If $j=\ell$, then $x_{i}x_{\ell}\in I$. Suppose $j\neq\ell$, then $\deg_{x_{j}}(x_{i}x_{j})>\deg_{x_{j}}(x_{\ell}^{2})$. By the exchange property, $x_{i}x_{\ell}\in I$, as desired. By equation (1), for all $k>0$, $\displaystyle\textup{HS}_{k}(I)\ $ $\displaystyle=\ \textup{HS}_{k}(J)+\sum_{\ell=1}^{t}x_{\ell}^{2}\cdot\textup{HS}_{k-1}((x_{i}:i\in[n]\setminus\\{\ell\\})).$ We set $J_{\ell}=(x_{i}:i\in[n]\setminus\\{\ell\\})$, $\ell=1,\dots,t$. Since $J$ is matroidal, $\textup{HS}_{k}(J)$ is matroidal by Corollary 4.4. Moreover, each ideal $J_{\ell}$ is generated by variables, and so it is matroidal. Hence all ideals $x_{\ell}^{2}\cdot\textup{HS}_{k-1}(J_{\ell})$ are polymatroidal. To verify that $\textup{HS}_{k}(I)$ is polymatroidal, we check the exchange property. Let $u,v\in G(\textup{HS}_{k}(I))$ and $i$ such that $\deg_{x_{i}}(u)>\deg_{x_{i}}(v)$. To achieve our goal, we note the following fact. Let $w\in S$ be any squarefree monomial of degree $k+1$ and let $\ell\in[t]$. Then $x_{\ell}w\in\textup{HS}_{k}(I)$. Indeed, if $x_{\ell}$ divides $w$, then $x_{\ell}w\in x_{\ell}^{2}\cdot\textup{HS}_{k-1}(J_{\ell})\subset\textup{HS}_{k}(I)$. Suppose $x_{\ell}$ does not divide $w$. For all $i$ such that $x_{i}$ divides $w$, $x_{i}x_{\ell}\in J$ because $i\neq\ell$. Fix a lexicographic order $\succ$ such that $x_{\ell}>x_{i}$ for all $i\in[n]\setminus\ell$. Up to relabeling, we can assume $\ell=1$ and that $\succ$ is induced by $x_{1}>x_{2}>\dots>x_{n}$. Writing $x_{\ell}w=x_{\ell}x_{j_{2}}\cdots x_{j_{k+2}}$ with $\ell=1<j_{2}<\dots<j_{k+2}\leq n$, then $x_{\ell}x_{j_{k+2}}\in J$, $x_{\ell}x_{j_{i}}\in J$ and $x_{\ell}x_{j_{i}}\succ x_{\ell}x_{j_{k+2}}$, for $i=2,\dots,k+1$. Hence $\\{j_{2},\dots,j_{k+1}\\}\subseteq\big{\\{}j\ |\ x_{j}\in(u\in G(J):u\succ x_{\ell}x_{j_{k+2}}):x_{\ell}x_{j_{k+2}}\big{\\}}.$ This shows that $x_{\ell}w\in\textup{HS}_{k}(J)\subset\textup{HS}_{k}(I)$, because by Theorem 4.2, $J$ has linear quotients with respect to $\succ$. If $u,v\in\textup{HS}_{k}(J)$ or $u,v\in x_{\ell}^{2}\cdot\textup{HS}_{k-1}(J_{\ell})$, we can find $j$ with $\deg_{x_{j}}(u)<\deg_{x_{j}}(v)$ such that $x_{j}(u/x_{i})\in\textup{HS}_{k}(I)$, because both $\textup{HS}_{k}(J),x_{\ell}^{2}\cdot\textup{HS}_{k-1}(J_{\ell})$ are polymatroidal. Suppose now $u\in\textup{HS}_{k}(J)$ and $v\in x_{\ell}^{2}\cdot\textup{HS}_{k-1}(J_{\ell})$. Then $\deg_{x_{\ell}}(u)<\deg_{x_{\ell}}(v)$ and $x_{\ell}(u/x_{i})\in\textup{HS}_{k}(I)$, because $u/x_{i}$ is a squarefree monomial of degree $k+1$. Suppose $u\in x_{\ell}^{2}\cdot\textup{HS}_{k-1}(J_{\ell})$ and $v\in\textup{HS}_{k}(J)$. Let $j$ such that $\deg_{x_{j}}(u)<\deg_{x_{j}}(v)$. Then $\deg_{x_{j}}(u)=0$. If $i=\ell$, then $x_{j}(u/x_{\ell})\in\textup{HS}_{k}(I)$ because it is the product of $x_{\ell}$ times a squarefree monomial of degree $k+1$. If $i\neq\ell$, then $x_{j}(u/x_{i})$ can also be written as such a product. In any case $x_{j}(u/x_{i})\in\textup{HS}_{k}(I)$. Finally, suppose $u\in x_{\ell}\cdot\textup{HS}_{k-1}(J_{\ell})$ and $v\in x_{h}^{2}\cdot\textup{HS}_{k-1}(J_{h})$ with $\ell\neq h$. Suppose $i=\ell$ and let $j$ such that $\deg_{x_{j}}(u)<\deg_{x_{j}}(v)$. Then $u^{\prime}=x_{j}(u/x_{i})$ is either $x_{\ell}$ times a squarefree monomial of degree $k+1$, or is equal to $x_{h}$ times a squarefree monomial of degree $k+1$. In both cases, $u^{\prime}\in\textup{HS}_{k}(I)$. Suppose now $i\neq\ell$. If there exist more than one $j$ with $\deg_{x_{j}}(u)<\deg_{x_{j}}(v)$, we can choose $j\neq h$. Then $\deg_{x_{j}}(v)=1$ and so $x_{j}$ does not divide $u$. Consequently $x_{j}(u/x_{i})$ is equal to $x_{\ell}$ times a squarefree monomial of degree $k+1$, and so $x_{j}(u/x_{i})\in\textup{HS}_{k}(I)$. If there is only one $j$ such that $\deg_{x_{j}}(u)<\deg_{x_{j}}(v)$, then $j=h$. We claim that $x_{h}$ does not divide $u$, then $x_{h}(u/x_{i})$ is equal to $x_{\ell}$ times a squarefree monomial of degree $k+1$, and so $x_{h}(u/x_{i})\in\textup{HS}_{k}(I)$, as wanted. Writing $v=x_{h}^{2}x_{j_{1}}\cdots x_{j_{k}}$, with $j_{p}\in[n]\setminus\\{h\\}$, $p=1,\dots,k$, then $\deg_{x_{j_{p}}}(v)=1\leq\deg_{x_{j_{p}}}(u)$, for all $p=1,\dots,k$. Then $x_{j_{1}}\cdots x_{j_{k}}$ divides $u/(x_{i}x_{\ell})$ because $\deg_{x_{\ell}}(u)>1\geq\deg_{x_{\ell}}(v)$ and $\deg_{x_{i}}(u)=1>\deg_{x_{i}}(v)$. This implies that $u=x_{i}x_{\ell}\cdot x_{j_{1}}\cdots x_{j_{k}}$. From this presentation it follows that $x_{h}$ does non divide $u$, because $i,\ell\neq h$ and $j_{p}\neq h$ for $p=1,\dots,k$, as well. The cases above show that the exchange property holds for all monomials of $G(\textup{HS}_{k}(I))$. Hence $\textup{HS}_{k}(I)$ is polymatroidal and the proof is complete. ## References * [1] S. Bandari, R. Rahmati-Asghar, On the Polymatroidal Property of Monomial Ideals with a View Towards Orderings of Minimal Generators, Bulletin of the Iranian Mathematical Society, 45 (2019), 657–666. * [2] S. Bayati, Multigraded shifts of matroidal ideals, Arch. Math., (Basel) 111 (2018), no. 3, 239–246. * [3] S. Bayati, I. Jahani, N. Taghipour, Linear quotients and multigraded shifts of Borel ideals, Bull. Aust. Math. Soc. 100 (2019), no. 1, 48–57. * [4] M. Crupi, A. Ficarra, Very well–covered graphs by Betti splittings, preprint 2022. * [5] G.A. Dirac, On rigid circuit graphs, Abh. Math. Sem. Univ. Hamburg, 38 (1961), 71–76. * [6] A. Ficarra. Homological shifts of polymatroidal ideals, available at arXiv preprint https://arxiv.org/abs/2205.04163[math.AC] (2022). * [7] A. Ficarra. HomologicalShiftsIdeals, Macaulay2 package (2022). * [8] C. A. Francisco, H. T. Ha, A. Van Tuyl, Splittings of monomial ideals, Proc. Amer. Math. Soc., 137 (10) (2009), 3271-3282. * [9] R. Fröberg, On Stanley-Reisner rings, Topics in algebra, Part 2 (Warsaw, 1988), 57–70, Banach Center Publ., 26, Part 2, PWN, Warsaw, 1990. * [10] D. R. Grayson, M. E. Stillman. Macaulay2, a software system for research in algebraic geometry. Available at http://www.math.uiuc.edu/Macaulay2. * [11] J. Herzog, T. Hibi, Cohen-Macaulay polymatroidal ideals, European J. Combin., 27 (2006), no. 4, 513–517. * [12] J. Herzog, T. Hibi. _Monomial ideals_ , Graduate texts in Mathematics 260, Springer–Verlag, 2011. * [13] J. Herzog, S. Moradi, M. Rahimbeigi, G. Zhu, Homological shift ideals. Collect. Math. 72 (2021), 157–174. * [14] J. Herzog, S. Moradi, M. Rahimbeigi, G. Zhu, Some homological properties of borel type ideals, 2021, to appear in Comm. Algebra. * [15] J. Herzog, Y. Takayama, Resolutions by mapping cones, in: The Roos Festschrift volume Nr.2(2), Homology, Homotopy and Applications 4, (2002), 277–294. * [16] A.S. Jahan, X. Zheng, Ideals with linear quotients, J. Combin. Theory Ser. A 117 (2010), 104–110. MR2557882 (2011d:13030) * [17] P.J. Looges and S. Olariu, Optimal greedy algorithms for indifference graphs, Comput. Math. Appl. 25, (1993), 15–25.
# Quantum reservoir computing in finite dimensions Rodrigo Martínez-Peña<EMAIL_ADDRESS>Instituto de Física Interdisciplinar y Sistemas Complejos (IFISC, UIB-CSIC), Campus Universitat de les Illes Balears, 07122 Palma de Mallorca, Spain Juan-Pablo Ortega juan- <EMAIL_ADDRESS>Division of Mathematical Sciences, Nanyang Technological University, 21 Nanyang Link, Singapore 637371 ###### Abstract Most existing results in the analysis of quantum reservoir computing (QRC) systems with classical inputs have been obtained using the density matrix formalism. This paper shows that alternative representations can provide better insights when dealing with design and assessment questions. More explicitly, system isomorphisms are established that unify the density matrix approach to QRC with the representation in the space of observables using Bloch vectors associated with Gell-Mann bases. It is shown that these vector representations yield state-affine systems (SAS) previously introduced in the classical reservoir computing literature and for which numerous theoretical results have been established. This connection is used to show that various statements in relation to the fading memory (FMP) and the echo state (ESP) properties are independent of the representation, and also to shed some light on fundamental questions in QRC theory in finite dimensions. In particular, a necessary and sufficient condition for the ESP and FMP to hold is formulated using standard hypotheses, and contractive quantum channels that have exclusively trivial semi-infinite solutions are characterized in terms of the existence of input-independent fixed points. Suggested keywords ## I Introduction The development of noisy intermediate-scale quantum (NISQ) devices is attracting a great deal of attention from the quantum community. Recent advancements in fields such as quantum computation [1], quantum simulation [2], and quantum communications [3] are just examples of the prosperous future that awaits these technologies. Nevertheless, NISQ devices are already demonstrating in the meantime that they can be very useful for diverse research fields like physics, chemistry, and optimization [4], even providing quantum advantage [1, 5, 6]. Machine learning (ML) is another example of thriving synergies with NISQ technologies. In this context, quantum machine learning (QML) aims to exploit the specific features of quantum mechanics to obtain an advantage over its classical counterparts when dealing with machine learning tasks, both with classical and quantum data [7]. There is already positive evidence in this direction, both from a theoretical point of view in the fault-tolerant picture [8] and in experiments [9]. The flexibility and range of action of ML and QML techniques is typically specified by the so-called universality approximation theorems. A universal approximation property takes place when a proposed restricted family of functions can approximate any function in a much larger class with arbitrary precision. There are many results of this type that are part of classical analysis dealing with, for instance, polynomials and Fourier series, and others that were added in the early days of ML like, for example, feed-forward neural networks [10, 11, 12]. Further results of this type have been proved for various ML paradigms like recurrent neural networks [13], support vector machines [14], extreme learning machines [15, 16], or kernel methods [17, 18, 19]. Universality results have also been obtained in the QML context, such as for one qubit algorithms [20, 21], and general quantum circuits [22, 23]. A framework in which we are particularly interested is quantum reservoir computing (QRC). As in classical reservoir computing (RC) [24, 25, 26], QRC harnesses the rich dynamics of (quantum) dynamical systems to solve tasks where memory and prediction capabilities are required. Examples of application of these techniques are found in the prediction of chaotic time-series [27, 28, 29] and complex spatiotemporal dynamics [30, 31, 32, 33]. Since the first work on QRC [34] many have followed (see [35, 36] for reviews). This includes both experimental implementations [37, 38, 39, 40, 41] as well as theoretical contributions on the universal approximation question [42, 37, 43]. The latter was inspired by the works in the classical framework [44, 45, 46, 47], where the discrete-time setting of RC theory fits well within the quantum dynamical map description. The universal approximation property in the RC context brings to the table some conditions that dynamical systems should meet to ensure it. The most prominent ones are the echo state property (ESP) [24] and the fading memory property (FMP) [48]. A system has the FMP if inputs that are close in the recent past produce outputs that are also close, independently of what happened in the distant past. The ESP guarantees that a well-defined input/output map can be associated with our system and amounts to an existence and uniqueness property with respect to the input sequence that is fed into the system. The ESP and FMP are standard requirements in many stochastic and deterministic learning paradigms since they mathematically encode the asymptotic decorrelation (and even independence) between physical states and initial conditions that most physical systems exhibit as time goes by. There are many physical mechanisms that lead to the declining relevance of a given initial condition in a system state as the temporal distance between them increases. For instance, two pervasive phenomena in applications in this direction are chaos (high sensitivity to initial conditions) and dissipation. Fading memory is an important modeling feature when using physical systems because a reservoir system can only store a finite amount of information in its trainable parameters [49]; this implies that information must be erased as time goes by in order to store new input information. Under very mild mathematical conditions, ESP and FMP are equivalent to the input-forgetting property [50] which mathematically encodes the information removal process that is needed to learn the newly fed one. The ESP property has been studied in the context of QRC in different works. Nurdin and Chen [42] provided sufficient conditions for the ESP (and FMP) to hold in terms of the contractivity of the quantum map acting on a restricted domain [42, 37, 51], while Tran and Nakajima also define the ESP using contractivity of the quantum map but without considering restrictions. They also numerically connect the ESP with the spectra of the quantum maps [32, 52]. Both approaches describe the ESP in the density matrix language, that is, the dynamical equations are defined in the space of quantum states. However, this description might be limited somehow, because, as we will see later on, more information can be extracted if we choose a description in terms of observables. This paper aims to fill some gaps in the description of QRC with finite dimensional systems and classical inputs. More precisely, we first unify the quantum ESP defined in previous works [42, 37, 51, 32, 52] using the norm of quantum maps, with the classical notion of ESP using observables, and show that they are equivalent for finite-dimensional quantum systems. We extend in passing these results to the FMP. More specifically, we establish explicit system isomorphisms between the description of QRC systems using the space of density matrices as state space and the one that uses the space of Bloch vectors associated to Gell-Mann bases. That choice of basis, which is customary in the study of quantum systems, happens to yield non-homogeneous affine state dynamics (the corresponding systems are called state-affine (SAS)) of the type introduced in [45]. The connection between QRC and SAS systems has important consequences. Our results in that context are contained in Proposition 3 and in Theorems 5 and 9. The proposition presents a necessary and sufficient condition for the ESP and FMP to hold in different representations with just a few requirements, such as the compactness of the input space. We emphasize that this hypothesis is satisfied in most RC tasks and when dealing with implementations of RC systems with dedicated hardware since the experimental ranges of the physical systems involved are always finite [26]. The theorems exhibit common situations that should be avoided in the design of quantum channels so that fully operational QRC systems are obtained. We shall work in an idealized framework in which observables are obtained after an infinite number of measurements, with no statistical error. Even in such an idealized setting, we expect that our results can contribute to the general understanding of QRC experiments in finite dimensional systems, as they have already been implemented in [37, 38, 40, 41]. The structure of the paper is as follows. Section II introduces the general framework and the definitions that will be needed along the paper. Definitions of the spaces of operators and quantum maps are included in Section II.1, while all the RC ingredients will be presented in Section II.2, together with some preliminary results. The main results are contained in Section III. Section IV includes a brief discussion on some of the consequences of Theorems 5 and 9, and Section V concludes the paper. ## II Definitions ### II.1 Quantum definitions We start by defining the space of quantum systems. See, for example, [53, 54] for further details. Consider a complex Hilbert space $\mathcal{H}$. The set of all bounded operators $\mathcal{B}(\mathcal{H})$ that act on $\mathcal{H}$ is a complex vector space under point-wise addition and scalar multiplication, and it forms an algebra under composition. If we add the involution $A\rightarrow A^{\dagger}$ given by the adjoint operation, then $\mathcal{B}(\mathcal{H})$ is also a $C^{*}$-algebra with respect to the operator norm ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{op}$ defined by ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|A\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{op}:=\sup_{||\psi||=1}\\{||A\psi||,\ \psi\in\mathcal{H}\\},$ (1) where $A\in\mathcal{B}(\mathcal{H})$. Along this manuscript, we will denote all the induced operator and matrix norms with the symbol ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$. The predual of $\mathcal{B}(\mathcal{H})$ is the Banach space $\mathcal{T}(\mathcal{H})$ of all trace-class operators on $\mathcal{H}$ that have finite trace norm $||A||_{1}:=\text{tr}\sqrt{AA^{\dagger}}$. The trace norm is a particular case (with $p=1$) of the Schatten norms defined by $||A||_{p}:=\left(\text{tr}\left(\left(\sqrt{AA^{\dagger}}\right)^{p}\right)\right)^{1/p}$. The space of quantum density matrices $\mathcal{S}(\mathcal{H})$ is a compact convex subset (see Section II.2 and the argument above (15)) of the normed vector space $\mathcal{T}(\mathcal{H})$ defined by $\mathcal{S}(\mathcal{H})=\\{\rho\in\mathcal{T}(\mathcal{H})\ |\ \rho^{\dagger}=\rho,\ \rho\geq 0,\ \text{tr}(\rho)=1\\}.$ (2) As $\mathcal{S}(\mathcal{H})$ is a closed subset of the Banach space $\mathcal{T}(\mathcal{H})$, it is then a complete metric space when using the distance induced by $\left\|\cdot\right\|_{1}$. All along this paper, we restrict ourselves to finite-dimensional Hilbert spaces, for which the spaces of bounded and trace class operators coincide, and we shall use the symbol $\mathcal{B}(\mathcal{H})$ to refer to both of them. Moreover, in that case, $\mathcal{S}(\mathcal{H})$ is a complete metric space with respect to any norm since all the norms are equivalent. We shall reserve the symbol $d\in{\mathbb{N}}$ for the dimension of ${\mathcal{H}}$. We now introduce the notion of quantum channel. All definitions and further properties of these maps can be found in [55, 56, 57, 58] and references therein. A quantum channel is a linear map $T:\mathcal{B}(\mathcal{H})\rightarrow\mathcal{B}(\mathcal{H})$ that is completely positive and trace preserving (CPTP). We recall here that a trace preserving map $T$ is the one that satisfies that $\text{tr}(T(A))=\text{tr}(A)$ for any $A\in\mathcal{B}({\mathcal{H}})$. Moreover, we say that $T$ is positive when it maps positive semi-definite operators to positive semi-definite operators. Finally, completely positive maps $T$ are those positive maps which, when extended to a larger space using the tensor map $T\otimes I_{k}$, with $I_{k}$ the identity map in dimension $k$, they also yield a positive map for any $k\in\mathbb{N}$. A linear map is CPTP if and only if it is possible to find a Kraus decomposition associated to a set of operators $\\{K_{i}\\}_{i\in X}$ such that for all $A\in\mathcal{B}(\mathcal{H})$: $T(A)=\sum_{i\in X}K_{i}AK^{\dagger}_{i},$ (3) where $\sum_{i\in X}K^{\dagger}_{i}K_{i}=I$ and $X$ is an index set of cardinality at most $d^{2}$, with $d$ the dimension of ${\mathcal{H}}$. CPTP maps obviously leave $\mathcal{S}(\mathcal{H})$ invariant and hence induce a restricted map $T:\mathcal{S}(\mathcal{H})\longrightarrow\mathcal{S}(\mathcal{H})$ that we shall denote with the same symbol and use interchangeably. Note that, unlike $\mathcal{B}({\mathcal{H}})$, the set $\mathcal{S}({\mathcal{H}})$ is not a vector space, and hence it is only when use the map $T:\mathcal{B}(\mathcal{H})\longrightarrow\mathcal{B}(\mathcal{H})$ that we can talk about matrix expressions and eigenvalues for the operator $T$. It can be shown that any CPTP map $T:\mathcal{S}(\mathcal{H})\longrightarrow\mathcal{S}(\mathcal{H})$ is non- expansive in the trace norm, which means that after applying $T$ to two input states $\rho_{1},\rho_{2}\in\mathcal{S}(\mathcal{H})$, the distance between these density matrices is either contracted or remains equal: $||T(\rho_{1})-T(\rho_{2})||_{1}\leq||\rho_{1}-\rho_{2}||_{1}.$ (4) We recall that the eigenvalues of a CPTP map $T:\mathcal{B}(\mathcal{H})\rightarrow\mathcal{B}(\mathcal{H})$ are the complex numbers $\lambda$ that make the map $T-\lambda\,\text{id}$ non- invertible. We will denote this set by $\text{spec}(T)$. The spectral radius $\rho(T):=\max\\{|\lambda|\ \mid\lambda\in\text{spec}(T)\\}$ of any CPTP map is one as a consequence of (4) and of the fact that $1$ is always an eigenvalue. Additionally, the set $\text{spec}(T)$ is invariant under complex conjugation (see [57]). On the other hand, since the finite-dimensional vector space $\mathcal{B}(\mathcal{H})$ is isomorphic to $\mathbb{C}^{d^{2}}$, we can represent $T:\mathcal{B}(\mathcal{H})\rightarrow\mathcal{B}(\mathcal{H})$ as a $d^{2}\times d^{2}$ matrix, which we write as $\widehat{T}$. This matrix representation can be obtained by fixing an orthonormal basis in $\mathcal{B}(\mathcal{H})$ for the Hilbert-Schmidt inner product, that is, $\\{B_{i}\in\mathcal{B}(\mathcal{H})\\}$ with $\text{tr}(B_{i}^{\dagger}B_{j})=\delta_{ij}$, and by setting $\widehat{T}_{ij}=\text{tr}(B_{i}^{\dagger}T(B_{j}))$. The spectrum of the matrix $\widehat{T}$ is given by the roots of the characteristic polynomial $\det(\widehat{T}-\lambda I)=0$ and, since we are working with finite dimensions, it coincides with $\text{spec}(T)$, as well as with the set of complex values $\lambda$ that satisfy $T(X)=\lambda X$, for some non-trivial eigenvector $X\in\mathcal{B}(\mathcal{H})$ [58]. We are particularly interested in the eigenvectors corresponding to the eigenvalue $\lambda=1$ and that we call fixed points, that is, they are elements $A\in\mathcal{B}(\mathcal{H})$ such that $T(A)=A$. The set of fixed points of a CPTP map is always non-empty in the space of density matrices. This is a consequence of Schauder’s Fixed-Point Theorem together with the continuity of $T$ and the compactness of $\mathcal{S}(\mathcal{H})$. We will work most of the time with quantum channels with single fixed points $\rho^{*}\in\mathcal{S}(\mathcal{H})$ in the space of density matrices. Such maps are called ergodic. Ergodicity of quantum channels is equivalent to the eigenspace of $T$ associated with the eigenvalue $\lambda=1$ in $\mathcal{B}(\mathcal{H})$ having dimension one, and to being made out of the complex multiples of $\rho^{*}\in\mathcal{S}(\mathcal{H})$. In that case, we obviously have (see Corollary 2 in [57]) that $T(A)=A$ with $A\in\mathcal{B}(\mathcal{H})$ if and only if $A=\text{tr}(A)\rho^{*}$. It is worth mentioning that the rest of the eigenvectors of an ergodic CPTP map must be traceless. Indeed, given an eigenvalue $\lambda\neq 1$ of the CPTP map $T$ and $A$ a corresponding eigenvector, we find that $\text{tr}(A)=\text{tr}(T(A))=\lambda\text{tr}(A)$, which implies that $(\lambda-1)\text{tr}(A)=0$, and hence that $\text{tr}(A)=0$. Therefore, the spectral set of a CPTP map can be decomposed as: $\text{spec}(T)=\\{1\\}\cup\text{spec}(T|_{\mathcal{B}_{0}(\mathcal{H})}),$ (5) where $\mathcal{B}_{0}(\mathcal{H})\subset\mathcal{B}(\mathcal{H})$ is defined as the vector subspace of traceless operators, where the restriction $T|_{\mathcal{B}_{0}(\mathcal{H})}$ and its corresponding matrix representation are obviously well defined. Repeated applications of an ergodic CPTP map do not necessarily converge to a fixed point (see [57] for an example). If convergence to a fixed point takes place, we say that the CPTP map is mixing. More specifically, a CPTP map is mixing if and only if its repeated applications converge in the trace norm, that is, $\lim_{n\rightarrow\infty}||T^{n}(\rho)-\rho^{*}||_{1}=0,\quad\forall\rho\in\mathcal{S}(\mathcal{H}).$ (6) Equation (6) implies that the sequence $\\{T^{n}(\rho)\\}$ converges to $\rho^{*}$ with respect to the trace norm, but in infinite-dimensional situations this does not necessarily imply that convergence takes place with respect to other norms (see Definition 5.4.1 in [59]). However, in the finite- dimensional case mixing becomes a topological property where $\lim_{n\rightarrow\infty}T^{n}(\rho)=\rho^{*}$ for any norm and for any $\rho\in\mathcal{S}(\mathcal{H})$ [56]. Although all mixing maps are ergodic, the converse is not true in general. However, for continuous-time Markovian evolution, both are equivalent and such maps are called relaxing [57]. Another important consequence of mixing condition is the following: a CPTP map is mixing if and only if the fixed point $\rho^{*}$ is the only eigenvector with eigenvalue $|\lambda|=1$. Then, it is straightforward to show that $\max\\{|\lambda|\mid\lambda\in\text{spec}(T|_{\mathcal{B}_{0}(\mathcal{H})})\\}<1.$ (7) We say that a CPTP map is called primitive when it is mixing and its unique fixed point $\rho^{*}>0$ has full rank. The smallest natural number $n$ for which $T^{n}$ sends positive semidefinite matrices to positive definite matrices is called the index of primitivity of $T$ and is denoted by $\omega(T)$. With that notation, we say that $T^{\omega(T)}$ is a strictly positive map. Bounds for this number are given by the quantum version of the Wietland inequality [60, 61]. Finally, we say that a quantum channel is strictly contractive when $||T(\rho_{1})-T(\rho_{2})||_{1}\leq r||\rho_{1}-\rho_{2}||_{1}$ (8) for all $\rho_{1},\rho_{2}\in\mathcal{S}(\mathcal{H})$, where $0\leq r<1$. As it is customary in the quantum channels literature, we reserve the term strictly contractive for the trace norm, unless a different norm is explicitly specified. It can be shown that strictly contractive channels with a full-rank fixed point must be primitive. Indeed, the contractivity condition, together with Banach’s Fixed Point Theorem (using that the convex closed subset of density matrices with the trace norm is a complete metric space) guarantees that the channel is mixing. A mixing channel with a strictly positive fixed point is a primitive channel. The converse holds when the primitive map becomes strictly positive, that is, it sends positive semidefinite matrices to positive definite ones: given a primitive channel $T$, $T^{\omega(T)}$ is strictly contractive (see Theorem VI.3 in [61]). Contractive maps can be also constructed by composing a strictly contractive channel with a general CPTP map. An important conclusion that can be drawn from all these considerations is that strictly contractive channels, which will be relevant along this work, can be constructed by either using a map like $T^{\omega(T)}$, where $T$ is a primitive channel, or by composing a strictly contractive channel with any other CPTP map, as it has been done in some examples in QRC [40, 39]. Examples of mixing/relaxing CPTP maps with full-rank single fixed points can be found in the Lindblad-Gorini-Kossakowski-Sudarshan equation (see [62] for a summary of the necessary and sufficient conditions). As these maps belong to the quantum Markov semigroup, iterative applications of the channels yield $T_{\Delta\tau}^{\omega(T)}=T_{\omega(T)\Delta\tau}$, where $\Delta\tau$ represents the time of action of the map $T_{\Delta\tau}$. Therefore, taking $\Delta\tau^{\prime}\geq\omega(T)\Delta\tau$ we obtain a strictly contractive map. ### II.2 RC definitions We now define quatum reservoir computing (QRC) systems in the density matrix formalism. Classical RC definitions and mathematical details can be found in, for instance, [44]. QRC maps are determined by two equations, namely, the state-space and the readout or observation equations. The state equation is given by a family of CPTP maps $T:\mathcal{B}(\mathcal{H})\times\mathbb{R}^{n}\rightarrow\mathcal{B}(\mathcal{H})$, with $n\in\mathbb{N}$ being the number of input features, which are taken to be real values (classical inputs). The maps $T$ and $h$ will be, most of the time, tacitly assumed to be continuous. The output is obtained from the readout map $h:\mathcal{B}(\mathcal{H})\rightarrow\mathbb{R}^{m}$, with $m\in\mathbb{N}$, which maps operators in $\mathcal{B}(\mathcal{H})$ to the Euclidean space $\mathbb{R}^{m}$. Inputs are typically bi-infinite discrete- time sequences of the form ${\bf z}=(\dots,{\bf z}_{-1},{\bf z}_{0},{\bf z}_{1},\dots)\in(\mathbb{R}^{n})^{\mathbb{Z}}$, and outputs $\textbf{y}\in(\mathbb{R}^{m})^{\mathbb{Z}}$ have the same structure. A QRC system is hence determined by the state-space transformations: $\begin{cases}&A_{t}=T(A_{t-1},{\bf z}_{t}),\\\ &\textbf{y}_{t}=h(A_{t}),\end{cases}$ (9) where $t\in\mathbb{Z}$ denotes the time index. Analogously, one can define the same setting for semi-infinite discrete-time sequences: $(\mathbb{R}^{n})^{\mathbb{Z}_{-}}=\\{{\bf z}=(\dots,{\bf z}_{-1},{\bf z}_{0})\ |\ {\bf z}_{i}\in\mathbb{R}^{n},i\in\mathbb{Z}_{-}\\}$ for left- infinite sequences and $(\mathbb{R}^{n})^{\mathbb{Z}_{+}}=\\{{\bf z}=({\bf z}_{0},{\bf z}_{1},\dots)\ |\ {\bf z}_{i}\in\mathbb{R}^{n},i\in\mathbb{Z}_{+}\\}$ for right-infinite sequences. Similar definitions apply to $(D_{n})^{\mathbb{Z}}$, $(D_{n})^{\mathbb{Z}_{-}}$, and $(D_{n})^{\mathbb{Z}_{+}}$ with elements in the subset $D_{n}\subset\mathbb{R}^{n}$. We can also construct sequence spaces $(\mathcal{B}(\mathcal{H}))^{\mathbb{Z}}$ for the space of bounded (trace- class) operators: $(\mathcal{B}(\mathcal{H}))^{\mathbb{Z}}=\\{\textbf{A}=(\dots,A_{-1},A_{0},A_{1},\dots)\\\ \mid A_{i}\in\mathcal{B}(\mathcal{H}),\ i\in\mathbb{Z}\\}.$ Analogous definitions for $(\mathcal{B}(\mathcal{H}))^{\mathbb{Z}_{-}}$,$(\mathcal{B}(\mathcal{H}))^{\mathbb{Z}_{+}}$, $(\mathcal{S}(\mathcal{H}))^{\mathbb{Z}}$, $(\mathcal{S}(\mathcal{H}))^{\mathbb{Z}_{-}}$, and $(\mathcal{S}(\mathcal{H}))^{\mathbb{Z}_{+}}$ follow immediately. A natural way to construct CPTP state-space transformations is to insert the input dependence using the Kraus decomposition that we introduced in (3), that is, $T(A,{\bf z})=\sum_{i\in X}K_{i}({\bf z})AK^{\dagger}_{i}({\bf z}),$ (10) and for any input ${\bf z}\in\mathbb{R}^{n}$, the matrices $\\{K_{i}({\bf z})\\}_{i\in X}$ satisfy that $\sum_{i\in X}K^{\dagger}_{i}({\bf z})K_{i}({\bf z})=I$. The by-design CPTP character of the map $T:\mathcal{B}(\mathcal{H})\times\mathbb{R}^{n}\rightarrow\mathcal{B}(\mathcal{H})$ in (10) implies that it naturally restricts to a state equation $T:\mathcal{S}(\mathcal{H})\times\mathbb{R}^{n}\rightarrow\mathcal{S}(\mathcal{H})$ with density matrices as state space, that we shall use interchangeably in the sequel and denote using the same symbol. The Echo State Property (ESP). Consider the QRC system defined in (9) or its analog for the subsets $\mathcal{S}(\mathcal{H})\subset\mathcal{B}(\mathcal{H})$ and $D_{n}\subset\mathbb{R}^{n}$, that is, $T:\mathcal{S}(\mathcal{H})\times D_{n}\rightarrow\mathcal{S}(\mathcal{H})$. Given an input sequence ${\bf z}\in(D_{n})^{\mathbb{Z}}$, we say that $\bm{\rho}\in(\mathcal{S}(\mathcal{H}))^{\mathbb{Z}}$ is a solution of (9) for the input ${\bf z}$ if the components of the sequences ${\bf z}$ and $\bm{\rho}$ satisfy the first relation in (9) for any $t\in\mathbb{Z}$. We say that the QRC system has the echo state property (ESP) when it has a unique solution for each input ${\bf z}\in(D_{n})^{\mathbb{Z}}$. More explicitly, for each ${\bf z}\in(D_{n})^{\mathbb{Z}}$, there exists a unique sequence $\bm{\rho}\in(\mathcal{S}(\mathcal{H}))^{\mathbb{Z}}$ such that $\rho_{t}=T(\rho_{t-1},{\bf z}_{t}),\ \text{for all}\ t\in\mathbb{Z}.$ (11) Filters and functionals. Let $\mathcal{S}(\mathcal{H})\subset\mathcal{B}(\mathcal{H})$ be the space of density matrices and let $D_{n}\subset\mathbb{R}^{n}$ be a subset in the input space. A map of the type $U:(D_{n})^{\mathbb{Z}}\rightarrow(\mathcal{S}(\mathcal{H}))^{\mathbb{Z}}$ is called a filter associated to the QRC system (9) when it satisfies that $U({\bf z})_{t}=T\left(U({\bf z})_{t-1},{\bf z}_{t}\right),\ \mbox{for all ${\bf z}\in(D_{n})^{\mathbb{Z}}$ and $t\in\mathbb{Z}$.}$ Filters induce what we call functionals $H:(D_{n})^{\mathbb{Z}}\rightarrow\mathcal{S}(\mathcal{H})$ via the relation $H({\bf z})=U({\bf z})_{0}$. It is clear that a uniquely determined filter can be associated with a QRC system that satisfies the ESP. The filter maps, in that case, any input sequence to the unique solution of the QRC system associated with it. A filter is called causal if it only produces outputs that depend on present and past inputs. More formally, causality means that for any two inputs ${\bf z},\textbf{v}\in(D_{n})^{\mathbb{Z}}$ that satisfy ${\bf z}_{\tau}=\textbf{v}_{\tau}$ for any $\tau\leq t$, for a given $t\in\mathbb{Z}$, we have $U({\bf z})_{t}=U(\textbf{v})_{t}$. The filter $U$ is called time-invariant if there is no explicit time dependence on the system that determines it, that is, it commutes with the time delay operator defined as $\mathcal{T}_{\tau}({\bf z})_{t}:={\bf z}_{t-\tau}$. Filters associated to QRC systems of the type (9) are always causal and time-invariant (Proposition 2.1 in [44]). As noted in previous works [48, 44, 45], there is a bijection between causal and time-invariant filters and functionals on $(D_{n})^{\mathbb{Z}_{-}}$. Then, we can restrict our work to causal and time- invariant filters with target and domain in spaces of left-infinite sequences. The Fading Memory Property (FMP). Infinite product spaces can be endowed with Banach space structures associated to the supremum norm and weighted norms. The supremum norm for input sequences is defined as $||{\bf z}||_{\infty}:=\sup_{t\in\mathbb{Z}}\\{||{\bf z}_{t}||\\}$, and for operator sequences as $||\textbf{A}||_{\infty}:=\sup_{t\in\mathbb{Z_{-}}}\\{||A_{t}||\\}$, where $||\cdot||$ represents a given vector and matrix norm, respectively. The symbols $l^{\infty}(\mathbb{R}^{n})$, $l_{\pm}^{\infty}(\mathbb{R}^{n})$, $l^{\infty}(\mathcal{B}(\mathcal{H}))$ and $l_{\pm}^{\infty}(\mathcal{B}(\mathcal{H}))$ denote the Banach spaces formed by the elements in the corresponding infinite product spaces with a finite supremum norm. We now define the weighted norm. Let $w:\mathbb{N}\rightarrow(0,1]$ be a decreasing sequence with zero limit and $w_{0}=1$. The weighted norm $||\cdot||_{w}$ on $(\mathbb{R})^{\mathbb{Z_{-}}}$ is defined as $||{\bf z}||_{w}:=\sup_{t\in\mathbb{Z}_{-}}\\{w_{-t}||{\bf z}_{t}||\\},$ (12) and the space $l^{w}_{-}(\mathbb{R}^{n})=\\{{\bf z}\in(\mathbb{R}^{n})^{\mathbb{Z_{-}}}|\ ||{\bf z}||_{w}<\infty\\},$ (13) with weighted norm $||\cdot||_{w}$ forms a Banach space (see Appendix A.2 in [44]). In the same vein, we can define $\begin{split}&||\textbf{A}||_{w}:=\sup_{t\in\mathbb{Z}_{-}}\\{w_{-t}||A||\\},\\\ &l^{w}_{-}(\mathcal{B}(\mathcal{H}))=\\{\textbf{A}\in(\mathcal{B}(\mathcal{H}))^{\mathbb{Z_{-}}}|\ ||\textbf{A}||_{w}<\infty\\}.\end{split}$ (14) It can be shown that $l^{w}_{-}(\mathcal{B}(\mathcal{H}))$ is a Banach space as well. We now turn to the space of density matrices $\mathcal{S}(\mathcal{H})$ and recall that since for positive semidefinite matrices the trace operator is submultiplicative (see Exercise 7.2.P26 in [59]) we can conclude that $\operatorname{tr}(\rho^{2})\leq 1$ for all $\rho\in\mathcal{S}(\mathcal{H})$. This observation implies that the elements in $\mathcal{S}(\mathcal{H})$ have Frobenius norms bounded by one. Since we are in finite dimensions, this statement holds for any other matrix norm (eventually with a bounding constant different from one) and allows us to conclude that $(\mathcal{S}(\mathcal{H}))^{\mathbb{Z}_{-}}\subset l^{\infty}_{-}(\mathcal{B}(\mathcal{H}))\subset l^{w}_{-}(\mathcal{B}(\mathcal{H})),$ (15) for any weighting sequence $w$. The previous boundedness consideration, together with the closedness of $\mathcal{S}(\mathcal{H})$ in $\mathcal{B}(\mathcal{H})$ implies that $\mathcal{S}(\mathcal{H})$ is necessary compact. An important consequence of this fact is that the relative topology induced by the $l^{w}_{-}(\mathcal{B}(\mathcal{H}))$ on $(\mathcal{S}(\mathcal{H}))^{\mathbb{Z}_{-}}$ coincides with the product topology (see Corollary 2.7 in [44]). Take now a subset $D_{n}\subset\mathbb{R}^{n}$ such that $(D_{n})^{\mathbb{Z}_{-}}\subset l^{w}_{-}(\mathbb{R}^{n})$ and consider a QRC system $T:\mathcal{S}(\mathcal{H})\times D_{n}\rightarrow\mathcal{S}(\mathcal{H})$ that has the ESP. We say that $T$ has the fading memory property (FMP) when the corresponding functional $H:(D_{n})^{\mathbb{Z}_{-}}\rightarrow\mathcal{S}(\mathcal{H})$ is a continuous map between the metric spaces $((D_{n})^{\mathbb{Z}_{-}},||\cdot||_{w})$ and $((\mathcal{S}(\mathcal{H}))^{\mathbb{Z}_{-}},||\cdot||_{w})$, for some weighting sequence $w$. If $D_{n}$ is compact, once $H$ is continuous for a given weighting sequence $w$, then it is continuous for all weighting sequences (see [44, Theorem 2.6]). The compactness of $\mathcal{S}(\mathcal{H})$ implies that we can apply with straightforward modifications Theorem 3.1 in [44] to prove the following statement. ###### Proposition 1. Let $D_{n}\subset\mathbb{R}^{n}$ be a compact subset of $\mathbb{R}^{n}$. Let $T:\mathcal{S}(\mathcal{H})\times D_{n}\rightarrow\mathcal{S}(\mathcal{H})$ be a continuous QRC system such that the CPTP maps $T(\cdot,{\bf z}):\mathcal{S}(\mathcal{H})\rightarrow\mathcal{S}(\mathcal{H})$ are strictly contractive for all ${\bf z}\in D_{n}$ as in (8) with a common contraction constant $0\leq r<1$ associated to some norm in $\mathcal{B}(\mathcal{H})$ (not necessarily $\left\|\cdot\right\|_{1}$). Then, the QRC system induced by $T$ has the ESP and the FMP. ## III Results Finding the expression for the QRC filter of a system that has the ESP is not straightforward in the language of density matrices since the dependence on the initial condition has to be addressed. More explicitly, given a general CPTP map expressed using its Kraus decomposition, the relation (10) can be iterated $n$-time steps into the past in order to obtain a quantum state $\rho_{t}^{n}(\rho^{0}_{t-n})\in\mathcal{S}({\mathcal{H}})$ at time $t$ out of an initial condition $\rho^{0}_{t-n}\in\mathcal{S}({\mathcal{H}})$ specified at time $t-n$ via the formula: $\footnotesize\begin{split}&\\!\\!\\!\\!\rho_{t}^{n}(\rho^{0}_{t-n})=\\\ &\sum_{i_{0},\ldots,i_{n-1}\in X}\left(\overleftarrow{\prod}^{n-1}_{l=0}K_{i_{l}}({\bf z}_{t-l})\right)\rho^{0}_{t-n}\left(\overrightarrow{\prod}^{n-1}_{l=0}K^{\dagger}_{i_{l}}({\bf z}_{t-l})\right),\end{split}$ (16) where $\overleftarrow{\prod}^{n-1}_{l=0}K_{i_{l}}({\bf z}_{t-l})=K_{i_{0}}({\bf z}_{t})\cdots K_{i_{n-1}}({\bf z}_{t-n-1})$ and $\overrightarrow{\prod}^{n-1}_{l=0}K^{\dagger}_{i_{l}}({\bf z}_{t-l})=K^{\dagger}_{i_{n-1}}({\bf z}_{t-n-1})\cdots K^{\dagger}_{i_{0}}({\bf z}_{t})$. If the QRC system has the ESP and hence a filter $U:(D_{n})^{\mathbb{Z}}\longrightarrow\left(\mathcal{S}({\mathcal{H}})\right)^{\mathbb{Z}}$ can be associated to it, it necessarily has to satisfy $U({\bf z})_{t}=\lim_{n\rightarrow\infty}\rho_{t}^{n}(\rho^{0}_{t-n}),$ (17) and this value has to be independent of the initial conditions $\rho^{0}_{t-n}$. This fact has been explicitly shown in [42] (and in Chapter 2 of [51] in more detail) in the case of strictly contractive CPTP maps with respect to the operator norm ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{p}$ associated to Schatten norms. We recall that given $T:\mathcal{B}({\mathcal{H}})\rightarrow\mathcal{B}({\mathcal{H}})$, we define ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|T\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{p}:=\sup_{||A||_{p}=1}\\{||T(A)||_{p}\mid A\in\mathcal{B}({\mathcal{H}})\\},$ (18) for some $p\in[1,\infty)$ and $\left\|\cdot\right\|_{p}$ the $p$-Schatten norm. Using this notation, the contractivity condition on the CPTP map $T$ is stated by requiring that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|T|_{\mathcal{B}_{0}(\mathcal{H})}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{p}<1$, for some $p\in[1,\infty)$. It is obvious that this condition ensures that the hypotheses of Proposition 1 are satisfied, which in turn implies that $T$ has the ESP and the FMP (notice that given $\rho_{1},\rho_{2}\in\mathcal{S}({\mathcal{H}})$ and ${\bf z}\in D_{n}$ arbitrary, the difference $T(\rho_{1},{\bf z})-T(\rho_{2},{\bf z})\in\mathcal{B}_{0}(\mathcal{H})$ and hence the hypothesis ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|T|_{\mathcal{B}_{0}(\mathcal{H})}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{p}<1$ implies that $\left\|T(\rho_{1},{\bf z})-T(\rho_{2},{\bf z})\right\|_{p}<\left\|\rho_{1}-\rho_{2}\right\|_{p}$). Since we are in a finite-dimensional context, it suffices to impose contractivity for just one norm ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{p}$, $p\in[1,\infty)$, in order to ensure that the limit in (17) exists and that its value is independent of the initial condition. The expression (16) shows that it is not possible to write down a closed-form expression for the filter of a QRC system with the ESP when using the density matrix representation, in which the initial dependence condition has been eliminated. This feature is ultimately due to the linearity of the setup. Already in the classical framework (see [45]), it has been shown that this can be avoided by working with affine instead of purely linear systems. In the quantum context, it can also be observed (see [37, 41]) that the non- homogeneous state-space system given by $\rho_{t}=(1-\epsilon)T(\rho_{t-1},{\bf z}_{t})+\epsilon\sigma$, where $T(\rho_{t-1},{\bf z}_{t})$ is a CPTP map, $\sigma$ an arbitrary density matrix, and $0<\epsilon<1$, defines a unique filter $U({\bf z})_{t}=\epsilon\sigma+\epsilon\sum^{\infty}_{j=1}(1-\epsilon)^{j}T(\sigma,{\bf z}_{t+1-j})$ in which the dependence on initial conditions has disappeared. In the following subsections, we shall circumvent this problem by showing that certain matrix representations of finite-dimensional QRC systems on density matrices have a built-in non-homogeneous affine structure that makes them into non-homogeneous state-affine systems (SAS) of the type introduced in [45]. More specifically, we shall be working with the Bloch vector representation of quantum finite dimensional systems [63, 64] associated to a given Gell-Mann basis. This idea is not new in QRC and it can already be seen in the seminal work [34] or, more recently, in the Methods section of [40]. In the paragraphs that follow, we shall explore in depth this representation, mostly in connection with the available literature on SAS systems (Section III.2), which will allow us later on in Section III.3 to identify various design constraints on quantum channels. ### III.1 Matrix representation of quantum channels We will start by introducing the notation necessary for the matrix representation of quantum channels. In the next section, we shall focus on a specific choice of basis adapted to density matrices. Let $\\{B_{i}\\}_{i\in\left\\{1,\ldots,d^{2}\right\\}}$ be an orthonormal basis for the vector space $\mathcal{B}(\mathcal{H})$, when endowed with the Hilbert-Schmidt inner product, that is, $\text{tr}(B^{\dagger}_{i}B_{j})=\delta_{ij}$. Using any such basis we can represent any operator $A\in\mathcal{B}(\mathcal{H})$ as $A=\sum_{i=1}^{d^{2}}a_{i}B_{i}$, with $a_{i}=\text{tr}(B_{i}^{\dagger}A)$. Analogously, we can express any linear map $T:\mathcal{B}(\mathcal{H})\rightarrow\mathcal{B}(\mathcal{H})$ as $T(A)=\sum^{d^{2}}_{i,j=1}\widehat{T}_{ij}a_{j}B_{i},\ \mbox{where $\widehat{T}_{ij}=\text{tr}(B^{\dagger}_{i}T(B_{j}))$.}$ (19) This observation implies that QRC systems $T:\mathcal{S}(\mathcal{H})\times D_{n}\rightarrow\mathcal{S}(\mathcal{H})$ admit an equivalent representation as a system $\widehat{T}:V\times D_{n}\rightarrow V$, where $V\subset\mathbb{R}^{d^{2}}$ is the subset of real Euclidean space that contains the coordinate representations of the elements in $\mathcal{S}(\mathcal{H})$ using the basis $\\{B_{i}\\}_{i\in\left\\{1,\ldots,d^{2}\right\\}}$. This statement can be formalized using the language of system morphisms (see [65, 66] for the standard definitions and elementary facts). Consider the state-space systems determined by the triples $({\cal X}_{i},F_{i},h_{i})$, $i\in\left\\{1,2\right\\}$, with $F_{i}:{\cal X}_{i}\times{\cal Z}\longrightarrow{\cal X}_{i}$ and $h_{i}:{\cal X}_{i}\longrightarrow{\cal Y}$. A map $f:{\cal X}_{1}\longrightarrow{\cal X}_{2}$ is a morphism between the systems $({\cal X}_{1},F_{1},h_{1})$ and $({\cal X}_{2},F_{2},h_{2})$ whenever it satisfies the following two properties: 1. (i) System equivariance: $f(F_{1}({\bf x}_{1},{\bf z}))=F_{2}(f({\bf x}_{1}),{\bf z})$, for all ${\bf x}_{1}\in{\cal X}_{1}$ and ${\bf z}\in{\cal Z}$. 2. (ii) Readout invariance: $h_{1}({\bf x}_{1})=h_{2}(f({\bf x}_{1}))$, for all ${\bf x}_{1}\in{\cal X}_{1}$. When the map $f$ has an inverse $f^{-1}$ and this inverse is also a morphism between the systems determined by $({\cal X}_{2},F_{2},h_{2})$ and $({\cal X}_{1},F_{1},h_{1})$, we say that $f$ is a system isomorphism and the systems $({\cal X}_{1},F_{1},h_{1})$ and $({\cal X}_{2},F_{2},h_{2})$ are isomorphic. We note that given a system $F_{1}:{\cal X}_{1}\times{\cal Z}\longrightarrow{\cal X}_{1},h_{1}:{\cal X}_{1}\longrightarrow{\cal Y}$ and a bijection $f:{\cal X}_{1}\longrightarrow{\cal X}_{2}$, the map $f$ is a system isomorphism with respect to the system $F_{2}:{\cal X}_{2}\times{\cal Z}\longrightarrow{\cal X}_{2},h_{2}:{\cal X}_{2}\longrightarrow{\cal Y}$ defined by $\displaystyle F_{2}({\bf x}_{2},{\bf z})$ $\displaystyle:=f(F_{1}(f^{-1}({\bf x}_{2}),{\bf z})),$ (20) $\displaystyle h_{2}({\bf x}_{2})$ $\displaystyle:=h_{1}(f^{-1}({\bf x}_{2})).$ (21) for all ${\bf x}_{2}\in{\cal X}_{2}$, ${\bf z}\in{\cal Z}$. Consider now the QRC system given by the quantum channel $T:\mathcal{S}(\mathcal{H})\times D_{n}\rightarrow\mathcal{S}(\mathcal{H})$ and the readout map $h:\mathcal{B}(\mathcal{H})\rightarrow\mathbb{R}^{m}$, $m\in\mathbb{N}$. Using the orthonormal basis $\mathcal{B}=\\{B_{i}\\}_{i\in\left\\{1,\ldots,d^{2}\right\\}}$ and the discussion above define the map $\begin{array}[]{cccc}G_{\mathcal{B}}:&\mathbb{C}^{d^{2}}&\longrightarrow&\mathcal{B}({\mathcal{H}})\\\ &\mathbf{a}&\longmapsto&\sum_{i=1}^{d^{2}}a_{i}B_{i}.\end{array}$ (22) This map is a linear homeomorphism. Define $V=G_{\mathcal{B}}^{-1}(\mathcal{S}(\mathcal{H}))\subset\mathbb{R}^{d^{2}}$ as well as the map (that we denote with the same symbol) $G_{\mathcal{B}}:V\longrightarrow\mathcal{S}(\mathcal{H})$ that we obtain by restriction of the domain and codomain in (22). This restricted map is also a homeomorphism when $V$ and $\mathcal{S}(\mathcal{H})$ are endowed with their relative topologies [67, Theorem 18.2]. With all these ingredients, it is straightforward to verify that the QRC system $(\mathcal{S}(\mathcal{H}),T,h)$ is system isomorphic to $(V,\widehat{T},\widehat{h})$ with $\widehat{T}:V\times D_{n}\longrightarrow V$ and $\widehat{h}:V\longrightarrow\mathbb{R}^{m}$ given by $\displaystyle\widehat{T}(\mathbf{a},{\bf z})$ $\displaystyle:=G_{\mathcal{B}}^{-1}\left(T\left(G_{\mathcal{B}}(\mathbf{a}),{\bf z}\right)\right),$ (23) $\displaystyle\widehat{h}(\mathbf{a})$ $\displaystyle:=h(G_{\mathcal{B}}(\mathbf{a})),$ (24) and that the isomorphism is given by the map $G_{\mathcal{B}}:V\longrightarrow\mathcal{S}(\mathcal{H})$. The procedure that we just spelled out can be reproduced for any other (orthonormal) basis $\mathcal{B}^{\prime}$ of $\mathcal{B}({\mathcal{H}})$, in which case we would obtain another system $(V^{\prime},\widehat{T}^{\prime},\widehat{h}^{\prime})$ which is obviously isomorphic to both $(V,\widehat{T},\widehat{h})$ and $(\mathcal{S}(\mathcal{H}),T,h)$. The system isomorphisms that we just defined and the compactness of the state spaces where they are defined allow us to establish important connections between the filters that they define. The following proposition describes those connections in detail. ###### Proposition 2. Let $T:\mathcal{S}(\mathcal{H})\times D_{n}\rightarrow\mathcal{S}(\mathcal{H})$ be a quantum channel, $h:\mathcal{S}(\mathcal{H})\longrightarrow\mathbb{R}^{m}$ a readout, and let $\mathcal{B}=\\{B_{i}\\}_{i\in\left\\{1,\ldots,d^{2}\right\\}}$ be an orthonormal basis for $\mathcal{B}(\mathcal{H})$. Let $\widehat{T}:V\times D_{n}\longrightarrow V$ be the isomorphic system defined in (22) and $\widehat{h}:V\longrightarrow\mathbb{R}^{m}$ the corresponding readout defined in (24). Then: 1. (i) Given an input ${\bf z}\in({\mathbb{R}}^{n})^{\mathbb{Z}}$, a sequence $\bm{\rho}\in\left(\mathcal{S}(\mathcal{H})\right)^{\mathbb{Z}}$ is a solution for that input for the system determined by $T$ if and only if the sequence ${\cal G}_{\mathcal{B}}^{-1}\left(\bm{\rho}\right)\in\left(V\right)^{\mathbb{Z}}$ is a solution for the system associated to $\widehat{T}$. The symbol ${\cal G}_{\mathcal{B}}=\prod_{\mathbb{Z}}G_{\mathcal{B}}:(V)^{\mathbb{Z}}\longrightarrow\left({\cal S}({\mathcal{H}})\right)^{\mathbb{Z}}$ stands for the product map. 2. (ii) $T$ has the ESP if and only if $\widehat{T}$ has the ESP. In that case, the filters $U_{T}$ and $U_{\widehat{T}}$ (respectively, $U_{T}^{h}$ and $U_{\widehat{T}}^{\widehat{h}}$) determined by $T$ and $\widehat{T}$ (respectively, by $(T,h)$ and $(\widehat{T},\widehat{h})$) satisfy that $U_{T}={\cal G}_{\mathcal{B}}\circ U_{\widehat{T}}$ (respectively, $U_{T}^{h}=U_{\widehat{T}}^{\widehat{h}}$). 3. (iii) $U_{T}$ has the FMP if and only if $U_{\widehat{T}}$ has the FMP. 4. (iv) Let $V^{\prime}$ be a set homeomorphic to $V$. The relations (20)-(21) determine in that case an isomorphic system $\widehat{T}^{\prime}:V^{\prime}\times D_{n}\longrightarrow V^{\prime}$, $\widehat{h}^{\prime}:V^{\prime}\rightarrow\mathbb{R}^{m}$ that has the ESP and the FMP if and only if $\widehat{T}$ and $\widehat{h}$ have that property. ###### Proof. Parts (i), (ii), and (iv) are a straightforward consequence of Proposition 2.2 in [44]. As to part (iii), given that by (ii) $U_{T}={\cal G}_{\mathcal{B}}\circ U_{\widehat{T}}$, it suffices to prove that the bijection ${\cal G}_{\mathcal{B}}=\prod_{\mathbb{Z}}G_{\mathcal{B}}:(V)^{\mathbb{Z}}\longrightarrow\left({\cal S}({\mathcal{H}})\right)^{\mathbb{Z}}$ is a homeomorphism when the domain and the target are endowed with a weighted norm. In order to prove that, recall first that, using the observation right under (15), the weighted norms in $(V)^{\mathbb{Z}}$ and $\left({\cal S}({\mathcal{H}})\right)^{\mathbb{Z}}$ induce the product topology due to the compactness of $V$ and $\mathcal{S}(\mathcal{H})$. This immediately implies that (Theorem 19.6 in [67]) ${\cal G}_{\mathcal{B}}=\prod_{\mathbb{Z}}G_{\mathcal{B}}$ is continuous due to the continuity of $G_{\mathcal{B}}$. The same argument can be immediately applied to the inverse map ${\cal G}_{\mathcal{B}}^{-1}$, which proves the statement. ∎ ### III.2 Non-homogenous state affine system representation The goal of this section is to choose a specific basis in $\mathcal{B}({\mathcal{H}})$ for the representation of the quantum channel $T:\mathcal{B}(\mathcal{H})\times D_{n}\rightarrow\mathcal{B}(\mathcal{H})$ (equivalently, $T:\mathcal{S}(\mathcal{H})\times D_{n}\rightarrow\mathcal{S}(\mathcal{H})$) in which the associated system isomorphic representation $\widehat{T}:\mathbb{C}^{d^{2}}\times D_{n}\longrightarrow\mathbb{C}^{d^{2}}$ (equivalently, $\widehat{T}:V\times D_{n}\longrightarrow V$) has a non-homogeneous state-affine form of the type introduced in [45]. More specifically, we choose a generalized Gell-Mann basis [63, 64, 68]. This is an orthonormal basis made of Hermitian operators in which, by convention, its first element is the normalized identity, namely, $B_{1}=I/\sqrt{d}$. The remaining $(d^{2}-1)$ traceless Hermitian operators are the generators $\mathcal{B}_{0}=\left\\{\sigma_{i}\right\\}_{i\in\left\\{1,\ldots,d^{2}-1\right\\}}$ (25) of the fundamental representation of the Lie algebra $\mathfrak{su}(d)$ of SU($d$). The case $d=2$ corresponds to the case of one qubit, and the Gell- Mann basis is made of the standard Pauli matrices. The orthonormality of the Gell-Mann basis is guaranteed by the product property of the fundamental representation of $\mathfrak{su}(d)$, namely, $\sigma_{a}\sigma_{b}=\delta_{ab}I/(2d)+\sum^{d^{2}-1}_{c}f_{c}\sigma_{c}$, where $\sigma_{a},\sigma_{b}$ are two elements of the Gell-Mann basis for $\mathfrak{su}(d)$, and $f_{c}$ are complex coefficients. The resulting Gell- Mann basis $\mathcal{B}=\left\\{B_{i}\right\\}_{i\in\left\\{1,\ldots,d^{2}\right\\}}$ of ${\cal B}({\mathcal{H}})$ is hence given by $B_{1}=I/\sqrt{d}$ and $B_{i}=\sigma_{i-1}$, $1<i\leq d^{2}$. Note that the subset $\mathcal{B}_{0}=\left\\{B_{i}\right\\}_{i\in\left\\{2,\ldots,d^{2}\right\\}}$ is a basis for the vector subspace ${\cal B}_{0}({\mathcal{H}})\subset{\cal B}({\mathcal{H}})$ of codimension $1$ made of traceless operators. If our system is made of $N$ $d$-dimensional systems (qudits), we can extend this basis to $\mathcal{B}\left({\mathcal{H}}^{\otimes^{N}}\right)$ by tensorization. The orthonormality of the tensorized basis with respect to the Hilbert-Schmidt inner product is preserved since $\text{tr}(A\otimes B)=\text{tr}(A)\text{tr}(B)$ for any two $A,B\in\mathcal{B}({\mathcal{H}})$. We now go back to the system constituted by one qudit and spell out the matrix expression $\widehat{T}:\mathbb{C}^{d^{2}}\times D_{n}\longrightarrow\mathbb{C}^{d^{2}}$ of a CPTP map $T:\mathcal{B}({\mathcal{H}})\times D_{n}\longrightarrow\mathcal{B}({\mathcal{H}})$ in the basis $\mathcal{B}$ that we just introduced, by using the prescription introduced in (19). We first note that the choice $B_{1}=I/\sqrt{d}$ and the trace-preserving character of $T(\cdot,{\bf z})$ for any ${\bf z}\in D_{n}$, imply that $\widehat{T}(\cdot,{\bf z})_{11}=1$. Analogously, $\widehat{T}(\cdot,{\bf z})_{1j}=\text{tr}(T(B_{j},{\bf z}))/\sqrt{d}=0$ for $1<j\leq d^{2}$, since $T$ is trace-preserving. This implies that the matrix $\widehat{T}(\cdot,{\bf z})$ can be written as $\widehat{T}(\cdot,{\bf z})=\left(\begin{matrix}1&{\bf 0}_{d^{2}-1}\\\ q({\bf z})&p({\bf z})\end{matrix}\right)$ (26) where $p({\bf z})$ is the square matrix of dimension $d^{2}-1$ with complex entries $p({\bf z})_{ij}:=\left(\widehat{T}(\cdot,{\bf z})|_{G_{\mathcal{B}}^{-1}\left(\mathcal{B}_{0}(\mathcal{H})\right)}\right)_{ij}\\\ =\left(\widehat{T}(\cdot,{\bf z})|_{G_{\mathcal{B}}^{-1}\left(\operatorname{span}\left\\{\mathcal{B}_{0}\right\\}\right)}\right)_{ij}=\text{tr}(B_{i}T(B_{j},{\bf z})),$ (27) $1<i,j\leq d^{2}$, and $q({\bf z})\in\mathbb{C}^{d^{2}-1}$ is given by $q({\bf z})_{i}=\text{tr}(B_{i}T(I))/\sqrt{d}$, $1<i\leq d^{2}$. The symbol ${\bf 0}_{d^{2}-1}$ denotes a zero-row of length $d^{2}-1$. Now, given that any element $\rho\in\mathcal{S}({\mathcal{H}})$ has coordinates in the Gell-Mann basis of the form $\left(1/\sqrt{d},\mathbf{x}^{\top}\right)^{\top}\in\mathbb{C}^{d^{2}}$ with $\mathbf{x}\in\mathbb{C}^{d^{2}-1}$ called the Bloch vector, the matrix form (26) implies that $\widehat{T}(\cdot,{\bf z})\left(\begin{matrix}1/\sqrt{d}\\\ \mathbf{x}\end{matrix}\right)=\left(\begin{matrix}1/\sqrt{d}\\\ p({\bf z})\textbf{x}+q({\bf z})\end{matrix}\right).$ (28) This expression implies that $\widehat{T}:V\times D_{n}\longrightarrow V$ admits a system-isomorphic representation $\widehat{T}_{0}:V_{0}\times D_{n}\longrightarrow V_{0}$ on the set $V_{0}=\left\\{\mathbf{x}\in\mathbb{C}^{d^{2}-1}\mid\left(1/\sqrt{d},\mathbf{x}^{\top}\right)^{\top}\in V\right\\}\subset\mathbb{C}^{d^{2}-1}$ given by $\begin{array}[]{cccc}\widehat{T}_{0}:&V_{0}\times D_{n}&\longrightarrow&V_{0}\\\ &(\mathbf{x},{\bf z})&\longmapsto&p({\bf z})\textbf{x}+q({\bf z}),\end{array}$ (29) with readout $\widehat{h}_{0}(\mathbf{x})=\widehat{h}\left(\left(1/\sqrt{d},\mathbf{x}^{\top}\right)^{\top}\right)$. The system isomorphism is in this case, given by the map $\begin{array}[]{cccc}i_{0}:&V_{0}&\longrightarrow&V\\\ &\mathbf{x}\in V_{0}&\longmapsto&\left(1/\sqrt{d},\mathbf{x}^{\top}\right)^{\top}.\end{array}$ (30) Part (iv) of Proposition 2 guarantees that the dynamical properties of the system $(\widehat{T},\widehat{h},V)$ (and hence those of the QRC system $(T,h,\mathcal{S}({\mathcal{H}}))$) are equivalent to those of $(\widehat{T}_{0},\widehat{h}_{0},V_{0})$. The importance of this observation is that it links $(T,h,\mathcal{S}({\mathcal{H}}))$ to the non-homogeneous state-affine system (SAS) introduced in [45] and for which various universality properties have been additionally proved in [46, 69]. SAS are defined as state equations that have the form spelled out in (29). Strictly speaking, the SAS systems studied in the above-cited references impose polynomial or trigonometric dependences of $q$ and $p$ on the inputs, while in our situation, the prescription introduced in (10) is capable of accommodating more general forms. It has been shown in [45] that when such a system has the ESP, the corresponding filter $U_{\widehat{T}_{0}}:\left(D_{n}\right)^{\mathbb{Z}}\rightarrow\left(V_{0}\right)^{\mathbb{Z}}$ can be written as $U_{\widehat{T}_{0}}({\bf z})_{t}=\sum^{\infty}_{j=0}\left(\prod^{j-1}_{k=0}p({\bf z}_{t-k})\right)q({\bf z}_{t-j}),$ (31) where $\prod_{k=0}^{j-1}p({\bf z}_{t-k}):=p({\bf z}_{t})\cdot p({\bf z}_{t-1})\cdots p({\bf z}_{t-j+1})$. A first sufficient condition for ESP and FMP has been formulated in [45] by imposing that $\sigma_{\text{max}}(p({\bf z}))<1$ for all ${\bf z}\in D_{n}$, where $\sigma_{\text{max}}$ is the maximum singular value of matrix $p({\bf z})$. Given a matrix $A$, the maximum singular value is equal to the 2-Schatten induced norm: ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|A\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{2}=\sigma_{\text{max}}(A)$, where the singular values of $A$ are the square-roots of the eigenvalues of $AA^{\dagger}$. An improved sufficient condition could be potentially found using other matrix norms as in [70]. The next proposition uses this hint and, moreover, spells out equivalent necessary and sufficient conditions for the ESP and FMP to hold in the three different equivalent representations for QRC systems that we have introduced in this section. More explicitly, the statement addresses the ESP and the FMP for the operator representation $T:\mathcal{B}(\mathcal{H})\times D_{n}\rightarrow\mathcal{B}(\mathcal{H})$, the matrix representation $\widehat{T}:\mathbb{C}^{d^{2}}\times D_{n}\longrightarrow\mathbb{C}^{d^{2}}$ associated to the Gell-Mann basis $\mathcal{B}$, and the SAS representation $\widehat{T}_{0}:\mathbb{C}^{d^{2}-1}\times D_{n}\longrightarrow\mathbb{C}^{d^{2}-1}$ introduced in (29). ###### Proposition 3. Let $T:\mathcal{B}(\mathcal{H})\times D_{n}\rightarrow\mathcal{B}(\mathcal{H})$ be a continuous QRC system. The following three statements are equivalent: 1. (i) There exists an operator norm ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ and $\epsilon>0$ such that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|T(\cdot,{\bf z})|_{\mathcal{B}_{0}(\mathcal{H})}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}<1-\epsilon,\ \mbox{for all ${\bf z}\in D_{n}$}.$ (32) 2. (ii) There exists a matrix norm ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ in the space of complex $d^{2}\times d^{2}$ matrices such that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widehat{T}(\cdot,{\bf z})|_{G_{\mathcal{B}}^{-1}\left(\operatorname{span}\left\\{\mathcal{B}_{0}\right\\}\right)}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}<1-\epsilon,\ \mbox{for all ${\bf z}\in D_{n}$},$ (33) with $\mathcal{B}_{0}$ the trace-zero elements in the basis $\mathcal{B}$ defined in (25) and $G_{\mathcal{B}}$ the isomorphism defined in (22) with respect to the Gell-Mann basis. 3. (iii) There is a matrix norm ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ in the space of complex $(d^{2}-1)\times(d^{2}-1)$ matrices such that the SAS representation $\widehat{T}_{0}:\mathbb{C}^{d^{2}-1}\times D_{n}\longrightarrow\mathbb{C}^{d^{2}-1}$ introduced in (29) satisfies that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|p({\bf z})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}<1-\epsilon,\ \mbox{for all ${\bf z}\in D_{n}$},$ (34) If any of these three equivalent statements hold and $D_{n}$ is compact, then: 1. 1. The isomorphic systems $T:\mathcal{S}(\mathcal{H})\times D_{n}\rightarrow\mathcal{S}(\mathcal{H})$, $\widehat{T}:V\times D_{n}\rightarrow V$, and $\widehat{T}_{0}:V_{0}\times D_{n}\rightarrow V_{0}$, have the ESP and the FMP, and hence continuous filters $U_{T}:(D_{n})^{\mathbb{Z}}\longrightarrow\left(\mathcal{S}(\mathcal{H})\right)^{\mathbb{Z}}$, $U_{\widehat{T}}:(D_{n})^{\mathbb{Z}}\longrightarrow V^{\mathbb{Z}}$, and $U_{\widehat{T}_{0}}:(D_{n})^{\mathbb{Z}}\longrightarrow V_{0}^{\mathbb{Z}}$ can be associated to them. 2. 2. In such case, the filter $U_{\widehat{T}_{0}}$ is then given by (31). $U_{\widehat{T}}$ is determined by $U_{\widehat{T}}={\cal I}_{0}\circ U_{\widehat{T}_{0}}$, with ${\cal I}_{0}=\prod_{\mathbb{Z}}i_{0}$ and $i_{0}$ as in (30). Finally, $U_{T}={\cal G}_{\mathcal{B}}\circ U_{\widehat{T}}$, with ${\cal G}_{\mathcal{B}}=\prod_{\mathbb{Z}}G_{\mathcal{B}}$ the map introduced in Proposition 2. 3. 3. The contraction conditions (32)-(34) are necessary for the ESP and the FMP to hold. ###### Proof. The equivalences between the contraction conditions (32)-(34) can be shown by using norms in $\mathbb{C}^{d^{2}}$ and $\mathcal{B}({\mathcal{H}})$ that make the map $G_{\mathcal{B}}$ in (22) into an isometry. More explicitly, let us start with an operator norm ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ for which the maps $T(\cdot,{\bf z}):\mathcal{B}({\mathcal{H}})\longrightarrow\mathcal{B}({\mathcal{H}})$ satisfy the condition (32). Assume that this operator norm is associated with a given norm $\left\|\cdot\right\|_{\mathcal{B}({\mathcal{H}})}$ in $\mathcal{B}({\mathcal{H}})$, that is, ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|T(\cdot,{\bf z})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}=\sup_{A\neq 0}\left\\{\left\|T(A,{\bf z})\right\|_{\mathcal{B}({\mathcal{H}})}/\left\|A\right\|_{\mathcal{B}({\mathcal{H}})}\right\\}$. Take now the norm $\left\|\cdot\right\|_{\mathbb{C}^{d^{2}}}$ in $\mathbb{C}^{d^{2}}$ with respect to which $G_{\mathcal{B}}$ is an isometry, that is, set $\left\|\mathbf{a}\right\|_{\mathbb{C}^{d^{2}}}=\left\|G_{\mathcal{B}}(\mathbf{a})\right\|_{\mathcal{B}({\mathcal{H}})}$, for all $\mathbf{a}\in\mathbb{C}^{d^{2}}$, and denote by ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widehat{T}(\cdot,{\bf z})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}^{\prime}=\sup_{\mathbf{a}\neq 0}\left\\{\left\|\widehat{T}(\mathbf{a},{\bf z})\right\|_{\mathbb{C}^{d^{2}}}/\left\|\mathbf{a}\right\|_{\mathbb{C}^{d^{2}}}\right\\}$. Using just these definitions, it is easy to see that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|T(\cdot,{\bf z})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}={\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widehat{T}(\cdot,{\bf z})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}^{\prime}$ and, moreover, that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|T(\cdot,{\bf z})|_{\mathcal{B}_{0}(\mathcal{H})}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}={\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widehat{T}(\cdot,{\bf z})|_{G_{\mathcal{B}}^{-1}\left(\operatorname{span}\left\\{\mathcal{B}_{0}\right\\}\right)}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}^{\prime},$ which can be easily used to prove the equivalence between (32) and (33). The equivalence between (33) and (34) follows from the fact that, as it can be seen in (26), $\widehat{T}(\cdot,{\bf z})|_{G_{\mathcal{B}}^{-1}\left(\operatorname{span}\left\\{\mathcal{B}_{0}\right\\}\right)}=p({\bf z}).$ The claims $1$ and $2$ in the second part of the proposition are straightforward consequences of Propositions 1 and 2. We now prove the last claim in point $3$ using the SAS representation (29) for the continuous QRC system $T$ that, this time, it is assumed to have the ESP and the FMP. If we iterate $n$ times this state equation using an arbitrary vector $\mathbf{x}_{0}\in\mathbb{C}^{d^{2}-1}$ as an initial condition, we obtain the vector $U^{n,\mathbf{x}_{0}}_{\widehat{T}_{0}}({\bf z})_{t}\in\mathbb{C}^{d^{2}-1}$ defined by: $U^{n,\mathbf{x}_{0}}_{\widehat{T}_{0}}({\bf z})_{t}=\sum^{n-1}_{j=0}\left(\prod^{j-1}_{k=0}p({\bf z}_{t-k})\right)q({\bf z}_{t-j})+\prod^{n-1}_{j=0}p({\bf z}_{t-j})\mathbf{x}_{0}.$ If we now assume that the system has the ESP, we necessarily have that $U_{\widehat{T}_{0}}({\bf z})_{t}=\lim_{n\rightarrow\infty}U^{n,\mathbf{x}_{0}}_{\widehat{T}_{0}}({\bf z})_{t}=\lim_{n\rightarrow\infty}U^{n,\mathbf{x}_{0}^{\prime}}_{\widehat{T}_{0}}({\bf z})_{t},$ (35) with $U_{\widehat{T}_{0}}$ the filter in (31) and $\mathbf{x}_{0},\mathbf{x}_{0}^{\prime}\in\mathbb{C}^{d^{2}-1}$ arbitrary vectors. The last equality in (35) and the arbitrary character of $\mathbf{x}_{0},\mathbf{x}_{0}^{\prime}\in\mathbb{C}^{d^{2}-1}$ imply that $\prod^{\infty}_{j=0}p({\bf z}_{t-j})=\lim_{n\rightarrow\infty}\prod^{n-1}_{j=0}p({\bf z}_{t-j})=0,$ necessarily. We now notice that since $T$ is a continuous map, then so is the dependence of $p({\bf z})$ on the inputs ${\bf z}$. Moreover, since, in this case, $D_{n}$ is assumed to be a compact set, then so is the matrix set $\left\\{p({\bf z})\mid{\bf z}\in D_{n}\right\\}$. Now, Corollary 6.4 in [71] guarantees the existence of matrix norm ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ for which (34) is satisfied, as required. ∎ ###### Remark 4. Using the characterization of the mixing property in (7), it is clear that a necessary condition for the QRC $T$ to satisfy the contractivity hypothesis in the previous proposition and, in passing, satisfy the ESP and the FMP, is that all the maps $T(\cdot,{\bf z})$ are mixing for all ${\bf z}\in D_{n}$. Indeed, since the spectral radius is a lower bound for any matrix norm (see [59, Theorem 5.6.9]) we have that $\lambda_{\text{max}}\left(\widehat{T}(\cdot,{\bf z})|_{G_{\mathcal{B}}^{-1}\left(\operatorname{span}\left\\{\mathcal{B}_{0}\right\\}\right)}\right)\leq{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widehat{T}(\cdot,{\bf z})|_{G_{\mathcal{B}}^{-1}\left(\operatorname{span}\left\\{\mathcal{B}_{0}\right\\}\right)}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$. Consequently, if ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\widehat{T}(\cdot,{\bf z})|_{G_{\mathcal{B}}^{-1}\left(\operatorname{span}\left\\{\mathcal{B}_{0}\right\\}\right)}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}<1-\epsilon$ then $\lambda_{\text{max}}\left(\widehat{T}(\cdot,{\bf z})|_{G_{\mathcal{B}}^{-1}\left(\operatorname{span}\left\\{\mathcal{B}_{0}\right\\}\right)}\right)\leq 1-\epsilon$, which is equivalent to $\lambda_{\text{max}}\left(T(\cdot,{\bf z})|_{\mathcal{B}_{0}(\mathcal{H})}\right)\leq 1-\epsilon$ and then all maps $T(\cdot,{\bf z})$ are mixing by (7). We emphasize that this mixing condition is necessary but not sufficient since, as the composition of two mixing maps is not necessarily mixing, the existence of the limit in (17) is not guaranteed even if each of the factor operators is mixing. ### III.3 Some constrains on CPTP maps for QRC The SAS representation allows us to easily characterize situations like the one in Proposition 3 in which a QRC system has the ESP and the FMP and, moreover, it allows us to write explicitly down the corresponding filter (31). As we shall now see in this section, more interesting facts can be derived from this representation having to do with design features that should be avoided, as they produce systems with only trivial solutions. The first one concerns unital quantum channels, that is, channels that satisfy $T(I,{\bf z})=I\quad\mbox{for all ${\bf z}\in D_{n}$}.$ That situation is studied in the next theorem, which will be generalized in Theorem 9 to the case of QRC systems that exhibit an input-independent fixed point. ###### Theorem 5. Let $T:\mathcal{B}(\mathcal{H})\times D_{n}\rightarrow\mathcal{B}(\mathcal{H})$ be a QRC system for which there exists an operator norm ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ and $\epsilon>0$ such that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|T(\cdot,{\bf z})|_{\mathcal{B}_{0}(\mathcal{H})}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}<1-\epsilon$, for all ${\bf z}\in D_{n}$. Then, the corresponding filter $U_{T}$ is constant with $U_{T}({\bf z})_{t}=I/d$ for all ${\bf z}\in(D_{n})^{\mathbb{Z}}$ (equivalently $U_{\widehat{T}_{0}}({\bf z})_{t}={\bf 0}$) if and only if $T$ is unital. ###### Proof. We prove this statement by using the SAS representation associated to the Gell-Mann basis $\mathcal{B}$. If $T$ is unital, that is, $T(I,{\bf z})=I$, then the expressions (19) and (26) imply that $q({\bf z})_{i}=\text{tr}(B_{i}T(I,{\bf z}))/\sqrt{d}=\text{tr}(B_{i})/\sqrt{d}=0$, for all $i\in\left\\{2,\ldots,d^{2}\right\\}$. Therefore, the SAS state equation becomes, in this case, $\textbf{x}_{t}=p({\bf z}_{t})\textbf{x}_{t-1}$, which is a homogeneous equation whose only solution under the contractivity hypotheses is the trivial one, that is, $\textbf{x}_{t}={\bf 0}$ for all $t\in\mathbb{Z}$. Consequently, $U_{\widehat{T}_{0}}={\bf 0}$. Proposition 3 implies then that $U_{T}={\cal G}_{\mathcal{B}}\circ{\cal I}_{0}\circ U_{\widehat{T}_{0}}=I/d$. Conversely, suppose that $U_{T}({\bf z})_{t}=I/d$ for all ${\bf z}\in(D_{n})^{\mathbb{Z}}$. Since the filter $U_{T}$ is determined by the recursions $U_{T}({\bf z})_{t}=T\left(U_{T}({\bf z})_{t-1},{\bf z}_{t}\right)$, then, this relation implies that $T(I,{\bf z})=I$, for all ${\bf z}\in D_{n}$. ∎ ###### Remark 6. There is a close relation between unital quantum channels and contractivity. Indeed, it has been shown that unital maps are contractive for all Schatten $p$-norms, that is, ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|T(\cdot,{\bf z})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{p}\leq 1$ (see Theorem 2.4 in [72]) when $T(\cdot,{\bf z})$ is unital. ###### Example 7. Consider the depolarizing channel $\mathcal{E}:\mathcal{S}({\mathcal{H}})\longrightarrow\mathcal{S}({\mathcal{H}})$ defined by $\mathcal{E}(\rho)=(1-\lambda)\rho+\lambda\frac{I}{d},$ (36) where $0\leq\lambda\leq 1$ denotes the probability of finding the system at the maximally mixed state $I/d$. If we arrange the previous equation as an input-dependent channel, we can write the following state equation: $\rho_{t}=\mathcal{E}(\rho_{t-1},{\bf z}_{t})=(1-\lambda_{t})\rho_{t-1}+\lambda_{t}\frac{I}{d},$ (37) where $\lambda_{t}:=\lambda({\bf z}_{t})$ is a function of the input at each time step. This map is strictly contractive whenever $r=\sup_{t\in\mathbb{Z}}(1-\lambda_{t})<1$ in which case $r$ is the contraction rate. Let us compute the state after $n$ backwards iterations: $\begin{split}\rho^{n}_{t}&=\left(\prod^{n-1}_{i=0}(1-\lambda_{t-i})\right)\rho_{t-n}\\\ &+\sum^{n-1}_{i=0}\left(\lambda_{t-i}\left(\prod^{i-1}_{j=0}(1-\lambda_{t-j})\right)\right)\frac{I}{d}.\end{split}$ (38) Taking the limit $n\rightarrow\infty$, it is easy to see that the first summand vanishes because $1-\lambda_{t-i}\leq r<1$ for all $t\in\mathbb{Z}$, and hence all possible dependence on the initial condition disappears. Regarding the second summand, since the map is strictly contractive, the limit $\lim_{n\rightarrow\infty}\rho^{n}_{t}$ exists and must yield a density matrix, which means that $\sum^{\infty}_{i=0}\left(\lambda_{t-i}\left(\prod^{i-1}_{j=0}(1-\lambda_{t-j})\right)\right)=1$ because of the normalization. This limit hence defines the filter $U_{\mathcal{E}}({\bf z})_{t}=I/d$, which is consistent with the conclusion of Theorem 5 since (37) is a strictly contractive unital map. We could have found the same result by directly using (29), for which we need to define the extension of the depolarizing channel to the whole space of bounded operators $\mathcal{B}(\mathcal{H})$. We define then the CPTP map $\mathcal{E}^{\prime}:\mathcal{B}({\mathcal{H}})\longrightarrow\mathcal{B}({\mathcal{H}})$ $\mathcal{E}^{\prime}(A)=(1-\lambda)A+\lambda\text{tr}(A)\frac{I}{d}.$ (39) The SAS representation associated with the input-dependent version of this map is such that $q({\bf z}_{t})=0$ because the map is unital. It is also easy to see that the system equation becomes $\textbf{x}_{t}=p({\bf z}_{t})\textbf{x}_{t-1}=(1-\lambda_{t})\textbf{x}_{t-1}$. The filter $U_{\widehat{\mathcal{E}}^{\prime}_{0}}$ of this QRC equation satisfies that $U_{\widehat{\mathcal{E}}^{\prime}_{0}}({\bf z})_{t}={\bf 0}$, for all $t\in\mathbb{Z}$. ###### Example 8. The next case is an example of “poorly engineered” QRC system which can be detected using Theorem 5. Indeed, we shall introduce a model of dissipation using tuneable local losses, but Theorem 5 will discard its long-term applicability because it is unital. Let us define the Markovian master equation that governs the dynamics between input injections: $\dot{\rho}=-i[H({\bf z}_{t}),\rho]+\gamma L\rho L^{\dagger}-\frac{\gamma}{2}\\{L^{\dagger}L,\rho\\},$ (40) where $L$ is the jump operator and $\\{L^{\dagger}L,\rho\\}$ denotes the anticommutator. We define the input-dependent Hamiltonian as $H({\bf z}_{t})=h({\bf z}_{t})\sigma^{x}/2$, where $h({\bf z}_{t})$ will be an arbitrary function of the input, and the jump operator as $L=\sigma^{z}$. This is a single qubit under the influence of an external magnetic field in the $x$ direction of the real space (whose intensity varies between inputs) with a local dephasing. Notice that the Hamiltonian is considered as time-independent when integrating the dynamics since it is constant between input injections. Going from the density matrix language to the real variable linear description requires to find the CPTP map representation of (40). Since Markovian master equations are CPTP linear transformations on their own, we just need to find the Kraus decomposition that represents the dynamics of this master equation as a map. We will follow the procedure as explained in [73] (see Section 2 in the reference for the details of the algorithm). In particular, the Kraus decomposition of a single qubit can be written in the following form: $T(\rho)=\sum_{i,j}S^{(U)}_{ij}B_{i}\rho B^{\dagger}_{j},$ (41) where $B_{i}$ are the basis elements of a single qubit in the operator space ($\\{I,\sigma^{x},\sigma^{y},\sigma^{z}\\}$) and $S^{(U)}_{ij}:=U^{\dagger}SU$ is a unitary transformation of the Choi matrix $S$. Applying (41) to the definition of matrix $T$ in (19), we can find the map of the single qubit observables: $\left(\begin{matrix}1\\\ \braket{\sigma^{x}}_{t}\\\ \braket{\sigma^{y}}_{t}\\\ \braket{\sigma^{z}}_{t}\end{matrix}\right)=\left(\begin{matrix}1&0&0&0\\\ 0&\widehat{T}_{22}&0&0\\\ 0&0&\widehat{T}_{33}&\widehat{T}_{34}\\\ 0&0&\widehat{T}_{43}&\widehat{T}_{44}\end{matrix}\right)\left(\begin{matrix}1\\\ \braket{\sigma^{x}}_{t-1}\\\ \braket{\sigma^{y}}_{t-1}\\\ \braket{\sigma^{z}}_{t-1}\end{matrix}\right),$ (42) where $\braket{\sigma^{a}}:=\text{tr}(\sigma^{a}\rho)$ is the expected value of the spin projection in the $a$ direction of the real space. The expressions for each matrix element are shown below: $\begin{matrix}\widehat{T}_{22}=e^{-2\gamma\Delta\tau},\\\ \widehat{T}_{33}=e^{-\gamma\Delta\tau}\left(\cosh\left(\Delta\tau\sqrt{\gamma^{2}-h_{t}^{2}}\right)-\frac{\gamma}{\sqrt{\gamma^{2}-h_{t}^{2}}}\sinh\left(\Delta\tau\sqrt{\gamma^{2}-h_{t}^{2}}\right)\right),\\\ \widehat{T}_{44}=e^{-\gamma\Delta\tau}\left(\cosh\left(\Delta\tau\sqrt{\gamma^{2}-h_{t}^{2}}\right)+\frac{\gamma}{\sqrt{\gamma^{2}-h_{t}^{2}}}\sinh\left(\Delta\tau\sqrt{\gamma^{2}-h_{t}^{2}}\right)\right),\\\ \widehat{T}_{34}=\frac{h_{t}e^{-\gamma\Delta\tau}}{\sqrt{\gamma^{2}-h_{t}^{2}}}\sinh\left(\Delta\tau\sqrt{\gamma^{2}-h_{t}^{2}}\right),\\\ \widehat{T}_{43}=-\widehat{T}_{34},\end{matrix}$ (43) where have shortened notation by setting $h_{t}=h({\bf z}_{t})$. The eigenvalues of matrix $\widehat{T}$ can be computed analytically: $\lambda_{1}=1$, $\lambda_{2}=e^{-2\gamma\Delta\tau}$, $\lambda_{3}=e^{-(\gamma+\sqrt{\gamma^{2}-h^{2}_{t}})\Delta\tau}$ and $\lambda_{4}=e^{-(\gamma-\sqrt{\gamma^{2}-h^{2}_{t}})\Delta\tau}$. The moduli of the eigenvalues are $|\lambda_{1}|=1$, $|\lambda_{2}|=e^{-2\gamma\Delta\tau}<1$ and $|\lambda_{3}|=e^{-(\gamma+\sqrt{\gamma^{2}-h^{2}_{t}})\Delta\tau}<1$. Eigenvalue $|\lambda_{4}|=e^{-(\gamma-\sqrt{\gamma^{2}-h^{2}_{t}})\Delta\tau}$ is smaller than one if and only if $h_{t}\neq 0$. Under that condition, the map is a mixing channel with a single fixed point. It can be checked that the map is unital so the fixed point is the maximally mixed state: $\rho^{*}=\left(\begin{matrix}1/2&0\\\ 0&1/2\end{matrix}\right).$ (44) Let us prove that there exists some norm where ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|p(\textbf{z})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}<1$ for all inputs. A straightforward induced norm to evaluate is the 2-Schatten induced norm. The singular values of the restriction $p(\textbf{z})=\left(\begin{matrix}\widehat{T}_{22}&0&0\\\ 0&\widehat{T}_{33}&\widehat{T}_{34}\\\ 0&\widehat{T}_{43}&\widehat{T}_{44}\end{matrix}\right)$ (45) are $\sigma_{1}=e^{-2\gamma\Delta\tau}<1$, $\sigma_{2}=e^{-\gamma\Delta\tau}\sqrt{f_{+}}$ and $\sigma_{3}=e^{-\gamma\Delta\tau}\sqrt{f_{-}}$, where $\begin{split}f_{\pm}&=\frac{1}{\gamma^{2}-h^{2}_{t}}\left(-h^{2}_{t}+\gamma^{2}\cosh\left(2\Delta\tau\sqrt{\gamma^{2}-h_{t}^{2}}\right)\right.\\\ &\left.\pm\gamma\sqrt{\sinh^{2}\left(\Delta\tau\sqrt{\gamma^{2}-h_{t}^{2}}\right)}\sqrt{-4h^{2}_{t}+2\gamma^{2}+2\gamma^{2}\cosh\left(2\Delta\tau\sqrt{\gamma^{2}-h_{t}^{2}}\right)}\right).\end{split}$ (46) Figure 1: Density plot of the singular values (a) $\sigma_{2}$ and (b) $\sigma_{3}$. The diagonal elements $h_{t}=\gamma$ are not determined because of the denominator $1/(\gamma^{2}-h^{2}_{t})$. Figure 1 numerically shows that for $\gamma\neq h_{t}$ and away from the axes $h_{t}=0$ and $\gamma=0$, we find $\sigma_{2},\sigma_{3}<1$. Therefore, the system has the ESP and the FMP. As we showed in Theorem 5, the corresponding filter (17) is necessarily trivial and given by $U_{T}({\bf z})_{t}=\rho^{*}=\left(\begin{matrix}1/2&0\\\ 0&1/2\end{matrix}\right).$ (47) Since the Bloch vector for this constant matrix is $(0,0,0)^{\top}$, this shows that in the SAS representation $U_{\widehat{T}_{0}}({\bf z})_{t}=(0,0,0)^{\top}$. Unital quantum channels are very common in the quantum information literature because of their practical advantages and well-known mathematical properties, see for example [74, 75], and references therein. However, Theorem 5 discards the possibility of relying on unital contractive channels for QRC for long input sequences (see also Section IV for more details). The physical explanation for this behavior stems from the fact that the fixed point of these maps does not depend on the input. Then, after each application of the channel, decoherence always leads to the same stationary state (the maximally mixed state), which does not keep track of these inputs. This hinders any possibility of storing the input information into the degrees of freedom of the quantum system since all coherences fade out and the diagonal elements of the density matrix become equal. We believe that the observation that we just made is very relevant, namely that quantum channels with input-independent fixed points become memoryless in the long-term. This fact is proved in the next theorem that generalizes Theorem 5. ###### Theorem 9. Let $T:\mathcal{B}(\mathcal{H})\times D_{n}\rightarrow\mathcal{B}(\mathcal{H})$ be a QRC system for which there exists an operator norm ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|\cdot\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}$ and $\epsilon>0$ such that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|T(\cdot,{\bf z})|_{\mathcal{B}_{0}(\mathcal{H})}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}<1-\epsilon$, for all ${\bf z}\in D_{n}$. Then, $T$ has an input-independent fixed point $\rho^{\ast}\in\mathcal{S}({\mathcal{H}})$, that is, $T(\rho^{*},{\bf z})=\rho^{*}$, for all ${\bf z}\in D_{n}$, if and only if the corresponding filter $U_{T}$ is constant, that is, $U_{T}({\bf z})_{t}=\rho^{\ast}\in\mathcal{S}({\mathcal{H}})$, for all ${\bf z}\in(D_{n})^{\mathbb{Z}}$ (equivalently, $U_{\widehat{T}_{0}}({\bf z})_{t}=i_{0}^{-1}\left(G_{\mathcal{B}}^{-1}(\rho^{\ast})\right)$) . ###### Proof. We first note that the contractivity hypothesis implies by Proposition 3 that the system associated to $T$ has the ESP and hence has a unique solution for each input. The hypothesis $T(\rho^{*},{\bf z})=\rho^{*}$, for all ${\bf z}\in D_{n}$, obviously implies that the constant sequence equal to $\rho^{*}$ is a solution for any input ${\bf z}\in(D_{n})^{\mathbb{Z}}$ and hence $U_{T}({\bf z})_{t}=\rho^{\ast}\in\mathcal{S}({\mathcal{H}})$, for all ${\bf z}\in(D_{n})^{\mathbb{Z}}$. Conversely, since the filter $U_{T}$ is determined by the recursions $U_{T}({\bf z})_{t}=T\left(U_{T}({\bf z})_{t-1},{\bf z}_{t}\right),$ (48) then, if we always have that $U_{T}({\bf z})_{t}=\rho^{\ast}$, the relation (48) implies that $T(\rho^{*},{\bf z})=\rho^{*}$, for all ${\bf z}\in D_{n}$. ∎ ###### Remark 10. The differences between the hypotheses in Theorems 5 and 9 on fixed points are apparent when the QRC system is expressed using the SAS representation in terms of the functions $q$ and $p$. Indeed, as we saw in the proof of Theorems 5, $q({\bf z})=0$ for and ${\bf z}\in D_{n}$, in that case, and the input dependence takes place only through $p$. This is the case in Example 7. Theorem 9 allows for an input dependence through $q$ too. ###### Example 11. We define a Markovian master equation for the dynamics between input injections: $\dot{\rho}=-i[H({\bf z}_{t}),\rho]+\gamma L\rho L^{\dagger}-\frac{\gamma}{2}\\{L^{\dagger}L,\rho\\},$ (49) where $H({\bf z}_{t})=h({\bf z}_{t})\sigma^{z}/2$ and $L=\sigma^{-}$. This is a single qubit under the influence of an external magnetic field in the $z$ direction with local dissipation. The matrix expression $\widehat{T}$ for the associated system is: $\left(\begin{matrix}1\\\ \braket{\sigma^{x}}_{t}\\\ \braket{\sigma^{y}}_{t}\\\ \braket{\sigma^{z}}_{t}\end{matrix}\right)=\left(\begin{matrix}1&0&0&0\\\ 0&e^{-\frac{\gamma\Delta\tau}{2}}\cos(h_{t}\Delta\tau)&e^{-\frac{\gamma\Delta\tau}{2}}\sin(h_{t}\Delta\tau)&0\\\ 0&-e^{-\frac{\gamma\Delta\tau}{2}}\sin(h_{t}\Delta\tau)&e^{-\frac{\gamma\Delta\tau}{2}}\cos(h_{t}\Delta\tau)&0\\\ e^{-\gamma\Delta\tau}-1&0&0&e^{-\gamma\Delta\tau}\end{matrix}\right)\left(\begin{matrix}1\\\ \braket{\sigma^{x}}_{t-1}\\\ \braket{\sigma^{y}}_{t-1}\\\ \braket{\sigma^{z}}_{t-1}\end{matrix}\right).$ (50) The eigenvalues of $\widehat{T}$ can be computed analytically: $\lambda_{1}=1$, $\lambda_{2}=e^{-\gamma\Delta\tau}$, $\lambda_{3}=e^{-\frac{\gamma\Delta\tau}{2}-ih_{t}\Delta\tau}$ and $\lambda_{4}=e^{-\frac{\gamma\Delta\tau}{2}+ih_{t}\Delta\tau}$. The moduli of the eigenvalues are $|\lambda_{1}|=1$, $|\lambda_{2}|=e^{-\gamma\Delta\tau}<1$ and $|\lambda_{3}|=|\lambda_{4}|=e^{-\frac{\gamma\Delta\tau}{2}}<1$, so $T$ is a mixing channel. In this case, the single fixed point is a pure input- independent state with density matrix $\rho^{*}=\left(\begin{matrix}0&0\\\ 0&1\end{matrix}\right).$ (51) Let us see if it is true that there exists some ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|p(\textbf{z})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}<1$ for all inputs. The singular values of the restriction $p(\textbf{z})=\left(\begin{matrix}e^{-\frac{\gamma\Delta\tau}{2}}\cos(h_{t}\Delta\tau)&e^{-\frac{\gamma\Delta\tau}{2}}\sin(h_{t}\Delta\tau)&0\\\ -e^{-\frac{\gamma\Delta\tau}{2}}\sin(h_{t}\Delta\tau)&e^{-\frac{\gamma\Delta\tau}{2}}\cos(h_{t}\Delta\tau)&0\\\ 0&0&e^{-\gamma\Delta\tau}\end{matrix}\right)$ (52) are $\sigma_{1}=e^{-\gamma\Delta\tau}<1$ and $\sigma_{2}=\sigma_{3}=e^{-\frac{\gamma\Delta\tau}{2}}<1$. Therefore, the system has the ESP and FMP. As we showed in Theorem 9, the associated filter is necessarily constant, and (17) necessarily yields the filter $U_{T}({\bf z})_{t}=\rho^{*}=\left(\begin{matrix}0&0\\\ 0&1\end{matrix}\right).$ (53) Since the Bloch vector for this constant matrix is $(0,0,-1)^{\top}$, this shows that in the SAS representation $U_{\widehat{T}_{0}}({\bf z})_{t}=(0,0,-1)^{\top}$. We can double-check the solution by explicitly computing the filter (31). Take $p({\bf z}_{t})$ as in (52) and $q({\bf z}_{t})^{\top}=(0,0,e^{-\gamma\Delta\tau}-1)$. Since $\sigma_{\text{max}}(p({\bf z}_{t}))<1$, the filter in (31) exists. Using now that $q({\bf z}_{t})$ is input-independent, we can write $\begin{split}U_{\widehat{T}_{0}}({\bf z})_{t}&=\sum^{\infty}_{j=0}\left(\prod^{j-1}_{k=0}p({\bf z}_{t-k})\right)q({\bf z}_{t-j})\\\ &=\left(\sum^{\infty}_{j=0}\prod^{j-1}_{k=0}p({\bf z}_{t-k})\right)q=(e^{-\gamma\Delta\tau}-1)M_{3},\end{split}$ (54) where $M_{3}$ is the third column of the matrix $M=\sum^{\infty}_{j=0}\prod^{j-1}_{k=0}p({\bf z}_{t-k})$. Notice that the third column of the product $\prod^{j-1}_{k=0}p({\bf z}_{t-k})$ is $(0,0,e^{-k\gamma\Delta\tau})^{\top}$. Then, the column $M_{3}$ equals $M_{3}=\left(\begin{matrix}0\\\ 0\\\ \sum^{\infty}_{j=0}e^{-k\gamma\Delta\tau}\end{matrix}\right)=\left(\begin{matrix}0\\\ 0\\\ \frac{1}{1-e^{-\gamma\Delta\tau}}\end{matrix}\right),$ (55) and it hence follows that $U_{\widehat{T}_{0}}({\bf z})_{t}=(0,0,-1)^{\top}$. To conclude, we show in Figure 2 a numerical example of input driving. We choose $h(z_{t})=z_{t}\frac{h}{2}\sigma^{z}$ as the external magnetic field function, where $h$ is a constant and $z_{t}$ is a unidimensional random input. The input will be drawn from a random uniform distribution in the interval $[0,1]$. As can be seen, the observables exhibit a transient time after which they converge to the input-independent stationary state. Figure 2: Dynamics of the spin projections $\braket{\sigma^{a}}_{t}$ for $a=x,y,z$ when driven with a random input sequence with Hamiltonian $H=z_{t}\frac{h}{2}\sigma^{z}$. The initial condition is the maximal coherent state $\rho=1/2\sum^{1}_{i,j=0}\ket{i}\bra{j}$ and the system parameters are $\Delta\tau=1$, $\gamma=1$ and $h=1$. ## IV Discussion Theorems 5 and 9 bring up a connection between long-term computation and noisy intermediate-scale quantum (NISQ) devices: they have a finite time of operation due to decoherence. Consider a model of the type $\rho_{t}=T(\rho_{t},{\bf z}_{t})=\mathcal{E}_{\text{deco}}(\mathcal{U}({\bf z}_{t})\rho_{t-1}\mathcal{U}^{\dagger}({\bf z}_{t})),$ (56) where $\mathcal{E}_{\text{deco}}$ represents the decoherence produced by the contact of the system with an external environment. This model is present in QRC experimental works like [40, 39], where the unitary dynamics $\mathcal{U}({\bf z}_{t})$ is given by a quantum circuit. Theorem 5 explains what happens in the extreme case in which the quantum noise of the device is unital. If the decoherence channel $\mathcal{E}_{\text{deco}}$ is a unital strictly contractive map then $T\left(\frac{I}{d},{\bf z}_{t}\right)=\mathcal{E}_{\text{deco}}\left(\mathcal{U}({\bf z}_{t})\frac{I}{d}\mathcal{U}^{\dagger}({\bf z}_{t})\right)=\frac{\mathcal{E}_{\text{deco}}(I)}{d}=\frac{I}{d},$ that is, $T(\cdot,{\bf z})$ becomes a unital strictly contractive map for all ${\bf z}\in D_{n}$. Theorem 5 shows, in this case, that the filter becomes trivial after the injection of long input sequences. Instances of unital decoherence can be found in depolarizing channels, like in Example 7, or in dephasing channels. A dephasing channel damps the coherences of the density matrix but does not affect the diagonal elements. More generally, Theorem 9 can be interpreted as a “common sense” warning: it explicitly states that the input codification must have a measurable influence on the attractor of the natural dynamics of the CPTP map. Otherwise, there is no possibility of storing input information in the long run. We emphasize that this is independent of the type of dissipation that the quantum channel produces. We can connect the implications of this theorem with the NISQ discussion and (56). If the input is codified in some particular coherences of the system, one must be careful that decoherence does not destroy those matrix elements, because then input-dependence would vanish, and the resulting filter would become trivial. This discussion does not imply that models like (56) are useless. Indeed, the opposite has been proven in previous works for short-term memory tasks [40, 39]. Theorems 5 and 9 only rigorously establish something that was already known about NISQ devices, that is, that there is a coherence time in which they can be exploited. Equivalently, models affected by these theorems have the ESP, but the resulting input/output dynamics becomes trivial for long input sequences. The coherence time limitation affects to all quantum platforms to a greater or lesser extent. Then, either QRC proposals are subject to operate on shorter time scales than the natural noise time scale (as done in [40, 39]), or the QRC system is carefully design to integrate it. For example, as we will see in Section IV.1, dephasing can be integrated as part of a QRC system that is not affected by Theorems 5 and 9. We could further extend this analysis to QRC models with measurements. As an example we take the model proposed in [76]. In this reference, the quantum measurement is applied at each time step after the input dependent CPTP map $T$. The measurement scheme is introduced by modeling an indirect measurement with a continuous-variable ancilla [77, 78, 79, 80, 81], producing a quantum reservoir with stochastic dynamics. For simplicity, we will restrict ourselves to the case of a single qubit, and we will average the quantum states over the limit of infinite measurements, yielding an unconditional state which is led by a deterministic CPTP quantum channel [82]. Besides, we choose, without loss of generality, to take measurements in the $z$ direction. Under these conditions, the CPTP map is $\rho_{t}=M\odot T(\rho_{t-1},\textbf{z}_{t}),$ (57) where $\odot$ represents the Hadamard or element-wise matrix product and $M$ is defined as $M=\begin{pmatrix}1&e^{-\frac{g^{2}}{2}}\\\ e^{-\frac{g^{2}}{2}}&1\end{pmatrix}.$ (58) The measurement strength $g$ allows us to quantify the decoherence introduced by sharp measurements ($g\gg 1$), while for $g\ll 1$ the state is weakly perturbed. It is straightforward to see that this model is introducing dephasing, such that we can rewrite (57) as $\rho_{t}=\mathcal{E}_{\text{deph}}\left(T(\rho_{t-1},\textbf{z}_{t})\right),$ (59) where the dephasing channel $\mathcal{E}_{\text{deph}}$ is defined as $\mathcal{E}_{\text{deph}}(\rho)=e^{-\frac{g^{2}}{2}}\rho+(1-e^{-\frac{g^{2}}{2}})\sigma^{z}\rho\sigma^{z}.$ (60) As we explained above, unitary dynamics for the map $T$ would lead to a memoryless reservoir in the long-term. However, one could engineer a mixing map $T$ such that there is a competition between the attractors of maps $T$ and $\mathcal{E}_{\text{deph}}$. The final fixed point of (59) would be somewhere between the original fixed point of $T$ and a diagonal state (which is the shape of the fixed points of $\mathcal{E}_{\text{deph}}$). We conclude by presenting an example of a qubit with tunable local dissipation that fulfills all the requirements to be a “properly engineered” QRC system. Then, we extend it with the measurement model of [76] to show that it still constitutes a proper QRC system. ### IV.1 A “properly engineered” QRC system We start by introducing the model without measurements. The Markovian master equation that governs the dynamics between input injections is: $\dot{\rho}=-i[H({\bf z}_{t}),\rho]+\gamma L\rho L^{\dagger}-\frac{\gamma}{2}\\{L^{\dagger}L,\rho\\},$ (61) where $H({\bf z}_{t})=h({\bf z}_{t})\sigma^{x}/2$, and $L=\sigma^{-}$. The corresponding matrix expression $\widehat{T}$ is $\left(\begin{matrix}1\\\ \braket{\sigma^{x}}_{t}\\\ \braket{\sigma^{y}}_{t}\\\ \braket{\sigma^{z}}_{t}\end{matrix}\right)=\left(\begin{matrix}1&0&0&0\\\ 0&\widehat{T}_{22}&0&0\\\ \widehat{T}_{31}&0&\widehat{T}_{33}&\widehat{T}_{34}\\\ \widehat{T}_{41}&0&\widehat{T}_{43}&\widehat{T}_{44}\end{matrix}\right)\left(\begin{matrix}1\\\ \braket{\sigma^{x}}_{t-1}\\\ \braket{\sigma^{y}}_{t-1}\\\ \braket{\sigma^{z}}_{t-1}\end{matrix}\right),$ (62) where the expressions for the matrix elements are shown below: $\begin{matrix}\widehat{T}_{22}=e^{-\frac{\gamma\Delta\tau}{2}},\\\ \widehat{T}_{33}=e^{-\frac{3\gamma\Delta\tau}{4}}\left(\cosh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)+\frac{\gamma}{\sqrt{\gamma^{2}-16h_{t}^{2}}}\sinh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)\right),\\\ \widehat{T}_{44}=e^{-\frac{3\gamma\Delta\tau}{4}}\left(\cosh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)-\frac{\gamma}{\sqrt{\gamma^{2}-16h_{t}^{2}}}\sinh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)\right),\\\ \widehat{T}_{34}=\frac{4h_{t}e^{-\frac{3\gamma\Delta\tau}{4}}}{\sqrt{\gamma^{2}-16h_{t}^{2}}}\sinh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right),\\\ \widehat{T}_{43}=-\widehat{T}_{34},\\\ \widehat{T}_{31}=\frac{2\gamma h_{t}}{\gamma^{2}+2h^{2}_{t}}\left\\{-1+e^{-\frac{3\gamma\Delta\tau}{4}}\left(\cosh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)+\frac{3\gamma}{\sqrt{\gamma^{2}-16h_{t}^{2}}}\sinh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)\right)\right\\},\\\ \widehat{T}_{41}=\frac{\gamma}{\gamma^{2}+2h^{2}_{t}}\left\\{-\gamma+e^{-\frac{3\gamma\Delta\tau}{4}}\left(\gamma\cosh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)-\frac{\gamma^{2}+8h^{2}_{t}}{\sqrt{\gamma^{2}-16h_{t}^{2}}}\sinh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)\right)\right\\}.\end{matrix}$ (63) Figure 3: Density plot of the singular values (a) $\sigma_{2}$ and (b) $\sigma_{3}$. The elements $\gamma=4h_{t}$ are not determined because of the denominator $1/(\gamma^{2}-16h^{2}_{t})$. The eigenvalues of matrix $\hat{T}$ are: $\lambda_{1}=1$, $\lambda_{2}=e^{-\frac{\gamma\Delta\tau}{2}}$, $\lambda_{3}=e^{-(3\gamma^{2}+\sqrt{\gamma^{2}-16h^{2}_{t}})\frac{\Delta\tau}{4}}$ and $\lambda_{4}=e^{-(3\gamma^{2}-\sqrt{\gamma^{2}-16h^{2}_{t}})\frac{\Delta\tau}{4}}$. The three eigenvalues $\lambda_{2}$, $\lambda_{3}$ and $\lambda_{4}$ have always modulus smaller than one when $\gamma\neq 0$. Then, the master equation fulfills the conditions for having a single full-rank fixed point [62], whose density matrix is $\rho^{*}=\frac{1}{\gamma^{2}+2h^{2}_{t}}\left(\begin{matrix}h^{2}_{t}&i\gamma h_{t}\\\ -i\gamma h_{t}&\gamma^{2}+h^{2}_{t}\end{matrix}\right).$ (64) Finally, we compute the singular values of the restriction to the traceless hyperplane. These values are $\sigma_{1}=e^{-\frac{\gamma\Delta\tau}{2}}<1$, $\sigma_{2}=e^{-\frac{3\gamma\Delta\tau}{4}}\sqrt{f_{+}}$ and $\sigma_{3}=e^{-\frac{3\gamma\Delta\tau}{4}}\sqrt{f_{-}}$, where $\begin{split}f_{\pm}&=\frac{1}{\gamma^{2}-16h^{2}_{t}}\left(-16h^{2}_{t}+\gamma^{2}\cosh\left(\frac{\Delta\tau}{2}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)\right.\\\ &\left.\pm\gamma\sqrt{\sinh^{2}\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)}\sqrt{-64h^{2}_{t}+2\gamma^{2}+2\gamma^{2}\cosh\left(\frac{\Delta\tau}{2}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)}\right).\end{split}$ (65) Figure 3 shows that for $\gamma\neq 4h_{t}$ and away from the axis $\gamma=0$, we find $\sigma_{2},\sigma_{3}<1$, demonstrating the ESP and the FMP. Since the fixed point $\rho^{*}$ is input dependent, this system exhibits the necessary ingredients to be a competent QRC system. As in Example 11, we show in Figure 4 a numerical example of input driving. We choose again $h(z_{t})=z_{t}\frac{h}{2}\sigma^{x}$ as the external magnetic field function, modifying its direction. Now, the observables $\braket{\sigma^{z}}$ and $\braket{\sigma^{y}}$ exhibit an explicit response to the driving, while $\braket{\sigma^{x}}$ converges to its input-independent stationary value (which can be predicted from (62)). Figure 4: Dynamics of the spin projections $\braket{\sigma^{a}}_{t}$ for $a=x,y,z$ when driven with a random input sequence with Hamiltonian $H=z_{t}\frac{h}{2}\sigma^{x}$. The initial condition is the maximal coherent state $\rho=1/2\sum^{1}_{i,j=0}\ket{i}\bra{j}$ and the system parameters are $\Delta\tau=1$, $\gamma=1$ and $h=1$. Now we further extend the model to incorporate the measurement formalism described in [76]. As the composition of CPTP maps can be described as the product of their matrix representations [58], we just need to obtain the Kraus operators of (60), which are $K_{0}=\sqrt{e^{-\frac{g^{2}}{2}}}I$, $K_{1}=\sqrt{1-e^{-\frac{g^{2}}{2}}}\ket{0}\bra{0}$ and $K_{2}=\sqrt{1-e^{-\frac{g^{2}}{2}}}\ket{1}\bra{1}$, where $\ket{0}$ and $\ket{1}$ are the basis states in the $z$ axis. The matrix representation of $\mathcal{E}_{\text{deph}}$ in the Pauli matrix basis is then $\widehat{\mathcal{E}}_{\text{deph}}=\left(\begin{matrix}1&0&0&0\\\ 0&e^{-\frac{g^{2}}{2}}&0&0\\\ 0&0&e^{-\frac{g^{2}}{2}}&0\\\ 0&0&0&1\end{matrix}\right).$ (66) It is straightforward to check that the maximum singular value of $\widehat{\mathcal{E}}_{\text{deph}}$ restricted to the traceless hyperplane is equal to one. The final matrix $\widehat{T}^{\prime}=\widehat{\mathcal{E}}_{\text{deph}}\widehat{T}$ is $\widehat{T}^{\prime}=\left(\begin{matrix}1&0&0&0\\\ 0&e^{-\frac{g^{2}}{2}}\widehat{T}_{22}&0&0\\\ e^{-\frac{g^{2}}{2}}\widehat{T}_{31}&0&e^{-\frac{g^{2}}{2}}\widehat{T}_{33}&e^{-\frac{g^{2}}{2}}\widehat{T}_{34}\\\ \widehat{T}_{41}&0&\widehat{T}_{43}&\widehat{T}_{44}\end{matrix}\right).$ (67) Given that the maximum singular value of $p(\textbf{z})$ is smaller than one (for $\gamma\neq 4h_{t}$ and $\gamma\neq 0$), we find that ${\left|\kern-1.07639pt\left|\kern-1.07639pt\left|p^{\prime}(\textbf{z})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{2}\leq{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|p_{\text{deph}}\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{2}\cdot{\left|\kern-1.07639pt\left|\kern-1.07639pt\left|p(\textbf{z})\right|\kern-1.07639pt\right|\kern-1.07639pt\right|}_{2}<1-\epsilon$ for some $\epsilon>0$ in the 2-Schatten norm, where $p(\textbf{z})$, $p^{\prime}(\textbf{z})$ and $p_{\text{deph}}$ are the restrictions to the traceless hyperplane of matrices $\widehat{T}$, $\widehat{T}^{\prime}$ and $\widehat{\mathcal{E}}_{\text{deph}}$ respectively (given by (27)). Then, the QRC system has the ESP and the FMP. The single fixed point of the map is given by $\rho^{*}=\frac{1}{\gamma^{2}+2h^{2}_{t}}\left(\begin{matrix}h^{2}_{t}-f_{1}&i\gamma h_{t}(1-f_{2})\\\ -i\gamma h_{t}(1-f_{2})&\gamma^{2}+h^{2}_{t}+f_{1}\end{matrix}\right),$ (68) where $\begin{split}&f_{1}=\frac{4\gamma h^{2}_{t}\sinh\left(\frac{g^{2}}{4}\right)\sinh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)}{\sqrt{\gamma^{2}-16h_{t}^{2}}\left(\cosh\left(\frac{g^{2}+3\gamma\Delta\tau}{4}\right)-\cosh\left(\frac{g^{2}}{4}\right)\cosh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)\right)+\gamma\sinh\left(\frac{g^{2}}{4}\right)\sinh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)},\\\ &f_{2}=\frac{\sinh\left(\frac{g^{2}}{4}\right)\left(e^{\frac{3\gamma\Delta\tau}{4}}-\cosh\left(\frac{g^{2}}{4}\right)\cosh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)+\frac{\gamma}{\sqrt{\gamma^{2}-16h_{t}^{2}}}\sinh\left(\frac{g^{2}}{4}\right)\sinh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)\right)}{\cosh\left(\frac{g^{2}+3\gamma\Delta\tau}{4}\right)-\cosh\left(\frac{g^{2}}{4}\right)\cosh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)+\frac{\gamma}{\sqrt{\gamma^{2}-16h_{t}^{2}}}\sinh\left(\frac{g^{2}}{4}\right)\sinh\left(\frac{\Delta\tau}{4}\sqrt{\gamma^{2}-16h_{t}^{2}}\right)}.\end{split}$ (69) This fixed point is produced by the competition between the original fixed point in (64) ($g\rightarrow 0$) and a diagonal state ($g\rightarrow\infty$). With this we can conclude that this engineered model, even including the measurement protocol, leads to an operational QRC system in the long-term run. ## V Conclusions In this paper, we have unified the density matrix approach of previous works in QRC with the Bloch vector representation. Moreover, we have shown that these representations are linked by system isomorphisms and that various results concerning the ESP and FMP are independent of the chosen representation. We have also observed that the QRC dynamics in the Bloch vectors representation amounts to that of a state-affine system (SAS) of the type introduced in [45] and for which numerous theoretical results have been established. We have capitalized on this connection to shed some light on fundamental questions in QRC theory in finite dimensions. In particular, we found a necessary and sufficient condition for the ESP and FMP in terms of the existence of an induced norm that bounds the CPTP map for all inputs, determining a guideline for its election. The necessity of this boundedness hypothesis emerges out of the compactness of the input space, which is a common requirement in the RC literature. If the input space is not compact, sufficient conditions can still be found in terms of the weighting sequence [50]. Besides, we described common situations in which QRC systems become useless in long term runs which can be summarized by saying that quantum channels that exhibit input-independent fixed points yield trivial input/output dynamics. Our work sets the grounds for further analysis and exploration of the QRC theory. Future work can follow several paths, such as studying the connection between spectral properties of QRC models and their performance in memory and information processing tasks, studying infinite-dimensional quantum reservoirs, or including generalized measurements (positive operator-valued measures) and the effect of a finite number of measurements in the statistics of expected values, given the effect that they imprint in the resources of QRC algorithms [76]. ## VI Acknowledgments We thank A. Sannia for useful discussion and inspiration for this work, and D. Burgarth and M. Rahaman for answering our questions regarding their work and its relation to our study. We also thank the editor and two anonymous referees for input that has significantly improved the paper. R.M.-P. and J.-P.O. acknowledge partial financial support from the Swiss National Science Foundation (Grant No. 200021 175801/1). R.M.-P. acknowledges the Spanish State Research Agency for support through the Severo Ochoa and María de Maeztu Program for Centers and Units of Excellence in R&D (Grant No. MDM-2017-0711) and through the QUARESC project (Projects No. 2019-109094GB-C21 and 2019-109094GB-C22/AEI/10.13039/501100011033). Part of this work was funded by MICINN/AEI/FEDER and the University of the Balearic Islands through a predoctoral fellowship (Grant No. MDM-2017-0711-18-1) for R.M.-P. R.M.-P. also is grateful for the hospitality of the Division of Mathematical Sciences of the Nanyang Technological University, where most of these results were obtained. ## References * Arute _et al._ [2019] F. Arute, K. Arya, R. Babbush, D. Bacon, J. C. Bardin, R. Barends, R. Biswas, S. Boixo, F. G. Brandao, D. A. Buell, _et al._ , Quantum supremacy using a programmable superconducting processor, Nature 574, 505 (2019). * Choi _et al._ [2016] J.-y. Choi, S. Hild, J. Zeiher, P. Schauß, A. Rubio-Abadal, T. Yefsah, V. Khemani, D. A. Huse, I. Bloch, and C. Gross, Exploring the many-body localization transition in two dimensions, Science 352, 1547 (2016). * Chen _et al._ [2021] Y.-A. Chen, Q. Zhang, T.-Y. Chen, W.-Q. Cai, S.-K. Liao, J. Zhang, K. Chen, J. Yin, J.-G. Ren, Z. Chen, _et al._ , An integrated space-to-ground quantum communication network over 4,600 kilometres, Nature 589, 214 (2021). * Bharti _et al._ [2022] K. Bharti, A. Cervera-Lierta, T. H. Kyaw, T. Haug, S. Alperin-Lea, A. Anand, M. Degroote, H. Heimonen, J. S. Kottmann, T. Menke, _et al._ , Noisy intermediate-scale quantum algorithms, Reviews of Modern Physics 94, 015004 (2022). * Zhong _et al._ [2020] H.-S. Zhong, H. Wang, Y.-H. Deng, M.-C. Chen, L.-C. Peng, Y.-H. Luo, J. Qin, D. Wu, X. Ding, Y. Hu, _et al._ , Quantum computational advantage using photons, Science 370, 1460 (2020). * Madsen _et al._ [2022] L. S. Madsen, F. Laudenbach, M. F. Askarani, F. Rortais, T. Vincent, J. F. Bulmer, F. M. Miatto, L. Neuhaus, L. G. Helt, M. J. Collins, _et al._ , Quantum computational advantage with a programmable photonic processor, Nature 606, 75 (2022). * Biamonte _et al._ [2017] J. Biamonte, P. Wittek, N. Pancotti, P. Rebentrost, N. Wiebe, and S. Lloyd, Quantum machine learning, Nature 549, 195 (2017). * Liu _et al._ [2021] Y. Liu, S. Arunachalam, and K. Temme, A rigorous and robust quantum speed-up in supervised machine learning, Nature Physics 17, 1013 (2021). * Huang _et al._ [2022] H.-Y. Huang, M. Broughton, J. Cotler, S. Chen, J. Li, M. Mohseni, H. Neven, R. Babbush, R. Kueng, J. Preskill, _et al._ , Quantum advantage in learning from experiments, Science 376, 1182 (2022). * Cybenko [1989] G. Cybenko, Approximation by superpositions of a sigmoidal function, Mathematics of control, signals and systems 2, 303 (1989). * Hornik _et al._ [1989] K. Hornik, M. Stinchcombe, and H. White, Multilayer feedforward networks are universal approximators, Neural networks 2, 359 (1989). * Hornik [1991] K. Hornik, Approximation capabilities of multilayer feedforward networks, Neural networks 4, 251 (1991). * Schäfer and Zimmermann [2006] A. M. Schäfer and H. G. Zimmermann, Recurrent neural networks are universal approximators, in _International Conference on Artificial Neural Networks_ (Springer, 2006) pp. 632–640. * Hammer and Gersmann [2003] B. Hammer and K. Gersmann, A note on the universal approximation capability of support vector machines, neural processing letters 17, 43 (2003). * Huang _et al._ [2006] G.-B. Huang, L. Chen, C. K. Siew, _et al._ , Universal approximation using incremental constructive feedforward networks with random hidden nodes, IEEE Trans. Neural Networks 17, 879 (2006). * Gonon _et al._ [2022a] L. Gonon, L. Grigoryeva, and J.-P. Ortega, Approximation error estimates for random neural networks and reservoir systems, To appear in the Annals of Applied Probability (2022a). * Steinwart [2001] I. Steinwart, On the influence of the kernel on the consistency of support vector machines, Journal of machine learning research 2, 67 (2001). * Micchelli _et al._ [2006] C. A. Micchelli, Y. Xu, and H. Zhang, Universal Kernels, Journal of Machine Learning Research 7, 2651 (2006). * Gonon _et al._ [2022b] L. Gonon, L. Grigoryeva, and J.-P. Ortega, Reservoir kernels and Volterra series, Preprint (2022b). * Pérez-Salinas _et al._ [2020] A. Pérez-Salinas, A. Cervera-Lierta, E. Gil-Fuster, and J. I. Latorre, Data re-uploading for a universal quantum classifier, Quantum 4, 226 (2020). * Pérez-Salinas _et al._ [2021] A. Pérez-Salinas, D. López-Núñez, A. García-Sáez, P. Forn-Díaz, and J. I. Latorre, One qubit as a universal approximant, Physical Review A 104, 012405 (2021). * Schuld _et al._ [2021] M. Schuld, R. Sweke, and J. J. Meyer, Effect of data encoding on the expressive power of variational quantum-machine-learning models, Physical Review A 103, 032430 (2021). * Goto _et al._ [2021] T. Goto, Q. H. Tran, and K. Nakajima, Universal approximation property of quantum machine learning models in quantum-enhanced feature spaces, Physical Review Letters 127, 090506 (2021). * Jaeger [2001] H. Jaeger, The “echo state” approach to analysing and training recurrent neural networks-with an erratum note, Bonn, Germany: German National Research Center for Information Technology GMD Technical Report 148, 13 (2001). * Lukoševičius and Jaeger [2009] M. Lukoševičius and H. Jaeger, Reservoir computing approaches to recurrent neural network training, Computer Science Review 3, 127 (2009). * Tanaka _et al._ [2019] G. Tanaka, T. Yamane, J. B. Héroux, R. Nakane, N. Kanazawa, S. Takeda, H. Numata, D. Nakano, and A. Hirose, Recent advances in physical reservoir computing: A review, Neural Networks 115, 100 (2019). * Jaeger and Haas [2004] H. Jaeger and H. Haas, Harnessing nonlinearity: Predicting chaotic systems and saving energy in wireless communication, science 304, 78 (2004). * Grigoryeva _et al._ [2014] L. Grigoryeva, J. Henriques, L. Larger, and J.-P. Ortega, Stochastic time series forecasting using time-delay reservoir computers: performance and universality, Neural Networks 55, 59 (2014). * Shahi _et al._ [2022] S. Shahi, F. H. Fenton, and E. M. Cherry, Prediction of chaotic time series using recurrent neural networks and reservoir computing techniques: A comparative study, Machine learning with applications 8, 100300 (2022). * Pathak _et al._ [2018] J. Pathak, A. Wikner, R. Fussell, S. Chandra, B. R. Hunt, M. Girvan, and E. Ott, Hybrid forecasting of chaotic processes: Using machine learning in conjunction with a knowledge-based model, Chaos 28, 10.1063/1.5028373 (2018), arXiv:1803.04779 . * Vlachas _et al._ [2020] P. R. Vlachas, J. Pathak, B. R. Hunt, T. P. Sapsis, M. Girvan, E. Ott, and P. Koumoutsakos, Backpropagation algorithms and reservoir computing in recurrent neural networks for the forecasting of complex spatiotemporal dynamics, Neural Networks 126, 191 (2020). * Tran and Nakajima [2020] Q. H. Tran and K. Nakajima, Higher-order quantum reservoir computing, arXiv preprint arXiv:2006.08999 (2020). * Wikner _et al._ [2021] A. Wikner, J. Pathak, B. R. Hunt, I. Szunyogh, M. Girvan, and E. Ott, Using data assimilation to train a hybrid forecast system that combines machine-learning and knowledge-based components, Chaos: An Interdisciplinary Journal of Nonlinear Science 31, 53114 (2021). * Fujii and Nakajima [2017] K. Fujii and K. Nakajima, Harnessing disordered-ensemble quantum dynamics for machine learning, Physical Review Applied 8, 024030 (2017). * Ghosh _et al._ [2021] S. Ghosh, K. Nakajima, T. Krisnanda, K. Fujii, and T. C. Liew, Quantum neuromorphic computing with reservoir computing networks, Advanced Quantum Technologies 4, 2100053 (2021). * Mujal _et al._ [2021] P. Mujal, R. Martínez-Peña, J. Nokkala, J. García-Beni, G. L. Giorgi, M. C. Soriano, and R. Zambrini, Opportunities in quantum reservoir computing and extreme learning machines, Advanced Quantum Technologies 4, 2100027 (2021). * Chen _et al._ [2020] J. Chen, H. I. Nurdin, and N. Yamamoto, Temporal information processing on noisy quantum computers, Physical Review Applied 14, 024065 (2020). * Dasgupta _et al._ [2020] S. Dasgupta, K. E. Hamilton, and A. Banerjee, Designing a nisq reservoir with maximal memory capacity for volatility forecasting, arXiv preprint arXiv:2004.08240 (2020). * Suzuki _et al._ [2022] Y. Suzuki, Q. Gao, K. C. Pradel, K. Yasuoka, and N. Yamamoto, Natural quantum reservoir computing for temporal information processing, Scientific reports 12, 1 (2022). * Kubota _et al._ [2022] T. Kubota, Y. Suzuki, S. Kobayashi, Q. H. Tran, N. Yamamoto, and K. Nakajima, Quantum noise-induced reservoir computing, arXiv preprint arXiv:2207.07924 (2022). * Molteni _et al._ [2022] R. Molteni, C. Destri, and E. Prati, Optimization of the memory reset rate of a quantum echo-state network for time sequential tasks, arXiv preprint arXiv:2210.01052 (2022). * Chen and Nurdin [2019] J. Chen and H. I. Nurdin, Learning nonlinear input–output maps with dissipative quantum systems, Quantum Information Processing 18, 1 (2019). * Nokkala _et al._ [2021] J. Nokkala, R. Martínez-Peña, G. L. Giorgi, V. Parigi, M. C. Soriano, and R. Zambrini, Gaussian states of continuous-variable quantum systems provide universal and versatile reservoir computing, Communications Physics 4, 1 (2021). * Grigoryeva and Ortega [2018a] L. Grigoryeva and J.-P. Ortega, Echo state networks are universal, Neural Networks 108, 495 (2018a). * Grigoryeva and Ortega [2018b] L. Grigoryeva and J.-P. Ortega, Universal discrete-time reservoir computers with stochastic inputs and linear readouts using non-homogeneous state-affine systems, Journal of Machine Learning Research 19, 1 (2018b). * Gonon and Ortega [2019] L. Gonon and J.-P. Ortega, Reservoir computing universality with stochastic inputs, IEEE transactions on neural networks and learning systems 31, 100 (2019). * Gonon and Ortega [2021] L. Gonon and J.-P. Ortega, Fading memory echo state networks are universal, Neural Networks 138, 10 (2021). * Boyd and Chua [1985] S. Boyd and L. Chua, Fading memory and the problem of approximating nonlinear operators with volterra series, IEEE Transactions on circuits and systems 32, 1150 (1985). * Wright and McMahon [2020] L. G. Wright and P. L. McMahon, The capacity of quantum neural networks, in _CLEO: Science and Innovations_ (Optica Publishing Group, 2020) pp. JM4G–5. * Grigoryeva and Ortega [2019] L. Grigoryeva and J.-P. Ortega, Differentiable reservoir computing, Journal of Machine Learning Research 20, 1 (2019). * Chen [2022] J. Chen, _Nonlinear Convergent Dynamics for Temporal Information Processing on Novel Quantum and Classical Devices_ , Ph.D. thesis, UNSW Sydney (2022). * Tran and Nakajima [2021] Q. H. Tran and K. Nakajima, Learning temporal quantum tomography, Physical Review Letters 127, 260401 (2021). * Rédei and Summers [2007] M. Rédei and S. J. Summers, Quantum probability theory, Studies in History and Philosophy of Science Part B: Studies in History and Philosophy of Modern Physics 38, 390 (2007). * Bengtsson and Życzkowski [2017] I. Bengtsson and K. Życzkowski, _Geometry of quantum states: an introduction to quantum entanglement_ (Cambridge university press, 2017). * Raginsky [2002] M. Raginsky, _Dynamical aspects of information storage in quantum-mechanical systems_ (Northwestern University, 2002). * Burgarth and Giovannetti [2007] D. Burgarth and V. Giovannetti, The generalized lyapunov theorem and its application to quantum channels, New Journal of Physics 9, 150 (2007). * Burgarth _et al._ [2013] D. Burgarth, G. Chiribella, V. Giovannetti, P. Perinotti, and K. Yuasa, Ergodic and mixing quantum channels in finite dimensions, New Journal of Physics 15, 073045 (2013). * Wolf [2012] M. M. Wolf, Quantum channels & operations: Guided tour, Lecture notes available at http://www-m5. ma. tum. de/foswiki/pub M 5 (2012). * Horn and Johnson [2013] R. A. Horn and C. R. Johnson, _Matrix Analysis_ , 2nd ed. (Cambridge University Press, 2013) p. 643. * Sanz _et al._ [2010] M. Sanz, D. Perez-Garcia, M. M. Wolf, and J. I. Cirac, A quantum version of wielandt’s inequality, IEEE Transactions on Information Theory 56, 4668 (2010). * Rahaman [2019] M. Rahaman, A new bound on quantum wielandt inequality, IEEE Transactions on Information Theory 66, 147 (2019). * Nigro [2019] D. Nigro, On the uniqueness of the steady-state solution of the lindblad–gorini–kossakowski–sudarshan equation, Journal of Statistical Mechanics: Theory and Experiment 2019, 043202 (2019). * Kimura [2003] G. Kimura, The bloch vector for n-level systems, Physics Letters A 314, 339 (2003). * Byrd and Khaneja [2003] M. S. Byrd and N. Khaneja, Characterization of the positivity of the density matrix in terms of the coherence vector representation, Physical Review A 68, 062322 (2003). * Gonon _et al._ [2020] L. Gonon, L. Grigoryeva, and J.-P. Ortega, Memory and forecasting capacities of nonlinear recurrent networks, Physica D 414, 1 (2020). * Grigoryeva and Ortega [2021] L. Grigoryeva and J.-P. Ortega, Dimension reduction in recurrent networks by canonicalization, Journal of Geometric Mechanics 13, 647 (2021). * Munkres [2000] J. R. Munkres, _Topology_ , Vol. 2 (Prentice Hall Upper Saddle River, 2000). * Siewert [2022] J. Siewert, On orthogonal bases in the hilbert-schmidt space of matrices, Journal of Physics Communications 6, 055014 (2022). * Cuchiero _et al._ [2021] C. Cuchiero, L. Gonon, L. Grigoryeva, J.-P. Ortega, and J. Teichmann, Discrete-time signatures and randomness in reservoir computing, IEEE Transactions on Neural Networks and Learning Systems 10.1109/TN, 1 (2021). * Buehner and Young [2006] M. Buehner and P. Young, A tighter bound for the echo state property, IEEE transactions on neural networks 17, 820 (2006). * Hartfiel [2002] D. J. Hartfiel, _Nonhomogeneous matrix products_ (World Scientific, 2002). * Perez-Garcia _et al._ [2006] D. Perez-Garcia, M. M. Wolf, D. Petz, and M. B. Ruskai, Contractivity of positive and trace-preserving maps under $l^{p}$ norms, Journal of Mathematical Physics 47, 083506 (2006). * Andersson _et al._ [2007] E. Andersson, J. D. Cresser, and M. J. Hall, Finding the kraus decomposition from a master equation and vice versa, Journal of Modern Optics 54, 1695 (2007). * Mendl and Wolf [2009] C. B. Mendl and M. M. Wolf, Unital quantum channels–convex structure and revivals of birkhoff’s theorem, Communications in Mathematical Physics 289, 1057 (2009). * Watrous [2018] J. Watrous, _The theory of quantum information_ (Cambridge university press, 2018). * Mujal _et al._ [2022] P. Mujal, R. Martínez-Peña, G. L. Giorgi, M. C. Soriano, and R. Zambrini, Time series quantum reservoir computing with weak and projective measurements, arXiv preprint arXiv:2205.06809 (2022). * Naghiloo [2019] M. Naghiloo, Introduction to experimental quantum measurement with superconducting qubits, arXiv preprint arXiv:1904.09291 (2019). * Murch _et al._ [2013] K. W. Murch, S. J. Weber, C. Macklin, and I. Siddiqi, Observing single quantum trajectories of a superconducting quantum bit, Nature 502, 211 (2013). * Weber _et al._ [2016] S. J. Weber, K. W. Murch, M. E. Kimchi-Schwartz, N. Roch, and I. Siddiqi, Quantum trajectories of superconducting qubits, Comptes Rendus Physique 17, 766 (2016), quantum microwaves / Micro-ondes quantiques. * Lecocq _et al._ [2021] F. Lecocq, F. Quinlan, K. Cicak, J. Aumentado, S. A. Diddams, and J. D. Teufel, Control and readout of a superconducting qubit using a photonic link, Nature 591, 575 (2021). * Hatridge _et al._ [2013] M. Hatridge, S. Shankar, M. Mirrahimi, F. Schackert, K. Geerlings, T. Brecht, K. M. Sliwa, B. Abdo, L. Frunzio, S. M. Girvin, R. J. Schoelkopf, and M. H. Devoret, Quantum back-action of an individual variable-strength measurement, Science 339, 178 (2013). * Wiseman and Milburn [2009] H. M. Wiseman and G. J. Milburn, _Quantum measurement and control_ (Cambridge university press, 2009).
# Fusion systematics for weakly bound nuclei using neutron flow and collective degrees of freedom S<EMAIL_ADDRESS>V. V<EMAIL_ADDRESS>V<EMAIL_ADDRESS>S<EMAIL_ADDRESS>1Department of Nuclear Physics, Andhra University, Visakhapatnam - 530003, India 2Nuclear Physics Division, Bhabha Atomic Research Centre, Mumbai - 400085, India 3Homi Bhabha National Institute, Anushaktinagar, Mumbai - 400094, India 4UM-DAE Centre for Excellence in Basic Sciences, Mumbai - 400098, India ###### Abstract A systematic analysis of the fusion cross sections around the Coulomb barrier energies with stable weakly bound (6Li,7Li,9Be) and strongly bound 12C projectiles on various targets was performed by using neutron flow model and coupled channels approach. The analysis show that both the models are successful in explaining the near barrier fusion data. Further, it is also observed that the collective degrees of freedom as well as the neutron flow influence the near barrier fusion process involving weakly bound projectiles. ## I Introduction Fusion of two nuclei at energies below the Coulomb barrier is a topic of great interest due to its relation in understanding nucleosynthesis of elements and probing underlying rich nuclear dynamics. It is well established that the fusion cross-sections are significantly enhanced at sub-barrier energies compared to the one dimensional barrier penetration model (1DBPM) Dasgupta et al. (1998); Beckerman (1988); Balantekin and Takigawa (1998). The enhancement of the fusion cross-sections is explained due to coupling of relative motion with the various internal degrees of freedom of the interacting nuclei Zagrebaev (2003); Stephanini et al. (1995); Bierman et al. (1996). In addition to this, the role of neutron rearrangement process on the fusion cross- sections has been also explored for various projectile-target systems Zagrebaev (2003). There are several theoretical prescriptions proposed to describe the enhancement in the fusion cross-sections at near barrier energies and it is observed that two different models are very much successful in explaining the enhancement in the cross-sections Dasso et al. (1983); Broglia et al. (1983); Stelson (1988); Stelson et al. (1990). The first model is the Coupled Channels (CC) model, which is based on the effect of couplings of the incident projectile and the target which leads to lowering the barrier height of the interacting system. Because of reduction of barrier height, an enhancement of the fusion cross-section compared to the 1DBPM Dasso et al. (1983); Broglia et al. (1983) is observed. The second model is the Stelson model, which is based on the neutron flow due to the exchange of neutrons between the interacting projectile and target nuclei Stelson (1988); Stelson et al. (1990). It is observed that both these models predict different barrier distributions for various reactions studied. In the past, there were only a few studies carried out in order to understand the enhancement of the sub barrier fusion cross- sections by using both the models Kailas and Navin (1993); Vinodkumar et al. (1996); Shapira and Stelson (1993); Vandenbosch (1992) and make a comparison between these models. Systematic study of several reactions using Stelson model indicates that there is an early onset of free neutron flow between the interacting projectile and the target at relatively larger inter nuclear distances. This leads to an enhancement of the fusion cross-sections at below barrier energies and the correlation of barrier shifts for various systems can also be successfully explained by this model Shapira and Stelson (1993); Vinodkumar et al. (1996). On the other hand, the coupled channel approach is well established and widely used method of calculating sub-barrier fusion, which is very successful in explaining the experimental data for a variety of nuclear systems Kailas and Navin (1993). There have been only a few attempts in literature, in order to make comparative study of these two methods and identify which of these mechanisms is more appropriate to explain the experimental data on sub barrier fusion. In one of the studies, it was reported that the coupled channels method is better correlated with the experimental data, when compared to the neutron flow model Kailas and Navin (1993). In the past, reactions involving strongly bound stable projectiles were investigated by using both the models discussed above. These investigations show that there is a very good correlation between the experimental data and the theoretical calculations Kailas and Navin (1993); Vinodkumar et al. (1996); Shapira and Stelson (1993); Vandenbosch (1992). It is very interesting to make similar study for the reactions involving weakly bound projectiles (WBP) and compare the results with the systematics of strongly bound projectile (SBP) on various targets. It is well known that for the reactions involving WBPs such as, 6Li,7Li and 9Be, due to their small breakup thresholds, there is an enhancement in the fusion cross-sections at sub barrier energies. There are several reports in literature showing that the reaction cross-sections involving WBPs were very high around the barrier energies Jha et al. (2020); Canto et al. (2006). In this context, it is very much interesting to investigate the systematic of WBPs on various targets by using both the formalisms discussed above Kailas and Navin (1993). In this paper, we investigate the systematics of WBPs on various targets by using both the coupled channel and Stelson model and compare the results with SBP 12C on various targets. The paper is organized as follows: In section II, the methodology for neutron flow based model and coupled channel models for calculation of fusion cross-sections are described. In section III, the results from WBP and SBP 12C projectile on various targets has been presented. The summary and conclusions are given in section IV. ## II Methodology In this paper, the complete fusion (CF) cross-sections for 6Li,7Li, 9Be and 12C projectiles on various targets was analyzed by using neutron flow model and coupled channels model. For completeness of the paper, a brief introduction about the two models was discussed here. The well known expression for the fusion cross-section for reactions with projectile energies greater than the barrier (B) can be written as $\sigma_{fus}=\pi R_{b}^{2}(1-\frac{B}{E})$ (1) where B, Rb, E are the Coulomb barrier, Coulomb radius and energy in the c.m. frame respectively Frobrich and Lipperheide (1996). According to the Stelson model, at near barrier energies the fusion barriers can be explained by a flat distribution of barriers with a threshold energy cutoff (Texp) Stelson (1988); Stelson et al. (1990). This barrier corresponds to the energy at which the interacting nuclei come sufficiently close to each other for neutrons to flow freely between target and projectile. So, the above expression transforms at near barrier energies to $\sigma_{fus}=\pi R_{b}^{2}\frac{(E-T_{exp})^{2}}{4E(B-T_{exp})}$ (2) The maximum value of the merged neutron potential Vmax can be calculated by assuming the neutron shell potential centered on each of the interacting nuclei. In this configuration, the distance between the interacting nuclei is given by Rt, the distance at which the threshold barrier (Tcal) is reached. According to this model, if the merged neutron potential (Vmax) is equal or lower than the binding energy of the valence neutron of the two interacting nuclei then only the neutron flow takes place. Further the extent of the barriers discussed above B - Texp are correlated by the difference between the Coulomb barrier and the threshold barrier B - Tcal. In the coupled channel formalism, it is well established that the coupling between the incident projectile and target channels (vibrational, rotational, transfer) can modify the barrier heights Dasso et al. (1983); Broglia et al. (1983). By including the transmission probabilities and strength of the couplings (F) through the modified barriers, the fusion cross-section can be calculated. If only inelastic couplings are considered in the analyses, the channel couplings (F) may lead to the decrease or increase in the barrier height and it is expected that the B - Texp values will be related to F. The coupling strength for inelastic excitations to collective states can be calculated from the deformation parameter $\beta_{\lambda,k}$, where $\lambda$ is the multi-polarity of the transition and k is the excited state of the nuclei (target or projectile) Kailas and Navin (1993). By using the above discussed formalism, complete fusion data of 6Li,7Li, 9Be and 12C projectiles on various targets was analyzed. The inelastic states of the target were considered in the coupled channel calculations for all the WBP (6Li,7Li, 9Be) and SBP (12C) systems, while the first projectile inelastic state for 12C (4.4 MeV) and 7Li (0.48 MeV) were considered. In the case 6Li and 9Be, inelastic excitation corresponding to 2.18 MeV and 2.43 MeV resonance states were considered respectively in the coupled channel calculations. Depending upon the correlation plots between B - Texp versus F and B - Texp versus B - Tcal, one can estimate which method is more reliable to explain the fusion cross-sections around the barrier energies. The values of B and $R_{b}$ are calculated by fitting the fusion data at above barrier energies by using the Eq.1 and the value of Texp was calculated by fitting the data at near barrier energies by using Eq.2. In order to calculate the effective value of the strength of the couplings (F), CCFULL code Hagino et al. (1999) was used. Fusion cross section was calculated with and without coupling at deep sub- barrier energies and equated the ratio to exp(F.$\epsilon$) where $\epsilon$ is the barrier curvature. Knowing $\epsilon$, F can be determined. The deformation, multipolarity and transition strengths of the excited states of the projectile/ target used in the CCFULL code are taken from the literature Raman et al. (1987); Kibedi and Spear (2002); Pritychenkov et al. (2016). Previously, it was shown that at sub-barrier energies, this formalism is very much valid in order to extract the effective value of F Kailas and Navin (1993); Landowne and Pieper (1984). B - Tcal values were calculated by the following procedure. The average one/two neutron separation energies for various projectiles and targets were taken from the mass table Wapstra et al. (1988). As the Sn values of WBPs 6,7Li and 9Be are smaller than those of the targets, the neutron flow is from the projectile to the target. However, in the case of 12C as the Sn value is nearly 19 MeV, the neutron flow is from the target to the projectile. By using the neutron potential values given in Ref. Stelson et al. (1990) for both the target and the projectile, the interacting distance Rt between the two has been optimized such that the merged neutron potential Vmax at this distance is equal to the Sn value of the target or projectile (whichever is lower). The Tcal value has been computed using Rt as discussed in Ref. Kailas and Navin (1993). The merged neutron potentials for the two reactions 12C + 208Pb and 9Be + 208Pb calculated are shown in Fig.1. One can observe that, the Vmax values match with the Sn values of 208Pb and 9Be respectively. Here, we have used S2n/2 for the targets to take care of the odd – even effects Kailas and Navin (1993). Figure 1: Merged neutron potentials at Rt, for the reactions 12C + 208Pb (upper panel) and 9Be + 208Pb (lower panel). ## III Results and Discussion By using the above discussed methodologies, the different parameters B, Rb, Texp, Sn, Rt,Tcal and F were extracted for 12C, 6,7Li and 9Be projectiles on various targets. The different parameters calculated by using both the methodologies are given in Table 1. Table 1: Summary of near barrier fusion analysis for 12C, 6,7Li and 9Be projectiles on various targets. Reaction | B | R0 | Texp (err.) | S2n/2 | Rt | Tcal | B - Tcal | B - Texp (err.) | F | Refs. ---|---|---|---|---|---|---|---|---|---|--- | (MeV) | (fm) | (MeV) | Sn (MeV) | (fm) | (MeV) | (MeV) | (MeV) | (MeV) | 12C + 46Ti | 21.09 (0.47) | 9.65 | 18.57 (0.10) | 11.36 | 10.07 | 18.87 | 2.22 | 2.52 (0.48) | 0.34 | Bozek et al. (1986) 12C + 48Ti | 20.32 (0.66) | 7.83 | 17.46 (0.22) | 10.25 | 10.29 | 18.46 | 1.86 | 2.86 (0.70) | 0.32 | Bozek et al. (1986) 12C + 50Ti | 19.32 (1.09) | 7.68 | 17.76 (0.16) | 9.54 | 10.46 | 18.16 | 1.16 | 1.56 (1.10) | 0.27 | Bozek et al. (1986) 12C + 92Zr | 31.90 (0.14) | 9.13 | 29.66 (0.02) | 7.91 | 11.65 | 29.66 | 2.28 | 2.24 (0.14) | 0.41 | Newton et al. (2001) 12C + 144Sm | 45.98 (0.46) | 10.49 | 44.23 (0.08) | 9.56 | 12.39 | 43.22 | 2.75 | 1.75 (0.47) | 0.47 | Abriola et al. (1992) 12C + 152Sm | 47.27 (0.92) | 11.30 | 42.0 (0.19) | 6.92 | 12.99 | 41.24 | 6.03 | 5.27 (0.94) | 1.54 | Broda et al. (1975) 12C + 181Ta | 52.97 (1.17) | 10.83 | 48.62 (0.1) | 7.11 | 13.35 | 47.26 | 5.17 | 4.35 (1.17) | 0.71 | Crippa et al. (1994) 12C + 194Pt | 54.99 (0.51) | 10.54 | 52.35 (0.04) | 7.30 | 13.47 | 50.03 | 4.96 | 2.64 (0.51) | 0.64 | Shrivastava et al. (2001) 12C + 198Pt | 54.86 (0.34) | 10.32 | 52.76 (0.02) | 6.70 | 13.64 | 49.39 | 5.47 | 2.1 (0.34) | 0.81 | Shrivastava et al. (2001) 12C + 204Pb | 54.15 (1.59) | 9.85 | 51.37(0.33) | 7.65 | 13.52 | 52.39 | 1.76 | 2.78 (1.62) | 0.64 | Sagaidak et al. (2003) 12C + 206Pb | 56.75 (2.29) | 11.20 | 53.19 (0.59) | 7.41 | 13.59 | 52.11 | 4.64 | 3.56 (2.37) | 0.65 | Sagaidak et al. (2003) 12C + 208Pb | 55.47 (0.69) | 9.04 | 54.17 (0.06) | 7.05 | 13.69 | 51.75 | 3.72 | 1.3 (0.50) | 0.64 | Mukherjee et al. (2007) 6Li + 28Si | 5.73 (0.7)) | 6.22 | 5.01 (0.12) | 5.66 | 9.84 | 5.27 | 0.46 | 0.72 (0.7) | 0.19 | Mandira Sinha et al. (2010) 6Li + 64Ni | 11.27 (0.9)) | 7.81 | 9.74 (0.06) | | 10.88 | 11.12 | 0.15 | 1.53 (0.9) | 0.29 | Md. Moin Shaikh et al. (2014) 6Li + 90Zr | 18.46 (0.3) | 8.24 | 15.14 (0.07) | | 11.63 | 14.85 | 3.61 | 3.32 (0.3) | 0.31 | Kumawat et al. (2012) 6Li + 124Sn | 20.84 (0.4)) | 8.67 | 17.08 (0.11) | | 12.26 | 17.62 | 3.22 | 3.76 (0.41) | 0.35 | Parkar et al. (2018a) 6Li + 144Sm | 26.03 (0.7)) | 7.43 | 22.71 (0.03) | | 12.58 | 21.29 | 4.73 | 3.32 (0.7) | 0.33 | Rath et al. (2009) 6Li + 152Sm | 25.71 (0.59) | 8.24 | 21.75 (0.04) | | 12.69 | 21.09 | 4.61 | 3.96 (0.6) | 0.33 | Rath et al. (2012) 6Li + 198Pt | 28.2 (1.01) | 8.71 | 24.3 (0.16) | | 13.30 | 25.32 | 2.87 | 3.9 (1.02) | 0.37 | Shrivastava et al. (2009) 6Li + 197Au | 28.42 (0.04) | 8.39 | 25.2 (0.09) | | 13.29 | 27.67 | 0.74 | 3.2 (0.1) | 0.34 | Palshetkar et al. (2014) 6Li + 208Pb | 27.72 (1.07) | 7.41 | 26.58 (0.09) | | 13.42 | 26.39 | 1.33 | 1.14 (1.1) | 0.35 | Wu et al. (2003) 6Li + 209Bi | 29.96 (0.83) | 8.78 | 26.57 (0.10) | | 13.44 | 28.69 | 1.27 | 3.39 (0.84) | 0.35 | Dasgupta et al. (2002) 7Li + 59Co | 11.61 (0.26) | 7.64 | 9.44 (0.1) | 7.25 | 10.66 | 10.94 | 0.68 | 2.17 (0.3) | 0.24 | Beck et al. (2003) 7Li + 124Sn | 21.46 (0.18) | 8.92 | 17.8 (0.03) | | 12.02 | 17.97 | 3.49 | 3.66 (0.18) | 0.29 | Parkar et al. (2018b) 7Li + 144Sm | 24.82 (0.15) | 8.71 | 21.7 (0.06) | | 12.28 | 21.81 | 3.01 | 3.12 (0.16) | 0.31 | Rath et al. (2013) 7Li + 152Sm | 24.18 (0.13) | 8.50 | 20.89 (0.04) | | 12.45 | 21.50 | 2.67 | 3.29 (0.13) | 0.40 | Rath et al. (2013) 7Li + 198Pt | 28.81 (0.45) | 9.71 | 26.09 (0.07) | | 13.06 | 25.79 | 3.02 | 2.71 (0.45) | 0.32 | Shrivastava et al. (2013) 7Li + 197Au | 28.45 (0.99) | 9.96 | 26.09 (0.34) | | 13.05 | 26.15 | 2.30 | 2.36 (1.05) | 0.31 | Palshetkar et al. (2014) 7Li + 209Bi | 29.52 (0.67) | 9.62 | 26.56 (0.10) | | 13.20 | 27.18 | 2.34 | 2.96 (0.68) | 0.33 | Dasgupta et al. (2002) 9Be + 89Y | 22.49 (0.46) | 7.78 | 20.77 (0.08) | 1.66 | 13.66 | 16.44 | 6.05 | 1.72 (0.47) | 0.38 | Palshetkar et al. (2010) 9Be + 124Sn | 26.97 (0.26) | 11.29 | 24.11 (0.04) | | 14.31 | 20.13 | 6.84 | 2.86 (0.26) | 0.29 | Parkar et al. (2010) 9Be + 144Sm | 32.13 (1.38) | 9.43 | 28.75 (0.17) | | 14.63 | 24.42 | 7.71 | 3.38 (1.39) | 0.41 | Gomes et al. (2005) 9Be + 208Pb | 40.19 (0.23) | 9.72 | 36.69 (0.05) | | 15.47 | 30.53 | 9.66 | 3.5 (0.24) | 0.51 | Dasgupta et al. (2004) 9Be + 209Bi | 37.84 (0.94) | 7.29 | 35.23 (0.06) | | 15.48 | 30.88 | 6.96 | 2.61 (0.94) | 0.49 | Dasgupta et al. (2004) As a typical example, the fusion cross section data for the reaction 9Be + 208Pb Dasgupta et al. (2004) along with the Stelson model calculations are shown in Fig. 2. Figure 2: Fusion cross section data available for 9Be + 208Pb system Dasgupta et al. (2004) along with the Stelson model calculations. Continuous and dashed lines are Stelson model fits from the Equations 1 and 2, respectively. Figure 3: The extracted values of B - Texp as a function of F (left panel) and B - Tcal (right panel) for the strongly bound 12C and weakly bound 6,7Li and 9Be projectiles on various targets Bozek et al. (1986); Newton et al. (2001); Abriola et al. (1992); Broda et al. (1975); Crippa et al. (1994); Shrivastava et al. (2001); Sagaidak et al. (2003); Mukherjee et al. (2007); Mandira Sinha et al. (2010); Md. Moin Shaikh et al. (2014); Kumawat et al. (2012); Parkar et al. (2018a); Rath et al. (2009, 2012); Shrivastava et al. (2009); Palshetkar et al. (2014); Wu et al. (2003); Beck et al. (2003); Parkar et al. (2018b); Dasgupta et al. (2002); Rath et al. (2013); Shrivastava et al. (2013); Palshetkar et al. (2010); Parkar et al. (2010); Gomes et al. (2005); Dasgupta et al. (2004). The lines are the best linear fits to the data. Figure 4: The extracted values of Texp values as a function of Tcal for the (a) strongly bound 12C and weakly bound (b) 9Be, (c) 7Li, and (d) 6Li projectiles on various targets. Lines are fit to the data (See text for details). Bozek et al. (1986); Newton et al. (2001); Abriola et al. (1992); Broda et al. (1975); Crippa et al. (1994); Shrivastava et al. (2001); Sagaidak et al. (2003); Mukherjee et al. (2007); Mandira Sinha et al. (2010); Md. Moin Shaikh et al. (2014); Kumawat et al. (2012); Parkar et al. (2018a); Rath et al. (2009, 2012); Shrivastava et al. (2009); Palshetkar et al. (2014); Wu et al. (2003); Beck et al. (2003); Parkar et al. (2018b); Dasgupta et al. (2002); Rath et al. (2013); Shrivastava et al. (2013); Palshetkar et al. (2010); Parkar et al. (2010); Gomes et al. (2005); Dasgupta et al. (2004). The values of the B - Texp as a function of F (left panel) and B - Tcal (right panel) are plotted in Fig. 3. From the present analysis (Fig.3), it can be observed that for the reactions induced by 12C on various targets, the B - Texp values on the average increase with increasing F and the B - Texp values also increase with increasing B - Tcal. The present results are very much similar to the correlation results for reactions induced by 16O projectiles on various targets Kailas and Navin (1993). It can also be observed that the correlation between B - Texp vs F is very much similar when compared to the correlation between B - Texp vs B -Tcal. Further, Fig.3 shows that there is a strong correlation for the data analyzed by using the two models, since the results are well reproduced by a simple linear fit. Followed by these results, the experimental data for the reactions induced by WBPs on various targets was analyzed in a similar way as discussed above for SBP 12C. From Fig.3, it is very interesting to observe a similar trend for WBPs as observed for 12C projectile on various targets. In case of WBPs also, one can observe that there is a good correlation between B - Texp vs B - Tcal and F and all the analyzed experimental data is very well reproduced by linear fits. The $\chi^{2}$ fits for the two models are similar. While both the Coupled Channels calculations and Stelson model describe the fusion data for a large number of systems well, there is an important difference between the two models. As per the Stelson model, for all isotopes of a given Z the Sn values are always lowest for the heaviest isotope, giving lowest T values for it irrespective of collectivity. In contrast, in the case of coupled channel effects due to collectivity, the isotope which is more collective has larger F value leading to lower T values. We have tried to look for this feature also in our data. However, as the B - Texp has large error bars and therefore we are not in a position to confirm the above expectation from our analysis. To understand the neutron transfer mechanism in Stelson model, further the Tcal values for 12C, 6,7Li and 9Be projectiles on various targets as a function of Texp have been plotted in Fig. 4 which shows a very good correlation between Texp and Tcal values. In general, the fusion cross section at near and below barrier energies is enhanced (for both strongly and weakly bound projectiles) when compared to the 1DBPM. Further, there is a small reduction at well above barrier for strongly bound projectiles, while there is a significant reduction 10 to 30 % in the case of weakly bound projectiles. The behavior of fusion excitation function is understood in terms of coupling to various degrees of freedom like collective, transfer and breakup, where the latter two processes are more important for weakly bound nuclei. Diffuseness parameter has been varied by 20 % to see its effect on the Rt and Tcal values. The change in Tcal values with diffuseness parameter varying from 0.65 to 0.8 fm is around 1 MeV. Further, it is observed that a 20 % change in fusion cross section changes the barrier radius by about 10 % and negligible change of the corresponding fusion barrier. Therefore, B-T values are not affected significantly. In order to understand the correlation between Texp and Tcal, the experimental data has been fitted by using linear function. From different kinds of linear fits, one can conclude that for the reactions induced by 12C, 6,7Li projectiles on various targets a simple linear fit function y = x is sufficient to show a very good correlation between the calculated and experimental results. But for the reactions induced by 9Be projectile on various targets a simple linear fit is not sufficient to explain the analyzed data. Good correlation between Texp and Tcal is obtained by using the function y = x + c, where c is a constant. From this analysis, we can clearly conclude that Tcal values are in good agreement with the values of Texp for the reactions induced by 12C, 6,7Li projectile on various targets and for the reactions induced by 9Be projectiles on various targets Texp values are higher by few MeV when compared to the values of Tcal. We need to understand, why this anomaly/different behavior exists for weakly bound 9Be projectile on various targets, when compared to the other reactions. Further, from this analysis it can be concluded that in the case of strongly bound stable projectile 12C and weakly bound 6,7Li, 9Be, both the models (Stelson model and coupled channels formalism) are very much successful in explaining the near barrier fusion data. However from the neutron transfer model, one has to understand why the Texp values are few MeV higher than the Tcal values for weakly bound 9Be projectiles. From this analysis, it can be observed that both these models (Stelson and coupled channels) describe the trend of B - Texp variation successfully for strongly bound and weakly bound projectiles, indicating the validity of these models for sub barrier fusion data. In the Coupled Channels calculations, the excited states of the targets have been included. In the case of 6Li and 9Be, the resonance states have been included as discussed above. Stelson neutron flow model considers only the transfer of neutrons between the interacting nuclei at distances close to the grazing distance. However, the weak or strong binding of the projectile is taken into account is through the neutron binding energy of the interacting nuclei. ## IV Summary and Conclusions In the present paper, the fusion data for weakly bound projectiles (6,7Li and 9Be) and strongly bound projectile 12C around the coulomb barrier energies has been analyzed on various targets. The present results suggest that both the neutron flow and coupled channels models are successful in explaining the near barrier fusion data. Further from the Stelson model, it has been observed that there is a good correlation between Texp and Tcal. For the reactions induced by 9Be projectiles on various targets, the Texp values are a few MeV higher than the corresponding Tcal values and this feature needs to be investigated. The present results show that for reactions induced by WBPs on various targets, the collective degrees of freedom as well as the neutron flow are important in influencing the near barrier fusion phenomenon. In the case of Sm isotopes considered in the present work, while the Sn (and hence the T values) values decrease with increase of the mass number of the isotopes, the collectivity increases with increase of mass number. Hence B-T ( Stelson model) and F (collectivity) values increase with increase of mass number and we can not choose as to which out of the above two models is more appropriate to describe the data. However, if a series of isotopes were choosen, which have both Sn and collectivity decrease with increase of mass number of the target, then it will be possible to select between these two models. In this case, if collectivity is the dominant mechanism, then B-T determined will increase with decrease of mass number as lower mass number has higher collectivity when compared to the heavier. If neutron flow is more important mechanism, then B-T will increase with increase of mass number as the heavier isotope has a lower value of Sn and hence T. It will be interesting to extend these studies to other WBPs and radioactive ions for a range of isotopes of a given target. ## V Acknowledgments The authors SK acknowledges INSA, India for the senior scientist fellowship and SA acknowledges the financial support from UGC, India. ## References * Dasgupta et al. (1998) M. Dasgupta, D. J. Hinde, N. Rowley, and A. M. Stefanini, Annu. Rev. Nucl. Part. Sci. 48, 401 (1998). * Beckerman (1988) M. Beckerman, Rep. Prog. Phys. 51, 1047 (1988). * Balantekin and Takigawa (1998) A. B. Balantekin and N. Takigawa, Rev. Mod. Phys. 70, 77 (1998). * Zagrebaev (2003) V. I. Zagrebaev, Phys. Rev. C 67, 061601[R] (2003). * Stephanini et al. (1995) A. M. Stephanini, D. Ackermann, L. Corradi, D. R. Napoli, C. Petrache, P. Spolaore, P. Bednarczyk, H. Q. Zhang, S. Beghini, G. Montagnoli, et al., Phys. Rev. Lett. 74, 864 (1995). * Bierman et al. (1996) J. D. Bierman, P. Chan, J. F. Liang, M. P. Kelly, A. A. Sonzogni, and R. Vandenbosch, Phys. Rev. Lett. 76, 1587 (1996). * Dasso et al. (1983) C. H. Dasso, S. Landowne, and A. Winther, Nucl. Phys. A 405, 381 (1983). * Broglia et al. (1983) R. A. Broglia, C. H. Dasso, S. Landowne, and G. Pollarolo, Phys. Lett. B 133, 34 (1983). * Stelson (1988) P. H. Stelson, Phys. Lett. B 205, 190 (1988). * Stelson et al. (1990) P. H. Stelson, H. J. Kim, M. Beckerman, D. Shapira, and R. L. Robinson, Phys. Rev. C 41, 1584 (1990). * Kailas and Navin (1993) S. Kailas and A. Navin, Pramana 41, 163 (1993). * Vinodkumar et al. (1996) A. M. Vinodkumar, A. K. Sinha, N. V. S. V. Prasad, K. M. Varier, and P. Sugathan, Phys. Rev. C 54, 791 (1996). * Shapira and Stelson (1993) D. Shapira and P. H. Stelson, Phys. Rev. C 47, 1666 (1993). * Vandenbosch (1992) R. Vandenbosch, Annu. Rev. Nucl. Part. Sci. 42, 447 (1992). * Jha et al. (2020) V. Jha, V. V. Parkar, and S. Kailas, Phys. Rep. 845, 1 (2020), and references therein. * Canto et al. (2006) L. F. Canto, P. R. S. Gomes, R. Donangelo, and M. S. Hussein, Phys. Rep. 424, 1 (2006). * Frobrich and Lipperheide (1996) P. Frobrich and R. Lipperheide, _Theory of Nuclear Reactions_ (Oxford University Press Inc., New York, 1996). * Hagino et al. (1999) K. Hagino, N. Rowley, and A. T. Kruppa, Comput. Phys. Commun. 123, 143 (1999). * Raman et al. (1987) S. Raman, C. H. Malarkey, W. T. Milner, C. W. Nestor, and P. H. Stelson, At. Data Nucl. Data Tables 36, 1 (1987). * Kibedi and Spear (2002) T. Kibedi and R. H. Spear, At. Data Nucl. Data Tables 80, 35 (2002). * Pritychenkov et al. (2016) B. Pritychenkov, M. Birch, B. Singh, and M. Horoi., At. Data Nucl. Data Tables 107, 1 (2016). * Landowne and Pieper (1984) S. Landowne and S. C. Pieper, Phys. Rev. C 29, 1352 (1984). * Wapstra et al. (1988) A. H. Wapstra, G. Audi, and H. Hoekstra, At. Data Nucl. Data Tables 39, 281 (1988). * Bozek et al. (1986) E. Bozek, D. M. D. Castro-Rizzo, S. Cavallaro, B. Delaunay, J. Delaunay, H.Dumont, A. D’onofrio, M. G. Saint-Laurent, L. Sperduto, and F. Terrasi, Nucl. Phys. A 451, 171 (1986). * Newton et al. (2001) J. O. Newton, C. R. Morton, M. Dasgupta, J. R. Leigh, J. C. Mein, D. J. Hinde, H. Timmers, and K. Hagino, Phys. Rev. C 64, 064608 (2001). * Abriola et al. (1992) D. Abriola, A. A. Sonzogni, M. di Tada, A. Etchegoyen, M. C. Etchegoyen, J. O. Fernández Niello, S. Gil, A. O. Macchiavelli, A. J. Pacheco, R. Piegaia, et al., Phys. Rev. C 46, 244 (1992). * Broda et al. (1975) R. Broda, M. Ishihara, B. Herskind, H. Oeschler, S. Ogaza, and H. Ryde, Nucl. Phys. A 248, 356 (1975). * Crippa et al. (1994) M. Crippa, E. Gadioli, P. Vergani, G. Ciavola, M. C., and M. Bonardi, Z. Phys. A 350, 121 (1994). * Shrivastava et al. (2001) A. Shrivastava, S. Kailas, A. Chatterjee, A. Navin, A. M. Samant, P. Singh, S. Santra, B. S. Mahata, K. Tomar, and G. Pollarolo, Phys. Rev. C 63, 054602 (2001). * Sagaidak et al. (2003) R. N. Sagaidak, G. N. Kniajeva, I. M. Itkis, M. G. Itkis, N. A. Kondratiev, E. M. Kozulin, I. V. Pokrovsky, A. I. Svirikhin, V. M. Voskressensky, A. V. Yeremin, et al., Phys. Rev. C 68, 014603 (2003). * Mukherjee et al. (2007) A. Mukherjee, D. J. Hinde, M. Dasgupta, K. Hagino, J. O. Newton, and R. D. Butt, Phys. Rev. C 75, 044608 (2007). * Mandira Sinha et al. (2010) Mandira Sinha, H. Majumdar, P. Basu, Subinit Roy, M. Biswas, M. K. Pradhan, R. Bhattacharya, R. Palit, I. Mazumdar, and S. Kailas, Eur. Phys. J. A 44, 403 (2010). * Md. Moin Shaikh et al. (2014) Md. Moin Shaikh, Subinit Roy, S. Rajbanshi, M. K. Pradhan, A. Mukherjee, P. Basu, S. Pal, V. Nanal, R. G. Pillay, and A. Shrivastava, Phys. Rev. C 90, 024615 (2014). * Kumawat et al. (2012) H. Kumawat, V. Jha, V. V. Parkar, B. J. Roy, S. K. Pandit, R. Palit, P. K. Rath, C. S. Palshetkar, Sushil K. Sharma, Shital Thakur, et al., Phys. Rev. C 86, 024607 (2012). * Parkar et al. (2018a) V. V. Parkar, S. K. Pandit, A. Shrivastava, R. Palit, K. Mahata, V. Jha, K. Ramachandran, Shilpi Gupta, S. Santra, Sushil K. Sharma, et al., Phys. Rev. C 98, 014601 (2018a). * Rath et al. (2009) P. K. Rath, S. Santra, N. L. Singh, R. Tripathi, V. V. Parkar, B. K. Nayak, K. Mahata, R. Palit, Suresh Kumar, S. Mukherjee, et al., Phys. Rev. C 79, 051601 (2009). * Rath et al. (2012) P. K. Rath, S. Santra, N. L. Singh, K. Mahata, R. Palit, B. K. Nayak, K. Ramachandran, V. V. Parkar, R. Tripathi, S. K. Pandit, et al., Nucl. Phys. A 874, 14 (2012). * Shrivastava et al. (2009) A. Shrivastava, A. Navin, A. Lemasson, K. Ramachandran, V. Nanal, M. Rejmund, K. Hagino, T. Ichikawa, S. Bhattacharyya, A. Chatterjee, et al., Phys. Rev. Lett. 103, 232702 (2009). * Palshetkar et al. (2014) C. S. Palshetkar, Shital Thakur, V. Nanal, A. Shrivastava, N. Dokania, V. Singh, V. V. Parkar, P. C. Rout, R. Palit, R. G. Pillay, et al., Phys. Rev. C 89, 024607 (2014). * Wu et al. (2003) Y. W. Wu, Z. H. Liu, C. J. Lin, H. Q. Zhang, M. Ruan, F. Yang, Z. C. Li, M. Trotta, and K. Hagino, Phys. Rev. C 68, 044605 (2003). * Dasgupta et al. (2002) M. Dasgupta, D. J. Hinde, K. Hagino, S. B. Moraes, P. R. S. Gomes, R. M. Anjos, R. D. Butt, A. C. Berriman, N. Carlin, C. R. Morton, et al., Phys. Rev. C 66, 041602 (2002). * Beck et al. (2003) C. Beck, F. A. Souza, N. Rowley, S. J. Sanders, N. Aissaoui, E. E. Alonso, P. Bednarczyk, N. Carlin, S. Courtin, A. Diaz-Torres, et al., Phys. Rev. C 67, 054602 (2003). * Parkar et al. (2018b) V. V. Parkar, Sushil K. Sharma, R. Palit, S. Upadhyaya, A. Shrivastava, S. K. Pandit, K. Mahata, V. Jha, S. Santra, K. Ramachandran, et al., Phys. Rev. C 97, 014607 (2018b). * Rath et al. (2013) P. K. Rath, S. Santra, N. L. Singh, B. K. Nayak, K. Mahata, R. Palit, K. Ramachandran, S. K. Pandit, A. Parihari, A. Pal, et al., Phys. Rev. C 88, 044617 (2013). * Shrivastava et al. (2013) A. Shrivastava, A. Navin, A. Diaz-Torres, V. Nanal, K. Ramachandran, M. Rejmund, S. Bhattacharyya, A. Chatterjee, S. Kailas, A. Lemasson, et al., Phys. Lett. B 718, 931 (2013). * Palshetkar et al. (2010) C. S. Palshetkar, S. Santra, A. Chatterjee, K. Ramachandran, Shital Thakur, S. K. Pandit, K. Mahata, A. Shrivastava, V. V. Parkar, and V. Nanal, Phys. Rev. C 82, 044608 (2010). * Parkar et al. (2010) V. V. Parkar, R. Palit, Sushil K. Sharma, B. S. Naidu, S. Santra, P. K. Joshi, P. K. Rath, K. Mahata, K. Ramachandran, T. Trivedi, et al., Phys. Rev. C 82, 054601 (2010). * Gomes et al. (2005) P. R. S. Gomes, I. Padron, E. Crema, O. A. Capurro, J. O. Fernández Niello, A. Arazi, G. V. Martí, J. Lubian, M. Trotta, A. J. Pacheco, et al., Phys. Rev. C 73, 064606 (2005). * Dasgupta et al. (2004) M. Dasgupta, P. R. S. Gomes, D. J. Hinde, S. B. Moraes, R. M. Anjos, A. C. Berriman, R. D. Butt, N. Carlin, J. Lubian, C. R. Morton, et al., Phys. Rev. C 70, 024606 (2004).
# Rethinking Two Consensuses of Transferability in Deep Learning Yixiong Chen 1, Jingxian Li 2, Chris Ding 1, Li Liu 1,3 Corresponding author: <EMAIL_ADDRESS> ###### Abstract Deep transfer learning (DTL) has formed a long-term quest toward enabling deep neural networks (DNNs) to reuse historical experiences as efficiently as humans. This ability is named knowledge transferability. A commonly used paradigm for DTL is firstly learning general knowledge (pre-training) and then reusing (fine-tuning) them for a specific target task. There are two consensuses of transferability of pre-trained DNNs: (1) a larger domain gap between pre-training and downstream data brings lower transferability; (2) the transferability gradually decreases from lower layers (near input) to higher layers (near output). However, these consensuses were basically drawn from the experiments based on natural images, which limits their scope of application. This work aims to study and complement them from a broader perspective by proposing a method to measure the transferability of pre-trained DNN parameters. Our experiments on twelve diverse image classification datasets get similar conclusions to the previous consensuses. More importantly, two new findings are presented, i.e., (1) in addition to the domain gap, a larger data amount and huge dataset diversity of downstream target task also prohibit the transferability; (2) although the lower layers learn basic image features, they are usually not the most transferable layers due to their domain sensitivity. ## Introduction Deep transfer learning (DTL) enables deep neural networks (DNNs) to learn faster and gain better generalization with limited data. This successful bionic design came from the knowledge transferability of human brains (Ausubel et al. 1968), which are capable of learning new concepts based on the original cognitive structure. DNNs transfer knowledge in a similar way. They learn general recognition abilities from large pre-training datasets (e.g., ImageNet (Deng et al. 2009)) and fine-tune the network parameters on specific downstream datasets (Tajbakhsh et al. 2016). Figure 1: The motivation of our transferability measurement. Pre-training on task A makes the fine-tuning only needs shorter optimization distance to reach the solution of task B than training from scratch. Previous literature (Yosinski et al. 2014; Fayek, Cavedon, and Wu 2018; Tajbakhsh et al. 2016) used semi-quantitative methods (i.e., usually by comparing the network performance with/without transfer learning) to judge the DNNs’ transferabilities. Two consensuses were drawn from these works. Firstly, the transferability decreases as the distance between the source and target dataset (domain gap) increases. We regard its root cause to be the poor transferability of DNNs to the samples dissimilar to pre-training data, which stimulates our interest in exploring other important factors leading to a greater extent of dissimilarity (e.g., dataset diversity). Secondly, the lowest layer (i.e., the layer that is close to the input) learns general knowledge like colors and edges, while the highest layer (i.e., the layer that is close to the output) learns semantic knowledge specific to the task, indicating there is a decreasing trend of transferability from low to high layers. Although this conclusion is generally correct, we conjecture that the pre-trained features of lower layers might not be that applicable to the downstream tasks. If the local pattern distributions of data change drastically (e.g., from natural images to cartoons or even medical data), the shallower local features might suffer more than the deeper semantic features. In these cases, the decreasing trend of layer-wise transferability may need further validations. Instead of using semi-quantitative measuring methods, this work revisits these consensuses with a quantitative method for more precise results. We mainly study the following two questions: * • In addition to the domain gap, what other critical factors about datasets affect the transferability? * • Do the transferabilities for sequential layers of a pre-trained DNN always follow a downward trend? To answer these two questions, we start with the nature of knowledge transfer, a procedure of reusing and adjusting learned abilities for new tasks, to design our measuring method. For DNNs, pre-trained parameters can be regarded as learned knowledge, and it offers a better parameter initialization than random initialization (Fig. 1). In this case, a few gradient steps of fine- tuning can produce good results on a new task (Finn, Abbeel, and Levine 2017). On this basis, we quantify transferability as how much the pre-training task A pushes the parameters towards the optimal point of the downstream task B. If the fine-tuning phase needs a shorter optimization distance than training from scratch, pre-training helps to make the downstream adaptation easier. This quantification method enables us to compare the transferabilities between a pre-training dataset and different downstream tasks under the same standard. It also provides a way to derive the transferabilities of different layers precisely. With this measurement, the above questions can be answered with solid justifications. For the under-explored factors that affect transferability, downstream data diversity (named domain width in this work) and data amount come to our mind for simple reasons: (1) When the domain width of a target dataset goes larger, there must exist more out-of-distribution samples that the DNN did not see in the pre-training phase, which impairs the transferability; (2) For data amount of the downstream task, it is shown by previous works (He, Girshick, and Dollár 2019; Newell and Deng 2020) that the greater amount of data decreases the benefit brought by pre-training. In the experiments, our results verify that domain width and data amount negatively affect the transferability even to a greater extent than the domain gap. For the layer-wise transferability, this work finds that the lowest layers may be domain-sensitive, which makes the trends downward when only excluding the first few layers. To conclude, the main contributions of this work are as follows: 1. 1. A novel method for quantifying the transferability of DNN parameters is proposed. 2. 2. By revisiting the previous consensuses through rich experimental validations on twelve datasets, we find two important phenomena in DTL: (1) the target domain width and data amount highly affect the transferability, even more than the domain gap; (2) the transferabilities of the lowest layers are usually prohibited by the domain shift. For all of our experiments with ResNet50, the lowest layer is not the most transferable among all layers. ## Related Work ### Transferability of DNNs Transferability of DNNs is the ability to gain transferable knowledge from the source domain and reuse the knowledge to decrease the generalization error on the target domain, under the distribution shift or the task discrepancy (Jiang et al. 2022). A common method to assess the transferability of DNNs is comparing the performance between fine-tuning and training from scratch (Yosinski et al. 2014; Matsoukas et al. 2022). Given a downstream task, more significant performance improvement means the DNN has better transferability. There are several key factors affecting how transferable the parameters of a DNN are. First, the performance of the DNN on the pre-training dataset is positively correlated to its transferability to downstream tasks (Kornblith, Shlens, and Le 2019) in most cases. Second, the more similar the source and target task are, the more transferable the DNNs would be (Yosinski et al. 2014; Dwivedi and Roig 2019). Third, the transferabilities of parameters are fragile to the large learning rate (Ro and Choi 2021). A drastic change of highly transferable parameters may destroy their original knowledge, making them less transferable. This work also tries to excavate the properties for DNNs’ transferabilities, but from the perspective of the dataset. ### Pre-training To improve DNNs’ transferabilities, many pre-training datasets and methods are proposed. The most dominant pre-training dataset for vision tasks is ImageNet (Deng et al. 2009) for its large-scale data with high-quality annotations. Earlier literature mainly pre-train models in a supervised way (He, Girshick, and Dollár 2019) with categorical labels. However, researchers found that ImageNet cannot cover the need for all visual applications because the large domain gap between natural and special images (e.g., medical images) leads to poor transferability. Pre-training datasets and methods for specific domains (Sermanet et al. 2018; Chen et al. 2021; Zhang et al. 2022) began to emerge for this reason. This work tests the transferability for large domain gaps, and tries to provide some practical guides when pre-trained models on specific domains are unavailable. Other than training schemes, models themselves are crucial for good transferabilities. Pre-training with big models like convolutional neural networks (CNNs) with hundreds of layers (Huang et al. 2017) and Vision Transformers (ViTs) (Dosovitskiy et al. 2020) are proved to bring more benefits to downstream tasks than smaller ones. The measurement proposed in this work also validates that larger pre-trained models have higher transferabilities (see the supplementary material). Figure 2: The method for measuring parameter transferability. (a) The transferability of the parameter $\theta_{A}$ is defined as $D(\theta_{r},\theta_{B})/D(\theta_{A},\theta_{B})$, but there is an ambiguity of $\theta_{B}$ between fine-tuning and training from scratch. We denote them as $\theta_{FB}$ and $\theta_{SB}$ respectively and use them to calculate two distance variants, $D(\theta_{r},\theta_{FB})$ and $D(\theta_{r},\theta_{SB})$. (b) The transferability of the whole network is calculated as the average transferability of each layer to avoid the parameter scale difference. ### Fine-tuning How to leverage the transferabilities of pre-trained DNNs for a new task has become a research focus in recent years. The practices to transfer feature extraction capability from an off-the-shelf network to a new network can be roughly categorized into two classes: initialization-based and teacher- student-based (Song et al. 2021). The former is the so-called fine-tuning method and is the default setting to measure transferabilities in this work. In addition to the common fine-tuning scheme that involves all layers, another powerful paradigm is tuning the last few layers of a pre-trained DNN (Tajbakhsh et al. 2016), so that the low-level general features inherent in the shallow layers can be kept intact. Likewise, many advanced fine-tuning schemes (Long et al. 2015; Li et al. 2019; Zhong et al. 2020; Ro and Choi 2021) assumed that the higher-level layers are more specific to the task, therefore, are less transferable and need more updates. However, recent research demonstrated some exceptions (Chen et al. 2022) when domain shift is large, where lower layers also need to be adjusted to adapt to the different visual feature distributions. This finding is consistent with our experimental results, which reveal that the first few layers are not necessarily as transferable as researchers previously considered. ## Methodology To answer the above two questions, a general transferability measurement, domain-agnostic, task-agnostic, and architecture-agnostic, is proposed to compare parameter transferabilities considering different domains, tasks, and layers. #### Definition of the Transferability. The aim for pre-training is to find a good initialization parameter $\theta_{A}$ through task A so that the optimal solution $\theta_{B}$ on task B can be obtained by fine-tuning as few steps as possible. We denote the distance between $\theta_{A}$ and $\theta_{B}$ as $D(\theta_{A},\theta_{B})$, which reflects the transferability of $\theta_{A}$ to task B. However, since the optimization difficulties for different parameters in a DNN on different tasks are different, $D(\theta_{A},\theta_{B})$ cannot be regarded as transferability directly. To tackle this problem, the distance between random initialized parameters $\theta_{r}$ and $\theta_{B}$, $D(\theta_{r},\theta_{B})$, is utilized for normalization. This work defines the transferability of random initialized parameters to be 1, and measures the transferability of parameter $\theta_{A}$ to task B as follows: $T_{B}(\theta_{A})=\frac{D(\theta_{r},\theta_{B})}{D(\theta_{A},\theta_{B})}.$ (1) When transferability is greater than 1, pre-training shortens the optimization distance, meaning that new knowledge is learned with the help of experiences. In practice, the distance metric can be calculated with mean absolute difference $D(\theta_{A},\theta_{B})=\sum_{i=1}^{K}|\theta_{A}^{i}-\theta_{B}^{i}|/K$, where $K$ is the number of parameters. #### Two Different Optimal Points. One crucial problem for this measurement is, that in most cases, the pre- trained DNNs would not converge to the same (local) optima as the random initialized ones do. Therefore, this work proposes two variants for measuring $D(\theta_{r},\theta_{B})$ (illustrated in Fig. 2 (a)). In our experiments, we use $\theta_{SB}$ to denote the converged parameters on B trained from scratch, and $\theta_{FB}$ to denote the fine-tuned parameters. Correspondingly, the transferabilities measured with $\theta_{SB}$ and $\theta_{FB}$ are denoted as $T_{SB}(\theta_{A})$ and $T_{FB}(\theta_{A})$ respectively. Table 1: Descriptions and statistics of the datasets used in this work. Dataset | Size (train/test) | Classes | Image description ---|---|---|--- ImageNet | 1,281,167/50,000 | 1000 | Photos of common objects CIFAR-10 | 50,000/10,000 | 10 | Photos of common objects, image sizes $32\times 32$ CIFAR-100 | 50,000/10,000 | 100 | Photos of common objects, image sizes $32\times 32$ Caltech-101 | 3,060/6,084 | 102 | Photos/paintings/sketches of common objects CUB-200 | 5,994/5,794 | 200 | Find-grained photos of birds FGVC Aircraft | 6,667/3,333 | 100 | Find-grained photos of aircrafts Flowers | 1,088/272 | 17 | Find-grained photos of flowers UC Merced Land Use | 1,680/420 | 21 | Remote sensing images of different land uses POCUS | 1,692/424 | 3 | Ultrasound images of COVID-19 DTD | 1,880/1,880 | 47 | Photos mainly contain textures of objects DomainNet real | 120,906/52,041 | 345 | Photos of common objects DomainNet painting | 50,416/21,850 | 345 | Oil Paintings, murals, drawings, tattoos DomainNet clipart | 33,525/14,604 | 345 | Clip art images #### Different Parameter Scales across Layers. Random initialization for modern DNNs considers the dimension of features in each layer an important factor. For the inputs following the standard Gaussian distribution, reasonable random initialization methods (Glorot and Bengio 2010; He et al. 2015) make the corresponding outputs of the layer to be also standard Gaussian. This work adapts Kaiming random initialization (He et al. 2015). It initializes the parameters with Gaussian distribution with mean value $\mu=0$ and variance $\sigma^{2}=2/n_{in}$, where $n_{in}$ is the input dimension. When measuring the transferability of a DNN, the scale difference of parameters between layers causes an imbalance. The shallow layers of DNNs usually have fewer channels, which leads to larger initialized parameters and a large distance after training. Therefore, the variation of shallow layers’ parameters affects $D(\theta_{A},\theta_{B})$ more than the deeper layers. This problem can be solved by calculating the average transferability of multiple layers (see Fig. 2 (b)): $T_{B}(\theta_{A})=\frac{1}{L}\sum_{l=1}^{L}T_{B}(\theta_{A}^{(l)}),$ (2) where $\theta_{A}^{(l)}$ and $L$ denote the parameters of layer $l$ and the total number of layers, respectively. It is worth noticing that, the transferability of layer $l$, $T_{B}(\theta_{A}^{(l)})$, can also be used to measure the trend of transferability between layers. ## Experiments and Results This section describes our experimental results in detail. We begin with the datasets and metrics to evaluate domain gaps and widths. Then we describe our implementation of transfer learning. Finally, we demonstrate and discuss the transferabilities of pre-trained parameters in different scenarios. ### Datasets and Metrics Our experiments include transfer learning from ImageNet to twelve image classification datasets (CIFAR-10/CIFAR-100 (Krizhevsky, Hinton et al. 2009), Caltech-101 (Fei-Fei, Fergus, and Perona 2004), CUB-200 (Wah et al. 2011), FGVC Aircraft (Maji et al. 2013), Flowers (Nilsback and Zisserman 2006), UC Merced Land Use (Yang and Newsam 2010), POCUS (Born et al. 2021), DTD (Sharan, Rosenholtz, and Adelson 2014), DomainNet (Peng et al. 2019)) from a wide range. This is because we find the experiments conducted by previous works (Yosinski et al. 2014; Azizpour et al. 2015) are primarily on natural images (i.e., photos), missing the images from farther domains (e.g., medical images, remote sensing images, paintings) or wider domains (e.g., DomainNet). But in many transfer learning applications, the distribution of the target domain is often significantly different from the source domain. The brief information of the datasets used in this work is summarized in Tab.1 To quantify the domain gaps of datasets to ImageNet as well as the domain widths, previous methods used trained DNNs to extract semantic features of different datasets (Liu and Zhang 2022; Stacke et al. 2020). An obvious drawback is that DNNs trained on specific tasks would have skewed performance on other tasks. In this work, we need the domain metrics to be domain-agnostic so that the gaps between different datasets can be more meaningful and comparable. The basic image color and texture features are desirable for evaluating the feature distributions of different datasets because they are not related to the categorical semantics (Saito et al. 2019). First, the RGB channels’ mean value and standard deviation are considered the color features. Second, Gray-Level Co-occurrence Matrix features (Haralick, Shanmugam, and Dinstein 1973) (Angular Second Moment, Entropy, Contrast, and Inverse Differential Moment) of 4 directions are adopted as the texture features. The total feature dimension is $3+3+4\times 4=22$. Given feature distributions, the domain gap is measured as the Maximum Mean Discrepancy (Gretton et al. 2012) between two datasets. Finally, the domain width is calculated as the maximum eigenvalue of the covariance matrix of features in a dataset. The domain gaps and widths of datasets used in this work are illustrated in Fig. 3, which is highly consistent with the common expectation (see supplementary material for the illustration of all datasets included in this work). Figure 3: The domain gaps with ImageNet and the domain widths of downstream datasets. For good illustration, the domain gap between ImageNet and itself is set to 0.1. Table 2: Dataset information (i.e., the domain gap compared with ImageNet, domain width, and training data amount), training performance (random initialization vs. fine-tuning), and transferabilities on different downstream tasks. | CIFAR-10 | CIFAR-100 | Caltech-101 | CUB-200 | Aircraft | Flowers | Land Use | POCUS | DTD | DomainNet-r | DomainNet-p | DomainNet-c ---|---|---|---|---|---|---|---|---|---|---|---|--- Domain gap | 2.27 | 2.39 | 0.63 | 1.62 | 2.95 | 1.64 | 1.89 | 13.12 | 0.74 | 1.45 | 0.47 | 6.56 Domain width | 186.0 | 242.9 | 385.0 | 220.0 | 114.6 | 145.3 | 181.2 | 46.9 | 822.3 | 467.7 | 455.5 | 823.2 Data amount | 50k | 50k | 3.1k | 6.0k | 6.7k | 1.1k | 1.7k | 1.7k | 1.9k | 120k | 50k | 34k Random init (%) | 84.9 | 54.6 | 64.6 | 32.4 | 41.5 | 72.8 | 83.6 | 85.9 | 29.3 | 71.4 | 48.5 | 53.5 Fine-tuning (%) | 90.0 | 66.3 | 94.6 | 76.4 | 81.6 | 98.9 | 98.8 | 94.6 | 73.1 | 81.7 | 69.0 | 74.9 Relative error rate (%) $\downarrow$ | 33.8 | 25.8 | 84.7 | 65.1 | 68.5 | 96.0 | 92.7 | 61.7 | 62.0 | 36.0 | 39.8 | 46.0 $T_{FB}(\theta_{A})$ | 7.79 | 5.59 | 23.92 | 21.96 | 12.45 | 106.30 | 39.09 | 53.58 | 24.15 | 6.32 | 7.12 | 8.91 $T_{SB}(\theta_{A})$ | 2.79 | 1.93 | 3.15 | 2.61 | 1.83 | 6.94 | 3.15 | 2.73 | 2.64 | 3.17 | 2.30 | 2.32 ### Implementation Details We utilize the ImageNet dataset to pre-train ResNet-50 (He et al. 2016) backbone (experiments with other backbones are demonstrated in the supplementary material, which shows similar conclusions). The hyperparameters are set to be the same as the original paper. For fine-tuning, we use $32\times 32$ CIFAR images and the input sizes of the other datasets are set to $224\times 224$. Batch size of 128, and training time of 100 epochs are used for all datasets. We use the initial learning rate (LR) of 0.01, SGD optimizer with the momentum of 0.9, and weight decay of 0.0001. The SGD optimizer is crucial for keeping learning rates for different layers consistent (adaptive optimizers control LRs in an unexpected way). The LRs decrease with a cosine LR scheduler. All images are augmented with random cropping and horizontal flipping with the probability of 0.5. When training DNNs from scratch, all hyperparameters are set to be identical as fine-tuning for a fair comparison. For all experiments with random initialization, initial parameters are also kept the same. When measuring the transferabilities of the whole DNN, the last fully connected layer will be excluded because it is meaningless to transfer parameters that only work for specific tasks. ### Verification of the Transferability Measurement In Tab. 2, the information of datasets, training performance from random initialization and fine-tuning, and the transferabilities on different tasks are demonstrated. The quantified transferabilities, $T_{FB}(\theta_{A})$ and $T_{SB}(\theta_{A})$, are highly correlated with the error rate decreasing ratio (p = 0.0001 and 0.0002 in F-test, respectively). This result indicates the rationality of the proposed transferability measurement because, with a higher transferability, a DNN can improve the downstream performance to a larger extent. Compared with the downstream performance, such as accuracy, the transferability defined in this work is less limited because it depends only on parameter variation, which is model-agnostic and task-agnostic. ### Factors Affecting Transferability In previous works, researchers usually regard DNNs to be less transferable when domain gaps are larger (Yosinski et al. 2014; Azizpour et al. 2015; Tajbakhsh et al. 2016). We find that this law does hold generally, and transferability is even more negatively affected by the domain width and data amount of downstream tasks. In Tab. 2, datasets having larger domain gaps with ImageNet tend to have lower transferabilities. For example, CIFAR-10 and CIFAR-100 have similar domain widths and data amounts, but CIFAR-100 has a larger domain gap than CIFAR-10, leading to lower transferability. The same phenomenon can also be observed between the Flowers and Land Use datasets. However, we can also see that CUB-200 and Flowers have similar domain gaps with ImageNet, but pre-trained DNNs have different transferabilities. This fact motivates our interest in exploring other crucial factors for transferability. Domain width and data amount are chosen as possible factors, as discussed in the introduction. Figure 4: Linear analysis of different factors. The correlation coefficient is 0.90. In this work, we validate the three factors by linear analysis. By fitting $log(T_{B}(\theta_{A}))$ to the domain gap $\mathcal{G}$, domain width $\mathcal{W}$, and log data amount $log(\mathcal{N})$, we find a high coefficient of determination for $log(T_{FB}(\theta_{A}))$ ($r^{2}$=0.90, see Fig. 4). The fitted linear relationship is $log(\hat{T}_{FB}(\theta_{A}))=-0.02\mathcal{G}-0.14\mathcal{W}-0.80log(\mathcal{N})+2.86.$ (3) We can see that, despite the slightly negative effect of the domain gap, there are stronger negative correlations between transferabilities and domain width and data amount. The reason will be elaborated on later. For $log(T_{SB}(\theta_{A}))$, linear regression performs poorly ($r^{2}=0.26$), indicating that the three factors can hardly interpret $T_{SB}(\theta_{A})$. This result may be due to the unstable convergence points of random initialized DNNs. Some convergence points are too bad to be considered in the “solution set” of the tasks (e.g., CUB and DTD). Therefore, $D(\theta_{r},\theta_{SB})$ is not a desirable metric for measuring transferability. $T_{FB}(\theta_{A})$ is mainly used for the following analysis. #### Domain Width. We try to remove distractions from domain gap and data amount, and explore the relationship between domain width and transferability $T_{FB}(\theta_{A})$. DTD and DomainNet-c are selected as the target datasets as they have the largest domain width. For DTD, we keep the training data amount to 1000 and use 12, 20, and 47 classes to calculate the transferabilities. It is worth noting that the domain gaps with ImageNet change slightly (from 6.56 to 6.44) when the number of classes decreases. For DomainNet-c, the domain gap is also stable when we randomly sample 2,000 images from 35, 100, or 345 classes. The results are shown in Fig. 5, where the transferabilities decrease as the domain widths increase, which validates Eqn. (3). #### Downstream Data Amount. This experiment verifies the negative correlation between transferabilities and data amounts. We uniformly sample data from all classes of CIFAR-100 and DomainNet-c with different amounts, and the transferability curves are shown in Fig. 6. It is worth mentioning that the random sampling scheme makes the domain gap and domain width stable in this experiment. With the increasing amount of data, the transferabilities decrease monotonically. Moreover, comparing Fig. 5 with Fig. 6, the effect of data amount seems greater than that of domain width, which is consistent with the Eqn. (3). Figure 5: The negative correlation between domain widths and transferabilities. Figure 6: The negative correlation between downstream data amounts and transferabilities. Figure 7: A conceptual visualization of source dataset and target dataset. “OOD” and “ID” are the abbreviations for “out-of-distribution” and “in- distribution”. #### A General Explanation. We try to explain the negative effects of domain gap, domain width, and data amount with a general framework (Fig. 7). In the pre-training phase, a DNN learns the basic knowledge (also called visual representations for visual tasks) from the source dataset. Within these representations, there are some in-distribution representations that can be used for the downstream task directly. However, there are still new visual representations that the DNN needs to learn from the downstream task for better performance. In the fine- tuning phase, the pre-trained DNN updates its parameters to learn these out- of-distribution representations. Such a larger parameter change corresponds to lower transferability. When the domain gap becomes greater or the domain width becomes larger, the more out-of-distribution representations the DNN needs to learn, and the lower transferability the DNN has. Similarly, given the fixed domain gap and domain width, higher downstream data amount makes the representations that need to be learned denser. The wider range and denser new representations decrease the DNN transferability by increasing knowledge that cannot be transferred from experiences. ### Layer-wise Transferability Figure 8: The layer-wise transferabilities on different datasets. The “Block” denotes the first convolutional layer, the bottle-neck (a three-conv-layer residual block), or the last fully connected layer of ResNet. Previous literature usually regards the lower layers to be more transferable (Yosinski et al. 2014), due to the generality of the low-level feature extraction abilities. This work shows that this is not always the case. The trend of transferabilities across layers is illustrated in Fig. 8. There are two noteworthy laws. First, nearly all layer-wise transferabilities are inverted U-shaped instead of downward trends, mainly caused by the first two poorly transferable layers. Second, discarding the first two layers, nine datasets have descending layer-wise transferabilities across layers consistent with our previous consensus. However, three datasets (CIFAR-10, CIFAR-100, and POCUS) still do not present any meaningful upward/downward trend. For these three datasets, we find that the low-level features may not be transferable as the low-level image pattern are not similar anymore. For CIFAR-10 and CIFAR-100, the image sizes are only $32\times 32$. The pre-trained ResNets have learned visual representations from $224\times 224$ ImageNet data, so these representations may be poorly transferable to the seriously different local pattern distributions. For the POCUS dataset, the ultrasound images also have quite different image patterns compared with natural images, making the pre-trained layers for extracting local features harder to transfer. All in all, the different pattern distributions of pre-training and downstream datasets make the layer-wise transferability no longer decreasing. #### Domain-sensitivity of the Lowest Layers. In Fig. 8, we can see a common phenomenon: the first layer has very low transferability on all datasets, including those natural image datasets with small domain gaps. This result may indicate that the first layer is more sensitive to the domain shift from ImageNet than the subsequent layers. We assume that when a pre-trained DNN “sees” a new sample with out-of- distribution basic visual patterns, the first layer weights would get large gradients and update drastically. These kinds of out-of-distribution basic visual patterns may be common in downstream datasets. To verify this assumption, we calculate the layer-wise gradients of pre-trained DNNs on two natural image datasets, i.e., CUB-100 and Flowers (see Fig. 9), and we find that both curves are U-shaped. It is reasonable for the last layers to have large gradients, as they are replaced and randomly initialized for specific tasks. But for the large gradients of the pre-trained lower layers, they suggest that the parameters of these layers are somehow “unfit” for the current tasks, which validates our assumption. Figure 9: The layer-wise mean absolute gradients on CUB and Flower datasets. Both curves are U-shaped. #### Visualization of Pre-trained/Fine-tuned Features. To further illustrate the weaker transferabilities of lower layers, we choose CUB-200 to show how first-layer features of a pre-trained DNN change after fine-tuning. In Fig. 10, the pre-trained first layer can successfully activate meaningful edges. But some features are not good enough (in red boxes), with unclear edges or activating the background gauze too much. After fine-tuning, these imperfect feature maps become clearer and more helpful for recognizing the bird. As reported in (Donahue et al. 2014), a unit’s activation is better be maximized to extract more meaningful image features, meaning that the fine- tuned parameters are more capable of finding key patterns on the images. The benefit of fine-tuning the first layer is increasing the activations of the target object in the image (mean value increases from 0.268 to 0.284). The activation distributions of pre-trained/fine-tuned DNNs significantly differ with the $t$ statistic of -16.99, $p=4.6\times 10^{-57}$, which shows significant improvement brought by updating the first layer. This result indicates that some previous methods of only tuning higher layers (Tajbakhsh et al. 2016) can still be improved concerning the first-layer transferability. Figure 10: Input image (left), first-layer features of pre-trained DNN (upper right), and the corresponding features after fine-tuning (lower right). For feature maps, lighter means more activated. ## Conclusion In this work, we have proposed a novel method for quantifying the transferabilities of DNNs. We conducted extensive experiments on twelve datasets with this method and showed how transferability is more negatively affected by domain width and downstream data amount than domain gap. This phenomenon suggests that downstream tasks with narrower domains and fewer data benefit more from pre-training. Transfer learning with a large domain gap (e.g., medical data) may reach higher performance considering this result. In addition, we observed that the layer-wise transferabilities of DNNs are not monotonically decreasing as the traditional consensus regarded. The root cause is that the first layers of DNNs are sensitive to domains. They have larger gradients than the middle layers and are usually imperfect for downstream tasks. Fine-tuning techniques with the least transferable layers from both ends may achieve better performance than previous works with only the last layers. ## References * Ausubel et al. (1968) Ausubel, D. P.; Novak, J. D.; Hanesian, H.; et al. 1968. _Educational psychology: A cognitive view_ , volume 6. holt, rinehart and Winston New York. * Azizpour et al. (2015) Azizpour, H.; Razavian, A. S.; Sullivan, J.; Maki, A.; and Carlsson, S. 2015. Factors of transferability for a generic convnet representation. _IEEE TPAMI_ , 38(9): 1790–1802. * Born et al. (2021) Born, J.; Wiedemann, N.; Cossio, M.; Buhre, C.; Brändle, G.; Leidermann, K.; Aujayeb, A.; Moor, M.; Rieck, B.; and Borgwardt, K. 2021. Accelerating Detection of Lung Pathologies with Explainable Ultrasound Image Analysis. _Applied Sciences_ , 11(2): 672. * Chen et al. (2022) Chen, Y.; Li, J.; Jiang, H.; Liu, L.; and Ding, C. 2022. MetaLR: Layer-wise Learning Rate based on Meta-Learning for Adaptively Fine-tuning Medical Pre-trained Models. _arXiv preprint arXiv:2206.01408_. * Chen et al. (2021) Chen, Y.; Zhang, C.; Liu, L.; Feng, C.; Dong, C.; Luo, Y.; and Wan, X. 2021. Uscl: Pretraining deep ultrasound image diagnosis model through video contrastive representation learning. In _MICCAI_ , 627–637. Springer. * Deng et al. (2009) Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; and Fei-Fei, L. 2009. Imagenet: A large-scale hierarchical image database. In _CVPR_ , 248–255. * Donahue et al. (2014) Donahue, J.; Jia, Y.; Vinyals, O.; Hoffman, J.; Zhang, N.; Tzeng, E.; and Darrell, T. 2014. Decaf: A deep convolutional activation feature for generic visual recognition. In _ICML_ , 647–655. PMLR. * Dosovitskiy et al. (2020) Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. 2020\. An image is worth 16x16 words: Transformers for image recognition at scale. _arXiv preprint arXiv:2010.11929_. * Dwivedi and Roig (2019) Dwivedi, K.; and Roig, G. 2019. Representation similarity analysis for efficient task taxonomy & transfer learning. In _CVPR_ , 12387–12396. * Fayek, Cavedon, and Wu (2018) Fayek, H. M.; Cavedon, L.; and Wu, H. R. 2018. On the transferability of representations in neural networks between datasets and tasks. _arXiv preprint arXiv:1811.12273_. * Fei-Fei, Fergus, and Perona (2004) Fei-Fei, L.; Fergus, R.; and Perona, P. 2004. Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. In _CVPR workshop_ , 178–178. IEEE. * Finn, Abbeel, and Levine (2017) Finn, C.; Abbeel, P.; and Levine, S. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In _ICML_ , 1126–1135. PMLR. * Glorot and Bengio (2010) Glorot, X.; and Bengio, Y. 2010. Understanding the difficulty of training deep feedforward neural networks. In _Proceedings of the thirteenth international conference on artificial intelligence and statistics_ , 249–256. JMLR Workshop and Conference Proceedings. * Gretton et al. (2012) Gretton, A.; Borgwardt, K. M.; Rasch, M. J.; Schölkopf, B.; and Smola, A. 2012\. A kernel two-sample test. _The Journal of Machine Learning Research_ , 13(1): 723–773. * Haralick, Shanmugam, and Dinstein (1973) Haralick, R. M.; Shanmugam, K.; and Dinstein, I. H. 1973. Textural features for image classification. _IEEE Transactions on systems, man, and cybernetics_ , (6): 610–621. * He, Girshick, and Dollár (2019) He, K.; Girshick, R.; and Dollár, P. 2019. Rethinking imagenet pre-training. In _ICCV_ , 4918–4927. * He et al. (2015) He, K.; Zhang, X.; Ren, S.; and Sun, J. 2015. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In _ICCV_ , 1026–1034. * He et al. (2016) He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016. Deep residual learning for image recognition. In _CVPR_ , 770–778. * Huang et al. (2017) Huang, G.; Liu, Z.; Van Der Maaten, L.; and Weinberger, K. Q. 2017. Densely connected convolutional networks. In _CVPR_ , 4700–4708. * Jiang et al. (2022) Jiang, J.; Shu, Y.; Wang, J.; and Long, M. 2022. Transferability in Deep Learning: A Survey. _arXiv preprint arXiv:2201.05867_. * Kornblith, Shlens, and Le (2019) Kornblith, S.; Shlens, J.; and Le, Q. V. 2019. Do better imagenet models transfer better? In _CVPR_ , 2661–2671. * Krizhevsky, Hinton et al. (2009) Krizhevsky, A.; Hinton, G.; et al. 2009. Learning multiple layers of features from tiny images. * Li et al. (2019) Li, X.; Xiong, H.; Wang, H.; Rao, Y.; Liu, L.; Chen, Z.; and Huan, J. 2019. Delta: Deep learning transfer using feature map with attention for convolutional networks. _arXiv preprint arXiv:1901.09229_. * Liu and Zhang (2022) Liu, X.; and Zhang, S. 2022. Who is closer: A computational method for domain gap evaluation. _Pattern Recognition_ , 122: 108293. * Long et al. (2015) Long, M.; Cao, Y.; Wang, J.; and Jordan, M. 2015. Learning transferable features with deep adaptation networks. In _ICML_ , 97–105. PMLR. * Maji et al. (2013) Maji, S.; Rahtu, E.; Kannala, J.; Blaschko, M.; and Vedaldi, A. 2013. Fine-grained visual classification of aircraft. _arXiv preprint arXiv:1306.5151_. * Matsoukas et al. (2022) Matsoukas, C.; Haslum, J. F.; Sorkhei, M.; Söderberg, M.; and Smith, K. 2022\. What Makes Transfer Learning Work For Medical Images: Feature Reuse & Other Factors. In _CVPR_ , 9225–9234. * Newell and Deng (2020) Newell, A.; and Deng, J. 2020. How useful is self-supervised pretraining for visual tasks? In _CVPR_ , 7345–7354. * Nilsback and Zisserman (2006) Nilsback, M.-E.; and Zisserman, A. 2006. A visual vocabulary for flower classification. In _CVPR_ , volume 2, 1447–1454. IEEE. * Peng et al. (2019) Peng, X.; Bai, Q.; Xia, X.; Huang, Z.; Saenko, K.; and Wang, B. 2019. Moment matching for multi-source domain adaptation. In _ICCV_ , 1406–1415. * Ro and Choi (2021) Ro, Y.; and Choi, J. Y. 2021. AutoLR: Layer-wise pruning and auto-tuning of learning rates in fine-tuning of deep networks. In _AAAI_ , volume 35, 2486–2494. * Saito et al. (2019) Saito, K.; Ushiku, Y.; Harada, T.; and Saenko, K. 2019. Strong-weak distribution alignment for adaptive object detection. In _CVPR_ , 6956–6965. * Sermanet et al. (2018) Sermanet, P.; Lynch, C.; Chebotar, Y.; Hsu, J.; Jang, E.; Schaal, S.; Levine, S.; and Brain, G. 2018. Time-contrastive networks: Self-supervised learning from video. In _ICRA_ , 1134–1141. IEEE. * Sharan, Rosenholtz, and Adelson (2014) Sharan, L.; Rosenholtz, R.; and Adelson, E. H. 2014. Accuracy and speed of material categorization in real-world images. _Journal of vision_ , 14(9): 12–12. * Song et al. (2021) Song, L.; Wu, J.; Yang, M.; Zhang, Q.; Li, Y.; and Yuan, J. 2021. Robust Knowledge Transfer via Hybrid Forward on the Teacher-Student Model. In _AAAI_ , 2558–2566. * Stacke et al. (2020) Stacke, K.; Eilertsen, G.; Unger, J.; and Lundström, C. 2020. Measuring domain shift for deep learning in histopathology. _IEEE journal of biomedical and health informatics_ , 25(2): 325–336. * Tajbakhsh et al. (2016) Tajbakhsh, N.; Shin, J. Y.; Gurudu, S. R.; Hurst, R. T.; Kendall, C. B.; Gotway, M. B.; and Liang, J. 2016. Convolutional neural networks for medical image analysis: Full training or fine tuning? _IEEE TMI_ , 35(5): 1299–1312. * Wah et al. (2011) Wah, C.; Branson, S.; Welinder, P.; Perona, P.; and Belongie, S. 2011. The caltech-ucsd birds-200-2011 dataset. * Yang and Newsam (2010) Yang, Y.; and Newsam, S. 2010. Bag-of-visual-words and spatial extensions for land-use classification. In _Proceedings of the 18th SIGSPATIAL international conference on advances in geographic information systems_ , 270–279. * Yosinski et al. (2014) Yosinski, J.; Clune, J.; Bengio, Y.; and Lipson, H. 2014. How transferable are features in deep neural networks? _NeurIPS_ , 27. * Zhang et al. (2022) Zhang, C.; Chen, Y.; Liu, L.; Liu, Q.; and Zhou, X. 2022. HiCo: Hierarchical Contrastive Learning for Ultrasound Video Model Pretraining. In _Proceedings of the Asian Conference on Computer Vision_ , 229–246. * Zhong et al. (2020) Zhong, J.; Wang, X.; Kou, Z.; Wang, J.; and Long, M. 2020. Bi-tuning of pre-trained representations. _arXiv preprint arXiv:2011.06182_.
# Free $\mathbb{Q}$-groups are residually torsion-free nilpotent Andrei Jaikin-Zapirain Departamento de Matemáticas, Universidad Autónoma de Madrid and Instituto de Ciencias Matemáticas, CSIC-UAM-UC3M-UCM <EMAIL_ADDRESS> (Date: November 2022) ###### Abstract. We develop a method to show that some (abstract) groups can be embedded into a free pro-$p$ group. In particular, we show that every finitely generated subgroup of a free $\mathbb{Q}$-group can be embedded into a free pro-$p$ group for almost all primes $p$. This solves an old problem raised by G. Baumslag: free $\mathbb{Q}$-groups are residually torsion-free nilpotent. ###### Key words and phrases: Free $\mathbb{Q}$-groups, free pro-$p$ groups, mod-$p$ $L^{2}$-Betti numbers, Lück approximation, universal division ring of fractions ###### 2010 Mathematics Subject Classification: Primary: 20E06, Secondary: 16K40, 20C07, 20E18, 20E26 ## 1\. Introduction A group $G$ is called a $\mathbb{Q}$-group if for any $n\in\mathbb{N}$ and $g\in G$ there exists exactly one $h\in G$ satisfying $h^{n}=g$. These groups were introduced by G. Baumslag in [4] under the name of $\operatorname{\mathcal{D}}$-groups. He observed that $\mathbb{Q}$-groups may be viewed as universal algebras, and as such they constitute a variety. Every variety of algebras contains free algebras (in that variety). In the variety of $\mathbb{Q}$-groups we call such free algebras free $\mathbb{Q}$-groups. G. Baumslag dedicated several papers to the study of residual properties of free $\mathbb{Q}$-groups [5, 7, 9]. For example, in [5] he showed that a free $\mathbb{Q}$-group is residually periodic-by-soluble and locally residually finite-by-soluble. He wrote in [5] “It is, of course, still possible that, locally, free $\operatorname{\mathcal{D}}$-groups are, say, residually finite $p$-groups” or in [7] “In particular it seems likely that free $\operatorname{\mathcal{D}}$-groups are residually torsion-free nilpotent. However the complicated nature of free $\operatorname{\mathcal{D}}$-groups makes it difficult to substantiate such a remark.” This conjecture is part of two main collections of problems in group theory ([10, Problem F12] and [38, Problem 13.39 (a),(c)]), and in addition to mentioned works of Baumslag, it was also studied in [15, 22]. In this paper we solve Baumslag’s conjecture. ###### Theorem 1.1. A free $\mathbb{Q}$-group is residually torsion-free nilpotent. The structure of a finitely generated subgroup of a free $\mathbb{Q}$-group was studied already in [4] (see also [53, Section 8] and Proposition 5.3). It was shown that it is the end result of repeatedly freely adjoining $n$th roots to a finitely generated free group. The key point of our proof of Theorem 1.1 is to show that any finitely generated subgroup of a free $\mathbb{Q}$-group can be embedded into a finitely generated free pro-$p$ group for some prime $p$. We actually prove the following more precise result. ###### Theorem 1.2. Let $p$ be a prime. Let $H_{0}$ be a finitely generated free group and let $H_{0}\hookrightarrow\mathbf{F}$ be the canonical embedding of $H_{0}$ into its pro-$p$ completion $\mathbf{F}$. Let $(H_{i})_{i\geq 0}$ be a sequence of subgroups of $\mathbf{F}$ such that for $i\geq 0$, 1. (1) $H_{i+1}=\langle H_{i},B_{i}\rangle$, where $B_{i}$ is a finitely generated abelian subgroup of $\mathbf{F}$ and 2. (2) $A_{i}=H_{i}\cap B_{i}$ is a maximal abelian subgroup of $H_{i}$. Then for every $i\geq 0$, the canonical map $H_{i}*_{A_{i}}B_{i}\to H_{i+1}$ is an isomorphism, Theorem 1.2 is actually an application of the slightly more technical Theorem 5.1. Let us make a few remarks about the groups $A_{i}$ and $B_{i}$. It is relatively easy to describe abelian subgroups of amalgamated products. In particular, the conclusion of the theorem implies that all abelian subgroups of $H_{i}$ are finitely generated. Thus, an implicit hypothesis, which appears in the theorem, that maximal abelian subgroups $A_{i}$ of $H_{i}$ are finitely generated, is automatically fulfilled. A maximal abelian subgroup of $\mathbf{F}$ is isomorphic to the additive group of the ring of $p$-adic numbers $(\mathbb{Z}_{p},+)$. Therefore, for any finitely generated (abstract) abelian subgroup $A$ of $\mathbf{F}$ and any finitely generated torsion-free abelian group $B$ which contains $A$ and such that $B/A$ has no $p$-torsion, it is posible to extend the embedding $A\hookrightarrow\mathbf{F}$ to an embedding $B\hookrightarrow\mathbf{F}$. This extension is unique if and only if $B/A$ is finite. Given a commutative ring $A$, we will introduce in Section 5 the notions of $A$-group and free $A$-group $F^{A}(X)$. For example, a free pro-$p$ group is an example of a $\mathbb{Z}_{p}$-group. We have the following consequence of Theorem 1.2. ###### Corollary 1.3. Let $F(X)$ be the free group on a finite free generating set $X$, let $\mathbf{F}$ be its pro-$p$ completion. Then the canonical homomorphism $\phi:F^{\mathbb{Z}_{p}}(X)\to\mathbf{F}$ is injective. Let $H$ be a group and $A$ the centralizer of a non-trivial element. Then the group $G=H*_{A}(A\times\mathbb{Z}^{k})$ is said to be obtained from $H$ by extension of a centralizer. A group is called an ICE group if it can be obtained from a free group using iterated centralizer extensions. A group $G$ is a limit group if and only if it is a finitely generated subgroup of an ICE group (see [37, 14]). All centralizers of non-trivial elements of an ICE group are abelian. Thus, Theorem 1.1 provides explicit realizations of ICE groups (and so limit groups) as subgroups of a non-abelian free pro-$p$-groups (for this application we only need the case where all $B_{i}/A_{i}$ are torsion- free). Non-explicit realizations of limit groups as subgroups of a non-abelian free pro-$p$ group (in fact, as subgroups of every compact group containing a non-abelian free group) was obtained in [2] (see also [12]). In Section 5 we recall the definition of the $\mathbb{Q}$-completion of a group $G$. For example, a free $\mathbb{Q}$-group is the $\mathbb{Q}$-completion of a free group. Theorem 1.2 allows also to show that the $\mathbb{Q}$-completion of a limit group is residually torsion-free nilpotent. ###### Theorem 1.4. The $\mathbb{Q}$-completion of a limit group is residually torsion-free nilpotent. A group $G$ is called parafree if it is residually nilpotent and for some free group $F$, we have that for all $i$, $G/\gamma_{i}(G)\cong F/\gamma_{i}(F)$ where $\gamma_{i}(G)$ denotes the terms of the lower central series of $G$. Baumslag introduced this family of groups and produced many examples of them [6]. In [35] we apply the method of the proof of Theorem 1.2 in order to construct new examples of finitely generated parafree groups. Our proof of Theorem 1.2 is by induction on $i$. In the inductive step argument we start with the following situation. We have a finitely generated subgroup $H$ of $\mathbf{F}$, a maximal abelian subgroup $A$ of $H$ and an abelian subgroup $B$ of $\mathbf{F}$ containing $A$. We want to show that the canonical homomorphism $H*_{A}B\to\langle H,B\rangle$ is an isomorphism. Unfortunately, we do not know how to show this statement in such a generality, but we prove it in Theorem 5.1 under an additional assumption that the embedding $H\hookrightarrow\mathbf{F}$ is strong (see Definition 3.8). Theorem 5.1 is the main result of the paper. Its proof uses in an essential way the results of [32], where we proved a particular case of the Lück approximation in positive characteristic. The paper is organized as follows. In Section 2 we give basic preliminaries. The proof of Theorem 5.1 uses the theory of mod-$p$ $L^{2}$-Betti numbers. In Section 3 we explain how to define them for subgroups $G$ of a free pro-$p$ group. In Section 4 we introduce a technical notion of $\mathcal{D}$-torsion- free modules and show that some relevant $\mathbb{F}_{p}[G]$-modules are $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free (see Proposition 4.10). In Section 5 we prove Theorem 5.1 and obtain all the results mentioned in the introduction. In Section 6 we discuss the following two well-known problems concerning linearity of free pro-$p$ groups and free $\mathbb{Q}$-groups: ###### Question 1.5. 1. (1) (I. Kapovich) Is a free $\mathbb{Q}$-group linear? 2. (2) (A. Lubotzky) Is a free pro-$p$ group linear? ## Acknowledgments I am very grateful to Pavel Zalesski who explained to me how a surface group can be embedded into a free pro-$p$ group. This was a starting point of Theorem 1.2. I would like to thank Warren Dicks for providing the references needed in the proof of Proposition 2.13, Yago Antolin, Ashot Minasyan and a referee for explaining to me the proof of Theorem 6.1, Emmanuel Breulliard for pointing out the paper [2] and Ismael Morales and anonymous referees for useful suggestions and comments. This paper is partially supported by the grants MTM2017-82690-P and PID2020-114032GB-I00 of the Ministry of Science and Innovation of Spain and by the ICMAT Severo Ochoa project CEX2019-000904-S4. ## 2\. Preliminaries ### 2.1. $R$-rings All rings in this paper are associative and have the identity element. All ring homomorphisms send the identity to the identity. We denote the invertible elements of a ring $R$ by $R^{*}$. An $R$-module means a left $R$-module. By an $R$-ring we understand a ring homomorphism $\varphi:R\to S$. We will often refer to $S$ as $R$-ring and omit the homomorphism $\varphi$ if $\varphi$ is clear from the context. Two $R$-rings $\varphi_{1}:R\to S_{1}$ and $\varphi_{2}:R\to S_{2}$ are said to be isomorphic if there exists a ring isomorphism $\alpha:S_{1}\to S_{2}$ such that $\alpha\circ\varphi_{1}=\varphi_{2}$. If $Y=\\{y_{i}\colon i\in I\\}$, we denote by $A\langle\\!\langle Y\rangle\\!\rangle$ the ring of of formal power series with coefficients in $A$ on the non-commuting indeterminates $Y$. ### 2.2. Left ideals in group algebras Let $G$ be a group and $k$ a commutative ring. We denote by $I_{G}$ the augmentation ideal of $k[G]$. If $H$ is a subgroup of $G$ we denote by $I_{H}^{G}$ the left ideal of $k[G]$ generated by $I_{H}$. The following lemma gives an alternative description of the $k[G]$-module $I_{H}^{G}$. ###### Lemma 2.1. Let $H\leq T$ be subgroups of $G$. Then the following holds. 1. (a) The canonical map $k[G]\otimes_{k[H]}I_{H}\to I_{H}^{G}$ sending $a\otimes b$ to $ab$, is an isomorphism of $k[G]$-modules. 2. (b) The canonical map $k[G]\otimes_{k[T]}(I_{T}/I_{H}^{T})\to I_{T}^{G}/I_{H}^{G}$, sending $a\otimes(b+I_{H}^{T})$ to $ab+I_{H}^{G}$, is an isomorphism of $k[G]$-modules. ###### Proof. (a) Consider an exact sequence $0\to I_{H}\to k[H]\to k\to 0.$ The freeness of $k[G]$ as $k[H]$-module implies that the sequence $0\to k[G]\otimes_{k[H]}I_{H}\xrightarrow{\alpha}k[G]\xrightarrow{\beta}k[G]\otimes_{k[H]}k\to 0$ is also exact. Here $\alpha$ sends $a\otimes b$ to $ab$ and $\beta$ sends $a$ to $a\otimes 1$. Thus, $\alpha$ establishes an isomorphisms of $k[G]$-modules between $k[G]\otimes_{k[H]}I_{H}$ and $\ker\beta=I_{H}^{G}$. This proves the first claim of the lemma. (b) Consider now the exact sequence $0\to I_{H}^{T}\to I_{T}\to I_{T}/I_{H}^{T}\to 0.$ Applying $k[G]\otimes_{k[T]}$ and taking again into account that $k[G]$ is a free $k[T]$-module, we obtain the exact sequence $0\to I_{H}^{G}\to I_{T}^{G}\to k[G]\otimes_{k[T]}(I_{T}/I_{H}^{T})\to 0.$ This proves the second claim. ∎ ### 2.3. Profinite modules over pro-$p$ groups In this paper the letters $\mathbf{F}$, $\mathbf{G}$, $\mathbf{H}$, etc. will denote pro-$p$ groups. When we speak about pro-$p$ groups, the finite generation, the finite presentation, the freeness etc. will always be considered in the category of pro-$p$ groups. For example, $d(\mathbf{G})$ denotes the minimal number of topological generators of $\mathbf{G}$. Almost all pro-$p$ groups that we consider are free pro-$p$ groups. Recall that a closed subgroup of a free pro-$p$ group is also free pro-$p$ ([56, Corollary 7.7.5]). As a consequence we obtain the following result which we will use often in this paper. ###### Lemma 2.2. Every maximal abelian subgroup of a non-trivial free pro-$p$ group is isomorphic to $(\mathbb{Z}_{p},+)$. Let $\mathbf{G}$ be a pro-$p$ group. We denote by $\mathbb{F}_{p}[[\mathbf{G}]]$ the inverse limit of $\mathbb{F}_{p}[\mathbf{G}/\mathbf{U}]$, where the limit is taken over all open normal subgroups $\mathbf{U}$ of $\mathbf{G}$. $\mathbb{F}_{p}[[\mathbf{G}]]$ is called the completed group algebra of $\mathbf{G}$ over $\mathbb{F}_{p}$. In the case where $\mathbf{G}$ is a free pro-$p$ group we have the following useful description of $\mathbb{F}_{p}[[\mathbf{G}]]$. ###### Proposition 2.3. [39, Section II, Proposition 3.1.4] Let $\mathbf{F}$ be a finitely generated free pro-$p$ group freely generated by $x_{1},\ldots,x_{d}$. Then the continuous $\mathbb{F}_{p}$-algebra homomorphism $\mathbb{F}_{p}\langle\\!\langle y_{1},\ldots,y_{d}\rangle\\!\rangle\to\mathbb{F}_{p}[[\mathbf{F}]]$ that sends $y_{i}$ to $x_{1}-1$ (for $1\leq i\leq d$) is an isomorphism. A discrete $\mathbb{F}_{p}[[\mathbf{G}]]$-module is an $\mathbb{F}_{p}[[\mathbf{G}]]$-module $M$ such that for any $m\in M$, $\operatorname{Ann}_{\mathbb{F}_{p}[[\mathbf{G}]]}(m):=\\{a\in\mathbb{F}_{p}[[\mathbf{G}]]\colon am=0\\}$ is open in $\mathbb{F}_{p}[[\mathbf{G}]]$. A profinite $\mathbb{F}_{p}[[\mathbf{G}]]$-module is an inverse limit of finite discrete $\mathbb{F}_{p}[[\mathbf{G}]]$-modules. ###### Lemma 2.4. Let $\mathbf{G}$ be a pro-$p$ group. Let $\alpha:M\to N$ be a homomorphism of profinite $\mathbb{F}_{p}[[\mathbf{G}]]$-modules. If $M$ is finitely generated as an $\mathbb{F}_{p}[[\mathbf{G}]]$-module, then $\alpha$ is continuous. ###### Remark 2.5. This lemma resembles a well-known result of Nikolov and Segal [55] that says that every homomorphism from a finitely generated profinite group to a profinite group is always continuous. This is equivalent to that every subgroup of finite index in a finitely generated profinite group is open. In the case of $\mathbb{F}_{p}[[\mathbf{G}]]$-modules, the situation is different: it is not always true that every left ideal of $\mathbb{F}_{p}[[\mathbf{G}]]$ of finite index is open. ###### Proof. Without loss of generality we may assume that $N$ is a finite discrete $\mathbb{F}_{p}[[\mathbf{G}]]$-module and we have to show that $\ker\alpha$ is open. We put $I=\operatorname{Ann}_{\mathbb{F}_{p}[[G]]}=\\{a\in\mathbb{F}_{p}[[\mathbf{G}]]\colon aN=\\{0\\}\\}$. Then $I$ is an open, and so also closed, ideal of $\mathbb{F}_{p}[[\mathbf{G}]]$. Let $IM$ be the submodule of $M$ generated by $\\{a\cdot m\colon a\in I,m\in M\\}$. Since $M$ is finitely generated we can write $M=\sum_{i=1}^{s}\mathbb{F}_{p}[[\mathbf{G}]]m_{i}.$ Therefore, $IM=\sum_{i=1}^{s}Im_{i}$. From the definition of a profinite $\mathbb{F}_{p}[[\mathbf{G}]]$-module it follows that the multiplicative map $\mathbb{F}_{p}[[\bf G]]\times M\to M$ is continuous. Hence, since $I$ is closed in $\mathbb{F}_{p}[[\mathbf{G}]]$, $IM=\sum_{i=1}^{s}Im_{i}$ is a closed submodule of $M$. But $IM$ is also of finite index. Hence, it is open in $M$. Since $IM\leq\ker\alpha$, $\ker\alpha$ is also open. ∎ If $M=\varprojlim_{i\in I}M_{i}$ and $N=\varprojlim_{j\in J}N_{i}$ are right and left, respectively, profinite $\mathbb{F}_{p}[[\mathbf{G}]]$-modules ($M_{i}$ and $N_{j}$ are finite discrete $\mathbb{F}_{p}[[\mathbf{G}]]$-modules), then the profinite tensor product is denoted by $\widehat{\otimes}$ and it is defined as the inverse limit of $M_{i}\otimes_{\mathbb{F}_{p}[[\mathbf{G}]]}N_{j}$. ###### Lemma 2.6. Let $\mathbf{H}$ be a closed subgroup of $\mathbf{G}$ and let $M$ be a finitely presented $\mathbb{F}_{p}[[\mathbf{H}]]$-module. Then $M$ is a profinite module and the canonical map $\gamma:\mathbb{F}_{p}[[\mathbf{G}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M\to\mathbb{F}_{p}[[\mathbf{G}]]\widehat{\otimes}_{\mathbb{F}_{p}[[\mathbf{H}]]}M$ is an isomorphism of $\mathbb{F}_{p}[[\mathbf{G}]]$-modules. ###### Remark 2.7. It is claimed in [56, Proposition 5.5.3 (d)] that it is enough to assume that $M$ is a finitely generated profinite $\mathbb{F}_{p}[[\mathbf{H}]]$-module. However, we want to warn the reader that the proof is incorrect. ###### Proof. Since $M$ is finitely presented, there exists an exact sequence of $\mathbb{F}_{p}[[\mathbf{H}]]$-modules (1) $\mathbb{F}_{p}[[\mathbf{H}]]^{r}\xrightarrow{\alpha}\mathbb{F}_{p}[[\mathbf{H}]]^{d}\to M\to 0.$ Thus, we can write $M\cong\mathbb{F}_{p}[[\mathbf{H}]]^{d}/I$, where $I=\operatorname{Im}\alpha$. By Lemma 2.4, $\alpha$ is continuous. Since $\mathbb{F}_{p}[[\mathbf{H}]]^{r}$ is compact and Hausdorff, $I$ is closed in $\mathbb{F}_{p}[[\mathbf{H}]]^{d}$. Hence $\mathbb{F}_{p}[[\mathbf{H}]]^{d}/I$, and so $M$, are profinite. It is clear that $\mathbb{F}_{p}[[\mathbf{G}]]\widehat{\otimes}_{\mathbb{F}_{p}[[\mathbf{H}]]}\mathbb{F}_{p}[[\mathbf{H}]]\cong\mathbb{F}_{p}[[\mathbf{G}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}\mathbb{F}_{p}[[\mathbf{H}]]\cong\mathbb{F}_{p}[[\mathbf{G}]].$ Thus, after applying $\mathbb{F}_{p}[[\mathbf{G}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}$ and $\mathbb{F}_{p}[[\mathbf{G}]]{\widehat{\otimes}}_{\mathbb{F}_{p}[[\mathbf{H}]]}$ to the sequence (1) we obtain a commutative diagram of $\mathbb{F}_{p}[[\mathbf{G}]]$-modules: $\begin{array}[]{cccccc}\mathbb{F}_{p}[[\mathbf{G}]]^{r}&\to&\mathbb{F}_{p}[[\mathbf{G}]]^{d}&\to&\mathbb{F}_{p}[[\mathbf{G}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M&\to 0\\\ ||&&||&&\downarrow^{\gamma}&\\\ \mathbb{F}_{p}[[\mathbf{G}]]^{r}&\to&\mathbb{F}_{p}[[\mathbf{G}]]^{d}&\to&\mathbb{F}_{p}[[\mathbf{G}]]\widehat{\otimes}_{\mathbb{F}_{p}[[\mathbf{H}]]}M&\to 0.\end{array}.$ Since $\otimes$ and $\widehat{\otimes}$ are right exact (see [57, Theorem 2.6.3] and [56, Proposition 5.5.3(a)]) the horizontal sequences are exact. This clearly implies that $\gamma$ is an isomorphism. ∎ Recall that if a pro-$p$ group $\mathbf{G}$ is finitely generated, then the augmentation ideal $I_{\mathbf{G}}$ is finitely generated as an $\mathbb{F}_{p}[[\mathbf{G}]]$-module. In particular, the trivial $\mathbb{F}_{p}[[\mathbf{G}]]$-module $\mathbb{F}_{p}$ is finitely presented. We can extend this to all finite discrete $\mathbb{F}_{p}[[\mathbf{G}]]$-modules. ###### Lemma 2.8. Let $\mathbf{G}$ be a finitely generated pro-$p$ group and let $M$ be a finite discrete $\mathbb{F}_{p}[[\mathbf{G}]]$-module. Then $M$ is finitely presented as an $\mathbb{F}_{p}[[G]]$-module. ###### Proof. We have to show that every open submodule of a finitely generated profinite $\mathbb{F}_{p}[[\mathbf{G}]]$-module $M$ is finitely generated. Arguing by induction on the index of the open submodule, we easily obtain this statement from the fact that $I_{\mathbf{G}}$ is finitely generated. ∎ ### 2.4. Left ideals in completed group algebras Now we apply the results of the previous subsection in a particular situation that interests us. Let $\mathbf{G}$ be a pro-$p$ group. The definition of free profinite $\mathbb{F}_{p}[[\mathbf{G}]]$-modules can be found in [56, Section 5.2]. We will need the the following fact. ###### Lemma 2.9. Let $\mathbf{G}$ be a pro-$p$ group and $\mathbf{H}$ a closed subgroup of $\mathbf{G}$. Then $\mathbb{F}_{p}[[\mathbf{G}]]$ is a free profinite $\mathbb{F}_{p}[[\mathbf{H}]]$-module. In particular the functor $\mathbb{F}_{p}[[\mathbf{G}]]\widehat{\otimes}_{\mathbb{F}_{p}[[\mathbf{H}]]}-$ is exact. ###### Proof. The first part follows from [56, Corollary 5.7.2] and the second one from [56, Proposition 5.5.3 (e)]. ∎ We denote by $I_{\mathbf{G}}$ the augmentation ideal of $\mathbb{F}_{p}[[\mathbf{G}]]$. The following lemma is well-known. We provide a proof for the convenience of the reader. ###### Lemma 2.10. Let $\mathbf{H}$ be a finitely presented pro-$p$ group. Then $I_{\mathbf{H}}$ is finitely presented as a $\mathbb{F}_{p}[[\mathbf{H}]]$-module. ###### Proof. Since $\mathbb{F}_{p}[[\mathbf{H}]]$ is a local ring, a profinite $\mathbb{F}_{p}[[\mathbf{H}]]$-module $M$ is finitely presented if and only if $H_{0}(\mathbf{H};M)$ and $H_{1}(\mathbf{H},M)$ are finite. Since $\mathbf{H}$ is finitely presented as a pro-$p$ group, [56, Theorem 7.8.1 and Theorem 7.8.3] imply that $H_{1}(\mathbf{H};\mathbb{F}_{p})$ and $H_{2}(\mathbf{H};\mathbb{F}_{p})$ are finite. However, we have that for $i\geq 0$, $H_{i}(\mathbf{H},I_{\mathbf{H}})=H_{i+1}(\mathbf{H};\mathbb{F}_{p})$. Thus, $I_{\mathbf{H}}$ is finitely presented as a $\mathbb{F}_{p}[[\mathbf{H}]]$-module. ∎ If $\mathbf{H}$ is a closed subgroup of $\mathbf{G}$, then $I_{\mathbf{H}}^{\mathbf{G}}$ denotes the closed left ideal of $\mathbb{F}_{p}[[\mathbf{G}]]$ generated by $I_{\mathbf{H}}$. ###### Lemma 2.11. Let $\mathbf{G}$ be a pro-$p$ group and let $\mathbf{H}\leq\mathbf{T}$ be closed subgroups of $\mathbf{G}$. Then the following holds. 1. (a) The continuous map $\displaystyle\mathbb{F}_{p}[[\mathbf{G}]]\widehat{\otimes}_{\mathbb{F}_{p}[[\mathbf{H}]]}I_{\mathbf{H}}\to I_{\mathbf{H}}^{\mathbf{G}}$ that sends $a\otimes b$ to $ab$, is an isomorphism of $\mathbb{F}_{p}[[\mathbf{G}]]$-modules. 2. (b) If $\mathbf{H}$ is finitely presented, the map $\displaystyle\mathbb{F}_{p}[[\mathbf{G}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}I_{\mathbf{H}}\to I_{\mathbf{H}}^{\mathbf{G}}$ that sends $a\otimes b$ to $ab$, is an isomorphism of $\mathbb{F}_{p}[[\mathbf{G}]]$-modules. 3. (c) If $\mathbf{T}$ is finitely presented and $\mathbf{H}$ is finitely generated, the map $\displaystyle\mathbb{F}_{p}[[\mathbf{G}]]\otimes_{\mathbb{F}_{p}[[\mathbf{T}]]}(I_{\mathbf{T}}/I_{\mathbf{H}}^{\mathbf{T}})\to I_{\mathbf{T}}^{\mathbf{G}}/I_{\mathbf{H}}^{\mathbf{G}}$ sending $a\otimes(b+I_{\mathbf{H}}^{\mathbf{T}})$ to $ab+I_{\mathbf{H}}^{\mathbf{G}}$, is an isomorphism of $\mathbb{F}_{p}[[\mathbf{G}]]$-modules. ###### Proof. (a) Consider an exact sequence $0\to I_{\mathbf{H}}\to\mathbb{F}_{p}[[\mathbf{H}]]\to\mathbb{F}_{p}\to 0.$ By Lemma 2.9, the sequence $0\to\mathbb{F}_{p}[[\mathbf{G}]]\widehat{\otimes}_{\mathbb{F}_{p}[[\mathbf{H}]]}I_{\mathbf{H}}\xrightarrow{\alpha}\mathbb{F}_{p}[[\mathbf{G}]]\xrightarrow{\beta}\mathbb{F}_{p}[[\mathbf{G}]]\widehat{\otimes}_{\mathbb{F}_{p}[[\mathbf{H}]]}\mathbb{F}_{p}\to 0$ is also exact. Thus, $\mathbb{F}_{p}[[\mathbf{G}]]\widehat{\otimes}_{\mathbb{F}_{p}[[\mathbf{H}]]}I_{\mathbf{H}}$ is isomorphic to $\ker\beta=I_{\mathbf{H}}^{\mathbf{G}}$. (b) Since $\mathbf{H}$ is finitely presented, by Lemma 2.10, $I_{\mathbf{H}}$ is finitely presented as $\mathbb{F}_{p}[[\mathbf{H}]]$-module, and so, by Lemma 2.6, $\mathbb{F}_{p}[[\mathbf{G}]]\widehat{\otimes}_{\mathbb{F}_{p}[[\mathbf{H}]]}I_{\mathbf{H}}\cong\mathbb{F}_{p}[[\mathbf{G}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}I_{\mathbf{H}}.$ Thus we conclude that the natural map from $\mathbb{F}_{p}[[\mathbf{G}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}I_{\mathbf{H}}$ to $I_{\mathbf{H}}^{\mathbf{G}}$ is also an isomorphism by (a). (c) Consider now the exact sequence $0\to I_{\mathbf{H}}^{\mathbf{T}}\to I_{\mathbf{T}}\to I_{\mathbf{T}}/I_{\mathbf{H}}^{\mathbf{T}}\to 0.$ Applying $\mathbb{F}_{p}[[\mathbf{G}]]\widehat{\otimes}_{\mathbb{F}_{p}[[\mathbf{T}]]}$ and taking again into account that $\mathbb{F}_{p}[[\mathbf{G}]]$ is a free profinite $\mathbb{F}_{p}[[\mathbf{T}]]$-module (Lemma 2.10), we obtain the exact sequence $0\to I_{\mathbf{H}}^{\mathbf{G}}\to I_{\mathbf{T}}^{\mathbf{G}}\to\mathbb{F}_{p}[[\mathbf{G}]]\widehat{\otimes}_{\mathbb{F}_{p}[[\mathbf{T}]]}(I_{\mathbf{T}}/I_{\mathbf{H}}^{\mathbf{T}})\to 0.$ Since $\mathbf{T}$ is finitely presented and $\mathbf{H}$ is finitely generated, by Lemma 2.10, $I_{\mathbf{T}}/I_{\mathbf{H}}^{\mathbf{T}}$ is finitely presented as $\mathbb{F}_{p}[[\mathbf{T}]]$-module. Thus, by Lemma 2.6, the canonical map $\mathbb{F}_{p}[[\mathbf{G}]]\otimes_{\mathbb{F}_{p}[[\mathbf{T}]]}(I_{\mathbf{T}}/I_{\mathbf{H}}^{\mathbf{T}})\to\mathbb{F}_{p}[[\mathbf{G}]]\widehat{\otimes}_{\mathbb{F}_{p}[[\mathbf{T}]]}(I_{\mathbf{T}}/I_{\mathbf{H}}^{\mathbf{T}})$ is an isomorphism. This proves the last claim. ∎ ###### Corollary 2.12. Let $\mathbf{H}$ be a finitely generated free subgroup of a pro-$p$ group $\mathbf{G}$. Then $I_{\mathbf{H}}^{\mathbf{G}}$ is a free $\mathbb{F}_{p}[[\mathbf{G}]]$-module of rank $d(\mathbf{H})$. ###### Proof. By Proposition 2.3, $I_{\mathbf{H}}$ is a free $\mathbb{F}_{p}[[\mathbf{H}]]$-module of rank $d(\mathbf{H})$. Therefore, by Lemma 2.11(b), $I_{\mathbf{H}}^{\mathbf{G}}$ is a free $\mathbb{F}_{p}[[\mathbf{G}]]$-module of rank $d(\mathbf{H})$. ∎ ### 2.5. On amalgamated products of groups Let $G$ be a group and $H_{1}$ and $H_{2}$ two subgroups that generate $G$ and have intersection $A=H_{1}\cap H_{2}$. The following result gives an algebraic condition for $G$ to be isomorphic to the amalgamated product of $H_{1}$ and $H_{2}$ over $A$. ###### Proposition 2.13. [40] Let $k$ be a non-trivial commutative ring. Then the canonical map $H_{1}*_{A}H_{2}\to G$ is an isomorphism if and only if $I_{H_{1}}^{G}\cap I_{H_{2}}^{G}=I_{A}^{G}$ in $k[G]$. ###### Proof. This is [40, Theorem 1], which is proved for $k=\mathbb{Z}$ but the proof works over an arbitrary nonzero commutative ring $k$. ∎ ### 2.6. On convergence of Sylvester rank functions Let $R$ be a ring. A Sylvester matrix rank function $\operatorname{rk}$ on $R$ is a function that assigns a non-negative real number to each matrix over $R$ and satisfies the following conditions. 1. (SMat1) $\operatorname{rk}(M)=0$ if $M$ is any zero matrix and $\operatorname{rk}(1)=1$ (where 1 denotes the identity matrix of size one); 2. (SMat2) $\operatorname{rk}(M_{1}M_{2})\leq\min\\{\operatorname{rk}(M_{1}),\operatorname{rk}(M_{2})\\}$ for any matrices $M_{1}$ and $M_{2}$ which can be multiplied; 3. (SMat3) $\operatorname{rk}\left(\begin{array}[]{cc}M_{1}&0\\\ 0&M_{2}\end{array}\right)=\operatorname{rk}(M_{1})+\operatorname{rk}(M_{2})$ for any matrices $M_{1}$ and $M_{2}$; 4. (SMat4) $\operatorname{rk}\left(\begin{array}[]{cc}M_{1}&M_{3}\\\ 0&M_{2}\end{array}\right)\geq\operatorname{rk}(M_{1})+\operatorname{rk}(M_{2})$ for any matrices $M_{1}$, $M_{2}$ and $M_{3}$ of appropriate sizes. We denote by $\mathbb{P}(R)$ the set of Sylvester matrix rank functions on $R$, which is a compact convex subset of the space of functions on matrices over $R$ considered with pointwise convergence. Many problems can be reinterpreted in terms of convergence in $\mathbb{P}(R)$. For example, if $G$ is group and $G\geq G_{1}\geq G_{2}\geq\ldots$ is a chain of normal subgroups of $G$ of finite index with trivial intersection, then the fact of the existence of the Lück approximation over $\mathbb{Q}$ and $\mathbb{C}$ can be viewed as the convergence of $\operatorname{rk}_{G/G_{i}}$ to $\operatorname{rk}_{G}$ in $\mathbb{P}(\mathbb{Q}[G])$ and $\mathbb{P}(\mathbb{C}[G])$ respectively (see [45, 30] for details). An alternative way to introduce Sylvester rank functions is via Sylvester module rank functions. A Sylvester module rank function $\dim$ on $R$ is a function that assigns a non-negative real number to each finitely presented $R$-module and satisfies the following conditions. 1. (SMod1) $\dim\\{0\\}=0$, $\dim R=1$; 2. (SMod2) $\dim(M_{1}\oplus M_{2})=\dim M_{1}+\dim M_{2}$; 3. (SMod3) if $M_{1}\to M_{2}\to M_{3}\to 0$ is exact then $\dim M_{1}+\dim M_{3}\geq\dim M_{2}\geq\dim M_{3}.$ By [50, Theorem 4] (see also [43, Proposition 1.2.8]), there exists a natural bijection between Sylvester matrix and module rank functions over a ring. Given a Sylvester module rank function $\dim$ on $R$ and a finitely presented $R$-module $M\cong R^{n}/R^{m}A$ ($A$ is a matrix over $R$), we define the corresponding Sylvester matrix rank function $\operatorname{rk}$ by means of $\operatorname{rk}(A)=n-\dim M$. By a recent result of Li [41], any Sylvester module rank function $\dim$ on $R$ can be extended to a unique function (satisfying some natural conditions) on arbitrary modules over $R$. We will call this extension, the extended Sylvester module rank function and denote it also by $\dim$. In this paper we will mostly use this extension for finitely generated modules $M$. In this case $\dim M$ is defined as (2) $\dim M=\inf\\{\dim\widetilde{M}:\ \widetilde{M}\textrm{\ is finitely presented and\ }M\textrm{\ is a quotient of\ }\widetilde{M}\\}.$ In the case of an arbitrary $R$-module $M$ the formula for $\dim M$ is more complex. In this case we put $\dim M=\sup_{M_{1}}\inf_{M_{2}}(\dim M_{2}-\dim(M_{2}/M_{1})),$ where $M_{1}\leq M_{2}$ are finitely generated $R$-submodules of $M$. Observe that we allow $+\infty$ to be a value of $\dim M$. ###### Remark 2.14. If $\operatorname{rk}$, $\operatorname{rk}_{i}\in\mathbb{P}(R)$ $(i\in\mathbb{N})$ are Sylvester matrix rank functions corresponding to Sylvester module rank functions $\dim$, $\dim_{i}$, respectively, then $\operatorname{rk}=\displaystyle\lim_{i\to\infty}\operatorname{rk}_{i}$ in the space $\mathbb{P}(R)$ if and only if for any finitely presented $R$-module $M$, $\dim M=\displaystyle\lim_{i\to\infty}\dim_{i}M$. However, the existence of the limit $\operatorname{rk}=\displaystyle\lim_{i\to\infty}\operatorname{rk}_{i}$ does not imply that $\dim M=\displaystyle\lim_{i\to\infty}\dim_{i}M$ for any finitely generated $R$-module $M$. This phenomenon is well-known. For example, it explains why the Lück approximation of the first $L^{2}$-Betti numbers is valid for finitely presented groups but not always valid for finitely generated groups (see [46]). If $M$ is a finitely generated $R$-module, we only always have that (3) $\dim M\geq\limsup_{i\to\infty}\dim_{i}M.$ Indeed, let $\mathcal{F}$ be the set of all finitely presented $R$-modules $\tilde{M}$ such that $M$ is a quotient of $\tilde{M}$. Then $\dim M=\inf_{\tilde{M}\in\mathcal{F}}\dim\tilde{M}=\inf_{\tilde{M}\in\mathcal{F}}\lim_{i\to\infty}\dim_{i}\tilde{M}\geq\\\ \limsup_{i\to\infty}\inf_{\tilde{M}\in\mathcal{F}}\dim_{i}\tilde{M}=\limsup_{i\to\infty}\dim_{i}M.$ In this subsection we will explain how to overcome this problem in some situations. For two Sylvester rank functions $\operatorname{rk}_{1}$ and $\operatorname{rk}_{2}\in\mathbb{P}(R)$ we write $\operatorname{rk}_{1}\geq\operatorname{rk}_{2}$ if $\operatorname{rk}_{1}(A)\geq\operatorname{rk}_{2}(A)$ for any matrix $A$ over $R$. If $\dim_{1}$ and $\dim_{2}$ are the Sylvester module rank functions on $R$ corresponding to $\operatorname{rk}_{1}$ and $\operatorname{rk}_{2}$, then the condition $\operatorname{rk}_{1}\geq\operatorname{rk}_{2}$ is equivalent to the condition $\dim_{1}\leq\dim_{2}$, meaning that $\dim_{1}M\leq\dim_{2}M$ for any finitely presented $R$-module $M$. ###### Proposition 2.15. Let $R$ be a ring and let $\operatorname{rk}$, $\operatorname{rk}_{i}\in\mathbb{P}(R)$ $(i\in\mathbb{N})$ be Sylvester matrix rank functions on $R$ corresponding to (extended) Sylvester module rank functions $\dim$, $\dim_{i}$ respectively. Assume that $\operatorname{rk}=\displaystyle\lim_{i\to\infty}\operatorname{rk}_{i}$ and for all $i$, $\operatorname{rk}\geq\operatorname{rk}_{i}$. Then, for any finitely generated $R$-module $M$, $\dim M=\displaystyle\lim_{i\to\infty}\dim_{i}M.$ ###### Proof. Fix $\varepsilon>0$. Let $k$ be such that $\liminf_{i\to\infty}\dim_{i}M\geq\dim_{k}M-\varepsilon.$ There exists a finitely presented $R$-module $\tilde{M}$ such that $M$ is a quotient of $\tilde{M}$ and $\dim_{k}M\geq\dim_{k}\tilde{M}-\varepsilon$. Since $\operatorname{rk}\geq\operatorname{rk}_{k}$, we have that $\dim_{k}\tilde{M}\geq\dim\tilde{M}$. Thus, we obtain $\liminf_{i\to\infty}\dim_{i}M\geq\dim_{k}M-\varepsilon\geq\dim_{k}\tilde{M}-2\varepsilon\geq\dim\tilde{M}-2\varepsilon\geq\dim M-2\varepsilon.$ Since $\varepsilon$ is arbitrary, we conclude that $\displaystyle\liminf_{i\to\infty}\dim_{i}M\geq\dim M$. In view of (3), we are done. ∎ ### 2.7. Epic division $R$-rings Let $R$ be a ring. An epic division $R$-ring is a $R$-ring $\phi:R\to\operatorname{\mathcal{D}}$ where $\operatorname{\mathcal{D}}$ is a division ring generated by $\phi(R)$. Moreover, we say that $\operatorname{\mathcal{D}}$ is a division $R$-ring of fractions if $\phi$ is injective. In this case we will normally omit $\phi$ and see $R$ as a subring of $\operatorname{\mathcal{D}}$. Each epic division $R$-ring $\operatorname{\mathcal{D}}$ induces a Sylvester module rank function $\dim_{\operatorname{\mathcal{D}}}$ on $R$: for every a finitely presented $R$-module $M$ we define $\dim_{\operatorname{\mathcal{D}}}M$ to be equal to the dimension of $\operatorname{\mathcal{D}}\otimes_{R}M$ as a $\operatorname{\mathcal{D}}$-module. The extended Sylvester rank function $\dim_{\operatorname{\mathcal{D}}}$ is calculated in the same way: for an $R$-module $M$, $\dim_{\operatorname{\mathcal{D}}}M$ is equal to the dimension of $\operatorname{\mathcal{D}}\otimes_{R}M$ as a $\operatorname{\mathcal{D}}$-module. We will use $\dim_{\operatorname{\mathcal{D}}}$ for the (extended) Sylvester module rank function on $R$ and for the $\operatorname{\mathcal{D}}$-dimension of $\operatorname{\mathcal{D}}$-spaces. This is a coherent notation because, since $\operatorname{\mathcal{D}}$ is epic, $\operatorname{\mathcal{D}}\otimes_{R}\operatorname{\mathcal{D}}$ is isomorphic to $\operatorname{\mathcal{D}}$ as $R$-bimodule (see [31, Section 4]). The following result of P. M. Cohn will be used several times in the paper. ###### Proposition 2.16. [18, Theorem 4.4.1] Two epic division $R$-rings $\operatorname{\mathcal{D}}_{1}$ and $\operatorname{\mathcal{D}}_{2}$ are isomorphic as $R$-rings if and only if $\dim_{\operatorname{\mathcal{D}}_{1}}M=\dim_{\operatorname{\mathcal{D}}_{2}}M$ for every finitely presented $R$-module $M$. ### 2.8. Natural extensions of Sylvester rank functions Let $G$ be a group with trivial element $e$. We say that a ring $R$ is $G$-graded if $R$ is equal to the direct sum $\oplus_{g\in G}R_{g}$ and $R_{g}R_{h}\subseteq R_{gh}$ for all $g$ and $h$ in $G$. If for each $g\in G$, $R_{g}$ contains an invertible element $u_{g}$, then we say that $R$ is a crossed product of $R_{e}$ and $G$ and we will write $R=S*G$ if $R_{e}=S$. Let $R=S*G$ be a crossed product. Let $\operatorname{rk}$ be a Sylvester matrix rank function on $S$ and $\dim$ its associated Sylvester module rank function. We say that $\operatorname{rk}$ (and $\dim$) are $R$-compatible if for any $g\in G$ and any matrix $A$ over $S$, $\operatorname{rk}(A)=\operatorname{rk}(u_{g}Au_{g}^{-1})$. If $G$ is finite and $M$ is a finitely presented $R$-module, then $M$ is also a finitely presented $S$-module. Thus, if $\dim$ is $R$-compatible, we can define (4) $\widetilde{\dim}\ M=\frac{\dim M}{|G|},$ where $M$ is a finitely presented $R$-module. This defines a Sylvester module rank function on $R$, called the natural extension of $\dim$. This notion was studied, for example, in [36]. We notice that the same formula (4) holds also for extended Sylvester module rank functions (that is, when $M$ is an arbitrary $R$-module). In this subsection we prove the following result. ###### Proposition 2.17. Let $R=S*G$ be a crossed product with $G$ finite and let $R\hookrightarrow\operatorname{\mathcal{D}}$ be a division $R$-ring of fractions. Denote by $\operatorname{\mathcal{D}}_{e}$ the division closure of $S$ in $\operatorname{\mathcal{D}}$. Then the following are equivalent. 1. (a) $\widetilde{\dim_{\operatorname{\mathcal{D}}_{e}}}=\dim_{\operatorname{\mathcal{D}}}$ as Sylvester functions on $R$. 2. (b) $\dim_{\operatorname{\mathcal{D}}_{e}}\operatorname{\mathcal{D}}=|G|$. 3. (c) $\operatorname{\mathcal{D}}$ is isomorphic to a crossed product $\operatorname{\mathcal{D}}_{e}*G$. 4. (d) $\operatorname{\mathcal{D}}$ is isomorphic to $\operatorname{\mathcal{D}}_{e}\otimes_{S}R$ as $(\operatorname{\mathcal{D}}_{e},R)$-bimodule. 5. (e) $\operatorname{\mathcal{D}}$ is isomorphic to $R\otimes_{S}\operatorname{\mathcal{D}}_{e}$ as $(R,\operatorname{\mathcal{D}}_{e})$-bimodule. ###### Proof. For any $h,g\in G$, define $\alpha(g,h)=u_{g}u_{h}(u_{gh})^{-1}\in R$. Since the conjugation by $u_{g}$ fixes $S$, it also fixes $\operatorname{\mathcal{D}}_{e}$. Therefore, we can define a ring structure on $T=\oplus_{g\in G}\operatorname{\mathcal{D}}_{e}v_{g}$ defining the multiplication on homogenous elements by means of $(d_{1}v_{g})\cdot(d_{2}v_{h})=(d_{1}u_{g}d_{2}(u_{g})^{-1}\alpha(g,h))v_{gh},\ d_{1},d_{2}\in\operatorname{\mathcal{D}}_{e},\ g,h\in G.$ It is clear that $T=\operatorname{\mathcal{D}}_{e}*G$ is a crossed product, it contains $R$ as a subring, and $T$ is isomorphic to $\operatorname{\mathcal{D}}_{e}\otimes_{S}R$ as $(\operatorname{\mathcal{D}}_{e},R)$-bimodule. There exists a natural map $\gamma:T\to\operatorname{\mathcal{D}}$, sending $\sum_{g\in G}d_{g}v_{g}$ ($d_{g}\in D_{e}$) to $\sum_{g\in G}d_{g}u_{g}$. Observe that $\gamma(T)$ is a domain and of finite dimension over $\operatorname{\mathcal{D}}_{e}$. Thus, $\gamma(T)$ is a division subring of $\operatorname{\mathcal{D}}$. Since $T$ contains $R$, $\operatorname{\mathcal{D}}=\gamma(T)$. This implies that (c) and (d) are equivalent. Since $\dim_{D_{e}}T=|G|$, (b) implies that $\gamma$ is an isomorphism, and so, (b) implies (c). Now, let us assume (d). Let $M$ be a finitely presented $R$-module. We have the following. $\widetilde{\dim_{\operatorname{\mathcal{D}}_{e}}}\ M=\frac{\dim_{\operatorname{\mathcal{D}}_{e}}(\operatorname{\mathcal{D}}_{e}\otimes_{S}M)}{|G|}=\frac{\dim_{\operatorname{\mathcal{D}}_{e}}(\operatorname{\mathcal{D}}_{e}\otimes_{S}(R\otimes_{R}M))}{|G|}=\\\ \frac{\dim_{\operatorname{\mathcal{D}}_{e}}((\operatorname{\mathcal{D}}_{e}\otimes_{S}R)\otimes_{R}M))}{|G|}\stackrel{{\scriptstyle\text{(d)}}}{{=}}\frac{\dim_{\operatorname{\mathcal{D}}_{e}}(\operatorname{\mathcal{D}}\otimes_{R}M)}{|G|}=\\\ \dim_{\operatorname{\mathcal{D}}}(\operatorname{\mathcal{D}}\otimes_{R}M)=\dim_{\operatorname{\mathcal{D}}}M.$ This proves (a). Now, we assume that (a) holds. Since $\operatorname{\mathcal{D}}_{e}$ es an epic $S$-ring $\operatorname{\mathcal{D}}_{e}\otimes_{S}\operatorname{\mathcal{D}}_{e}$ is isomorphic to $\operatorname{\mathcal{D}}_{e}$ as $\operatorname{\mathcal{D}}_{e}$-bimodule and by the same reason, $\operatorname{\mathcal{D}}\otimes_{R}\operatorname{\mathcal{D}}$ is isomorphic to $\operatorname{\mathcal{D}}$ as $\operatorname{\mathcal{D}}$-bimodule. Consider $M=\operatorname{\mathcal{D}}$ as an $R$-module and $N=\operatorname{\mathcal{D}}_{e}$ as a $S$-module. Then $1=\dim_{\operatorname{\mathcal{D}}}(\operatorname{\mathcal{D}}\otimes_{R}M)=\dim_{\operatorname{\mathcal{D}}}M\stackrel{{\scriptstyle\text{(a)}}}{{=}}\widetilde{\dim_{\operatorname{\mathcal{D}}_{e}}}M=\\\ \frac{\dim_{\operatorname{\mathcal{D}}_{e}}(\operatorname{\mathcal{D}}_{e}\otimes_{S}M)}{|G|}=\frac{\dim_{\operatorname{\mathcal{D}}_{e}}(\operatorname{\mathcal{D}}_{e}\otimes_{S}(N^{\dim_{\operatorname{\mathcal{D}}_{e}}\operatorname{\mathcal{D}}}))}{|G|}=\frac{\dim_{\operatorname{\mathcal{D}}_{e}}\operatorname{\mathcal{D}}}{|G|}$ This implies (b). Let $R^{op}$ denotes the opposite ring, that is the ring with the same elements and addition operation, but with the multiplication performed in the reverse order. Then $R^{op}\cong S^{op}*G$ and the condition (c) is equivalent to 1. (c’) $\operatorname{\mathcal{D}}^{op}$ is isomorphic to a crossed product $(\operatorname{\mathcal{D}}_{e})^{op}*G$. Our previous proof gives that (c’) is equivalent to 1. (d’) $\operatorname{\mathcal{D}}^{op}$ is isomorphic to $(\operatorname{\mathcal{D}}_{e})^{op}\otimes_{S^{op}}R^{op}$ as $((\operatorname{\mathcal{D}}_{e})^{op},R^{op})$-bimodule. Now, observe that (d’) is equivalent to (e). ∎ ## 3\. On mod-$p$ $L^{2}$-Betti numbers of subgroups of a free pro-$p$ groups ### 3.1. Universal division ring of fractions Given two epic division $R$-rings $\operatorname{\mathcal{D}}_{1}$ and $\operatorname{\mathcal{D}}_{2}$ the condition $\dim_{\operatorname{\mathcal{D}}_{1}}\leq\dim_{\operatorname{\mathcal{D}}_{2}}$ is equivalent to the existence of a specialization from $\operatorname{\mathcal{D}}_{1}$ to $\operatorname{\mathcal{D}}_{2}$ in the sense of P. Cohn ([18, Subsection 4.1]). We say that an epic division $R$-ring $\operatorname{\mathcal{D}}$ is universal if for every division $R$-ring $\mathcal{E}$, $\dim_{\operatorname{\mathcal{D}}}\leq\dim_{\mathcal{E}}$. If a universal epic division $R$-ring exists, it is unique up to $R$-isomorphism. We will denote it by $\operatorname{\mathcal{D}}_{R}$ and instead of $\dim_{\operatorname{\mathcal{D}}_{R}}$ we will simply write $\dim_{R}$. We say that a ring $R$ is a semifir if every finitely generated left ideal of $R$ is free and free modules of distinct finite rank are non-isomorphic. For example, if $K$ is a field, the ring $K\langle\\!\langle X\rangle\\!\rangle$ of non-commutative power series is a semifir ([17, Theorem 2.9.4]). By a theorem of P. M. Cohn [16] a semifir $R$ has a universal division $R$-ring. P. M. Cohn proved that in this case $\operatorname{\mathcal{D}}_{R}$ can be obtained from $R$ by formally inverting all full matrices over $R$. In particular, this implies the following result. ###### Proposition 3.1. Let $R$ be a semifir. Then $\dim_{R}$ is the smallest Sylvester module rank function among all the Sylvester module rank functions on $R$. We will need the following result. ###### Proposition 3.2. [32, Proposition 2.2] Let $R$ be a semifir. Then $\operatorname{Tor}^{R}_{1}(\operatorname{\mathcal{D}}_{R},M)=0$ for any $R$-submodule of a $\operatorname{\mathcal{D}}_{R}$-module. Let $G$ be a residually torsion-free nilpotent group (for example, $G$ is a subgroup of a free pro-$p$ group). Let $K$ be a field. Then the universal division ring of fractions $\operatorname{\mathcal{D}}_{K[G]}$ exists (see [33]). It can be constructed in the following way. Since $G$ is residually torsion free nilpotent, $G$ is bi-orderable. Fix a bi-invariant order $\preceq$ on $G$. A. Malcev [49] and B. Neumann [54] (following an idea of H. Hahn [24]) showed independently that the set $K((G,\preceq))$ of formal power series over $G$ with coefficients in $K$ having well-ordered support has a natural structure of a ring and, moreover, it is a division ring. $\operatorname{\mathcal{D}}_{K[G]}$ can be defined as the division closure of $K[G]$ in $K((G,\preceq))$. The universality of this division ring is shown in [33, Theorem 1.1]. If $A$ is a torsion-free abelian group, then $\operatorname{\mathcal{D}}_{K[A]}$ coincides with the classical ring of fractions $\mathcal{Q}(K[A])$ of $K[A]$. ### 3.2. The division ring $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$ If $\mathbf{F}$ is a free pro-$p$ group freely generated by $f_{1},\ldots,f_{d}$, then, by Proposition 2.3, the continuous homomorphism $\mathbb{F}_{p}\langle\\!\langle x_{1},\ldots,x_{d}\rangle\\!\rangle\to\mathbb{F}_{p}[[\mathbf{F}]]$ that sends $x_{i}$ to $f_{1}-1$, is an isomorphism. Thus, there exists a universal division ring of fraction $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$. Using results of [32] we establish the following formula for $\dim_{\mathbb{F}_{p}[[\mathbf{F}]]}$ which is one of main ingredients of the proof of Theorem 1.2. ###### Proposition 3.3. Let $\mathbf{F}=\mathbf{N}_{1}>\mathbf{N}_{2}>\mathbf{N}_{3}>\ldots$ be a chain of open normal subgroups of a finitely generated free pro-$p$ group $\mathbf{F}$ with trivial intersection. Let $M$ be a finitely generated $\mathbb{F}_{p}[[\mathbf{F}]]$-module. Then $\dim_{\mathbb{F}_{p}[[\mathbf{F}]]}M=\lim_{i\to\infty}\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}_{i}]\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}M)}{|\mathbf{F}:\mathbf{N}_{i}|}.$ ###### Proof. Let $\mathbf{N}$ be a normal open subgroup of $\mathbf{F}$ and let $\dim_{\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}]}$ be a Sylvester module rank function on $\mathbb{F}_{p}[[\mathbf{F}]]$ defined by $\dim_{\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}]}L=\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}]\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}L)}{|\mathbf{F}:\mathbf{N}|},$ where $L$ is a finitely presented $\mathbb{F}_{p}[[\mathbf{F}]]$-module. Let $M=\mathbb{F}_{p}[[\mathbf{F}]]^{n}/I$ be a finitely generated $\mathbb{F}_{p}[[\mathbf{F}]]$-module. Since ${\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}M\cong\mathbb{F}[[\mathbf{G}]]^{n}/(I+(I_{\mathbf{N}}^{\mathbf{G}})^{n})$ is finite, it is finitely presented by Lemma 2.10. Therefore, there exists a finitely generated $\mathbb{F}_{p}[[\mathbf{G}]]$-submodule $J$ of $I$ such that $I+(I_{N}^{G})^{n}=J+(I_{N}^{G})^{n}$. Put $\widetilde{M}=\mathbb{F}_{p}[[\mathbf{F}]]^{n}/J$. Then $\widetilde{M}$ is a finitely presented $\mathbb{F}_{p}[[\mathbf{F}]]$-module $\widetilde{M}$ satisfying the following conditions: 1. (1) $M$ is a quotient of $\widetilde{M}$ and 2. (2) ${\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}\widetilde{M}\cong{\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}M.$ Therefore, from (2) we obtain that $\dim_{\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}]}M=\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}]\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}\widetilde{M})}{|\mathbf{F}:\mathbf{N}|}=\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}]\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}M)}{|\mathbf{F}:\mathbf{N}|}.$ In the case where $M$ is finitely presented, [32, Theorem 1.4] and Remark 2.14 impliy that (5) $\dim_{\mathbb{F}_{p}[[\mathbf{F}]]}M=\lim_{i\to\infty}\dim_{\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}_{i}]}M.$ By Proposition 3.1, we also have that $\dim_{\mathbb{F}_{p}[[\mathbf{F}]]}M\leq\dim_{\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}_{i}]}M$ for all $i$. Therefore, Proposition 2.15 implies that (5) holds also when $M$ is finitely generated. Hence, $\dim_{\mathbb{F}_{p}[[\mathbf{F}]]}M=\lim_{i\to\infty}\dim_{\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}_{i}]}M=\lim_{i\to\infty}\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}_{i}]\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}M)}{|\mathbf{F}:\mathbf{N}_{i}|}.$ ∎ In the following proposition we collect some basic properties of $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$. ###### Proposition 3.4. Let $\mathbf{H}$ be a finitely generated closed subgroup of $\mathbf{F}$. The following holds. 1. (a) Let $\operatorname{\mathcal{D}}_{\mathbf{H}}$ be the division closure of $\mathbb{F}_{p}[[\mathbf{H}]]$ in $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$. Then $\operatorname{\mathcal{D}}_{\mathbf{H}}$ is isomorphic to $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}$ as an $\mathbb{F}_{p}[[\mathbf{H}]]$-ring. 2. (b) If $M$ is a finitely generated $\mathbb{F}_{p}[[\mathbf{H}]]$-module, then $\dim_{\mathbb{F}_{p}[[\mathbf{H}]]}(M)=\dim_{\mathbb{F}_{p}[[\mathbf{F}]]}(\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M).$ 3. (c) If $\mathbf{H}$ is open then, $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\cong\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}$ as $(\mathbb{F}_{p}[[\mathbf{F}]],\mathbb{F}_{p}[[\mathbf{H}]])$-bimodules. ###### Proof. (a) Fix a normal chain $\mathbf{F}=\mathbf{N}_{1}>\mathbf{N}_{2}>\mathbf{N}_{3}>\ldots$ of open normal subgroups of $\mathbf{F}$, and let $\mathbf{H}_{i}=\mathbf{N}_{i}\cap\mathbf{H}$. Let $M$ be a finitely generated $\mathbb{F}_{p}[[\mathbf{H}]]$-module. Observe first, that by Proposition 3.3, (6) $\dim_{\mathbb{F}_{p}[[\mathbf{H}]]}M=\lim_{i\to\infty}\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}[\mathbf{H}/\mathbf{H}_{i}]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M)}{|\mathbf{H}:\mathbf{H}_{i}|}.$ Considering $\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}_{i}]$ as a right $\mathbb{F}_{p}[[\mathbf{H}]]$-module, we obtain that $\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}_{i}]\cong\mathbb{F}_{p}[\mathbf{H}/\mathbf{H}_{i}]^{|\mathbf{F}:\mathbf{N}_{i}\mathbf{H}|}$ as right $\mathbb{F}_{p}[[\mathbf{H}]]$-modules. Thus, (7) $\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}[\mathbf{H}/\mathbf{H}_{i}]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M)=\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}_{i}]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M)}{|\mathbf{F}:\mathbf{N}_{i}\mathbf{H}|}.$ Therefore, from (6), (7) and Proposition 3.3, we conclude that (8) $\dim_{\mathbb{F}_{p}[[\mathbf{H}]]}M=\lim_{i\to\infty}\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}_{i}]\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}(\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M)}{|\mathbf{F}:\mathbf{N}_{i}|}=\\\ \dim_{\mathbb{F}_{p}[[\mathbf{F}]]}(\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M).$ On the other hand, we have $\dim_{\mathbb{F}_{p}[[\mathbf{F}]]}(\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M)=\\\ \dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}(\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M))=\\\ \dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M))=\dim_{\operatorname{\mathcal{D}}_{\mathbf{H}}}(\operatorname{\mathcal{D}}_{\mathbf{H}}\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M).$ Thus, we conclude that $\dim_{\mathbb{F}_{p}[[\mathbf{H}]]}M=\dim_{\operatorname{\mathcal{D}}_{\mathbf{H}}}(\operatorname{\mathcal{D}}_{\mathbf{H}}\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M),$ and so, by Proposition 2.16, $\operatorname{\mathcal{D}}_{\mathbf{H}}$ is isomorphic to $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}$ as an $\mathbb{F}_{p}[[\mathbf{H}]]$-ring. (b) This is the equality (8). (c) First assume that $\mathbf{H}$ is normal in $\mathbf{F}$. Observe that for large $i$, $\mathbf{N}_{i}\leq\mathbf{H}$. Let $M$ be a finitely presented $\mathbb{F}_{p}[[F]]$-module. Then we obtain that $\dim_{\mathbb{F}_{p}[[\mathbf{F}]]}M=\lim_{i\to\infty}\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}\otimes_{\mathbb{F}_{p}[[\mathbf{N}_{i}]]}M)}{|\mathbf{F}:\mathbf{N}_{i}|}=\\\ \lim_{i\to\infty}\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}\otimes_{\mathbb{F}_{p}[[\mathbf{N}_{i}]]}M)}{|\mathbf{F}:\mathbf{H}||\mathbf{H}:\mathbf{N}_{i}|}=\frac{\dim_{\mathbb{F}_{p}[[\mathbf{H}]]}M}{|\mathbf{F}:\mathbf{H}|}.$ Therefore, $\dim_{\mathbb{F}_{p}[[\mathbf{F}]]}=\widetilde{\dim_{\mathbb{F}_{p}[[\mathbf{H}]]}}$. Now, the result follows from Proposition 2.17. Now we assume that $\mathbf{H}$ is arbitrary, We argue by induction on $|\mathbf{F}:\mathbf{H}|$. If $\mathbf{H}$ has index $p$ in $\mathbf{F}$, then it is normal. If $|\mathbf{F}:\mathbf{H}|>p$, we find $\mathbf{H}_{1}$ of index $p$ in $\mathbf{F}$ containing $\mathbf{H}$. Then by induction, $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\cong\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}_{1}]]}\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}_{1}]]}\cong\\\ \mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}_{1}]]}(\mathbb{F}_{p}[[\mathbf{H}_{1}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]})\cong\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}.$ ∎ In view of the previous proposition, we will identify $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}$ and the division closure of $\mathbb{F}_{p}[[\mathbf{H}]]$ in $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$, and see $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}$ as a subring of $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$. ### 3.3. The division rings $\operatorname{\mathcal{D}}(\mathbb{F}_{p}[G],\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]})$ Let $G$ be a subgroup of $\mathbf{F}$. As we have explained in Subsection 3.1 there exists the universal division $\mathbb{F}_{p}[G]$-ring of fractions $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$. Let $\operatorname{\mathcal{D}}_{G}=\operatorname{\mathcal{D}}(\mathbb{F}_{p}[G],\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]})$ be the division closure of $\mathbb{F}_{p}[G]$ in $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$. In this subsection we will show that $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$ and $\operatorname{\mathcal{D}}_{G}$ are isomorphic as $\mathbb{F}_{p}[G]$-rings. In the case $G=F$ is a finitely generated free group and $\mathbf{F}$ is the pro-$p$ completion of $F$, this result follows from [17, Corollary 2.9.16]. ###### Proposition 3.5. Let $\mathbf{F}$ be a finitely generated free pro-$p$ group and let $G$ be a finitely generated subgroup of $\mathbf{F}$. Let $\mathbf{F}=\mathbf{N}_{1}>\mathbf{N}_{2}>\mathbf{N}_{3}>\ldots$ be a chain of open normal subgroups of $\mathbf{F}$ with trivial intersection. We put $G_{j}=G\cap\mathbf{N}_{j}$. Let $\operatorname{\mathcal{D}}_{G}=\operatorname{\mathcal{D}}(\mathbb{F}_{p}[G],\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]})$ be the division closure of $\mathbb{F}_{p}[G]$ in $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$. Then for every finitely generated $\mathbb{F}_{p}[G]$-module $M$, $\dim_{\operatorname{\mathcal{D}}_{G}}(\operatorname{\mathcal{D}}_{G}\otimes_{\mathbb{F}_{p}[G]}M)=\lim_{i\to\infty}\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}\otimes_{\mathbb{F}_{p}[G_{i}]}M)}{|G:G_{i}|}=\\\ \dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}\otimes_{\mathbb{F}_{p}[G]}M).$ In particular, the divison rings $\operatorname{\mathcal{D}}_{G}$ and $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$ are isomorphic as $\mathbb{F}_{p}[G]$-rings. ###### Proof. Without loss of generality we assume that $G$ is dense in $\mathbf{F}$. First observe that (9) $\dim_{\operatorname{\mathcal{D}}_{G}}(\operatorname{\mathcal{D}}_{G}\otimes_{\mathbb{F}_{p}[G]}M)=\dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[G]}M)=\\\ \dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}(\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[G]}M))=\dim_{\mathbb{F}_{p}[[\mathbf{F}]]}(\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[G]}M).$ Observe also that, since $G$ is dense in $\mathbf{F}$, $|\mathbf{F}:\mathbf{N}_{i}|=|G:G_{i}|$ and $\mathbb{F}_{p}\otimes_{\mathbb{F}_{p}[G_{i}]}M\cong\mathbb{F}_{p}[G/G_{i}]\otimes_{\mathbb{F}_{p}[G]}M\cong\\\ \mathbb{F}_{p}[\mathbf{F}/\mathbf{N}_{i}]\otimes_{\mathbb{F}_{p}[G]}M\cong\mathbb{F}_{p}[\mathbf{F}/\mathbf{N}_{i}]\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}(\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[G]}M)).$ Thus, Proposition 3.3 implies that (10) $\dim_{\operatorname{\mathcal{D}}_{G}}(\operatorname{\mathcal{D}}_{G}\otimes_{\mathbb{F}_{p}[G]}M)=\lim_{i\to\infty}\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}\otimes_{\mathbb{F}_{p}[G_{i}]}M)}{|G:G_{i}|}.$ Let $\mathbf{F}_{i}=\gamma_{i}(\mathbf{F})$ and we put $H_{i}=G\cap\mathbf{F}_{i}$. The ring $\mathbb{F}_{p}[G/H_{i}]$ is a Noetherian domain and its classical field of fractions $\mathcal{Q}(\mathbb{F}_{p}[G/H_{i}])$ is universal. Moreover, by [33, Theorem 1.2], we have that (11) $\dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}\otimes_{\mathbb{F}_{p}[G]}M)=\lim_{i\to\infty}\dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G/H_{i}]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G/H_{i}]}\otimes_{\mathbb{F}_{p}[G]}M).$ Observe that $\mathbb{F}_{p}[[\mathbf{F}/\mathbf{F}_{i}]]$ is also a Noetherian domain, and so the division closure of $\mathbb{F}_{p}[G/H_{i}]$ in $\mathcal{Q}(\mathbb{F}_{p}[[\mathbf{F}/\mathbf{F}_{i}]])$ is isomorphic to $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G/H_{i}]}$ (as a $\mathbb{F}_{p}[G]$-ring). Therefore, (12) $\dim_{\mathcal{Q}(\mathbb{F}_{p}[[\mathbf{F}/\mathbf{F}_{i}]])}(\mathcal{Q}(\mathbb{F}_{p}[[\mathbf{F}/\mathbf{F}_{i}]])\otimes_{\mathbb{F}_{p}[G]}M)=\dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G/H_{i}]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G/H_{i}]}\otimes_{\mathbb{F}_{p}[G]}M).$ Using Proposition 3.3 and arguing as in the proof of [29, Theorem 2.3], we obtain that (13) $\dim_{\mathbb{F}_{p}[[\mathbf{F}]]}(\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[G]}M)=\lim_{i\to\infty}\dim_{\mathcal{Q}(\mathbb{F}_{p}[[\mathbf{F}/F_{i}]])}(\mathcal{Q}(\mathbb{F}_{p}[[\mathbf{F}/\mathbf{F}_{i}]])\otimes_{\mathbb{F}_{p}[G]}M).$ Therefore, putting together (11), (12), (13), (9) and (10), we obtain that $\dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}\otimes_{\mathbb{F}_{p}[G]}M)=\dim_{\operatorname{\mathcal{D}}_{G}}(\operatorname{\mathcal{D}}_{G}\otimes_{\mathbb{F}_{p}[G]}M)=\\\ \lim_{i\to\infty}\frac{\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}\otimes_{\mathbb{F}_{p}[G_{i}]}M)}{|G:G_{i}|}.$ Applying Proposition 2.16, we obtain that the divison rings $\operatorname{\mathcal{D}}_{G}$ and $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$ are isomorphic as $\mathbb{F}_{p}[G]$-rings. ∎ An alternative approach of proving that $\operatorname{\mathcal{D}}_{G}$ is isomorphic to $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$ as $\mathbb{F}_{p}[G]$-ring is taken in [51, Lemma 7.5.5], where the result is proved by using a variation from [58, Theorem 6.3] of the uniqueness of Hughes-free division rings [27] (see also [19]). ### 3.4. Mod-$p$ $L^{2}$-Betti numbers $L^{2}$-Betti numbers play an important role in the solution of many problems in group theory. In the last years there was an attempt to develop a theory of mod-$p$ $L^{2}$-Betti numbers for different families of groups (see [31]). If $G$ is torsion-free and satisfies the Atiyah conjecture, P. Linnell [42] showed that $L^{2}$-Betti numbers of $G$ can be defined as $b_{i}^{(2)}(G)=\dim_{\operatorname{\mathcal{D}}(G)}H_{i}(G;\operatorname{\mathcal{D}}(G)),$ where $\operatorname{\mathcal{D}}(G)$ is the division ring obtained as the division closure of $\mathbb{Q}[G]$ in the ring of affilated operators $\mathcal{U}(G)$. It turns out that if $G$ is residually torsion-free nilpotent, $\operatorname{\mathcal{D}}(G)$ is isomorphic to the universal division ring of fractions $\operatorname{\mathcal{D}}_{\mathbb{Q}[G]}$ (see, for example, [34]). Therefore, we have $b_{i}^{(2)}(G)=\dim_{\operatorname{\mathcal{D}}_{\mathbb{Q}[G]}}H_{i}(G;\operatorname{\mathcal{D}}_{\mathbb{Q}[G]}).$ Thus, by analogy, if $G$ is a residually torsion-free nilpotent group, we define the $i$th mod-$p$ $L^{2}$-Betti number of $G$ as $\beta^{\operatorname{mod}p}_{i}(G)=\dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}}H_{i}(G;\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}).$ In the case, where $G$ is a subgroup of a free pro-$p$ group, we also obtain the following formula, which can be seen as a mod-$p$ analogue of the Lück approximation theorem [45]. ###### Proposition 3.6. Let $\mathbf{F}$ be a finitely generated free pro-$p$ group and let $G$ be a subgroup of $\mathbf{F}$ of type $FP_{k}$ for some $k\geq 1$. Let $\mathbf{F}=\mathbf{N}_{1}>\mathbf{N}_{2}>\mathbf{N}_{3}>\ldots$ be a chain of open normal subgroups of $\mathbf{F}$ with trivial intersection. We put $G_{j}=G\cap\mathbf{N}_{j}$. Then $\beta_{k}^{\operatorname{mod}p}(G)=\displaystyle\lim_{j\to\infty}\frac{\dim_{\mathbb{F}_{p}}H_{k}(G_{j};\mathbb{F}_{p})}{|G:G_{j}|}.$ ###### Proof. There exists a resolution of the $\mathbb{F}_{p}[G]$-module $\mathbb{F}_{p}$ $0\to R_{k}\to\mathbb{F}_{p}[G]^{n_{k}}\xrightarrow{\phi_{k}}\ldots\to\mathbb{F}_{p}[G]^{n_{1}}\xrightarrow{\phi_{1}}\mathbb{F}_{p}\xrightarrow{\phi_{0}}0$ with $R_{k}$ finitely generated. The relevant part of the sequence for calculation of $H_{k}(G;*)$ is the following exact sequence $0\to R_{k}\to\mathbb{F}_{p}[G]^{n_{k}}\to R_{k-1}\to 0,$ ($R_{k-1}=\operatorname{Im}\phi_{k}=\ker\phi_{k-1}$), because for any right $\mathbb{F}_{p}[G]$-module $M$ we have $0\to H_{k}(G;M)\to M\otimes_{\mathbb{F}_{p}[G]}R_{k}\to M^{n_{k}}\to M\otimes_{\mathbb{F}_{p}[G]}R_{k-1}\to 0$ Then we obtain that $\beta_{k}^{\operatorname{mod}p}(G)=\dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}}H_{k}(G;\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]})=\dim_{\mathbb{F}_{p}[G]}R_{k}-{n_{k}}+\dim_{\mathbb{F}_{p}[G]}R_{k-1}\textrm{\ and\ }$ $\dim_{\mathbb{F}_{p}}H_{k}(G_{j};\mathbb{F}_{p})=\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}\otimes_{\mathbb{F}_{p}[G_{j}]}R_{k})-n_{k}|G:G_{j}|+\dim_{\mathbb{F}_{p}}(\mathbb{F}_{p}\otimes_{\mathbb{F}_{p}[G_{j}]}R_{k-1}).$ Thus, Proposition 3.5 implies the proposition.∎ In this paper we will work only with $\beta_{1}^{\operatorname{mod}p}(G)$. Observe that in this case, if $G$ is infinite, we have the formula $\beta_{1}^{\operatorname{mod}p}(G)=\dim_{\mathbb{F}_{p}[G]}I_{G}-1.$ Also observe that if $A$ is a non-trivial torsion-free abelian group, then since $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[A]}=\mathcal{Q}(\mathbb{F}_{p}[A])$ is the field of fractions of $\mathbb{F}_{p}[A]$, (14) $\beta_{1}^{\operatorname{mod}p}(A)=\dim_{\mathbb{F}_{p}[A]}I_{A}-1=\dim_{\mathcal{Q}(\mathbb{F}_{p}[A])}(\mathcal{Q}(\mathbb{F}_{p}[A])\otimes_{\mathbb{F}_{p}[A]}I_{A})-1=0$ ### 3.5. Strong embeddings into free pro-$p$ groups Assume that a finitely generated group $G$ is a subgroup of a free pro-$p$ group $\mathbf{F}$. Since a closed subgroup of a free pro-$p$ group is free (see [56, Corollary 7.7.5]), changing $\mathbf{F}$ by the closure of $G$ in $\mathbf{F}$, we may assume that $G$ is dense in $\mathbf{F}$. Let $\mathbf{F}=\mathbf{N}_{1}>\mathbf{N}_{2}>\mathbf{N}_{3}>\ldots$ be a chain of open normal subgroups of $\mathbf{F}$ with trivial intersection and put $G_{j}=G\cap\mathbf{N}_{j}$. Observe that the closure of $G_{j}$ in $\mathbf{F}$ is equal to $\mathbf{N}_{j}$, and so (15) $|G_{j}:G_{j}^{p}[G_{j},G_{j}]|\geq|\mathbf{N}_{j}:\mathbf{N}_{j}^{p}[\mathbf{N}_{j},\mathbf{N}_{j}]|.$ Denote by $d(\mathbf{F})$ the number of profinite generators of $\mathbf{F}$. By [56, Theorem 7.8.1], $d(\mathbf{F})=\log_{p}|\mathbf{F}:\mathbf{F}^{p}[\mathbf{F},\mathbf{F}]|$ and by the index Schreier formula $d(\mathbf{N}_{j})=(d(\mathbf{F})-1)|\mathbf{F}:\mathbf{N}_{j}|+1.$ Therefore, we obtain $\dim_{\mathbb{F}_{p}}H_{1}(G_{j};\mathbb{F}_{p})=\log_{p}|G_{j}:G_{j}^{p}[G_{j},G_{j}]|\geq\log_{p}|\mathbf{N}_{j}:\mathbf{N}_{j}^{p}[\mathbf{N}_{j},\mathbf{N}_{j}]|=\\\ d(\mathbf{N}_{j})=(d(\mathbf{F})-1)|\mathbf{F}:\mathbf{N}_{j}|+1=(d(\mathbf{F})-1)|G:G_{j}|+1.$ Thus, Proposition 3.6 implies the following corollary. ###### Corollary 3.7. Let $G$ be a finitely generated dense subgroup of a free pro-$p$ group $\mathbf{F}$. Then (16) $\dim_{\mathbb{F}_{p}G}I_{G}=\beta_{1}^{\operatorname{mod}p}(G)+1\geq d(\mathbf{F}).$ This result suggests the following definition. ###### Definition 3.8. We say that an embedding $G\hookrightarrow\mathbf{F}$ of a finitely generated group $G$ into a free pro-$p$ group $\mathbf{F}$ is strong if $G$ is dense in $\mathbf{F}$ and $\dim_{\mathbb{F}_{p}G}I_{G}=\beta_{1}^{\operatorname{mod}p}(G)+1=d(\mathbf{F})$. A finitely generated group $G$ is called strongly embeddable in a free pro-$p$ group (SE($p$) for simplicity) if there are a free pro-$p$ group $\mathbf{F}$ and a strong embedding $G\hookrightarrow\mathbf{F}$. Let $G$ be a parafree group. Observe that $G$ is residually-$p$ for every prime $p$. Thus, if $G$ is finitely generated, its pro-$p$ completion $G_{\widehat{p}}$ is a finitely generated free pro-$p$ group and $G$ is a subgroup of $G_{\widehat{p}}$. In this case the inequality (15) is an equality, and so in the same way as we obtained the inequality (16), we obtain that $\beta_{1}^{\operatorname{mod}p}(G)=d(G_{\widehat{p}})-1$. Therefore, the embedding $G\hookrightarrow G_{\widehat{p}}$ is strong. Thus, all finitely generated parafree groups are SE($p$). On the other hand, not every subgroup of a parafree group is SE($p$). For example, the fundamental group of an oriented surface of genus greater than 1 is not SE($p$). However the fundamental group of an oriented surface of genus greater than 2 can be embedded in a parafree group (see [8, Section 4.1]). By [13, Proposition 7.5], if $G$ is a finitely generated dense subgroup of a finitely generated free pro-$p$ group $\mathbf{F}$, then $b_{1}^{(2)}(G)\geq d(\mathbf{F})-1$. On the other hand, by [21, Theorem 1.6] and Proposition 3.6 , $b_{1}^{(2)}(G)\leq\beta_{1}^{\operatorname{mod}p}(G)$. Thus, if $G\hookrightarrow\mathbf{F}$ is a strong embedding, we have $b_{1}^{(2)}(G)=\beta_{1}^{\operatorname{mod}p}(G)=d(\mathbf{F})-1$. ## 4\. On $\mathcal{D}$-torsion-free modules. ### 4.1. General results Let $R$ be a ring and let $R\hookrightarrow\mathcal{D}$ be an embedding of $R$ into a division ring $\mathcal{D}$. Let $M$ be a left $R$-module. We say that $M$ is $\mathcal{D}$-torsion-free if the canonical map $M\to\mathcal{D}\otimes_{R}M,\ m\mapsto 1\otimes m,$ is injective. The following lemma describes several equivalent definitions of $\mathcal{D}$-torsion-free modules. ###### Lemma 4.1. The following statements for a left $R$-module $M$ are equivalent. 1. (a) $M$ is $\mathcal{D}$-torsion-free. 2. (b) There are a $\operatorname{\mathcal{D}}$-module $N$ and an injective homomorphism $\varphi:M\to N$ of $R$-modules. 3. (c) For any $0\neq m\in M$, there exists a homomorphism of $R$-modules $\varphi:M\to\operatorname{\mathcal{D}}$, such that $\varphi(m)\neq 0$. ###### Proof. The proof is straightforward and we leave it to the reader.∎ Let $M$ be a left $R$-module. Recall that we use $\dim_{\operatorname{\mathcal{D}}}M$ to denote the dimension of $\operatorname{\mathcal{D}}\otimes_{R}M$ as a $\operatorname{\mathcal{D}}$-module. Observe that if $\dim_{\operatorname{\mathcal{D}}}M$ is finite, it is also equal to the dimension of $\operatorname{Hom}_{R}(M,\operatorname{\mathcal{D}})$ as a right $\mathcal{D}$-module. ###### Lemma 4.2. Let $0\to M_{1}\to M_{2}\to M_{3}\to 0$ be an exact sequence of $R$-modules. Assume that 1. (1) $M_{1}$ and $M_{3}$ are $\operatorname{\mathcal{D}}$-torsion-free, 2. (2) $\dim_{\operatorname{\mathcal{D}}}M_{1}$ and $\dim_{\operatorname{\mathcal{D}}}M_{3}$ are finite and 3. (3) $\dim_{\operatorname{\mathcal{D}}}M_{2}=\dim_{\operatorname{\mathcal{D}}}M_{1}+\dim_{\operatorname{\mathcal{D}}}M_{3}$. Then $M_{2}$ is also $\operatorname{\mathcal{D}}$-torsion-free. ###### Proof. We are going to use the third characterization of $\operatorname{\mathcal{D}}$-torsion-free modules from Lemma 4.1. Consider the following exact sequence of right $\mathcal{D}$-modules. $0\to\operatorname{Hom}_{R}(M_{3},\operatorname{\mathcal{D}})\to\operatorname{Hom}_{R}(M_{2},\operatorname{\mathcal{D}})\to\operatorname{Hom}_{R}(M_{1},\operatorname{\mathcal{D}}).$ Since $\dim_{\operatorname{\mathcal{D}}}M_{2}=\dim_{\operatorname{\mathcal{D}}}M_{1}+\dim_{\operatorname{\mathcal{D}}}M_{3}$, the last map is surjective. Let $m\in M_{2}$. If $m\not\in M_{1}$, then since $M_{3}$ is $\mathcal{D}$-torsion-free, there exists $\varphi\in\operatorname{Hom}_{R}(M_{2}/M_{1},\operatorname{\mathcal{D}})$ such that $\varphi(m+M_{1})\neq 0$. Hence there exists $\widetilde{\varphi}\in\operatorname{Hom}_{R}(M_{2},\operatorname{\mathcal{D}})$, such that $\widetilde{\varphi}(m)\neq 0$. If $m\in M_{1}$, then since $M_{1}$ is $\mathcal{D}$-torsion-free, there exists $\varphi\in\operatorname{Hom}_{R}(M_{1},\operatorname{\mathcal{D}})$ such that $\varphi(m)\neq 0$. Using that the restriction map $\operatorname{Hom}_{R}(M_{2},\operatorname{\mathcal{D}})\to\operatorname{Hom}_{R}(M_{1},\operatorname{\mathcal{D}})$ is surjective, we obtain again that there exists $\widetilde{\varphi}\in\operatorname{Hom}_{R}(M_{2},\operatorname{\mathcal{D}})$, such that $\widetilde{\varphi}(m)\neq 0$.∎ In the calculations of $\dim_{\operatorname{\mathcal{D}}}$ the following elementary lemma will be useful. ###### Lemma 4.3. Let $\operatorname{\mathcal{D}}$ be a division $R$-ring and $M$ be a $\operatorname{\mathcal{D}}$-torsion-free $R$-module of finite $\operatorname{\mathcal{D}}$-dimension. Let $L$ be a non-trivial $R$-submodule of $M$. Then $\dim_{\operatorname{\mathcal{D}}}(M/L)<\dim_{\operatorname{\mathcal{D}}}M$. Moreover, if $\dim_{\operatorname{\mathcal{D}}}L=1$, then $\dim_{\operatorname{\mathcal{D}}}(M/L)=\dim_{\operatorname{\mathcal{D}}}M-1$. ###### Proof. Since $M$ is $\operatorname{\mathcal{D}}$-torsion-free, $\operatorname{\mathcal{D}}\otimes_{R}(M/L)$ is a proper quotient of $\operatorname{\mathcal{D}}\otimes_{R}M$. Hence $\dim_{\operatorname{\mathcal{D}}}M/L<\dim_{\operatorname{\mathcal{D}}}M$. Now assume that $\dim_{\operatorname{\mathcal{D}}}L=1$. In this case $\dim_{\operatorname{\mathcal{D}}}(M/L)\geq\dim_{\operatorname{\mathcal{D}}}M-1$. Therefore, $\dim_{\operatorname{\mathcal{D}}}(M/L)=\dim_{\operatorname{\mathcal{D}}}M-1$.∎ ### 4.2. $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$-torsion- free modules Let $\mathbf{F}$ be a finitely generated free pro-$p$ group. The main purpose of this subsection is to prove the following result. ###### Proposition 4.4. Assume that $1\neq z\in\mathbf{F}$ is not a proper $p$-power of an element of $\mathbf{F}$. Denote by $\mathbf{Z}$ the closed subgroup of $\mathbf{F}$ generated by $z$. Then the module $I_{\mathbf{F}}/I_{\mathbf{Z}}^{\mathbf{F}}$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$-torsion-free. Before proving the proposition we have to establish several preliminary results. ###### Lemma 4.5. Let $\mathbf{H}$ be an open subgroup of $\mathbf{F}$ Let $M$ be a $\mathbb{F}_{p}[[\mathbf{H}]]$-module. Then $\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$-torsion-free if and only if $M$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}$-torsion-free. ###### Proof. Assume that $M$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}$-torsion-free. We have that the map $M\to\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M$ is injective. Then, since $\mathbb{F}_{p}[[\mathbf{F}]]$ is a free right $\mathbb{F}_{p}[[\mathbf{H}]]$-module, the map $\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M\xrightarrow{\alpha}\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M)$ is also injective. Consider the canonical isomorphism between tensor products $\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M)\xrightarrow{\beta}(\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]})\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M.$ By Propositopn 3.4(c), $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\cong\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}$ as $(\mathbb{F}_{p}[[\mathbf{F}]],\mathbb{F}_{p}[[\mathbf{H}]])$-bimodules. Thus, there exists an isomorphism of $\mathbb{F}_{p}[[\mathbf{F}]]$-modules $(\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H})]]})\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M\xrightarrow{\gamma}\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M.$ We put $\varphi=\gamma\circ\beta\circ\alpha$ and apply Lemma 4.1. Since $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M$ is a $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$-module and $\varphi$ is an injective $\mathbb{F}_{p}[[\mathbf{F}]]$-homomorphism, $\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$-torsion-free. Another direction of the proposition is clear because $M$ is a $\mathbb{F}_{p}[[\mathbf{H}]]$-submodule of $\mathbb{F}_{p}[[\mathbf{F}]]\otimes_{\mathbb{F}_{p}[[\mathbf{H}]]}M$.∎ ###### Lemma 4.6. Let $\mathbf{H}$ be an open normal subgroup of $\mathbf{F}$ and assume that $1\neq z\in\mathbf{H}$. Let $\mathbf{Z}$ be the closed subgroup of $\mathbf{H}$ generated by $z$. Then the following are equivalent. 1. (a) The $\mathbb{F}_{p}[[\mathbf{F}]]$-module $I_{\mathbf{H}}^{\mathbf{F}}/I^{\mathbf{F}}_{\mathbf{Z}}$ is not $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$-torsion-free. 2. (b) The $\mathbb{F}_{p}[[\mathbf{H}]]$-module $I_{\mathbf{H}}/I^{\mathbf{H}}_{\mathbf{Z}}$ is not $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}$-torsion-free. 3. (c) There are $a\in I_{\mathbf{H}}^{\mathbf{F}}$ and $b\in I_{\mathbf{F}}$ such that $ba=z-1$. 4. (d) There are $a\in I_{\mathbf{H}}^{\mathbf{F}}$ and $b\in I_{\mathbf{F}}$ such that $ab=z-1$. ###### Proof. (a)$\Longleftrightarrow$(b): This follows from Lemma 4.5 and Lemma 2.11. (c)$\Longrightarrow$(a): Put $N=\mathbb{F}_{p}[[\mathbf{F}]]a/I_{\mathbf{Z}}^{\mathbf{F}}$, Then $N=\mathbb{F}_{p}[[\mathbf{F}]]a/\mathbb{F}_{p}[[\mathbf{F}]](z-1)=\mathbb{F}_{p}[[\mathbf{F}]]a/\mathbb{F}_{p}[[\mathbf{F}]]ba.$ Since $b$ is not invertible in $\mathbb{F}_{p}[[\mathbf{F}]]$, $N$ is a non- trivial submodule of $I_{\mathbf{H}}^{\mathbf{F}}/I_{\mathbf{Z}}^{\mathbf{F}}$ and since $\mathbb{F}_{p}[[\mathbf{F}]]$ does not have non-trivial zero- divisors $\mathbb{F}_{p}[[\mathbf{F}]]a/\mathbb{F}_{p}[[\mathbf{F}]]ba\cong\mathbb{F}_{p}[[\mathbf{F}]]/\mathbb{F}_{p}[[\mathbf{F}]]b.$ Clearly $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}N=0$, and so (a) holds. (a)$\Longrightarrow$(c): We put $M=I_{\mathbf{H}}^{\mathbf{F}}/I^{\mathbf{F}}_{\mathbf{Z}}$ and let $\phi:M\to\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}M$ be the canonical map. Let $\overline{L}=L/I_{\mathbf{Z}}^{\mathbf{F}}$ be the kernel of $\phi$ and $\overline{M}$ the image of $\phi$. Hence we have the exact sequence $0\to L\to I_{\mathbf{H}}^{\mathbf{F}}\to\overline{M}\to 0.$ After applying $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}$ we obtain the exact sequence (17) $0\to\operatorname{Tor}_{1}^{\mathbb{F}_{p}[[\mathbf{F}]]}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]},\overline{M})\to\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}L\to\\\ (\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]})^{d(\mathbf{H})}\to\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}\overline{M}\to 0.$ By Corrollary 2.12, the $\mathbb{F}_{p}[[\mathbf{F}]]$-module $I_{\mathbf{H}}^{\mathbf{F}}$ is free of rank $d(\mathbf{H})$ and the $\mathbb{F}_{p}[[\mathbf{F}]]$-module $I_{\mathbf{Z}}^{\mathbf{F}}$ is cyclic. Thus, by Lemma 4.3, $\dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}M)=d(\mathbf{H})-1.$ Therefore, $\dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}\overline{M})=\dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}M)=d(\mathbf{H})-1.$ Observe also that by Proposition 3.2, $\operatorname{Tor}_{1}^{\mathbb{F}_{p}[[\mathbf{F}]]}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]},\overline{M})=0$. Therefore, (17) implies that $\dim_{\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}}(\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}\otimes_{\mathbb{F}_{p}[[\mathbf{F}]]}L)=1.$ By [29, Lemma 3.1] a profinite submodule of a free profinite $\mathbb{F}_{p}[[\mathbf{F}]]$-module is again free. Hence $L$ is a free profinite $\mathbb{F}_{p}[[\mathbf{F}]]$-module, and so $L$ should be a cyclic $\mathbb{F}_{p}[[\mathbf{F}]]$-module. We write $L=\mathbb{F}_{p}[[\mathbf{F}]]a$ for some $a\in I_{\mathbf{H}}^{\mathbf{F}}$. Then there exists $b\in\mathbb{F}_{p}[[\mathbf{F}]]$ such that $ba=z-1$. By our assumption $L\neq I_{\mathbf{Z}}^{\mathbf{F}}$. Thus, $b$ is not invertible, and so $b\in I_{\mathbf{F}}$. (c)$\Longrightarrow$(d): The map $g\mapsto g^{-1}$ on $\mathbf{F}$ can be extended to a continuous anti-isomorphism $\alpha:\mathbb{F}_{p}[[\mathbf{F}]]\to\mathbb{F}_{p}[[\mathbf{F}]]$. If $z-1=ba$, then $z^{-1}-1=\alpha(z-1)=\alpha(a)\alpha(b)$ and so $z-1=(-z\alpha(a))\alpha(b)$. Now note that $-z\alpha(a)\in I_{\mathbf{H}}\mathbb{F}_{p}[[\mathbf{F}]]$ and $\alpha(b)\in I_{\mathbf{F}}$. Since $\mathbf{H}$ is normal in $\mathbf{F}$, $I_{\mathbf{H}}\mathbb{F}_{p}[[\mathbf{F}]]=I_{\mathbf{H}}^{\mathbf{F}}$, and we obtain (d). (d)$\Longrightarrow$(c): It is proved in the same way as (c)$\Longrightarrow$(d).∎ Now we are ready to prove Proposition 4.4. ###### Proof of Proposition 4.4. If $\mathbf{F}$ is cyclic, then $I_{\mathbf{F}}=I_{\mathbf{Z}}^{F}$ and so $I_{\mathbf{F}}/I_{\mathbf{Z}}^{\mathbf{F}}$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$-torsion-free. Now we assume that $\mathbf{F}$ is not cyclic. There exists a normal open subgroup $\mathbf{N}$ of $\mathbf{F}$ such that $z\mathbf{N}$ is not a $p$-power in $\mathbf{F}/\mathbf{N}$. We will prove the proposition by induction on $|\mathbf{F}/\mathbf{N}|$. If $\mathbf{F}/\mathbf{N}$ is cyclic, then $z\not\in\Phi(\mathbf{F})$ and so $z$ is a member of a free generating system of $\mathbf{F}$. If $\\{z,x_{1},\ldots.x_{k}\\}$ is a free generating set of $F$, then $I_{\mathbf{F}}/I_{\mathbf{Z}}^{\mathbf{F}}=(\mathbb{F}_{p}[[\mathbf{F}]](z-1)\oplus(\oplus_{i=1}^{k}\mathbb{F}_{p}[[\mathbf{F}]](x_{i}-1)))/\mathbb{F}_{p}[[\mathbf{F}]](z-1)\cong\mathbb{F}_{p}[[\mathbf{F}]]^{k}$ is a free $\mathbb{F}_{p}[[\mathbf{F}]]$-module and we are done. Assume now that $\mathbf{F}/\mathbf{N}$ is not cyclic. Since $\Phi(\mathbf{F}/\mathbf{N})=\mathbf{N}\Phi(\mathbf{F})/\mathbf{N}$, the pro-$p$ group $\mathbf{F}/\mathbf{N}\Phi(\mathbf{F})$ is not cyclic as well. Let $\mathbf{M}$ be the closed subgroup of $\mathbf{F}$ containing the commutator subgroup $[\mathbf{F},\mathbf{F}]$ and the element $z$ and such that $\mathbf{M}/([\mathbf{F},\mathbf{F}]\mathbf{Z})$ is the torsion part of $\mathbf{F}/([\mathbf{F},\mathbf{F}]\mathbf{Z})$. Since $\mathbf{Z}$ is cyclic and $\mathbf{F}/([\mathbf{F},\mathbf{F}])$ is torsion-free and abelian, $\mathbf{M}\Phi(\mathbf{F})/\Phi(\mathbf{F})$ is non-trivial cyclic (if $z\not\in\Phi(\mathbf{F})$) or trivial (if $z\in\Phi(\mathbf{F})$). Therefore, since $\mathbf{F}/\mathbf{N}\Phi(\mathbf{F})$ is not cyclic, $\mathbf{M}\mathbf{N}$ is a proper subgroup of $\mathbf{F}$. By the construction of $\mathbf{M}$, $\mathbf{F}/\mathbf{M}\cong\mathbb{Z}_{p}^{k}$ for some $k\geq 1$. Since $\mathbf{M}\mathbf{N}$ is a proper subgroup of $\mathbf{F}$, $\mathbf{M}\mathbf{N}/\mathbf{M}$ is a proper subgroup of $\mathbf{F}/\mathbf{M}$. Therefore, there exists a surjective map $\sigma:\mathbf{F}\to\mathbb{Z}_{p}$ such that $\mathbf{M}\leq\ker\sigma$ and $\mathbf{N}\ker\sigma\neq\mathbf{F}$. We put $\mathbf{H}=\mathbf{N}\ker\sigma$ and extend $\sigma$ to the map $\widetilde{\sigma}:\mathbb{F}_{p}[[\mathbf{F}]]\to\mathbb{F}_{p}[[\mathbb{Z}_{p}]]$. Observe that $\ker\widetilde{\sigma}=I_{\ker\sigma}^{\mathbf{F}}$. By way of contradiction, assume that $I_{\mathbf{F}}/I_{\mathbf{Z}}^{\mathbf{F}}$ is not $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$-torsion-free. Then by Lemma 4.6, there are $a,b\in I_{\mathbf{F}}$ such that $ab=z-1$. Applying $\widetilde{\sigma}$ we obtain that $\widetilde{\sigma}(a)\widetilde{\sigma}(b)=0$. Since $\mathbb{F}_{p}[[\mathbb{Z}_{p}]]$ is a domain, either $a$ or $b$ lie in $\ker\widetilde{\sigma}=I_{\ker\sigma}^{\mathbf{F}}\subset I_{\mathbf{H}}^{\mathbf{F}}$. Applying again Lemma 4.6, we conclude that $I_{\mathbf{H}}/I_{\mathbf{Z}}^{\mathbf{H}}$ is not $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}$-torsion-free. However, observe that $\mathbf{N}$ is also a normal subgroup of $\mathbf{H}$, $z\mathbf{N}$ is not a $p$-power in $\mathbf{H}/\mathbf{N}$ and $|\mathbf{H}/\mathbf{N}|<|\mathbf{F}/\mathbf{N}|$. Thus, we can apply the inductive assumption and conclude that $I_{\mathbf{H}}/I_{\mathbf{Z}}^{\mathbf{H}}$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{H}]]}$-torsion-free. We have arrived to a contradiction. Thus, $I_{\mathbf{F}}/I_{\mathbf{Z}}^{\mathbf{F}}$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$-torsion-free. ∎ ### 4.3. $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free modules In this subsection we assume that $\mathbf{F}$ is a finitely generated free pro-$p$ group and $G$ is an (abstract) dense finitely generated subgroup of $\mathbf{F}$. First we prove the following analogue of Lemma 4.5. ###### Lemma 4.7. Let $H$ be a subgroup of $G$ and let $M$ be a $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[H]}$-torsion-free left $\mathbb{F}_{p}[H]$-module. Then $\mathbb{F}_{p}[G]\otimes_{\mathbb{F}_{p}[H]}M$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free. ###### Proof. Let $\operatorname{\mathcal{D}}_{H}$ be the division closure of $\mathbb{F}_{p}[H]$ in $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$. Observe that $\operatorname{\mathcal{D}}_{H}$ and $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[H]}$ are isomorphic as $\mathbb{F}_{p}[H]$-rings (it follows, for example, from Proposition 3.5). We have that the map $M\to\operatorname{\mathcal{D}}_{H}\otimes_{\mathbb{F}_{p}[H]}M$ is injective. Then, since $\mathbb{F}_{p}[G]$ is a free right $\mathbb{F}_{p}[H]$-module, the map $\mathbb{F}_{p}[G]\otimes_{\mathbb{F}_{p}[H]}M\xrightarrow{\alpha}\mathbb{F}_{p}[G]\otimes_{\mathbb{F}_{p}[H]}(\operatorname{\mathcal{D}}_{H}\otimes_{\mathbb{F}_{p}[H]}M)$ is also injective. Consider the canonical isomorphism between tensor products $\mathbb{F}_{p}[G]\otimes_{\mathbb{F}_{p}[H]}(\operatorname{\mathcal{D}}_{H}\otimes_{\mathbb{F}_{p}[H]}M)\xrightarrow{\beta}(\mathbb{F}_{p}[G]\otimes_{\mathbb{F}_{p}[H]}\operatorname{\mathcal{D}}_{H})\otimes_{\mathbb{F}_{p}[H]}M.$ By [23], $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$ is strongly Hughes- free. This means that the canonical map of $(\mathbb{F}_{p}[G],\mathbb{F}_{p}[H])$-bimodules $\mathbb{F}_{p}[G]\otimes_{\mathbb{F}_{p}[H]}\operatorname{\mathcal{D}}_{H}\to\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$ is injective. Moreover, the image of $\mathbb{F}_{p}[G]\otimes_{\mathbb{F}_{p}[H]}\operatorname{\mathcal{D}}_{H}$ is a direct summand of $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$ as a right $\operatorname{\mathcal{D}}_{H}$-submodule (and so, it is also a direct summand as a right $\mathbb{F}_{p}[H]$-submodule). Thus, the following canonical map of $\mathbb{F}_{p}[G]$-modules $(\mathbb{F}_{p}[G]\otimes_{\mathbb{F}_{p}[H]}\operatorname{\mathcal{D}}_{H})\otimes_{\mathbb{F}_{p}[H]}M\xhookrightarrow{\gamma}\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}\otimes_{\mathbb{F}_{p}[H]}M$ is injective. We put $\varphi=\gamma\circ\beta\circ\alpha$ and apply Lemma 4.1. Since $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}\otimes_{\mathbb{F}_{p}[H]}M$ is a $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-module and $\varphi$ is an injective $\mathbb{F}_{p}[G]$-homomorphism, $\mathbb{F}_{p}[G]\otimes_{\mathbb{F}_{p}[H]}M$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free. ∎ Now we can present our first example of a $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free $\mathbb{F}_{p}[G]$-module. ###### Proposition 4.8. Let $H$ be a non-trivial subgroup of $G$ and $A$ a maximal abelian subgroup of $H$. Then the $\mathbb{F}_{p}[G]$-module $I^{G}_{H}/I^{G}_{A}$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free. ###### Proof. From Lemma 2.1 we know that $I^{G}_{H}/I^{G}_{A}\cong\mathbb{F}_{p}[G]\otimes_{\mathbb{F}_{p}[H]}(I_{H}/I^{H}_{A}).$ Thus, in view of Lemma 4.7, it is enough to show that $I_{H}/I^{H}_{A}$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[H]}$-torsion-free. Let $\mathbf{Z}=C_{\mathbf{F}}(A)$. Since $\mathbf{F}$ is a free pro-$p$ group, a centralizator of a non-trivial element is maximal cyclic pro-$p$ subgroup. Therefore, since $A$ is abelian and non-trivial, $\mathbf{Z}$ is a maximal cyclic pro-$p$ subgroup of $\mathbf{F}$. ###### Claim 4.9. The canonical map $\mathbb{F}_{p}[H/A]\to\mathbb{F}_{p}[[\mathbf{F}/\mathbf{Z}]]$ is injective. ###### Proof. Since $A$ is maximal abelian in $H$, we have that $A=\mathbf{Z}\cap H$. Hence the obvious map $\mathbb{F}_{p}[H/A]\to\mathbb{F}_{p}[\mathbf{F}/\mathbf{Z}]$ in injective. The map $\mathbb{F}_{p}[\mathbf{F}/\mathbf{Z}]\to\mathbb{F}_{p}[[\mathbf{F}/\mathbf{Z}]]$ is also injective, because $\mathbf{Z}$ is closed in $\mathbf{F}$. This finishes the proof of the claim.∎ Observe that $\mathbb{F}_{p}[H/A]\cong\mathbb{F}_{p}[H]/I_{A}^{H}$ and $\mathbb{F}_{p}[[\mathbf{F}]]/I_{\mathbf{Z}}^{\mathbf{F}}\cong\mathbb{F}_{p}[[\mathbf{F}/\mathbf{Z}]]$. Therefore, by Claim 4.9, $I_{H}/I^{H}_{A}$ is a $\mathbb{F}_{p}[H]$-submodule of $I_{\mathbf{F}}/I_{\mathbf{Z}}^{\mathbf{F}}$. By Proposition 4.4, we can embed $I_{\mathbf{F}}/I_{\mathbf{Z}}^{\mathbf{F}}$ in a $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$-module. By Proposition 3.5, every $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[[\mathbf{F}]]}$-module is also a $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[H]}$-module. Therefore, by Lemma 4.1, $I_{H}/I^{H}_{A}$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[H]}$-torsion-free. ∎ The following proposition shows another example of a $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free $\mathbb{F}_{p}[G]$-module. This is the main result of this section. ###### Proposition 4.10. Let $\mathbf{F}$ be a finitely generated free pro-$p$ group and let $G$ be an (abstract) dense finitely generated subgroup of $\mathbf{F}$. Let $H$ be a non-trivial subgroup of $G$ and $A$ a maximal abelian subgroup of $H$. Let $B$ be an abelian subgroup of $G$ containing $A$. We put $J=\\{(x,-x)\in I^{G}_{H}\oplus I^{G}_{B}:\ x\in I^{G}_{A}\\}.$ Then $M=(I^{G}_{H}\oplus I^{G}_{B})/J$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free and $\dim_{\mathbb{F}_{p}[G]}M=\dim_{\mathbb{F}_{p}[G]}I^{G}_{H}$. ###### Proof. Let $L=(I^{G}_{A}\oplus I^{G}_{B})/J\leq M$. Then $L\cong I^{G}_{B}$ is a submodule of $\mathbb{F}_{p}[G]$, and so it is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free. The $\mathbb{F}_{p}[G]$-module $M/L$ is isomorphic to $I_{H}^{G}/I^{G}_{A}$, and so it is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free by Proposition 4.8. By (14), $\dim_{\mathbb{F}_{p}[A]}I_{A}=1$. Therefore, by Proposition 3.4(b), $\dim_{\mathbb{F}_{p}[G]}I^{G}_{A}=\dim_{\mathbb{F}_{p}[A]}I_{A}=1.$ In the same way we obtain that $\dim_{\mathbb{F}_{p}[G]}I^{G}_{B}=1$. Since $\dim_{\mathbb{F}_{p}[G]}(I^{G}_{H}\oplus I^{G}_{B})=\dim_{\mathbb{F}_{p}[G]}I^{G}_{H}+1,$ by Lemma 4.3, $\dim_{\mathbb{F}_{p}[G]}M=\dim_{\mathbb{F}_{p}[G]}I^{G}_{H}+1-1=\dim_{\mathbb{F}_{p}[G]}I^{G}_{H}\textrm{\ and}\\\ \dim_{\mathbb{F}_{p}[G]}(I_{H}^{G}/I^{G}_{A})=\dim_{\mathbb{F}_{p}[G]}I^{G}_{H}-1=\dim_{\mathbb{F}_{p}[G]}M-1.$ Therefore, $\dim_{\mathbb{F}_{p}[G]}(M/L)=\dim_{\mathbb{F}_{p}[G]}(I_{H}^{G}/I^{G}_{A})=\dim_{\mathbb{F}_{p}[G]}M-1.$ Thus, we have obtained that $M/L$ and $L$ are $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free and $\dim_{\mathbb{F}_{p}[G]}M=\dim_{\mathbb{F}_{p}[G]}(M/L)+\dim_{\mathbb{F}_{p}[G]}L.$ Applying Lemma 4.2, we conclude that $M$ is also $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free. ∎ ## 5\. Proof of main results ### 5.1. The inductive step in the proof of Theorem 1.2 The following theorem is the main result of the paper. Theorem 1.2 follows from it directly. ###### Theorem 5.1. Let $\mathbf{F}$ be a finitely generated free pro-$p$ group and let $H\hookrightarrow\mathbf{F}$ be a strong embedding of a finitely generated group $H$. Let $A$ be a maximal abelian subgroup of $H$ and let $B$ be an abelian finitely generated subgroup of $\mathbf{F}$ containing $A$. Put $G=\langle H,B\rangle$. Then the canonical homomorphism $H*_{A}B\to G$ is an isomorphism, and the embedding $G\hookrightarrow\mathbf{F}$ is strong. ###### Proof. In view of Proposition 2.13 we have to show that $I^{G}_{H}\cap I_{B}^{G}=I_{A}^{G}$ in $\mathbb{F}_{p}[G]$. Let $J=\\{(x,-x)\in I^{G}_{H}\oplus I^{G}_{B}:\ x\in I^{G}_{A}\\}\textrm{\ and\ }M=(I^{G}_{H}\oplus I^{G}_{B})/J.$ Then by Proposition 4.10, $\dim_{\mathbb{F}_{p}[G]}M=\dim_{\mathbb{F}_{p}[G]}I^{G}_{H}$. Therefore, $\dim_{\mathbb{F}_{p}[G]}M=\dim_{\mathbb{F}_{p}[G]}I^{G}_{H}=\dim_{\mathbb{F}_{p}[H]}I_{H}=\beta_{1}^{\operatorname{mod}p}(H)+1=d(\mathbf{F}).$ Since $I_{G}=I_{H}^{G}+I_{B}^{G}$, we have that the natural map $\alpha:M\to I_{G}$ is surjective. In particular $\beta^{\operatorname{mod}p}_{1}(G)=\dim_{\mathbb{F}_{p}[G]}I_{G}-1\leq\dim_{\mathbb{F}_{p}[G]}M-1=d(\mathbf{F})-1.$ Thus, using (16) we obtain that $\beta^{\operatorname{mod}p}_{1}(G)=d(\mathbf{F})-1$ and $\dim_{\mathbb{F}_{p}[G]}I_{G}=\dim_{\mathbb{F}_{p}[G]}M$. This shows that the embedding $G\hookrightarrow\mathbf{F}$ is strong. By Proposition 4.10, $M$ is $\operatorname{\mathcal{D}}_{\mathbb{F}_{p}[G]}$-torsion-free. Therefore, by Lemma 4.3, for any proper quotient $\overline{M}$ of $M$, $\dim_{\mathbb{F}_{p}[G]}\overline{M}<\dim_{\mathbb{F}_{p}[G]}M$. This implies that $\alpha$ is an isomorphism, and so $I^{G}_{H}\cap I_{B}^{G}=I_{A}^{G}$. Hence, Proposition 2.13 implies that $G\cong H*_{A}B$.∎ Another direct consequence of Theorem 5.1 is the following corollary. ###### Corollary 5.2. Let $\mathbf{F}$ be a finitely generated free pro-$p$ group and let $H\hookrightarrow\mathbf{F}$ be a strong embedding of finitely generated group $H$. Let $A$ be a maximal abelian subgroup of $H$. Assume that $A$ is finitely generated. Let $B$ be a finitely generated torsion-free abelian group containing $A$ and such that $B/A$ has no $p$-torsion. Then there exists an embedding of $H*_{A}B$ into $\mathbf{F}$ that extends $H\hookrightarrow\mathbf{F}$. In particular, $H*_{A}B$ is SE($p$) (see Definition 3.8). ###### Proof. Let $H\hookrightarrow\mathbf{F}$ be a strong embedding. Since $B/A$ has no $p$-torsion and $B$ is torsion-free abelian, the embedding of $A$ into $\mathbf{F}$ can be extended to an embedding of $B$ into $\mathbf{F}$. Now, we can apply Theorem 5.1.∎ ### 5.2. $A$-groups Let $A$ be a commutative ring. An $A$-group is a group $G$ along with a map $G\times A\to G$, called “action”, satisfying the following axioms: 1. (a) $g^{1_{A}}=g$, $g^{0_{A}}=1$, $1^{\alpha}=1$; 2. (b) $g^{\alpha}g^{\beta}=g^{\alpha+\beta}$, $(g^{\alpha})^{\beta}=g^{\alpha\beta}$; 3. (c) $(h^{-1}gh)^{\alpha}=h^{-1}g^{\alpha}h$; 4. (d) If $gh=hg$, $(gh)^{\alpha}=g^{\alpha}h^{\alpha}$; for all $g,h\in G$ and $\alpha,\beta\in A$. This definition generalizes the definition of $\mathbb{Q}$-group which appears in the introduction. Given a group $G$ and a commutative ring $A$, an $A$-completion of $G$ is an $A$-group $G^{A}$ with a group homomorphism $\lambda:G\to G^{A}$ such that $G^{A}$ is generated by $\lambda(G)$ as an $A$-group and for any $A$-group $H$ and any homomorphism $\phi:G\to H$ there exists a unique $A$-homomorphism $\psi:G^{A}\to H$ such that $\phi=\psi\circ\lambda$. It is shown in [52, Theorems 1 and 2] that an $A$-completion of $G$ exists and it is unique up to an $A$-isomorphism. An $A$-group $F^{A}(X)$ with the set of $A$-generators $X$ is said to be a free $A$-group with base $X$, if for every $A$-group $G$ an arbitrary mapping $\phi_{0}:X\to G$ can be extended to an $A$-homomorphism $\phi:F^{A}(X)\to G$. Thus, $F^{A}(X)$ is the $A$-completion of the ordinary free group $F(X)$ with free generating set $X$. A CSA-group is a group in which the centralizer of every nontrivial element is abelian and malnormal. In [53, Theorem 5] it is shown that every torsion-free extension of centralizer of a torsion-free CSA group is again CSA. This is used in [53, Theorem 8] to describe the group $F^{A}(X)$. In the following proposition we extract the information that we will need later. ###### Proposition 5.3. Let $X$ be a finite set and $A$ a commutative ring with a torsion-free additive group. Then there are subgroups $\\{W_{n}\\}_{n\geq 0}$ of $F^{A}(X)$ such that 1. (a) $W_{0}=F(X)$; 2. (b) $W_{n+1}=\langle W_{n},z_{n}^{A}\rangle$, where $z_{n}\in W_{n}$ generates a maximal abelian subgroup in $W_{n}$ and the canonical map $\displaystyle W_{n}*_{z_{n}=1_{A}}A\to W_{n+1}$ is an isomorphism; 3. (c) $F(X)^{A}=\bigcup_{i=0}^{\infty}W_{n}$. Now we are ready to prove Corollary 1.3. ###### Proof of Corollary 1.3. If $\phi$ is not injective, then there exists a finitely generated subgroup $G$ of $F^{\mathbb{Z}_{p}}(X)$ such that $G\cap\ker\phi$ is not trivial. By Proposition 5.3, there exists a non-negative integer $k$ and a sequence $G_{0}\leq G_{1}\leq\ldots\leq G_{k}$ of subgroups of $F^{A}(X)$, where 1. (a) $G_{0}=F(X_{0})$ is the group generated by a finite subset $X_{0}$ of $X$; 2. (b) for $0\leq i\leq k-1$ there exists an element $z_{i}\in G_{i}$, generating a maximal abelian subgroup in $G_{i}$, and a finitely generated subgroup $T_{i}$ of $(A,+)$ containing $1_{A}$, such that $z_{i}^{T_{i}}\leq G_{i+1}$ and the canonical map $G_{i}*_{z_{i}=1_{A}}T_{i}\to G_{i+1}$ is an isomorphism; 3. (c) $G$ is a subgroup of $G_{k}$. Put $H_{i}=\phi(G_{i})$ $A_{i}=\langle\phi(z_{i})\rangle$ and $B_{i}=\phi(z_{i}^{T_{i}})$. Let us show by induction on $i$ that $\phi_{|G_{i}}:G_{i}\to H_{i}$ is an isomorphism. It is clear for $i=0$. Assume we have proved it for $i<k$. Observe that, since $\phi(z_{i})\neq 1$, $B_{i}=\phi(z_{i}^{T_{i}})=\phi(z_{i})^{T_{i}}\cong T_{i}\textrm{\ and \ }A_{i}=\phi(\langle z_{i}\rangle)=\phi(G_{i}\cap z_{i}^{T_{i}})=B_{i}\cap H_{i}.$ Then by Theorem 1.2, $\phi_{|G_{i+1}}:G_{i+1}\to H_{i+1}$ is also an isomorphism. Thus, $\ker\phi\cap G_{k}=\\{1\\}$. This is a contradiction. ∎ We will also need the following consequence of Proposition 5.3. ###### Corollary 5.4. [53, Corollary 5] Let $B$ be a ring with a torsion-free additive group and $A$ a subring of $B$. Then the canonical map $F^{A}(X)\to F^{B}(X)$ is injective. ### 5.3. Proof of Theorem 1.1 In this subsection we will prove that $F^{A}(X)$ is residually nilpotent for every commutative ring $A$ with a torsion-free additive group. The proof is based on the following general result communicated to us by a referee. ###### Proposition 5.5. Let $G$ be a group. Assume that for each $n\geq 0$ there are subgroups $G_{n}$ and $B_{n}$ of $G$ such that 1. (a) $G_{n+1}=\langle G_{n},B_{n}\rangle$; 2. (b) $B_{n}$ is abelian and $G_{n}$ is residually torsion-free nilpotent; 3. (c) $A_{n}=G_{n}\cap B_{n}$ and the canonical map $G_{n}*_{A_{n}}B_{n}\to G_{n+1}$ is an isomorphism; 4. (d) $G=\bigcup_{i=0}^{\infty}G_{n}$. Then $G$ is residually torsion-free nilpotent. ###### Proof. Let $x\in G$. We want to show that $G$ has a torsion-free nilpotent quotient such that the image of $x$ in this quotient is not trivial. Let $x\in G_{m}$ for some $m\geq 0$. Since $G_{m}$ is residually torsion-free nilpotent, there exists a torsion-free nilpotent group $N$ and the map $\phi_{m}:G_{m}\to N$ such that $\phi_{m}(x)\neq 1$. We can embed $N$ in its Malcev $\mathbb{Q}$-completion. Thus, without loss of generality, we may assume that $N$ is also a $\mathbb{Q}$-group. We will show that for each $n\geq m$ there exist $\phi_{n}:G_{n}\to N$ such that the restriction of $\phi_{n}$ on $G_{n-1}$ is $\phi_{n-1}$. We construct $\phi_{n}$ by induction. Assume that we constructed $\phi_{n}$ for $m\leq n\leq k$. Since $N$ is a $\mathbb{Q}$-group, we can extend $\phi_{k}$ from $A_{k}$ to $A_{k}^{\mathbb{Q}}$ and so to $B_{k}^{\mathbb{Q}}$ and $B_{k}$. Thus we have a homomorphism $\tilde{\phi}_{k}:B_{k}\to N$ that extends $(\phi_{k})_{|A_{k}}:A_{k}\to N$. Now, by the universal property of the amalgamated product, there exists a homomorphism $\phi_{k+1}:G_{k+1}\to N$ whose restriction on $G_{k}$ is $\phi_{k}$ and on $B_{k}$ is $\tilde{\phi}_{k}$. Let $\phi:G\to N$ be the homomorphism satisfying $\phi(g)=\phi_{n}(g)$ if $g\in G_{n}$. Then $\phi$ is well-defined and $\phi(x)\neq\\{1\\}$. ∎ ###### Corollary 5.6. Let $X$ be a set and $A$ a countable commutative ring with a torsion-free additive group. Then $F^{A}(X)$ is residually torsion-free nilpotent. ###### Proof. First assume that $X$ is finite. Since $A$ is countable, by Proposition 5.3 there are subgroups $\\{G_{n}\\}_{n\geq 0}$ and $\\{B_{n}\\}_{n\geq 0}$ of $F^{A}(X)$ such that 1. (a) $G_{n+1}=\langle G_{n},B_{n}\rangle$; 2. (b) $B_{n}$ is abelian and finitely generated; 3. (c) $A_{n}=G_{n}\cap B_{n}$ is a maximal abelian subgroup of $G_{n}$ and the canonical map $G_{n}*_{A_{n}}B_{n}\to G_{n+1}$ is an isomorphism; 4. (d) $F^{A}(X)=\bigcup_{i=0}^{\infty}G_{n}$. By Theorem 1.2, $G_{n}$ are residually torsion-free nilpotent. Thus, Proposition 5.5 implies that $F^{A}(X)$ is residually torsion-free nilpotent. Now assume that $X$ is an arbitrary set. Given an element $1\neq g$ of $F^{A}(X)$, it belongs to $F^{A}(X_{0})$ for some finite subset $X_{0}$ of $X$, and $F^{A}(X_{0})$ is a retract of $F^{A}(X)$. As we have already proved, there exists a homomorphism $\phi:F^{A}(X_{0})\to N$, where $N$ is torsion- free nilpotent such that $\phi(g)\neq 1$. This finishes the proof. ∎ ###### Remark 5.7. The proof of the previous corollary can also be adapted to the case where $A$ is not countable. We prove it, using a different method, in Corollary 5.9. Let $X=\\{x_{i}\colon i\in I\\}$ and $Y=\\{y_{i}\colon i\in I\\}$ be two sets indexed by the elements of a set $I$. Let $A$ be a commutative ring. We say that $A$ is a binomial domain if $A$ is a domain and $a\choose n$ belongs to $A$ for every $a\in A$. For example $\mathbb{Z}_{p}$ and $\mathbb{Q}$-algebras are binomial domains. Assume that $A$ is a binomial domain. If $\Delta_{A}$ denotes the ideal of $A\langle\\!\langle Y\rangle\\!\rangle$ generated by $Y$ then $1+\Delta_{A}$ is a subgroup of the group of units of $A\langle\\!\langle Y\rangle\\!\rangle$. We can define an action of $A$ on $1+\Delta_{A}$ in the following way: $(1+f)^{a}=1+\sum_{n=1}^{\infty}{a\choose n}f^{n}\ (a\in A,f\in\Delta_{A}).$ Then, by [60], $1+\Delta_{A}$ provided with this action is an $A$-group. Therefore, the map $x_{i}\mapsto 1+y_{i}$ can be uniquely extended to an $A$-homomorphism $\phi_{A}:F^{A}(X)\to 1+\Delta_{A}$ called the Magnus representation of $F^{A}(X)$. Magnus (see, for example, [47]) proved that $\phi_{\mathbb{Z}}$ is injective. Now we prove the main result of this subsection. ###### Theorem 5.8. Let $X$ be an arbitrary set and $A$ be a binomial domain.Then the map $\phi_{A}:F^{A}(X)\to 1+\Delta_{A}$ is injective. ###### Proof. By Corollary 5.4, we can assume that $A$ is a $\mathbb{Q}$-algebra. If $\ker\phi_{A}$ is not trivial, then there exists a finite subset $X_{0}$ of $X$ and countable $\mathbb{Q}$-subalgebra of $A$ such that the kernel of the map $F^{A_{0}}(X_{0})\to 1+\Delta_{A_{0}}$ is not trivial. Hence, we can also assume that $X$ is finite and $A$ is countable. Let $F_{n}=F(X)/\gamma_{n}(F(X))$. Denote by $K_{n}$ the kernel of the canonical map $F^{A}(X)\to F_{n}^{A}$. Using [52, Property 2], we obtain that $F_{n}^{A}$ is nilpotent of nilpotency class $n-1$. By Proposition 5.6, $F^{A}(X)$ is residually torsion-free nilpotent. Hence $\cap_{n}K_{n}=\\{1\\}$. The map $F_{n}\to 1+\Delta_{A}/1+(\Delta_{A})^{n}$ induces the map $\phi_{n}:F_{n}^{A}\to 1+\Delta_{A}/1+(\Delta_{A})^{n}$. It is known that $\phi_{n}$ is injective (see, for example, [22, Theorem 23]). Thus $K_{n}$ is the kernel of the canonical map $F^{A}(X)\to 1+\Delta_{A}/1+(\Delta_{A})^{n}$. This implies that $\phi_{A}$ is injective. ∎ The following definition has been suggested to us by a referee. We say that an $A$-group $G$ is $A$-torsion-free if for every $1\neq g\in G$, the map $A\to G$ that sends $a$ to $g^{a}$ is injective. ###### Corollary 5.9. Let $A$ be a commutative ring with a torsion-free additive group. Then the group $F^{A}(X)$ is residually-($A$-torsion-free nilpotent). ###### Proof. By Corollary 5.4, we can assume that $A$ is a $\mathbb{Q}$-algebra. Hence the corollary follows from Theorem 5.8 because the $A$-groups $1+\Delta_{A}/1+(\Delta_{A})^{n}$ are $A$-torsion-free and nilpotent. ∎ ### 5.4. Proof of Theorem 1.4 The proof of Theorem 1.4 uses a particular case of [53, Theorem 8] that we describe now. ###### Proposition 5.10. Let $G$ be a countable torsion-free CSA-group. Then the $\mathbb{Q}$-completion $G^{\mathbb{Q}}$ of $G$ is a direct union of subgroups $W_{0}\leq W_{1}\leq\ldots$ such that 1. (a) $W_{0}=G$; 2. (b) for $i\geq 0$ there exists a maximal abelian subgroup $A_{i}$ of $W_{i}$ such that $W_{i+1}$ is the image of the canonical map $W_{i}*_{A_{i}}A_{i}^{\mathbb{Q}}\to G^{\mathbb{Q}}$. Moreover, if $H$ is a subgroup $G^{\mathbb{Q}}$ and $A$ is a maximal abelian subgroup of $H$, then the canonical map $H*_{A}A^{\mathbb{Q}}\to G{{}^{\mathbb{Q}}}$ is injective. ###### Proof of Theorem 1.4. Our definition of a limit group from the introduction and Proposition 5.3 show that the limit groups are exactly finitely generated subgroups of $F^{\mathbb{Z}[t]}(X)$. Since $F^{\mathbb{Z}[t]}(X)$ is a CSA-group, its $\mathbb{Q}$-completion can be calculated using Proposition 5.10. In particular, $G^{\mathbb{Q}}$ is a subgroup of the $\mathbb{Q}$-completion of $F^{\mathbb{Z}[t]}(X)$ and so it is a subgroup of $F^{\mathbb{Q}[t]}(X)$. Hence by Corollary 5.9, it is residually torsion-free nilpotent. ∎ ## 6\. Linearity of free $\mathbb{Q}$-groups and free pro-$p$ groups We finish this paper with a discussion on another two well-known problems concerning linearity of free $\mathbb{Q}$-groups and free pro-$p$ groups. The problem of whether a free $\mathbb{Q}$-group $F^{\mathbb{Q}}(X)$ is linear appears in [10, Problem F13] and it is attributed to I. Kapovich (see also [38, Problem 13.39(b)]). The problem of whether a free pro-$p$ group $\mathbf{F}$ is linear is usually attributed to A. Lubotzky (for example, we discussed this question in Jerusalem in November, 2001). In the context of profinite groups, one can consider two kinds of linearity (see, for example, [28]). On one hand, we say that a profinite group $G$ is linear if it is linear as an abstract group, that is it has a faithful representation by matrices of fixed degree over a field. On the other hand, the concept of $t$-linear profinite group takes into account the topology of $G$ and means that $G$ can be faithfully represented as a closed subgroup of the group of invertible matrices of fixed degree over a profinite commutative ring. It is commonly believed that a non-abelian free pro-$p$ group is not $t$-linear (see, [44, Conjecture 3.8], the discussion after [3, Theorem 1.1] and [59, Section 5.3]). An equivalent reformulation of this statement is that a $p$-adic analytic pro-$p$ group satisfies a non-trvial pro-$p$ identity. A. Zubkov [64] proved that if $p>2$, then a non-abelian free pro-$p$ group cannot be represented by 2-by-2 matrices over a profinite commutative ring. E. Zelmanov announced that given a fixed $n$, a non-abelian free pro-$p$ group cannot be represented by $n$-by-$n$ matrices over a profinite commutative ring for every large enough prime $p>>n$ (see [62, 63]). Recently, D. El-Chai Ben- Ezra, E. Zelmanov showed that a free pro-2 group cannot be represented by 2-by-2 matrices over a profinite commutative ring of characteristic 2 [20]. Recall that by a result of A. Malcev [48, Theorem IV], a group can be represented by matrices of degree $n$ over a field if and only if every one of its finitely generated subgroup has this property. Thus, in order to decide whether $F^{\mathbb{Q}}(X)$ or $\mathbf{F}$ are linear, we have to analyze the structure of their finitely generated (abstract) subgroups. In order to apply the Malcev criterion one should find a uniform $n$ which does not depend on a finitely generated subgroup. We may ask a weaker question of whether $F^{\mathbb{Q}}(X)$ and $\mathbf{F}$ are locally linear. Using recent advances in geometric group theory one can answer this positively in the case of $F^{\mathbb{Q}}(X)$. ###### Theorem 6.1. The groups $F^{\mathbb{Q}}(X)$ are locally linear over $\mathbb{Z}$. ###### Proof. Let $G$ be a finitely generated subgroup of $F^{\mathbb{Q}}(X)$. Then by Proposition 5.3, there exists $k\geq 0$ and a sequence $G_{0}\leq G_{1}\leq\ldots\leq G_{k}$ of subgroups of $F^{\mathbb{Q}}(X)$, such that $G_{0}$ is free, $G_{i+1}$ is obtained from $G_{i}$ by adjoining a root and $G\leq G_{k}$. By [53, Theorem 5], the groups $G_{i}$ are CSA. Thus, from a corollary on the page 100 of [11] we obtain that $G$ is hyperbolic. Therefore, [26, Corollary C] ensures, by induction on $i$, that each $G_{i}$ acts properly and cocompactly on a CAT(0) cube complex. Hence, by [1, Theorem 1.1], $G_{k}$ has a finite index subgroup acting faithfully and specially on a CAT(0) cube complex. Finally from [25, Theorem 1.1] it follows that $G_{k}$, and so $G$, are linear over $\mathbb{Z}$.∎ I wouldn’t be surprised if the groups $F^{\mathbb{Q}[t]}(X)$ are also locally linear. However, to decide whether $F^{\mathbb{Q}}(X)$ and $F^{\mathbb{Q}[t]}(X)$ are linear will require a completely new approach. ## References * [1] I. Agol, The virtual Haken conjecture. With an appendix by Agol, Groves, and Manning. Doc. Math. 18 (2013), 1045–1087. * [2] J. Barlev, T. Gelander, Compactifications and algebraic completions of limit groups. J. Anal. Math. 112 (2010), 261–287. * [3] Y. Barnea, M. Larsen, A non-abelian free pro-$p$ group is not linear over a local field. J. Algebra 214 (1999), 338–341. * [4] G. Baumslag, Some aspects of groups with unique roots. Acta Math. 104 (1960), 217–303. * [5] G. Baumslag, On free $\mathcal{D}$-groups. Comm. Pure Appl. Math. 18 (1965), 25-30. * [6] G. Baumslag, Groups with the same lower central sequence as a relatively free group. I. The groups. Trans. Amer. Math. Soc. 129 (1967), 308–321. * [7] G. Baumslag, On the residual nilpotence of certain one-relator groups. Comm. Pure Appl. Math. 21 (1968), 491–506. * [8] G. Baumslag, Groups with the same lower central sequence as a relatively free group. II. Properties. Trans. Amer. Math. Soc. 142 (1969), 507–538. * [9] G. Baumslag, Some reflections on proving groups residually torsion-free nilpotent. I. Illinois J. Math. 54 (2010), 315–325. * [10] G. Baumslag, A. Myasnikov, V. Shpilrain, Open problems in combinatorial group theory. Second edition. Combinatorial and geometric group theory (New York, 2000/Hoboken, NJ, 2001), 1–38, Contemp. Math., 296, Amer. Math. Soc., Providence, RI, 2002. * [11] M. Bestvina, M. Feighn, A combination theorem for negatively curved groups. J. Differential Geom. 35 (1992), 85–101. * [12] E. Breuillard, T. Gelander, J. Souto, P. Storm, Dense embeddings of surface groups. Geom. Topol. 10 (2006), 1373–1389. * [13] M. Bridson, A. Reid, Nilpotent completions of groups, Grothendieck pairs, and four problems of Baumslag. Int. Math. Res. Not. IMRN 2015, 2111–2140. * [14] C. Champetier, V. Guirardel, Limit groups as limits of free groups. Israel J. Math., 146 (2005), 1–75. * [15] B. Chandler, The representation of a generalized free product in an associative ring. Comm. Pure Appl. Math. 21 (1968), 271–288. * [16] P. M. Cohn, Localization in semifirs. Bull. London Math. Soc. 6 (1974), 13–20. * [17] P. M. Cohn, Free rings and their relations. Second edition. London Mathematical Society Monographs, 19. Academic Press, Inc., London, 1985. * [18] P. M. Cohn, Skew fields. Theory of general division rings. Encyclopedia of Mathematics and its Applications, 57. Cambridge University Press, Cambridge, 1995. * [19] W. Dicks, D. Herbera, J. Sánchez, On a theorem of Ian Hughes about division rings of fractions. Comm. Algebra 32 (2004), 1127–1149. * [20] D. El-Chai Ben-Ezra, E. Zelmanov, On pro-2 identities of $2\times 2$ linear groups, Trans. Amer. Math. Soc. 374 (2021), 4093–4128 * [21] M. Ershov, W. Lück, The first $L^{2}$-Betti number and approximation in arbitrary characteristic, Doc. Math. 19 (2014), 313–332. * [22] A. Gaglione, A. Myasnikov, V. Remeslennikov, D. Spellman, Formal power series representations of free exponential groups. Comm. Algebra 25 (1997), 631–648. * [23] J. Gräter, Free Division Rings of Fractions of Crossed Products of Groups With Conradian Left-Orders, Forum Math. 32 (2020), 739–772. . * [24] H. Hahn, Über die nichtarchimedischen Grössensysteme, Sitzungsberichte der K. Akademie der Wissenschaften, Vienna 116 (1907), 601–655. * [25] F. Haglund, D. Wise, Special cube complexes. Geom. Funct. Anal. 17 (2008), 1551–1620. * [26] T. Hsu, D. Wise, Cubulating malnormal amalgams. Invent. Math. 199 (2015), 293–331. * [27] I. Hughes, Division rings of fractions for group rings. Comm. Pure Appl. Math., 23 (1970), 181–188. * [28] A. Jaikin-Zapirain, On linear just infinite pro-$p$ groups. J. Algebra 255 (2002), 392–404. * [29] A. Jaikin-Zapirain, Approximation by subgroups of finite index and the Hanna Neumann conjecture, Duke Math. J. 166 (2017), 1955–1987. * [30] A. Jaikin-Zapirain, The base change in the Atiyah and the Lück approximation conjectures. Geom. Funct. Anal. 29 (2019), 464–538. * [31] A. Jaikin-Zapirain, $L^{2}$-Betti numbers and their analogues in positive characteristic, Groups St Andrews 2017 in Birmingham, 346–406, London Math. Soc. Lecture Note Ser., 455, Cambridge Univ. Press, Cambridge, 2019. * [32] A. Jaikin-Zapirain, An explicit construction of the universal division ring of fractions of $E\langle\\!\langle x_{1},\ldots,x_{d}\rangle\\!\rangle$, J. Comb. Algebra 4 (2020), 369–395. * [33] A. Jaikin-Zapirain, The universality of Hughes-free division rings, Sel. Math. New Ser. 27 (2021), 74. * [34] A. Jaikin-Zapirain, D. López-Álvarez, The strong Atiyah and Lück approximation conjectures for one-relator groups, Math. Ann. 376 (2020), 1741–1793. * [35] A. Jaikin-Zapirain, I. Morales, Parafree fundamental groups of graph of groups with cyclic edge subgroups, preprint (2021), arXiv:2110.11655. * [36] B. Jiang, H. Li, Sylvester rank functions for amenable normal extensions, Journal of Functional Analysis 280 (2021), 108913, 47 pp. * [37] O. Kharlampovich, A. Miasnikov. Elementary theory of free nonabelian groups, J. Algebra, 302 (2006), 451–552. * [38] The Kourovka notebook. Unsolved problems in group theory. Nineteenth edition. March 2019 update. Edited by E. I. Khukhro and V. D. Mazurov. Sobolev Institute of Mathematics. Russian Academy of Sciences. Siberian Branch, Novosibirsk, 2018. * [39] M. Lazard, Groupes analytiques p-adiques, Inst. Hautes Études Sci. Publ. Math. No. 26 (1965), 389–603. * [40] J. Lewin, On the intersection of augmentation ideals. J. Algebra 16 (1970), 519–522. * [41] H. Li, Bivariant and extended Sylvester rank functions, J. London Math. Soc. (2021), 222–249. * [42] P. Linnell, Division rings and group von Neumann algebras. Forum Math. 5 (1993), 561–576. * [43] D. López-Álvarez, Sylvester rank functions, epic division rings, and the strong Atiyah conjecture for locally indicable groups, Ph.D. Thesis 2021, Universidad Autónoma de Madrid, https://www.icmat.es/Thesis/2021/Tesis_Diego_Lopez_Alvarez.pdf. * [44] A. Lubotzky, A. Shalev, On some $\Lambda$-analytic pro-$p$ groups. Israel J. Math. 85 (1994), 307–337. * [45] W. Lück, Approximating $L^{2}$-invariants by their finite-dimensional analogues. Geom. Funct. Anal. 4 (1994), 455–481. * [46] W. Lück, D. Osin, Approximating the first $L^{2}$-Betti number of residually finite groups. J. Topol. Anal. 3 (2011), 153–160. * [47] W. Magnus, A. Karrass, D. Solitar, Combinatorial group theory. Presentations of groups in terms of generators and relations. Dover Publications, Inc., Mineola, NY, 2004. * [48] A. Malcev, On isomorphic matrix representations of infinite groups. (Russian) Rec. Math. [Mat. Sbornik] N.S. 8 (50), (1940), 405–422. * [49] A. I. Malcev, On the embedding of group algebras in division algebras. (Russian) Doklady Akad. Nauk SSSR (N.S.) 60 (1948). 1499–1501. * [50] P. Malcolmson, Determining homomorphisms to skew fields. J. Algebra 64 (1980), 399–413. * [51] I. Morales, Embeddings into pro-$p$ groups and the construction of parafree groups, Master Thesis, Universidad Autónoma de Madrid (2021), arXiv:2109.12341. * [52] A. Myasnikov, V. Remeslennikov, Exponential groups I: foundations of the theory and tensor completion. Siberian Math. J., 5 (1994), 986–996. * [53] A. Myasnikov, V. Remeslennikov, Exponential groups. II. Extensions of centralizers and tensor completion of CSA-groups. Internat. J. Algebra Comput. 6 (1996), 687–711. * [54] B. H. Neumann, On ordered division rings. Trans. Amer. Math. Soc. 66, (1949). * [55] N. Nikolov, D. Segal, On finitely generated profinite groups. I. Strong completeness and uniform bounds. Ann. of Math. (2) 165 (2007), no. 1, 171–238. * [56] L Ribes, P. Zalesskii, Profinite groups. Second edition. A Series of Modern Surveys in Mathematics, 40. Springer-Verlag, Berlin, 2010. * [57] J. J. Rotman, An introduction to homological algebra. Second edition. Universitext. Springer, New York, 2009. * [58] J. Sánchez, On division rings and tilting modules, Ph. D. Thesis, Universitat Autònoma de Barcelona (2008), www.tdx.cat/bitstream/handle/10803/3107/jss1de1.pdf. * [59] A. Shalev, Lie methods in the theory of pro-$p$ groups. New horizons in pro-$p$ groups, 1–54, Progr. Math., 184, Birkhäuser Boston, Boston, MA, 2000. * [60] R. B. Warfield, Nilpotent groups. Lecture Notes in Mathematics, Vol. 513. Springer-Verlag, Berlin-New York, 1976. * [61] D. Wise, Research announcement: the structure of groups with a quasiconvex hierarchy. Electron. Res. Announc. Math. Sci. 16 (2009), 44–55. * [62] E. Zelmanov, Infinite algebras and pro-$p$ groups, Infinite groups: geometric, combinatorial and dynamical aspects, Progr. Math. 248, Birkhäuser, Basel, 2005, 403–413. * [63] E. Zelmanov, Groups with identities, Note. Mat. 36, 2016, 101–113. * [64] A. N. Zubkov, Nonrepresentability of a free nonabelian pro-$p$-group by second-order matrices. (Russian) Sibirsk. Mat. Zh. 28 (1987), 64–69, English translation: Siberian Math. J. 28 (1987), 742–746.
# Fast convolution kernels on pascal GPU with high memory efficiency Qiong Chang Grad. of Systems and Information Engineering University of Tsukuba Tsukuba, Japan, 305-8577 <EMAIL_ADDRESS> &Masaki Onishi National Institute of Advanced Industrial Science and Technology (AIST) Tsukuba, Japan, 305-8560 <EMAIL_ADDRESS> &Tsutomu Maruyama Grad. of Systems and Information Engineering University of Tsukuba Tsukuba, Japan, 305-8577 <EMAIL_ADDRESS> ###### Abstract The convolution computation is widely used in many fields, especially in CNNs. Because of the rapid growth of the training data in CNNs, GPUs have been used for the acceleration, and memory-efficient algorithms are focused because of thier high performance. In this paper, we propose two convolution kernels for single-channel convolution and multi-channel convolution respectively. Our two methods achieve high performance by hiding the access delay of the global memory efficiently, and achieving high ratio of floating point Fused Multiply- Add operations per fetched data from the global memory. In comparison to the latest Cudnn library developed by Nvidia aimed to accelerate the deep-learning computation, the average performance improvement by our research is 2.6X for the single-channel, and 1.4X for the multi-channel. ## 1 Introduction Convolution is widely used as a fundamental operation in many applications such as computer vision, natural language processing, signal processing. Especially, the Convolution Neural Network (CNN), a popular model used for deep-learning, is widely used in many applications such as image recognition [2][6], video analysis[3], natural language processing[7], and has yielded remarkable results. Recently, many CNN models have been proposed, such as AlexNet [15], GoogleNet [11], VGGNet [6], ResNet [9], etc. They are used in many areas, and are improved steadily. The sizes of their networks have grown larger, and this leads to the increase of their processing time. The CNNs have several layers, but a large portion of the total processing time is used for the convolution layers. Because of the high inherent parallelism of the convolution algorithms, many researchers are exploring to use GPUs to accelerate them. They can be divided into four categories: 1) Direct-based method, 2) FFT-based method, 3) Winograd-based method and 4) General matrix multiplication (GEMM) based method. Recently, many of the algorithms and their modified versions have been aggregated into a public library called Cudnn by Nvidia, which aims to accelerate the deep-learning platforms like Chainer[4], Caffe[14], etc. As the volume of the processing data used in deep-learning increases, the memory- efficient algorithms play an increasingly more important role, resulting in the constant proposal of many improved versions. Among them, Implicit-GEMM[12] has been included in the Cudnn because of its high efficiency. It divides the feature map and the filter data into many sub-blocks and converts them to many sub-matrices by using only on-chip memory of GPU, not using global memory. This method is very memory-efficient, and achieved a high ratio of floating point Fused Multiply-Add (FMA) operations per data transferred from the global memory. In [1], two memory-efficient methods were proposed. Both of them are faster than the Implicit-GEMM, however their performances are negatively affected when the feature map size is smaller than 32, because it fixes the amount of the data assigned to each $SM$, which sometimes is not suitable to the small feature map. In CNN models such as [15][11][6][9], more than half of the convolution layers are used for the calculation of the images smaller than 32 (such as 28, 14, 7). This means that [1] cannot handle the modern CNN models efficiently. In this paper, we propose two methods to accelerate the convolution kernels for single-channel and multi-channel respectively. In these methods, the memory access is optimized to achieve higher performance. For the single- channel kernel, our approach follows the computation method proposed in [1], however, the data are divided and assigned to each $SM$ carefully to hide the access delay of the global memory considering the input data size and the hardware features of our target GPUs (Pascal GPUs). For the multi-channel kernel, we propose a stride-fixed block method. This method aims to maximize the number of FMA operations per loaded data because the total amount of data that have to be loaded to each $SM$ is much larger than the single-channel convolution, and the access delay can be hidden by data prefetching. ## 2 Optimization Methods In this section, we first introduce the CNN models, and then discuss what kind of optimization method is applicable on Pascal GPUs. ### 2.1 The Convolution Models The multi-channel convolution can be defined as follows: $\begin{split}O^{m}(x,y)=\sum_{ch=1}^{C}\sum_{i=0}^{K-1}\sum_{j=0}^{K-1}I^{ch}(x+i,y+j)\cdot F^{ch,m}(i,j),\\\ where\hskip 14.22636ptx\in[0,W_{x}-K+1),y\in[0,W_{y}-K+1),m\in[1,M].\end{split}$ (1) Here, $I$ is the input feature map set and $F$ is the input filter set. $O$ is the output feature map set which is generated from $I$ and $F$. $x$ and $y$ are the coordinates of the pixels of the feature maps. $W_{x}$ is the width and $W_{y}$ is the height of the input feature map. $i$ and $j$ are the offsets, and are added to the coordinates. Their upper bound $K$ decides the size of filter. $ch$ represents the channel of the input in the range of $[1,C]$ ($C$ is the number of channels), and all of the convolution results are added along the dimension $ch$. $m$ represents the filter number, and each filter has $C$ channels. $M$ is the number of filters, and it is defined in each convolution layer. When $C=1$, it is called single-channel convolution, and its definition is given the following equation. $\begin{split}O^{m}(x,y)=\sum_{i=0}^{K-1}\sum_{j=0}^{K-1}I(x+i,y+j)\cdot F^{m}(i,j)\\\ \end{split}$ (2) ### 2.2 Acceleration models on GPU Here, we discuss the acceleration methods of the convolution calculation. However, this discussion is not restricted to the convolution, and can be applied to other applications. In GPUs, the on-chip memory including registers and shared memory, and the off-chip memory, mainly the global memory are supported. The register is the fastest, and the global memory is the slowest and largest. To fully utilize this hierarchy, many studies such as [1] have been proposed. However, throughout all computation, data loading time from the global memory to the on-chip memory is most critical, and hiding the latency of the global memory is the most important point for the acceleration. To hide the latency of the global memory, two methods can be considered: 1. 1. keep the operation units busy (mainly Fused Multiply-Add (FMA) operation units in convolution) by executing more than $N_{FMA}$ operations (the lowest value to make the units busy) in each $SM$ for the current data set until the next data set arrive from the global memory by data prefetching, and 2. 2. transfer a large volume of data ($V_{s}$) from the global memory continuously. In most cases, the first approach is preferable, because the data loading overhead from the global memory can be relatively reduced more by executing more number of operations per loaded data. In the multi-channel convolution, the data size is large enough, and it is possible to find the division of the feature maps and filters that makes it possible to execute more than $N_{FMA}$ operations in each $SM$. However, in the single-channel convolution, when the size of feature maps is small, the number of executable operations becomes less than $N_{FMA}$ even with the data prefetching, and the second approach is required. Thus, it is necessary to make it clear under what conditions which method shows better performance. Table 1 shows several parameters of GTX 1080Ti and its performance for accessing single precision data. As shown in Table 1, in GTX 1080Ti, 2 FMA operations can be executed in one clock cycle in each core, namely 256 FMA operations in each $SM$ (each $SM$ has $N_{cores}$ = 128 cores). According to the method proposed in [5], the global memory latency of the GTX 1080Ti is 258 clock cycles. In order to hide this 258 clock cycles, $N_{FMA}=66,048$ FMA operations ($66,048=258\times N_{cores}\times 2$) are required in each $SM$ for the current data set (the set of divided feature maps and filters). The volume size $V_{s}$ can be calculated as follows. The Geforce GTX 1080Ti has a base clock of 1480 MHz and the bandwidth of 484 GB/s, which means the transfer rate is roughly 327 bytes per clock cycle. Therefore, the volume size which is needed to hide the latency (258 clock cycles) becomes 84,366 = 327 $\times$ 258 bytes. To realize the data transfer of this size, 21,120 ($=$ 84,366 / 4) threads are required because each thread fetches a 4 byte data in single precision. Thus, in each of 28 $SMs$ ($N_{sm}=28$ is the total number of $SMs$ in the GTX 1080Ti), 768 threads are required to fetch one 4-byte word respectively (in total, it becomes 768 $\times 4\times 28=$ 86,016 > 84,366). This means that the minimum volume size to make the global memory busy is $V_{s}=$ 86,016 bytes. For dividing the feature maps and filters, and assigning them to each $SM$, the following procedure should be taken: 1. 1. Divide the feature maps and filters so that the total size of data that are assigned to each $SM$ is smaller than the size of the shared memory $S_{shared}$ (96KB in GTX 1080Ti). 2. 2. Evaluate the number of FMA operations that can be executed for the data in each $SM$. 3. 3. If it is larger than $N_{FMA}$, use the the first method which is based on the data prefetching. 4. 4. If not, redivide the feature maps and filters so that the total size of data that are transferred to all $SMs$ becomes larger than $V_{s}$, and use the second approach. Additionally, for accessing global memory, it is necessary to confirm that the starting address and the size of the sequential accessing segment is a multiple of 32-byte. In Pascal GPU, a multiple of 128-byte shows better performance than that of 32-byte and 64-byte, but the performance for 32-byte and 64-byte is acceptable. Table 1: Parameters to access single precision data | GTX 1080Ti ---|--- Architecture | Pascal Global Memory Latency (clock cycles) | 258 Bandwidth (Gb/s) | 484 Base clock cycle (MHz) | 1480 SM | 28 Transmission Rate (Byte/clock cycle) | 327 Data Requirement (bytes) | 84,366 Thread Requirement/SM | 768 Warp Requirement/SM | 24 Data Requirement/SM (bytes) | 3072 Flops/clock cycle/core | 2 ### 2.3 Data Mapping As shown in Fig.1(a), in the single-channel convolution ($C=1$), the size of each filter is $K\times K\times{4\mbox{-}byte}$, and they are stored in the global memory continuously. With this data mapping, the filters can be divided only along the dimension $m$, and the filters can be loaded from the global memory efficiently because they are stored continuously. Three approaches for dividing the feature maps and filters can be considered. 1. 1. Only the filters are divided. They are assigned to each $SM$, and in each $SM$, the assigned filters are applied to the whole feature maps (the feature map is processed sequentially against each filter). 2. 2. Only the feature maps are divided. They are assigned to each $SM$, and in each $SM$, the assigned feature maps are processed by all filters (the filters are applied sequentially). 3. 3. Both feature maps and filters are divided, and they are assigned to each $SM$ (the combination of the first and the second approach). By using different approach, the amount of data that has to be loaded to the shared memory memory from the global memory, and the number of FMA operations that can be executed in parallel become different. Therefore, finding a good balance between the size of divided feature maps and filters becomes a key point. (a) Single-Channel (b) Multi-Channel --- Figure 1: Memory storage form of the Filter (a) (b) --- (c) (d) (e) Figure 2: Assignment of the input data In the multi-channel convolution ($C>1$), which is the typical case in the convolution layers of the CNN except for the first one, the data size becomes much larger than the single-channel convolution. Fig.1(b) shows how the filters are stored in the global memory. They are stored along the dimension $ch$ first, and then along the dimension $m$. In this case, the dividing method of the filters along the dimension $m$ as used for the single-channel convolution can not be applied as it is, because the data size of each filter is normally not a multiple of 32-byte. Especially when $K=1$ (the filter size is only 4 bytes), the filters are accessed as 4-byte segments, and it causes serious performance reduction because of non-coalescing memory access.To solve this problem, several approaches can be considered. Fig.2(a) shows the whole data structure before the data division. 1. 1. In Fig.2(b), both the filters and the feature maps are divided along the dimension $ch$, and the data for $C^{\prime}=C/N_{sm}$ channels are assigned to each $SM$ ($N_{sm}$ is the total number of $SMs$). With this division, the data calculated in each thread have to be summed along dimension $ch$. This means that add operations among $SMs$ are required, and $W_{x}\times{W_{y}\times{C^{\prime}}\times{4\mbox{-}byte}}$ byte in the global memory are used for this summation. The global memory accesses to this area and the synchronous operations required for this summation considerably reduce the overall performance. 2. 2. In Fig.2(c), only the filters are divided along the dimension $m$. $M^{\prime}\times C=M/N_{sm}\times C$ filters are assigned to each $SM$, and the whole feature map are loaded to $SMs$ from the global memory. In this approach, if the total size of the filters is less than the total size of all shared memory ($K\times K\times C\times M\times 4\mbox{-}byte<N_{sm}\times S_{shared}$), the divided filters can be cached in each $SM$, and no additional access to the global memory is required. 3. 3. On the other hand, in Fig.2(d), only the feature maps are divided along dimension $y$. The divided feature maps are assigned to each $SM$, and the whole filters are loaded to $SMs$ from the global memory. In this case, if the total size of the feature maps is smaller than the total size of the shared memory, the divided feature maps can be cached in each $SM$, and no additional access to the global memory is required. 4. 4. However, the total size of the filters and feature maps are larger than the total size of the shared memory in general. Thus, as shown in Fig.2(e), both the filters and feature maps have to be divided respectively, and each divided segments is cached in each $SM$ or loaded from the global memory. In this case, there exist many alternatives for how to divide the filters and features maps. According to our preliminary evaluation, the performance with the data dividing method along the dimension $ch$ (Fig.2(b)) is obviously slower than other dividing methods because of the additional access to the global memory for the addition. For achieving higher performance, it is necessary to choose other dividing methods considering the data size and the hardware features of the target GPUs so that in each $SM$ the number of FMA operations that can be executed per loaded data from the global memory is maximized. ## 3 GPU Implementation According to the discussion in Section 2, in both the single-channel and multi-channel convolution, it is important to make the number of FMA operations per pre-fetched data higher than $N_{FMA}$ in order to achieve higher performance. However, in some cases of the single-channel convolution, for example when the size of feature maps is small, the number of FMA operations cannot be kept high enough by data prefetching. This means that, in case of single channel convolution, according to the size of input data, we need to choose one of the two methods described in Section 2.2: data prefetching or data transfer larger than $V_{s}$. In the multi-channel convolution, the size of input data is large enough, and the number of FMA operations can be kept high enough by data prefetching. However, the performance can be improved more by achieving higher FMA operation ratio for the fetched data, because to fetch the data from the global memory, each thread has to issue the instruction to read data, and the clock cycles are spent for issuing these read instructions. Therefore, in the multi-channel convolution, to find the data dividing method that maximizes the number of FMA operations for each divided data is the key to achieve higher performance. ### 3.1 Single-Channel Convolution Here, we describe how to divide the input data to achieve higher performance in the single channel convolution. As for the convolution calculation in each $SM$, we follow the method proposed in [1]. As shown in equation (4), the total amount of the input data is given by: $\begin{split}D_{input}=D_{filter}+D_{map}=(K\times{K}\times{M}+W_{x}\times{W_{y}})\times{4}\,Bytes.\end{split}$ (3) Let $N_{sm}$ be the number of $SMs$. There are two ways to divide the input data and assign them to each $SM$. In the first method, the input data is divided along the dimension $m$ of the filter. $D_{1}$, the size of input data assigned to each $SM$, becomes $\begin{split}D_{1}=\frac{D_{filter}}{N_{sm}}+D_{map}=(K\times{K}\times{\lceil\frac{M}{N_{sm}}\rceil}+W_{x}\times{W_{y}})\times{4}\,Bytes\end{split}$ (4) In general, $D_{map}$ is too large to be stored in the on-chip memory of each $SM$. Thus, $D_{map}$ is divided into $P$ pieces along the dimension $y$. The size of each piece becomes $D_{Inc1}=D_{map}/{P}$. Here, for each line of feature map, since the convolution requires additional $K-1$ lines, the amount of data that have to be held in the on-chip memory becomes $\begin{split}D_{1}=\frac{D_{filter}}{N_{sm}}+D_{Inc1}+(K-1)*W_{x}=(K\times{K}\times{\lceil\frac{M}{N_{sm}}\rceil}+(\lceil\frac{W_{y}}{P}\rceil+K-1)\times{W_{x}})\times{4}\,Bytes\end{split}$ (5) and the number of FMA operation that can be executed for these data in each $SM$ is given by $\begin{split}Th_{1}=\frac{D_{filter}}{N_{sm}}\times{D_{Inc1}}=K\times{K}\times{\lceil\frac{M}{N_{sm}}\rceil}\times{\lceil\frac{W_{y}}{P}\rceil\times{W_{x}}}.\end{split}$ (6) In the second method, the input data is divided along the dimension $y$ of the feature map. In this case, $D_{2}$, the amount of the input data assigned to each $SM$, becomes $\begin{split}D_{2}=D_{filter}+\frac{D_{map}}{N_{sm}}=(K\times{K}\times{M}+({\lceil\frac{W_{y}}{N_{sm}}\rceil+K-1})\times{W_{x}})\times{4}\,Bytes.\end{split}$ (7) $D_{filter}$ is too large to be stored in the on-chip memory in general, and it is divided into $Q$ pieces. The size of each piece becomes $D_{Inc2}=D_{filter}/{Q}$. Then, $D_{2}$ becomes $\begin{split}D_{2}=D_{Inc2}+\frac{D_{map}}{N_{sm}}=\frac{D_{filter}}{Q}+\frac{D_{map}}{N_{sm}}=(K\times{K}\times{\lceil\frac{M}{Q}\rceil}+(\lceil\frac{W_{y}}{N_{sm}}\rceil+K-1)\times{W_{x}})\times{4}\,Bytes\end{split}$ (8) and the number of FMA operation that can be executed for these data in each $SM$ is given by $\begin{split}Th_{2}=D_{Inc2}\times\frac{D_{map}}{N_{sm}}=K\times{K}\times{\lceil\frac{M}{Q}\rceil}\times({\lceil\frac{W_{y}}{N_{sm}}\rceil})\times{W_{x}}.\end{split}$ (9) The values of $P$ and $Q$ are decided considering if $D_{1}$ or $D_{2}$ is smaller than $S_{shared}$, and if $Th_{1}$ or $Th_{2}$ is larger than $N_{FMA}$. If $P=1$ or $Q=1$, the feature maps or the filters are not divided, and they are transferred to the on-chip memory at a time. If $P>1$ or $Q>1$, the feature maps or the filters are divided into several pieces, and the pieces are transferred to each $SM$ by using the data prefetching. With smaller $P$ and $Q$, $D_{1}$, $D_{2}$ and $Th_{1}$, $Th_{2}$ become larger. The lower bound of $P$ and $Q$ is given by the requirement that $D_{1}$ and $D_{2}$ have to be smaller than $S_{shared}$, and the upper bound is given by the requirement that $Th_{1}$ and $Th_{2}$ should be larger than $N_{FMA}$. $P$ and $Q$ should be chosen so that these requirements can be satisfied. In our implementation, $P$ and $Q$ are decided as follows. 1. 1. $Th_{1}$ or $Th_{2}$ should be larger than the number of FMA operation $N_{FMA}$. $Th_{1}\geq N_{FMA}$ and $Th_{2}\geq N_{FMA}$ Thus, the upper bound of $P$ and $Q$ (they must be smaller than $W_{y}$ and $M$ respectively) is given as follows $P\leq\frac{K\times{K}\times{\lceil\frac{M}{N_{sm}}\rceil}\times{W_{y}}\times{W_{x}}}{N_{FMA}}\,and\,P\leq{W_{y}}$, $Q\leq\frac{K\times{K}\times{M}\times{\lceil\frac{W_{y}}{N_{sm}}\rceil}\times{W_{x}}}{N_{FMA}}\,and\,Q\leq{M}$ 2. 2. $D_{1}$ and $D_{2}$ must be smaller than the size of on-chip memory. The lower bound of $P$ and $Q$ is given as follows. $P\geq\frac{4\times{W_{y}}\times{W_{x}}}{S_{shared}-4\times{K\times{K\times{\lceil\frac{M}{N_{SM}}}\rceil}}+(1-K)\times{4}\times{W_{x}}}$, $Q\geq\frac{4\times{M}\times{K}\times{K}}{S_{shared}-4\times{W_{x}}\times{(\lceil\frac{W_{y}}{N_{SM}}+K-1)}}$ Actually, there exist one more requirement to decide this lower bound. The number of required registers for the computation must be smaller than that supported in each $SM$. Its detail is not shown here, but considering this requirement, the lower bound is calculated. 3. 3. If there exist $P$ and $Q$ ($P$ and $Q$ must be an integer) in the range specified by (2) and (3), Any of them can be used. In our current implementation, the minimum ones are chosen as $P$ and $Q$, because the smaller values means less number of division, and make the processing sequence simpler. If no value exists, $P$ and $Q$ are set to 1. 4. 4. Using the obtained $P$ and $Q$, $D_{1}$ and $D_{2}$ are calculated and compared. If $D_{1}$ is smaller than $D_{2}$, $Q$ is reset to 1 to use the first dividing method described above, and otherwise, $P$ is reset to 1 to use the second one. Both methods can be used because they both satisfy the requirements, but for the safety (for leaving more memory space on the on-chip memory), the smaller one is chosen. Following this procedure, the input data are divided and allocated to each $SM$ in the best balance. ### 3.2 Multi-Channel Convolution As described above, in the multi-channel convolution, both feature maps and filters are divided, and prefetching is used to transfer them to each $SM$ from the global memory. Recently, the block-based methods show high performance in convolution due to their continuous and simple memory access sequence. Fig.3 shows the data mapping of the filters and feature map, and how they are divided and calculated in each $SM$. In the block-based method, as shown in Fig.3(a), the following data are loaded to the on-chip memory in each $SM$. 1. 1. $S$ bytes of each filter along the dimension $ch$ (called segment in the following discussion) of $M^{\prime}$ filters ($S\times{M^{\prime}}$ bytes in total), and 2. 2. a part of feature map, $W^{\prime}_{x}\times{W^{\prime}_{y}}\times{4\,bytes}$ in the same channel ($W^{\prime}_{x}$ is an arbitrary value that is decided by the size of on-chip memory, but $W^{\prime}_{y}$ is specified as $\lceil\frac{S}{K\times{4bytes}}\rceil$, because when $S$ bytes are fetched along the dimension $ch$, $\lceil\frac{S}{K\times{4bytes}}\rceil$ lines in the feature map are required to apply the filter). Then, the convolution is calculated for these data, and the next data (next $S\times M^{\prime}$ bytes of filters and $W^{\prime}_{y}\times W^{\prime}_{x}$ bytes of feature maps are loaded by data prefetching. In [1], the filter size is chosen as $S$ ($S=K\times{K}\times{4\,bytes}$), and only the filters of the target channel and a part of feature map of the same $channel$ are loaded to the on-chip memory. However, the filter size $K\times K$ is usually odd and often small, and the performance is seriously degraded because of non-coalescing memory access. [16] tried to solve this problem by extending $S$ to 128-bytes. By fetching continuous 128 bytes on the global memory, the highest memory throughput can be achieved in GPUs. In this method, the filters of several channels (and a part of the next channel) are fetched at the same time, and they are kept in the on-chip memory. First, only the filters of the first channel are used for the computation, and then, the filters of the next channel are used. With this larger $S$, $M^{\prime}$ has to be kept small because of the limited size of on-chip memory, and smaller $M^{\prime}$ means less parallelism ($M^{\prime}$ filters are applied in parallel to the feature map of the same channel). In [1], higher parallelism comes first, while in [16], lower access delay has a higher priority. Figure 3: Multi-Channel Convolution Kernel Here, we propose a stride-fixed block method not only to maintain the efficient global memory access, but also to achieve high parallelism in each $SM$. 1. 1. $S$ is set to a multiple of 32-bytes. Actually, 32 or 64 is used. Small $S$ allows larger $M^{\prime}$, namely higher parallelism, under the limited size of on-chip memory. When $S$ is 32 or 64 bytes, the memory throughput from the global memory becomes a bit worse than $S=128$ bytes (the highest throughput), but it is acceptable. $S=32$ is the minimum value to maintain efficient global memory access. 2. 2. Next, we fix $W^{\prime}_{x}$. $W^{\prime}_{x}$ pixels in the feature map are fetched along the dimension $x$ from the global memory. Thus, $W^{\prime}_{x}$ should be a multiple of 128-bytes to achieve the highest memory throughput. Larger $W^{\prime}_{x}$ is preferable because it increases the Instruction Level Parallelism (ILP), which can improve the performance of the convolution. 3. 3. After deciding the values of $S$ and $W^{\prime}_{x}$, the most suitable $M^{\prime}$ can be found by the requirements of the number of FMA operations. $M^{\prime}\geq\frac{N_{FMA}\times{4\mbox{-}bytes}}{S\times{W^{\prime}_{x}}}$. 4. 4. Because the data prefetching is used to fetch the next data set while the current data set is being used for the current calculation, the size of data set cannot exceeds the half of the size of shared memory. Thus, $(S\times{M^{\prime}}+\lceil\frac{S}{K\times{4bytes}}\rceil\times{W^{\prime}_{x}})\leq\frac{S_{shared}}{2}$ Here, $\lceil\frac{S}{K\times{4bytes}}\rceil=W^{\prime}_{y}$ is the number of feature maps required for the calculation. With this approach, for given $S$, $W^{\prime}_{x}$ and $M^{\prime}$ to achieve high performance based on block method can be obtained. From here, we describe how the convolution calculation is executed in each $SM$. As shown in Fig.3(a)(b), first, each $SM$ loads $S\,bytes$ of $M^{\prime}$ filters to the shared memory. At the same time, $W^{\prime}_{x}$ pixels on $\lceil\frac{S}{K\times{4bytes}}\rceil=W^{\prime}_{y}$ lines of the feature maps are also loaded. After the first round loading of these data, the same size of data for the next round are pre-fetched: the next $S\times M^{\prime}$ bytes along the dimension $ch$, and the next $W^{\prime}_{x}$ pixels of the $W^{\prime}_{y}$ lines. During the second round loading, the convolutions for the first round data set are calculated on the chip as shown in Fig.3(b)(c). On the chip, each thread corresponds to one target pixel of the feature map as shown in Fig.3(b). Because the accessing speed of registers is faster than that of shared memory, it is required to transfer each data in the shared memory to the registers in order to achieve high performance. In the convolution computation, the target pixel and its neighbors in the feature map are sent to the corresponding registers by each thread. Here, one important point is that only $S/{4bytes}$ pixels in the feature map have to be loaded onto the registers. The rest pixels, the red ones in Fig.3(b), are just held in the shared memory for the next round. The filter data are also transferred to the registers by the corresponding thread, but in this case, all data are transferred to the registers because all of them are used. After that, each pixel data is multiplied by the corresponding filter data, and their products are added. When all computations for the data stored in on-chip memory has been finished, data prefetching for the third round is started. During this loading, the convolution for the second round data is calculated. By using this method, the size of $S$ can be kept small, and the number of filters $M^{\prime}$ can be increased. This ensures that more filters can be applied in parallel to the same feature map. This does not increase the number of data loading of the feature maps, and hides the latency caused by global memory access. ## 4 Experimental analysis We implemented our two convolution kernels on Pascal series GPU Geforce GTX 1080Ti by using the CUDA 8.0. Their performances were evaluated using many convolutions which are commonly used in popular CNN models[15][9][6][11], and compared with the latest public library Cudnn v7.1[12]. In the single-channel convolution, we changed the sample size of the feature maps from 28 to 1K and the size of the corresponding channels from 512 to 32\. The filter size is 1, 3 or 5, which is usually used in many CNN models. In CUDA programming, by assigning more number of blocks to each $SM$, the $SMs$ can be kept busy. In our current implementation, $N_{block}=2\times{N_{SM}}=2\times{28}$ blocks are used. Two blocks are assigned to each $SM$, and 512 threads are assigned to each block. Thus, the maximum number of registers for each thread is constrained to 128. For each tested case, $P$ and $Q$ are decided following the method described in Section 3.1. Fig.5 shows the results of the single-channel convolution. Our method is faster than Cudnn v7.1 in all tested cases. The performance gain is 1.5X to 5.6X, and its average is 2.6X. In the multi-channel convolution, we changed the sample size of the feature maps from 7 to 512, and the size of the corresponding channels from 64 to 512. The filter size is also 1, 3, or 5. As discussed in Section 3.2, larger $M^{\prime}$ is preferable for making data prefetching more effective. Therefore, we fixed the segment size $S$ as 32 or 64 bytes, and then $M^{\prime}$ and $W^{\prime}_{x}$ are decided following the method described in Section 3.2. According to our preliminary evaluation, when $M^{\prime}=64$ and $W^{\prime}_{x}=128$, the performance becomes best, and we used these values for this comparison. As shown in Fig.5, our method is faster than Cudnn in all tested cases, and the throughput has been increased by 1.05X to 2X, with an average increase of 1.39X. In [1], a different GPU is used, and a direct comparision is not possible. However, when $K=3$, our performance is 4X faster than [1] on GPU the peak performance of which is 2.4X faster than that used in [1]. We also implemented our two kernels on Maxwell series GPU GTX Titan X, and it also showed that our performance is faster than Cudnn on the same GPU by 1.3X to 3.7X in the single-channel convolution and 1.08X to 1.8X in the multi- channel convolution. Filter Size = 1 Filter Size = 3 Filter Size = 5 --- Figure 4: Performance of the Single-Channel Convolution Kernel Filter Size = 1 Filter Size = 3 Filter Size = 5 --- Figure 5: Performance of the Multi-Channel Convolution Kernel ## 5 CONCLUSIONS In this paper, we proposed two convolution kernels on Pascal series GPUs for single-channel and multi-channel respectively. For single-channel convolution, we introduced an effective method of data mapping, which can hide the access delay of the global memory efficiently. For multi-channel convolution, we introduced a method that not only guarantees the memory access efficiency, but also achieves high FMA operation ratio per loaded data. Performance comparison with the public library Cudnn shows that our approaches are faster in all tested cases: 1.5X to 5.5X in the single-channel convolution and 1.05X to 2X in the multi-channel convolution. Our approaches was designed assuming Pascal architecture, but the performance is also faster than Cudnn on Maxwell architecture. This practice shows that our approaches can be applied to the wide range of CNN models on various GPUs. In our current implementation, the throughput is still lower than the theoretical maximum. It means that the convolution kernel still has the room for improvement, and this is our main future work. ## References * [1] Xiaoming Chen, Jianxu Chen, Danny Z. Chen, Xiaobo Sharon Hu. “Optimizing Memory Efficiency for Convolution Kernels on Kepler GPUs,”In DAC, 2017. * [2] A. Poznanski, L. Wolf, “Cnn-n-gram for handwriting word recognition,”in: CVPR, pp. 2305-2314, 2016. * [3] D. Yu, W. Xiong, J. Droppo, A. Stolcke, G. Ye, J. Li, and G. Zweig, “Deep convolutional neural networks with layer-wise context expansion and attention,”in Proc.Interspeech, 2016. * [4] Eddie Bell. “A implementation of squeezenet in chainer.”https://github.com/ejlb/squeezenet-chainer, 2016. * [5] X. Mei, X. Chu, “Dissecting GPU memory hierarchy through microbenchmarking,”IEEE Trans. Parallel Distrib. Syst, 2016. * [6] K. Simonyan and A. Zisserman. “Very deep convolutional networks for large-scale image recognition,”In ICLR, 2015. * [7] Zhang, Y. and Wallace, B. “A sensitivity analysis of (and practitioners guide to) convolutional neural networks for sentence classification.”arXiv:1510.03820, 2015. * [8] A. Lavin. “Fast algorithms for convolutional neural networks.”arXiv:1509.09308, 2015. * [9] K. He, X. Zhang, et al. “Deep residual learning for image recognition,”arXiv:1512.03385, 2015. * [10] S. Li, Y. Zhang, C. Xiang, and L. Shi. “Fast Convolution Operations on Many-Core Architectures.”In HPCC, pages 316-323, 2015. * [11] Szegedy, C. et al. “Going deeper with convolutions.”Preprint at http://arxiv.org/abs/1409.4842, 2014. * [12] Sharan Chetlur, Cliff Woolley, et al. “cuDNN: Efficient primitives for deep learning.”arXiv:1410.0759, 2014. * [13] M. Mathieu, M. Henaff, et al. “Fast training of convolutional networks through ffts.”In CoRR, 2013. * [14] Jia, Y. “Caffe: An open source convolutional architecture for fast feature embedding.” http://caffe.berkeleyvision.org/, 2013. * [15] Krizhevsky, A., Sutskever, I., and Hinton, G. E. “ImageNet classification with deep convolutional neural networks,”In NIPS, pp. 1106-1114, 2012. * [16] Guangming Tan, Linchuan Li, et al. “Fast implementation of DGEMM on Fermi GPU.”In Supercomputing 2011, pages 35:15:11, New York, NY, USA, 2011. ACM. * [17] R. Nath, S. Tomov, and J. Dongarra. “An improved magma gemm for fermi gpus.”Technical Report 227, LAPACK Working Note, 2010.
# Quasidisorder Induced Topology M. F. Madeira Departamento de Física and CeFEMA, Instituto Superior Técnico, Universidade de Lisboa, Av. Rovisco Pais, 1049-001 Lisboa, Portugal P. D. Sacramento Departamento de Física and CeFEMA, Instituto Superior Técnico, Universidade de Lisboa, Av. Rovisco Pais, 1049-001 Lisboa, Portugal (August 28, 2024) ###### Abstract We study the effects of quasidisorder and Anderson disorder on a two dimensional topological superconductor with an applied external magnetic field. The cases of a $p$-wave superconductor and a noncentrosymmetric superconductor with mixed $p$ and $s$-wave pairings and Rashba spin-orbit coupling are studied. We show that, for a perpendicular magnetic field, the introduction of quasidisorder leads to the appearance of topological phases in new regions, characterised by an integer value of the Chern number. For a parallel magnetic field, we identify regimes with the appearance of new Majorana flat bands and also new unidirectional Majorana edge states, as quasidisorder is introduced. We show that the Majorana flat bands have a quantized Berry phase of $\pi$ and identify it as a topological invariant. Two topological transitions are identified and the values of the critical exponents $z$ and $\nu$ are obtained. The fractal nature of the eigenstates is discussed both for Anderson disorder and Aubry-André disorder. ††preprint: APS/123-QED ## I Introduction The search and study of topological properties of matter has proved fruitful in recent years in research in materials science and condensed matter physics. Superconductors have long been a focus of interest due to their promising applications. Superconductors with intrinsic topological properties, in particular, have recently attracted theoretical and experimental interest due to phenomena associated with surface or edge Majorana modes, which appear from an interplay between topology and bulk-boundary correspondence [1, 2, 3]. These Majorana zero modes emerge with non-Abelian exchange statistics and are sought after due to their promising expected applications in quantum computing, being candidates for the building blocks of a quantum qubit [4, 5]. It has been theoretically predicted that Majorana states appear as flat dispersion bands in gapless superconducting phases, such as in the $d_{xy}$+$p$-wave pairing noncentrosymmetric superconductor in two dimensions with preserved time-reversal symmetry [6, 7], or for a $p$-wave topological superconductor in two dimensions, with broken time reversal symmetry by an applied magnetic field parallel to the two dimensional plane of the system [8]. Flat bands also emerge on the surface of three dimensional noncentrosymmetric superconductors, with spin-orbit coupling and which preserve time-reversal symmetry [9, 10]. It is predicted that flat bands can increase the critical temperature for superconductivity, and even give rise to room-temperature superconductivity [11, 12]. Similar behavior has been found when one has finite-size systems (with increased fluctuations of the density of states) [13, 14], non-homogeneous order parameters [15, 16, 17, 18, 19, 20], or fractal (critical) states [21, 22, 23, 24, 25, 26, 27, 28, 29] with corresponding spatial fluctuations of the amplitude of the wave functions. The difference between an isolated flat band and a flat band with band touchings has also been recently discussed [30]. It was shown that isolated flat bands are not needed to achieve a higher superconducting temperature, and that band touchings can actually increase it. Flat electronic bands can also be found in some Kagome-type superconductors [31]. A growing interest has been seen in these types of materials, AV3Sb5 (with A$=$K, Rb, Cs), which can host exotic quantum properties, displaying topological phases, an unconventional charge density wave, and evidence of time-reversal symmetry breaking [32, 33, 34, 35]. The study of perturbations in condensed matter systems, namely through the introduction of disorder, is a central issue. On one hand, introducing disorder can destroy some phases and their properties, preventing their experimental observation. In this sense, the study of their robustness becomes crucial. On the other hand, disorder can by itself lead to new phenomena or stabilize previously existing phases. One type of disorder that has been attracting interest in the research field is quasiperiodic disorder. These systems are somewhat in between periodic and truly random systems, and exhibit interesting phenomena, in transport [36, 37], topological properties [38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51], and critical behaviour [36, 52, 53, 54]. It is possible to realize these types of systems in experimental setups of ultracold atoms [55, 56], in optical lattices [57, 58] or in photonics systems [59]. In addition to systems subject to quasiperiodic potentials, as in the Aubry-André model [60], there has been growing interest in Moiré systems in which two incommensurate lattices are connected, or in which layers of lattices are put in contact and rotated, such as the 2d twisted bilayer graphene [61, 62, 63, 64, 65]. In such systems, a superlattice potential is created from proximity coupling between the two lattices, which, depending on the angle of rotation between the two, may exhibit quasiperiodicity. An example of the study of coexisting quasidisorder and superconductivity, which is significant in the context of this work, is the one dimensional Kitaev chain with Aubry-André modulation [66, 67, 68, 69]. Without superconductivity the model has a topological nature revealed by its mapping to a $2d$ quantum Hall system [70], maintaining a topological nature as we add superconducting pairing. In general, the mappings involve a corresponding model in a higher dimension in the form of some parent Hofstadter generalized Hamiltonian. Topology in quasicrystals may be understood considering mappings to higher dimensions, typically of the types $1d$ to $2d$ and $2d$ to $4d$. In $1d$ with no superconductivity the model is self-dual (position and momentum space) and there is a single transition from an extended state phase to a phase where all the states are localized. At the transition point the system has critical states. Generalized models show the existence of mobility edges, such that there is a separation as a function of energy between extended and localized states [71, 72, 73, 74, 75, 76, 77, 78] and the existence of hidden dualities leads to a rich class of systems, where such edges appear [79]. The introduction of $p$-wave pairing in the Aubry-André model leads to the appearance of a finite extent region of critical (fractal) states, between the regions of extended and localized states. Remarkably, the transitions between localised and critical regimes have been studied and were found to deviate from the known Aubry-André universality class [68, 69]. In this work we study a model of a two-dimensional superconductor with spin triplet $p$-wave pairing, or mixed $p$ and $s$-wave pairings with Rashba spin- orbit coupling, in the presence of a time reversal symmetry breaking magnetic field. Some materials which are candidates for realizing triplet pairing superconductivity include Sr2RuO4 [80], UPt3 [81] and CuxBi2Se3 [82]. In the presence of $s$-wave pairing and Rashba spin orbit coupling, the model describes a noncentrosymmetric superconductor, of which are examples $\text{CePt}_{3}\text{Si}$ [83], $\text{CeIrSi}_{3}$ [84] and $\text{CeRhSi}_{3}$ [85]. In the noncentrosymmetric regime the breaking of inversion symmetry allows for the mixture of spin-triplet and spin-singlet pairings. This mixing is expected to lead to novel phenomena such as higher than usual values of the upper critical field [86, 87]. The clean model has been studied, in both the centrosymmetric and the noncentrosymmetric regimes, and is known to possess diverse topological properties. If time-reversal symmetry is preserved, the model displays gapless Majorana edge states and is characterised by a $\mathbf{Z}_{2}$ invariant. The observed properties when time-reversal symmetry is broken by an external magnetic field are found to be very dependent on its direction in relation to the two-dimensional superconducting plane. If the magnetic field is such that it is perpendicular to the plane of the superconductor, the model has a rich phase diagram indexed by the Chern number [88]. When the magnetic field is parallel to the plane of the system, interesting phenomena, such as Majorana flat bands or Majorana unidirectional states, appear on phases with a gapless bulk [8, 89]. The effect of disorder may be considered in different ways. One possibility is to consider a non-homogeneous magnetic field, achieved by inserting magnetic impurities in the clean superconductor [90] which may give rise to or change topological properties in the system. Examples include the addition of chains of magnetic adatoms [91, 92], islands of magnetic impurities [93] or fully random distributions of impurities [94]. Another possibility is to consider potential scattering impurities on the superconductor in the presence of a constant magnetic field, either perpendicular or parallel to the system. We are interested in studying the effects of quasidisorder in these regimes. Besides Aubry-André disorder, we will also consider Anderson disorder as a comparison to the effects of quasi-periodicity. Anderson localization does not require full randomness. If differences in potential between sites are large enough compared to hoppings, one may expect a transition to localized states. In addition to full randomness, a quasidisordered potential leads to localization if the disorder amplitude is large enough [95]. One expects that the Aubry-André quasiperiodic potential should affect the long-range nature of states, and in particular topological states that are by themselves of long-range nature. Aubry-André is expected to be naturally of a multifractal nature. Anderson and Aubry-André are different and, in particular, critical states due to Anderson appear at the transition to localization while in the Aubry-André added to the Kitaev $1d$ model one finds phases with this behavior (or in $2d$ a mixture of critical states in the crossover to localization). Multifractality probes long distances and therefore one expects that it may enhance superconductivity due to Chalker scaling [96, 97], as expected and observed with other inhomogeneities. Multifractal wave functions have larger spatial overlap and stronger state to state correlations for states with similar energies. As stated previously, quasiperiodicity may also lead to topological properties [98, 99]. A $2d$ topological insulator plus quasiperiodic potential shows a transition from a trivial insulator to a topological insulator. Flat topological bands and eigenstate criticality have also been shown as a result of a quasiperiodic perturbation in the context of the Bernevig-Hughes-Zhang model plus $2d$ quasiperiodic potential [100]. The presence of gapless states in a system may also be associated with long- distance behavior. For instance, nodal points of Weyl semimetals may lead to interesting behavior in the presence of disorder. It has been shown that they survive the presence of moderate disorder [101]. On the other hand, in the case of gapless states of the form of nodal loops, any amount of disorder mixes states. Disorder-driven multifractality has been shown in Weyl nodal loops [102]. In the case of magic angle semimetals quasiperiodicity generically leads to flat bands in nodal, semi-metallic structrures. A transition from a Weyl semimetal to metal driven by quasiperiodic potential has been found in $3d$ [52, 103]. While the influence of disorder, either Anderson or quasidisorder has been extensively considered in the case of one-dimensional systems, including in the presence of superconductivity, it is interesting to consider their effects on a two-dimensional $p$-wave superconductor, and in particular in the presence of a magnetic field. In the clean system the topology is influenced by the orientation of the magnetic field and, in particular, the gapped or gapless nature of the states may lead to different responses to disorder. As stated before, the difference of symmetry classes plus disorder gives rise to new universality classes. Also, topology may be induced by quasiperiodicity, which leads to the expectation of new universality classes (beyond the usual classification), as found in the one-dimensional case. In particular, one may expect interesting effects with the interplay of quasiperiodicity due to the presence of critical bulk states, and the existence of Majorana flat bands. The long-range nature of the quasiperiodic potential and the intrinsic long- range nature of the gapless states may lead to an interesting competition. A distinction between Anderson disorder (with moderate intensity) and quasidisorder is therefore interesting to consider, as shown in non- superconducting systems, where for instance nodal points and nodal loops are affected differently by Anderson disorder, or on a semimetal where imposing a quasiperiodic potential leads to flat bands. The rest of the paper is organized as follows. Section II introduces the model of the Hamiltonian and the topological properties of the clean system are discussed, first under a perpendicular and second under a parallel magnetic field, respectively in subsections II.1 and II.2. In subsection II.2 we derive the regions where the model is topological, and show that the topological regions are characterized by a Berry phase of $\pi$. In section III we present the results for the disordered model under a perpendicular magnetic field. We show that the introduction of Aubry-André disorder leads to the appearance of topological phases in new regions. In IV we present the results for the disordered model under a parallel magnetic field. First we discuss the localization properties of the system in real space under different types of disorder, using the inverse participation ratio (IPR). We then turn to a mixed space description and discuss the evolution of the system as Anderson or Aubry-André disorder are introduced. We show that the introduction of Aubry- André disorder leads to the appearance of new regimes: for the $p$-wave superconductor, new gapless regimes with Majorana flat bands appear, and for the noncentrosymmetric superconductor, new regimes with unidirectional edge states appear. We then obtain the Berry phase using twisted boundary conditions and show it is quantized to a value of $\pi$ for the quasidisorder induced flat bands. Identifying it as a topological invariant, we study two topological transitions and obtain the critical exponents $z$ and $\nu$, which we find to deviate from the known universality classes. Finally, using the IPR we study the nature of the eigenfunctions distinguishing between localized, single-fractal and multifractal regimes in the thermodynamic limit for both Anderson and Aubry-André disorder. We conclude in section V. Three appendices discuss some further results on the disorder driven transitions under a perpendicular magnetic field in Appendix A, the influence of the dimensionality of the quasidisorder potential in Appendix B and the energy spectra and density of states for the disordered noncentrosymmetric superconductor in Appendix C. ## II Model Hamiltonian In momentum space, the Bogoliubov-de Gennes (BdG) Hamiltonian matrix of the two dimensional model is written as $\mathcal{H}(\mathbf{k})=\left(\begin{array}[]{cc}\xi(\mathbf{k})+\mathbf{B}\cdot\bm{\sigma}&\Delta(\mathbf{k})\\\ \Delta^{\dagger}(\mathbf{k})&-\xi^{T}(-\mathbf{k})-\mathbf{B}\cdot\bm{\sigma}^{*}\end{array}\right)$ (1) in a basis $(\bm{c}_{\mathbf{k}}^{\dagger},\bm{c}_{-\mathbf{k}})=(c_{\mathbf{k}\uparrow}^{\dagger},c_{\mathbf{k}\downarrow}^{\dagger},c_{-\mathbf{k}\uparrow},c_{-\mathbf{k}\downarrow})$ with $c_{\mathbf{k}\sigma}^{\dagger}$ ($c_{\mathbf{k}\sigma}$) the creation (annihilation) operator for an electron with momentum $\bm{k}=(k_{x},k_{y})$ and spin projection $\sigma$. In the BdG Hamiltonian, $\xi(\bm{k})=\epsilon_{\bm{k}}\sigma_{0}+\mathbf{s}\cdot\bm{\sigma}$, where $\epsilon_{\mathbf{k}}=\left[-2t\left(\cos k_{x}+\cos k_{y}\right)-\mu\right]\sigma_{0}$ is the kinetic term, with $t$ the nearest- neighbour hopping integral and $\mu$ the chemical potential, $\mathbf{s}\cdot\bm{\sigma}=-\alpha(-\sin k_{y},\sin k_{x},0)\cdot\bm{\sigma}=-\alpha\left[-\sin k_{y}\sigma_{x}+\sin k_{x}\sigma_{y}\right]$ is the Rashba spin-orbit term with $\mathbf{s}$ the spin-orbit vector. The term $\mathbf{B}\cdot\bm{\sigma}$ describes the Zeeman coupling of the electrons with an external magnetic field $\mathbf{B}$ and $\hat{\Delta}(\mathbf{k})=\left[\Delta_{s}+\mathbf{d}(\mathbf{k})\cdot\bm{\sigma}\right]\left(i\sigma_{y}\right)$ is the superconducting gap function. The pairing vector is chosen as $\mathbf{d}=d(-\sin k_{y},\sin k_{x},0)$, so that $d$ is the $p$-wave pairing amplitude and $\Delta_{s}$ is the $s$-wave pairing amplitude. The simultaneous existence of $s$ and $p$-wave terms is possible with a nonzero spin-orbit term, which breaks the parity symmetry. The case of study is that of a system with periodic boundary conditions along the $x$ direction and open boundary conditions in the $y$ direction, such as in a cylinder geometry. Thus we can also write the Hamiltonian in a mixed space, $(k_{x},y)$, where a Fourier transform to the reciprocal space is only done in the $x$ direction. In this case, for each value of $k_{x}$ the Hamiltonian matrix has a dimension $(4\times N_{y})\times(4\times N_{y})$, where $N_{y}$ is the number of sites in $y$. It is also of interest to write the Hamiltonian in real space. In this case the Hamiltonian matrix has dimension $(4\times N)\times(4\times N)$ with $N=N_{x}\times N_{y}$ the total number of sites and $N_{x}$, $N_{y}$ the number of sites in the $x$ and $y$ directions, respectively. When $\mathbf{B}=0$, the system respects the time-reversal symmetry (TRS) $\mathcal{T}=(\sigma_{0}\otimes\mathrm{i}\sigma_{y})$ and the particle-hole symmetry (PHS) $\mathcal{P}=(\sigma_{x}\otimes\sigma_{o})$ such that $\begin{split}&\mathcal{P}\mathcal{H}(\mathbf{k})\mathcal{P^{\dagger}}=-\mathcal{H}^{*}(-\mathbf{k}),\\\ &\mathcal{T}\mathcal{H}(\mathbf{k})\mathcal{T}^{\dagger}=\mathcal{H}^{*}(-\mathbf{k}),\end{split}$ (2) and $\mathcal{T}^{2}=-1$, $\mathcal{P}^{2}=1$. Therefore the Hamiltonian belongs to the DIII symmetry class, and if $|d|>|\Delta_{s}|$ the system has a nontrivial $\mathbf{Z}_{2}$ number, displaying gapless counterpropagating Majorana edge states [88, 8]. For $\mathbf{B}\neq 0$ the time-reversal symmetry is broken. The system exhibits different topological properties whether the applied magnetic field is perpendicular or parallel to the plane of the system, as will be now discussed. ### II.1 Perpendicular Magnetic Field Let us first consider the case in which the external magnetic field is perpendicular to the plane of the system, $\mathbf{B}=(0,0,B_{z})$. We have a gap closing point if one of the equations is satisfied [88]: $\begin{split}(-4t-\mu)^{2}+\Delta^{2}_{s}=B^{2}_{z},\\\ \mu^{2}+\Delta^{2}_{s}=B^{2}_{z},\\\ (4t-\mu)^{2}+\Delta^{2}_{s}=B^{2}_{z}.\end{split}$ (3) Eqs. 3 define the boundaries between regions in which the system has different topological properties. At the gap closing points the $D$ class system with broken time reversal symmetry undergoes topological transitions between gapped phases with different Chern numbers. The phase diagram of the system (indexed by the Chern number) is presented in Fig. 1(a) for $t=1$, $\Delta_{s}=0$, and $d>0$. Figure 1: Phase diagram for a) Chern number and b) winding number $I(k_{y}=0,\pi)$ as a function of $\mu$ and $B_{z}$, for $t=1$, $\Delta_{s}=0$, $d>0$. The regimes with a Chern number of zero and $B_{z}<2$, $0<|\mu|<4t$ exhibit edge states, besides having $C=0$. This can be explained by one additional topological invariant. It can be defined noting that the Hamiltonian obeys a particle hole symmetry $\mathcal{P}=(\sigma_{x}\otimes\sigma_{0})$ with $\mathcal{P}\mathcal{H}(\mathbf{k})\mathcal{P}^{\dagger}=-\mathcal{H}^{*}(-\mathbf{k}).$ (4) For the values $k_{y}=0$ and $k_{y}=\pi$, the Hamiltonian obeys $\mathcal{H}^{*}(-\mathbf{k})=\mathcal{H}(\mathbf{k})$ and thus anticommutes with $\mathcal{P}$, $\\{\mathcal{H}(\mathbf{k}),\mathcal{P}\\}=0$. Therefore the basis which diagonalizes $\mathcal{P}$ anti-diagonalizes the Hamiltonian. A winding number $I(k_{y})$ can then be defined as [88] $\begin{split}I\left(k_{y}\right)=\frac{1}{4\pi i}\int_{-\pi}^{\pi}dk_{x}\operatorname{tr}\big{[}q^{-1}(k_{x})\partial_{k_{x}}q(k_{x})-\\\ q^{\dagger-1}(k_{x})\partial_{k_{x}}q^{\dagger}(k_{x})\big{]},\quad k_{y}=0,\pi,\end{split}$ (5) with $q(k_{x})=\left(\begin{array}[]{cc}-\epsilon_{\mathbf{k}}-B_{z}+id\sin k_{x}&\Delta_{s}-i\alpha\sin k_{x}\\\ -\Delta_{s}+i\alpha\sin k_{x}&-\epsilon_{\mathbf{k}}+B_{z}+id\sin k_{x}\end{array}\right)$ (6) the anti-diagonal block of the Hamiltonian matrix. The values of $I(0)$ and $I(\pi)$ inside each phase are represented in Fig. 1(b). The invariant $I(k_{y})$ loses its meaning if a finite magnetic field in the $y$ direction, $B_{y}$, is applied. However, we found that this is not true for the Chern number. Fig. 2 shows phase diagrams indexed by the Chern number as a function of $B_{z}$ and $B_{y}$ for three different values of $\mu$. In this case the Chern number depends only on the value of $\sqrt{B^{2}_{y}+B^{2}_{z}}$. Also note that the diagrams only concern values of $B_{z}>0$, excluding the points where $B_{z}=0$ and $B_{y}\neq 0$. In Fig. 3 we present the phase diagram of the system as a function of $\mu$ and $B_{z}$ for constant values of $B_{y}$. Figure 2: Phase diagrams for ($B_{z}>0,B_{y}$), indexed by the Chern number, obtained numerically for a) $\mu=0$, b) $\mu=1$ and c) $\mu=-3.5$ for $\Delta_{s}=0$. Figure 3: Phase diagram indexed by the Chern number as a function of $\mu$ and $B_{z}$ (with $B_{z}>0$), for $t=1$, $\Delta_{s}=0$, $d>0$, and a) $B_{y}=2$ and b) $B_{y}=4.5$. ### II.2 Parallel Magnetic Field Now let us consider the case in which the applied magnetic field is parallel to the system, $\mathbf{B}=(B_{x},B_{y},0)$. This could be realized, for instance, by threading a wire through the center of the superconductor in a cylindrical geometry. Taking first the $s$-wave term $\Delta_{s}$ and the spin-orbit term $\alpha$ to be zero, the eigenvalues of the Hamiltonian are given by $E(\mathbf{k})=\pm\sqrt{z_{1}\pm 2\sqrt{z_{2}}},$ (7) with $\begin{split}z_{1}=\mathbf{d}\cdot\mathbf{d}+\epsilon^{2}_{\mathbf{k}}+\mathbf{B}\cdot\mathbf{B},\\\ z_{2}=\epsilon^{2}_{\mathbf{k}}(\mathbf{B}\cdot\mathbf{B})+(\mathbf{B}\cdot\mathbf{d})^{2}.\end{split}$ (8) The gap closing points are solutions of the equation $z_{1}=2\sqrt{z_{2}}$, which is equivalent to the two equations being simultaneously satisfied: $\begin{split}&\mathbf{d}\cdot\mathbf{d}+\epsilon^{2}_{\mathbf{k}}=\mathbf{B}\cdot\mathbf{B},\\\ &(\mathbf{B}\cdot\mathbf{B})(\mathbf{d}\cdot\mathbf{d})=(\mathbf{B}\cdot\mathbf{d})^{2}.\end{split}$ (9) Eqs. 9 simplify if we consider the magnetic field aligned with one of the axes. Let us then take the magnetic field aligned with the $y$ direction, $\mathbf{B}=(0,B_{y},0)$. In this case, the second equation simplifies to $\sin{k_{y}}=0$ which implies the bulk gap will close at $k_{y,0}=n\pi,n\in\mathbb{Z}$, provided there are values of $k_{x}$ that satisfy the equations $d^{2}\sin^{2}{k_{x}}+(-2t(\cos{k_{x}}\pm 1)-\mu)^{2}=B^{2}_{y}.$ (10) When the $p$-wave superconductor is in a gapless phase, and for a certain range of magnetic field, Majorana flat bands (MFBs) will appear in the system. This will be discussed next. When finite spin-orbit $\alpha$ and $s$-wave pairing $\Delta_{s}$ terms are also considered, the flat bands will (for certain values of the magnetic field) acquire a slope, giving origin to unidirectional Majorana edge states (MESs). The appearance of such states is only possible with a gapless bulk, where a counter-propagating bulk current is created to cancel the edge current [8]. #### II.2.1 Flat bands: winding number and Berry phase quantization When the system is subject to an applied magnetic field, it no longer respects time-reversal symmetry. If the applied field has a generic form $\mathbf{B}=(B_{x},B_{y},0)$ we can, however, take $k_{x}$ as a fixed parameter of the Hamiltonian and find a set of symmetries that are only satisfied in the $y$ direction. It is found that the Hamiltonian respects the symmetries: $\begin{split}\mathcal{T}_{k_{y}}^{-1}\mathcal{H}(k_{x},k_{y})\mathcal{T}_{k_{y}}=\mathcal{H}(k_{x},-k_{y}),\\\ \mathcal{P}_{k_{y}}^{-1}\mathcal{H}(k_{x},k_{y})\mathcal{P}_{k_{y}}=-\mathcal{H}(k_{x},-k_{y}),\\\ \end{split}$ (11) where $\mathcal{T}_{k_{y}}=(\sigma_{z}\otimes\sigma_{z})K$ and $\mathcal{P}_{k_{y}}=(\sigma_{y}\otimes\sigma_{y})K$ are, respectively, defined as a ”time-reversal-like” symmetry and a ”particle-hole-like” symmetry [8] with $\mathcal{T}_{k_{y}}^{2}=\mathcal{P}_{k_{y}}^{2}=1$ ($K$ is the complex conjugate operator). From these we can define a third chiral-like symmetry $\mathcal{S}_{k_{y}}=\mathcal{T}_{k_{y}}\mathcal{P}_{k_{y}}$: $\begin{split}\mathcal{S}_{k_{y}}^{-1}\mathcal{H}(k_{x},k_{y})\mathcal{S}_{k_{y}}=-\mathcal{H}(k_{x},k_{y}).\\\ \end{split}$ (12) Since we have that $\mathcal{T}_{k_{y}}^{2}=\mathcal{P}_{k_{y}}^{2}=1$, the Hamiltonian belongs to the BDI symmetry class and, since the problem is effectively reduced to one dimension, the system can be characterized by an integer topological invariant. We can then write the Hamiltonian in the basis where $\mathcal{S}_{k_{y}}$ is diagonal, in which the Hamiltonian takes an anti-diagonal form. From here it is possible to obtain a winding number $\mathcal{W}$ at each value of $k_{x}$. It can be shown [8] that the winding number is calculated as $\mathcal{W}(k_{x})=\frac{i}{\pi}\left[\log{\left(\frac{sgn(\mathcal{M}(k_{y}=0))}{sgn(\mathcal{M}(k_{y}=\pi))}\right)}\right]$ (13) with $\begin{split}&\mathcal{M}\left(k_{x},k_{y}\right)=\\\ &\left[\mu+2t\left(\cos k_{x}+\cos k_{y}\right)\right]^{2}+d^{2}\sin^{2}k_{x}-B_{y}^{2}+B_{x}^{2}.\end{split}$ (14) In the regimes with $|\mathcal{W}|=1$ the system has a topological nature and Majorana flat bands appear, as is shown in Fig. 4. These are protected by the chiral symmetry $\mathcal{S}_{k_{y}}$ as defined in Eq. 12. Figure 4: Energy spectrum, absolute value of the winding number $\mathcal{W}$ and Berry phase $\gamma$ normalized by $2\pi$, as a function of $k_{x}/\pi$. The values of the parameters are $t=1$, $d=1/6$, $\mu=-3.5$ and a) $B_{y}=d$, b) $B_{y}=3.5d$. The existence of topological flat bands may also be identified by a non- trivial Berry phase. In general, the Berry phase can take any real value. In the presence of certain symmetry constraints, the Berry phase can become quantized to 0 or $\pi$ and carry topological information (at the value of $\pi$). This quantization can happen in the presence of inversion or chiral symmetries, also leading to the quantization of polarization [104]. As the problem is reduced to one dimension, we can obtain a Berry phase $\gamma_{B}$ at each value of $k_{x}$, given by: $\gamma_{B}(k_{x})=i\int_{0}^{2\pi}dk_{y}\langle\Psi(k_{x},k_{y})\mid\frac{\partial}{\partial k_{y}}\Psi(k_{x},k_{y})\rangle$ (15) with $\Psi$ the ground-state wavefunction. The calculation is done numerically by discretizing the Brillouin zone [104, 105, 106, 107] in the $y$ direction. As is shown in Fig. 4, we have found that in the regimes with $|\mathcal{W}|=1$, the Berry phase is also quantized to a value of $\pi$. #### II.2.2 Domain of flat band existence: topological and gapless regions From Eq. 13 it is found that $|\mathcal{W}|=1$ in the regimes where $\mathcal{M}\left(k_{x},k_{y}=0\right)$ and $\mathcal{M}\left(k_{x},k_{y}=\pi\right)$ have opposite signs. This is only possible if $|B_{y}|>|B_{x}|$, thus this is a necessary condition for the appearance of MFBs. The flat band regions can be summarized in (with $\tilde{B}^{2}=B_{y}^{2}-B_{x}^{2}$): * • $(1)$ $\mu\geq 2t$ $\mathcal{D}_{+}>\tilde{B}^{2}>\mathcal{D}_{-}$ (16) * • $(2)$ $\mu\leq-2t$ $\mathcal{D}_{-}>\tilde{B}^{2}>\mathcal{D}_{+}$ (17) * • $(3)$ $-2t<\mu<2t$ $(\mathcal{D}_{+}>\tilde{B}^{2}>\mathcal{D}_{-})\vee(\mathcal{D}_{-}>\tilde{B}^{2}>\mathcal{D}_{+})$ (18) where $\mathcal{D}_{\pm}=\left[\mu+2t\left(\cos k_{x}\pm 1\right)\right]^{2}+d^{2}\sin^{2}k_{x}.$ (19) Eqs. 16, 17 and 18 define the regions where the superconductor is in a nontrivial regime with $|\mathcal{W}|=1$, for a certain value of $k_{x}$. Furthermore, since MFBs can only appear in a gapless phase, the equations also define the regions where the bulk is gapless, as a function of the in-plane magnetic field. Figure 5: a) Domain of existence of Majorana flat bands (shaded region) for $B_{y}$ vs. $k_{x}$ for the parameters $t=1$, $d=1/6$, $\mu=-3.5$. b) Closeup of a) in the region $B_{y}\in$ [$-1,1$] and $k_{x}\in$ [$-1.5,1.5$]. Note that the chiral-like symmetry that protects the flat bands is broken by either a non-zero $s$-wave pairing term $\Delta_{s}$ or a non-zero spin-orbit term $\alpha$. A finite perpendicular magnetic field $B_{z}$ is also found to break the chiral-like symmetry, leading to the absence of flat bands. If the flat band includes the point $k_{x}=0$, the addition of a finite $B_{z}$ will lead to the appearance of bands with a finite slope that cross at zero energy at $k_{x}=0$. Otherwise, the bands will be lifted to finite energy. Figure 6: Phase diagrams indexed by the Chern number $C$ for a system with 20x20 sites, for several values of disorder strength $\lambda$ and perpendicular magnetic field $B_{z}$, obtained for an average over 10 disorder configurations. For Aubry-André disorder, each random disorder configuration is obtained by selecting a random value of $\phi$. The first row with panels a)-c) concerns the case of Anderson disorder (2d), the second row with panels d)-f) concerns the case of Anderson disorder (1d along $y$, uniform along $x$), the third row with panels g)-i) concerns Aubry-André disorder (1d along $y$, uniform along $x$), and the fourth row with panels j)-l) concerns Aubry- André disorder (2d). The values of the parameters are $t=1$, $d=0.6$ and $\mu=0$ (left), $d=0.6$ and $\mu=1$ (middle), $d=1/6$ and $\mu=3d-4t=-3.5$ (right). ## III Disordered model under a perpendicular magnetic field We first want to investigate the effects of quasidisorder and disorder on the system subject to an applied magnetic field in the perpendicular direction, $\mathbf{B}=(0,0,B_{z})$. Here we limit ourselves to the study of the system in real space and, to classify the topological nature of the system, the Chern number is obtained numerically [108]. We consider four different types of disorder potentials: 1. 1. Anderson disorder (2d), where the disorder term is random at each site and varies with uniform probability within an interval: $\Lambda(x,y)\in[-\lambda,\lambda].$ (20) 2. 2. Anderson disorder (1d along $y$, uniform along $x$), where the potential is of the same type as described above but varies only along the $y$ direction, being uniform along the $x$ direction: $\Lambda(x,y)=\Lambda(y)\in[-\lambda,\lambda].$ (21) 3. 3. Aubry-André disorder (1d along $y$, uniform along $x$), where the disorder term is a quasiperiodic potential of the form: $\Lambda(x,y)=\Lambda(y)=\lambda\cos(2\pi\beta f(x,y)+\phi)$ (22) with $f(x,y)$ a function of the lattice sites, $\beta=\frac{\sqrt{5}-1}{2}$ the inverse golden ratio, and $\phi$ a phase between $0$ and $2\pi$. Here we take $f(x,y)=y$, so that the considered quasiperiodic potential is uniform in the $x$ direction. 4. 4. Aubry-André disorder (2d), where the disorder term is a sum of two quasiperiodic potentials of the form: $\Lambda(x,y)=\lambda\cos(2\pi\beta x+\phi)+\lambda\cos(2\pi\beta y+\phi)$ (23) so that disorder potentials are introduced in both the $x$ and $y$ directions. In Fig. 6 we show the phase diagrams indexed by the Chern number, for three different values of $\mu$ and $d$ (with $t=1$ in all cases) and for a system with size $20\times 20$. When Anderson disorder is introduced in the system (first row), the topological regimes are destroyed as the disorder strength is increased. There is, however, some difference in robustness as a function of the magnetic field. This is noticeable in Figs. 6(a) and (b), where we see that the robustness of the topological phases increases with the increase of $B_{z}$. In Fig. 6(a) and for a small region of magnetic field (for $B_{z}>4$) we observe reentrant topology as disorder is increased, as in Fig. 6(b), for lower values of magnetic field ($B_{z}<1$). The second row of the figure is obtained when disorder is considered with uniformity in the $x$ direction. Unexpectedly, the topological regions are to be less robust if compared with the previous case where Anderson disorder was considered with no modulation. Small traces of induced topology are observed for $B_{z}<1$ in Fig. 6(e) and $B_{z}<0.5$ in Fig. 6(f). For Aubry-André disorder uniform in the $x$ direction (third row) we obtain phase diagrams with well defined boundaries, and with induced topological regions. Here, the topological phases show an interesting and unexpected response to the increase of quasidisorder. There is a clear difference in robustness for different values of $B_{z}$, which originates the seemingly effect of “peaks” and “valleys” in the phase diagram, respectively at more robust and more vulnerable values of $B_{z}$. Induced topology is visible in panels g)-i), with topological transitions to finite values of $C$ happening at low and high values of the magnetic field with the increase of disorder. The last row of Fig. 6 concerns the case of two-dimensional Aubry-André disorder. The introduction of disorder leads to the appearance of new topological regions, where several are characterized by values of $C$ that are not seen in the clean system, in the range of $[-4,4]$. Also, some regions appear where the Chern number oscillates within an interval between two integer values, without tending clearly to one of them. We may argue that by adding disorder, local fluctuations of $\mu$ may lead to changes of the Chern number. This is particularly seen in the presence of quasidisorder. This suggests that the long-range quasiperiodicity resonates more with the calculation of the Chern number, that reflects the global structure of the states. Further details on the effect of disorder are shown in Appendix A. Figure 7: Average IPR of the whole system as a function of disorder strength, for a) Anderson disorder (2d), b) Anderson disorder (1d along $y$, uniform along $x$), c) Aubry-André disorder (1d along $y$, uniform along $x$), d) Aubry-André disorder (2d). The IPR is averaged over all eigenstates of the system in a given disorder configuration, and averaged over 10 disorder configurations. In a) fits are done to functions of the form $IPR=C_{1}\exp{C_{2}\lambda}$ in the range $\lambda\in[1.5,3]$, giving the values $(C_{1},C_{2})=(5.6\times 10^{-4},1.13)$ for $B_{y}=0.5d$ and $B_{y}=3.5d$, and $(C_{1},C_{2})=(3.6\times 10^{-4},1.22)$ for $B_{y}=4d$, $\alpha=0.2d$, $\Delta_{s}=0.5d$. In b) fits are done to functions of the form $IPR=C_{1}\lambda+C_{2}$ in the range $\lambda\in[0.5,3]$, giving the values $(C_{1},C_{2})=(4.4\times 10^{-3},5\times 10^{-5})$ for $B_{y}=0.5d$ and $B_{y}=3.5d$, and $(C_{1},C_{2})=(2.8\times 10^{-3},1.4\times 10^{-4})$ for $B_{y}=4d$, $\alpha=0.2d$, $\Delta_{s}=0.5d$. ## IV Disordered model under a parallel magnetic field We now introduce disorder on the system with an applied parallel magnetic field in the $y$ direction, $\mathbf{B}=(0,B_{y},0)$. In section A we consider the system in real space, where the disorder term takes the same form as in the previous section (Eqs. 20, 21, 22,23). From section B onwards we study the system in a mixed $(k_{x},y)$ space. In this case the disorder term is either of the form of Eq. 21 or Eq. 22, with the potential varying in the $y$ direction and being the same for all $k_{x}$, as: 1. 1. Anderson disorder: $\Lambda(y)\in[-\lambda,\lambda],$ (24) 2. 2. Aubry-André disorder: $\Lambda(y)=\lambda\cos(2\pi\beta y+\phi),$ (25) where, as before, $\beta=\frac{\sqrt{5}-1}{2}$ is the inverse golden ratio and $\phi$ is a phase between $0$ and $2\pi$. ### IV.1 Localization properties in real space Here we briefly consider the system in real space and study its localization properties in three different regimes. We fix the parameter values as $t=1$, $d=1/6$, $\mu=-3.5$ and consider three different cases: the case of a $p$-wave superconductor for which a magnetic field $B_{y}=0.5d$ is added, such that the system is in a phase with a gapped bulk but gapless edge states; a $p$-wave superconductor with an added magnetic field of $B_{y}=3.5d$, where the system has a gapless bulk and is in the MFB regime; and a case of the noncentrosymmetric superconductor, with $B_{y}=4d$ and added $s$-wave pairing and spin-orbit terms, $\Delta_{s}=0.3d$ and $\alpha=0.2d$, where the system has a gapless bulk and unidirectional MESs. To quantify the effects of disorder on the system’s localization we use the inverse participation ratio, IPR. For a given eigenstate labeled by $m$, the IPR is defined as: $\text{IPR}_{m}=\sum_{i}\left|\psi_{i}^{m}\right|^{4},$ (26) with $\psi_{i}^{m}$ the wavefunction of the eigenstate $m$ at a site $i$. For perfectly localized states we have that $\text{IPR}_{m}\sim 1$ and for delocalized states $\text{IPR}_{m}\sim 1/N$. In Fig. 7 we present results for the average IPR as a function of disorder for a system of size $N=N_{x}\times N_{y}=41\times 41$ [109] and for the same types of disorder as before: Fig. 7(a) Anderson disorder (2d), Fig. 7(b) Anderson disorder (1d along $y$, uniform along $x$), Fig. 7(c) Aubry-André disorder (1d along $y$, uniform along $x$), and Fig. 7(d) Aubry-André disorder (2d). From observation of Figs. 7(a)-(d) we find four qualitatively different behaviours. In Fig. 7(a) (Anderson disorder) we see that the IPR shows an exponential-like behaviour for $\lambda>1.5$. A fit of the form $\text{IPR}=C_{1}\exp{C_{2}\lambda}$ is done in the range $\lambda\in[1.5,3]$, and is presented in Fig. 7(a) in dashed lines. We find that for $\lambda>1.5$ the IPR follows an exponential behaviour closely, while for $\lambda<1.5$ there is a deviation from it. As disorder is increased, the low energy states become increasingly localized inside the bulk. From inspection of the wavefunctions we observed that the edge states quickly lose their structure for low values of disorder, although they do not become as quickly localized as the remaining bulk states. Accordingly, the IPR of these low energy states shows a slower increase than what is observed in Fig. 7(a). Figure 8: Energy spectra evolution with a) Anderson disorder and b) Aubry- André disorder, for $B_{y}=0.5d$. For Anderson disorder along $y$ and with $x$ uniformity, we find a different localization behaviour. The IPR grows linearly with the increase of disorder, although with some fluctuations and a deviation for $\lambda<0.5$. A fit of the form $\text{IPR}=C_{1}\lambda+C_{2}$ is done to the range $\lambda\in[0.5,3]$ and presented in red dashed lines. The change in behaviour in relation to 7(a) is a result of imposing periodicity in the $x$ direction on the disorder term, which unexpectedly causes the behaviour of the IPR to become linear. The IPR of the low energy states follows a similar behaviour to what is seen for the average IPR. By increasing disorder the low energy edge states are removed from the edges and localize inside the bulk, while remaining periodic in the $x$ direction. In Fig. 7(c) (Aubry-André disorder with $x$ uniformity) we see a threshold behaviour where a transition happens around $\lambda=2$. For $\lambda<2$ (approximately) there is a slow increase of the IPR, while for $\lambda>2$ the IPR greatly increases. This resembles some known results: in the one- dimensional Aubry-André model, where the system undergoes an extended- localized transition at $\lambda=2t$, after which the average IPR shows a marked increase; for a one dimensional $p$-wave superconductor with an Aubry- André potential this transition point changes to $\lambda$ = 2(t+d) with $d$ the $p$-wave pairing amplitude (when the chemical potential is taken as zero) [41, 42]. For values of $\lambda$ before the transition, we observed that some bulk states acquire a critical like behaviour in the $y$ direction, while remaining periodic in the $x$ direction. The low energy states are more robust to disorder if compared with the Anderson disorder cases, and are only removed from the system at the transition: after the threshold value of $\lambda$ there are no edge states in the system. Fig. 7(d) concerns the case of two-dimensional Aubry-André disorder. The IPR shows again a threshold behaviour, and as in c) a transition is seen slightly below $\lambda=2$. However, the transition between two different regimes is abrupt in the IPR, and more closely resembles that of the one dimensional Aubry-André chain. Also, unlike cases a)-c), the IPR follows the same behaviour for the three regimes considered. By a closer inspection of the IPR we see that this is only true for values of disorder over $\lambda=0.7$, as for $\lambda<0.7$ the noncentrosymmetric regime shows a consistently lower IPR, as before. As disorder is increased for $\lambda>2$, states localize along both the $x$ and $y$ directions. The comparison between $1d$ Aubry-André disorder and $2d$ Aubry-André disorder, as well as a comparison between a perpendicular and a parallel magnetic field, is detailed in Appendix B, with particular emphasis on the existence of critical states and the apparent absence of a transition between extended and critical states, in contrast to what is found in the one- dimensional case. Although results are not explicitly shown, the effect of edge disorder was also briefly studied, extending previous results obtained for a time-reversal invariant system [110]. We considered both Anderson and Aubry-André disorder potentials which were introduced locally at the edges at $y=0$ and $y=N_{y}$, varying along the $x$ direction (along the edge) also for a system of size $N=41\times 41$. We found that the bulk states and the system as a whole are almost unaffected by edge disorder, and the average IPR of the system remains nearly constant. However, the edge states are affected, and their behaviour depends on the type of disorder introduced. For Anderson disorder, the states localize continuously along the edges, while for Aubry-André disorder there is also a threshold behaviour, similarly to what is presented in Fig. 7(c). Figure 9: Energy spectra evolution with a) Anderson disorder and b) Aubry- André disorder, for $B_{y}=d$. ### IV.2 Energy spectra evolution and density of states We now consider the system in a mixed $(k_{x},y)$ space, with finite width along $y$ and OBC. We fix the parameter values as $t=1$, $d=1/6$, $\mu=-3.5$, and $B_{y}=0.5d$ or $B_{y}=d$ (such that the system describes a $p$-wave superconductor) and obtain the evolution of the energy spectra for several values of disorder strength $\lambda$ for both Anderson and Aubry-André disorder. Since the values of $t$, $d$ and $\mu$ will be kept constant we will now omit them. In Fig. 8 we show the energy spectra for $B_{y}=0.5d$ with a) Anderson disorder and b) Aubry-André disorder. The clean system has gapless edge states and the bulk gap is not closed by $B_{y}$. As Anderson disorder is increased, the edge states lose their structure and the bulk gap is closed. Accordingly, there is an increase in the density of states at $E=0$ and around zero energy as it can be seen in Fig. 10(a). Introducing quasidisorder, as seen in panel b), leads to a closing of the bulk gap with the appearance of new Majorana flat bands. As disorder is increased, the flat band then splits in two and disappears as a gap opens in the system for around $\lambda=1.8$. The appearance of MFBs leads to an increase of the density of states at zero energy, as can be seen in Fig. 10(b) for the value of $\lambda=1.4$. At higher values of disorder, the system is gapped and the DOS at $E=0$ goes to zero. The reopening of the gap contrasts with what was found for Anderson disorder, where the bulk remains gapless as disorder is increased. We observed that the edge states inside the quasidisorder induced flat bands appear localized at both edges simultaneously. While the edge states of the clean system are localized symmetrically on both edges, the flat band states lose this symmetry and localize more near one of the edges if quasidisorder is present. Near the edge on which a given state appears less localized, there is also a deviation from the edge, and the state mostly localizes on the subsequent sites in $y$. Figure 10: Density of states evolution with a) Anderson disorder and b) Aubry-André (AA) disorder, for $B_{y}=0.5d$ and with c) Anderson disorder and d) Aubry-André disorder for $B_{y}=d$. In Fig. 9 the clean system with $B_{y}=d$ is in a gapless phase with both edge states and a range of $k_{x}$ supporting Majorana flat bands. As Anderson disorder is increased, the bulk remains gapless and there is a sharp increase in the density of states at zero energy, as the bulk states come from finite energies to lower energies. The sharp peak in the DOS observed at $E=0$ is reminiscent of the characteristic behaviour of a two-dimensional disordered superconductor with broken time-reversal invariance in the thermal metal regime [111] in which the density of states displays a logarithmic divergence at zero energy. In Fig. 9(b) when Aubry-André disorder is introduced, the edge states appear to be robust up until around $\lambda\approx 0.8$. However, the MFBs which are present at $\lambda=0$ are more robust if compared with the edge states, with the band staying at zero energy but the initial range of $k_{x}$ hosting flat bands decreasing as $\lambda$ increases. Simultaneously, flat bands appear for new values of $k_{x}$, as is can be seen in the figure for $\lambda=1.4$, and accordingly, the density of states at zero energy increases. At higher values of disorder there is a collapse of states to lower energies and the density of states exhibits a peak at $E=0$ which is reminiscent of the behaviour found for Anderson disorder for the same parameter values. Contrary to what is observed in for a lower magnetic field, there is no opening of the bulk gap for larger values of $\lambda$. When quasiperiodic disorder is introduced, a gap will only open for larger values of $\lambda$ if the bulk was gapped prior to introducing disorder, as in Fig. 8, otherwise the bulk will remain gapless. Let us now consider the addition of finite values of $\alpha$ and $\Delta_{s}$. The addition of finite values of spin-orbit coupling and $s$-wave pairing potential breaks the chiral-like symmetry $\mathcal{S}_{k_{y}}$ (defined in Eq. 12) that protects the flat bands. As a result, the latter are lifted to a finite energy and the spectrum acquires a tilt. For certain regimes of $B_{y}$, the noncentrosymmetric superconductor in the clean system shows unidirectional edge states. In such regimes, the addition of Aubry-André disorder leads to the appearance of ”flipped” unidirectional states in the system. This can be seen in Fig. 11 for $\lambda=1.4$. Figure 11: Energy spectra evolution with Aubry-André disorder for $B_{y}=4d$, $\alpha=0.2d$, $\Delta_{s}=0.3d$ and a) $\lambda=0$, b) $\lambda=1.4$. ### IV.3 Topological nature of quasidisorder induced flat bands We want to investigate if the Majorana flat bands that arise in the presence of a quasiperiodic potential have a topological nature, such as is the case of the flat bands in the ordered system. Since the Berry phase was found to be quantized to a value of $\pi$ in the clean system in the region of flat bands, we calculate it here for the disordered case. The Berry phase $\gamma_{B}$ is obtained in real space using twisted boundary conditions. Considering a twisted boundary phase $\theta_{y}$ we have: $\gamma_{B}(k_{x})=i\int_{0}^{2\pi}d\theta_{y}\langle\Psi(k_{x},\theta_{y})\mid\frac{\partial}{\partial\theta_{y}}\Psi(k_{x},\theta_{y})\rangle$ (27) where $\Psi$ denotes the ground-state many body wavefunction, which is given by the Slater determinant of the single particle wavefunctions. We can represent the ground state wavefunction by an $M\times N$ matrix $\mathbf{\Psi}^{\theta_{y}}$ where $N$ is the number of sites in $y$ and $M$ is the number of occupied states (negative energy states). Numerically, the twist variable is discretized into $L$ points between $0$ and $2\pi$, such that $\theta_{y}$ is constrained to take the values $\theta_{y,n}=\frac{2\pi}{L}n$, with $n$ an integer that goes from $0$ to $L-1$. A link variable can then be defined as $U(\theta_{y,n})=\text{det}\left[\mathbf{\Psi^{\dagger}}_{\theta_{y,n}}\mathbf{\Psi}_{\theta_{y,n+1}}\right]$, and the Berry phase is obtained as $\gamma_{B}=-\mathrm{i}\sum_{n=1}^{L}\log{U(\theta_{y,n})}.$ (28) Figure 12: Energy spectrum and Berry phase $\gamma$ normalized by $2\pi$, as a function of $k_{x}$. The values of the parameters are $t=1$, $d=t/6$, $\mu=3d-4t$ and a) $B_{y}=0.5d$, $\lambda=1.4$ b) $B_{y}=d$, $\lambda=1.6$, with $\lambda$ the strength of the quasiperiodic Aubry-André potential. We find that at the values of $k_{x}$ where Majorana flat bands appear the Berry phase is quantized to $\pi$, as shown in Fig. 12. We have also found that, for the considered system sizes, the values of $k_{x}$ at which the Berry phase is quantized to $\pi$ are independent of the phase $\phi$ in the Aubry-André potential. To quantify the induced bands at zero energy and study the transition to a $\pi$-quantized Berry phase, we use the concept of Majorana pair density, defined as [112]: $\rho_{\gamma}=\frac{N_{\gamma}}{N_{k}},$ (29) where $N_{k}$ is the number of discrete points of $k_{x}$ taken inside the interval $\left[-\pi,\pi\right]$, and $N_{\gamma}$ is the number of such points which support MFBs at the edges. Numerically it is more convenient to consider the number of $k_{x}$ points for which the Berry phase is quantized to $\pi$, $N_{\pi}$, since it was found that $N_{\pi}=N_{\gamma}$. A transition from $\rho_{\gamma}=0$ to $\rho_{\gamma}\neq 0$ then signals a transition from a trivial to a topological regime ($\pi$-quantized Berry phase). Fig. 13(a) shows the evolution of $\rho_{\gamma}$ as a function of quas-idisorder strength for the case $t=1$, $d=t/6$, $\mu=-3.5$ and $B_{y}=0.5d$, and for the range $\lambda\in[1,2]$. A transition $\rho_{\gamma}=0\rightarrow\rho_{\gamma}\neq 0$ occurs between $\lambda=1.22$ and $\lambda=1.23$ at a certain critical value $\lambda_{C,1}$. The value of $\rho_{\gamma}$ grows until $1.49\pm 0.01$ when the flat band splits in two and the behaviour of $\rho_{\gamma}$ changes, with an abrupt change in the sign of the second derivative. A second transition occurs between $1.79$ and $1.8$, at a critical value $\lambda_{C,2}$, where $\rho_{\gamma}$ becomes zero. Figure 13: a) Values of $\rho_{\gamma}$ for the case $t=1$, $d=t/6$, $\mu=-3.5$ and $B_{y}=0.5d$ vs. quasidisorder strength $\lambda$. Obtained for a system with $76$ sites in $y$. b) Value of the DOS at $E=0$ for the same parameter values as in a), vs. quasidisorder strength $\lambda$, and the contribution for $\rho(E=0)$ which comes from the Majorana flat bands in the corresponding regime. In Fig. 13(b) the density of states at zero energy $\rho(E=0)$ (normalized by the system size) is shown, for the same parameters as in Fig. 13(a) and for $N_{y}=76$, along with the corresponding contribution for the zero energy density of states which comes from the MFB, $\rho(E=0)_{\gamma}$. Inside the topological phase, which is highlighted, we can see that the finite value of $\rho(E=0)$ observed in the system with OBC comes almost entirely from the presence of flat bands. $N_{y}$ | $\lambda_{C,1}$ | $\lambda_{C,2}$ ---|---|--- 76 | $1.225\pm 0.005$ | $1.800\pm 0.005$ 100 | $1.215\pm 0.005$ | $1.775\pm 0.005$ 175 | $1.230\pm 0.005$ | $1.805\pm 0.005$ 200 | $1.220\pm 0.005$ | $1.800\pm 0.005$ 400 | $1.230\pm 0.005$ | $1.805\pm 0.005$ 800 | $1.225\pm 0.005$ | $1.805\pm 0.005$ Table 1: Values of the critical points $\lambda_{C,1}$ and $\lambda_{C,2}$, for the parameter values $t=1$, $d=t/6$, $\mu=-3.5$ and $B_{y}=0.5d$ and for the system sizes $\\{76,100,175,200,400,800\\}$. Table 1 shows the values of $\lambda_{C,1}$ and $\lambda_{C,2}$ for several system sizes, obtained for random values of the phase $\phi$ in the Aubry- André potential, where the uncertainty is taken as the minimum interval considered between values of $\lambda$. Is is found that the values of the critical points show little variation with the system size, and we also found that the critical points are independent of $\phi$ for the system sizes considered. ### IV.4 Scaling of the density of states: critical exponents #### IV.4.1 A detour to the clean system Let us first briefly consider the clean system, without disorder. For the clean case, it is possible to obtain the values of the dynamical exponent $z$ and of the critical exponent $\nu$ analytically, for the transition that occurs as $B_{y}$ is increased, corresponding to a transition from a winding number of 0 to 1 or a Berry phase of 0 to $\pi$. Here we consider the case of $\mu<-2t$ (such that the topological phase is within the region described by Eq. 17). At the topological transition to a gapless phase, the gap closing points in $k_{x}$, $k_{x,0}$, are given by $k_{x,0}=\pm\arccos{\left[-\frac{2(t\mu+2t^{2})}{-d^{2}+4t^{2}}\right]}+2n\pi,n\in\mathbb{Z}.$ (30) The values of $k_{y}$ for which the gap closes are given by $k_{y,0}=n\pi$, $n\in\mathbb{Z}$ (general solution). In this case the transition happens at $k_{y,0}=2n\pi$, $n\in\mathbb{Z}$. The gap closes at a critical value of the magnetic field, $B_{y_{C}}$, which, fixing $k_{y}=k_{y,0}$, is defined from the value of $k_{x,0}$ as $B^{2}_{y_{C}}=\left[\mu+2t\left(\cos k_{x,0}+1\right)\right]^{2}+d^{2}\sin^{2}k_{x,0}.$ (31) We can now first expand the expressions for the bulk energy around $k_{x,0}$ to find the dependence of the energy on $k_{x}$. We only need to consider the first positive energy band, $E_{+}(k_{x})$. Taking $k_{y}=k_{y,0}$ and expanding around $k_{x}=k_{x,0}$ we find $\begin{split}E_{+}(k_{x})\propto(k_{x}-k_{x,0}),\end{split}$ (32) implying a value of the dynamical exponent $z=1$ for the transition. We can now take $k_{x}=k_{x,0}$ and see how the gap closes as a function of $B_{y}$. We find $\begin{split}E_{+}(k_{x}=k_{x,0})=\left||B_{y_{C}}|-|B_{y}|\right|.\end{split}$ (33) Near a quantum phase transition as a critical point $\lambda_{C}$ is approached, the gap behaves as $\Delta\sim|\lambda-\lambda_{C}|^{z\nu}$, therefore at $k_{x}=k_{x,0}$ the gap vanishes linearly, with an exponent $z\nu=1$. Since $z=1$, this implies $\nu=1$, and $z=1,\quad\nu=1.$ (34) #### IV.4.2 Quasidisorder: numerical calculation of the critical exponents Around a critical point, the density of states $\rho(E)$ follows [113] $\rho(E)=\delta^{(D-z)\nu}f(|E|\delta^{-z\nu}),$ (35) with $D$ the dimension of the system (here $D=2$), $\delta=\frac{|\lambda-\lambda_{C}|}{\lambda_{C}}$ the normalized distance to the critical point $\lambda_{C}$, and $f$ a scaling function. Right at the critical point, when $\delta=0$, the DOS behaves as $\rho(E)\sim|E|^{\frac{D}{z}-1}.$ (36) From the behaviour of the density of states near the phase transition and using Eqs. 35 and 36 it is possible to obtain the values of the critical exponents numerically. Here we study a system with $N_{y}=800$ sites in $y$ and consider the obtained critical values $\lambda_{C,1}=1.225$ and $\lambda_{C,2}=1.805$ (as shown in table 1 for this system size). A fit of the form of Eq. 36 for the density of states at the critical points, done in the interval E $\in[0.005,0.025]$, gives the values of the critical exponents $z=1.27\pm 0.04$ for the first transition and $z=1.23\pm 0.03$ for the second transition. To determine the value of $\nu$ we take values of $\lambda$ inside the topological (gapless) phase, $\lambda>1.225$ and $\lambda<1.805$, and obtain the density of states close to zero energy. For small values of $\delta$ and close to zero energy a collapse of the scaled values of the density of states according to Eq. 35 is expected. Figure 14: Density of states for $E\in[0.005,0.025]$ and several values of $\lambda$ close to the critical values, for a) $\lambda_{C,1}=1.225$ and b) $\lambda_{C,2}=1.805$, scaled according to Eq. 35 for a) $z=1.27$ and $\nu=0.95$ and b) $z=1.23$ and $\nu=1.00$. In Fig. 14 we show the results for the scaled density of states for: a) values close to the first transition at $\lambda_{C,1}=1.225$, and b) values close to the second transition at $\lambda_{C,2}=1.805$. The density of states shows a collapse for a) $z=1.27$ and $\nu=0.95$ and b) $z=1.23$ and $\nu=1.00$. The quantum phase transitions in the disordered regime are therefore in a different universality class than that of the clean case, which was found to behave with $z=\nu=1$. The obtained values also differ significantly from the known results for the Anderson or the Aubry-André transitions in one dimension, the first belonging to an universality class with with $\nu=2$ and $z=2/3$, and the second case with critical exponents $\nu=1$ and $z=2.375$ [114]. Recent results show that for a one dimensional system with $p$-wave superconductivity subject to an Aubry-André potential the quasidisorder driven transitions also deviate from the normal Aubry-André class. For the localized- critical transition line and when the $p$-wave pairing term is finite, the correlation length exponent has been obtained as $\nu=0.997$ and the dynamical exponent as $z=1.373$ in [68], and as $\nu=1.000$, $z=1.388$ in [69]. Note, however, that the referred results are for $D=1$ while we are studying a two dimensional system, and concern systems with no applied magnetic field. Nevertheless, one could say that the aforementioned results make it so that deviations from the known universality classes are also expected for transitions in the system at study. Up to numerical errors, the values of $\nu$ obtained for the disordered driven transitions coincide with that of the Aubry-André transition; nevertheless the value of $z$ deviates from that of the known classes, which suggests these transitions belong to new universality classes. The identified transitions, where Marojana flat bands appear as a result of a quasidisorder induced gap closing, and the subsequent opening of the bulk gap, are found to happen for other values of the imposed parameters. Considering the values of the parameters $\mu$, $t$ and $d$ are such that the topological regions of the superconductor are described by Eq. 17, then as long as $B_{y}<B_{y,C}$ (when the bulk is gapless) with $B_{y,C}$ defined as in Eq. 31) the same type of transitions will take place with the increase of $\lambda$. $\lambda_{C}$ | $z$ | $\nu$ ---|---|--- 1.225 | $1.27\pm 0.04$ | $0.95\pm 0.05$ 1.805 | $1.23\pm 0.03$ | $1.00\pm 0.05$ Table 2: Values of $z$ and $\nu$ obtained numerically for the topological transitions for $N_{y}=800$. ### IV.5 Fractal Analysis Figure 15: Results of $\tau$ vs. $q$, for several values of disorder strength, $\lambda$, for $k_{x}=0.02\pi$ and $k_{x}=0.2\pi$, for a) Anderson disorder and b) Aubry-André disorder. In all cases, the IPR is averaged for the states within the energy range $E\in[0.05,1]$. One of the effects of Anderson transitions is the emergence of multifractality, which is characterized by fluctuations of eigenstates. These fluctuactions are manifested in the generalized inverse participation ratio. For a given eigenstate labeled by $m$, the generalized IPR is defined as: $\text{IPR}(q)_{m}=\sum_{i}\left|\psi_{i}^{m}\right|^{2q},$ (37) where, as before, $\psi_{i}^{m}$ is the wavefunction of the eigenstate $m$ at a site $i$. At criticality, the generalized IPR behaves as [115] $\text{IPR}(q)\sim L^{\tau(q)}$ (38) where $L$ is the system size and the exponent $\tau(q)$ is defined in terms of a generalized dimension $D(q)$ as $\tau(q)=D(q)(q-1)$. In a metallic phase, $D(q)=d$ and for an insulating phase $D(q)=0$. Wavefunction multifractality is characterized by a $q$ dependent value of $D(q)$, whereas the cases of a constant $D(q)$ are single fractals [116]. Here we want to make a simple fractal analysis of the system both for disorder and quasidisorder. We take $k_{x}$ at fixed values, such that system is reduced to an effective one dimension. The IPR as a function of $q$ is calculated and averaged within the energy range $E\in[0.05,1]$. We fix the parameters $t=1$, $d=t/6$, $\mu=3d-4t$ and $B_{y}=0.5d$ and consider both the cases of Aubry-André disorder and Anderson disorder. The following subintervals of $L$ are considered, to which a fit of an equation of the form of Eq. 38 is done: $\begin{split}&L_{1}=\\\ &\\{75,100,150,175,200,255,275,400,475,600,675,800\\},\\\ &L_{2}=\\{150,175,200,255,275,400,475,600,675,800\\},\\\ &L_{3}=\\{200,255,275,400,475,600,675,800\\},\\\ &L_{4}=\\{275,400,475,600,675,800\\}.\end{split}$ (39) The obtained results are presented in Figs. 15 and 16. #### IV.5.1 Anderson disorder Fig. 15(a) shows the values of $\tau(q)$ for $k_{x}=0.02\pi$ and $k_{x}=0.2\pi$, for several values of $\lambda$ and considering the system size interval $L_{1}$. One thing that can be immediately noticed is that for the clean system, $\lambda=0$, the values of $\tau(q)$ closely follow the line $\tau(q)=(q-1)$, indicating that $D(q)$ is $q-$independent and equal to $1$. This is the expected behaviour of the clean system (taking a fixed $k_{x}$ where the system is reduced to one dimension) and reveals that the bulk states are extended in the $y$ direction. For higher values of disorder, $\tau(q)$ approaches the line $\tau(q)=0$, where $D(q)=0$, suggesting the states are localized. For other values of disorder strength, starting at $\lambda=0.1$, $\tau(q)$ does not follow a behaviour characteristic either of $D(q)=1$ or $D(q)=0$. In order to take a conclusion, it is necessary to evaluate $\tau(q)$ as the system size tends to infinity. To do this, the subintervals of $L$ in Eq. 39 are considered, to which a fit of equation of the form of Eq. 38 is done. The results are presented in Fig. 16(a). We find that for the values $\lambda=0.1$ and above, as larger values of $L$ are considered, the curves $\tau(q)$ approach $\tau(q)=0$. This confirms a localization of the bulk states in the thermodynamic limit for small values of disorder. Figure 16: a) Values of $\tau$ at different values of $q$ and Anderson disorder strength $\lambda$, for $k_{x}=0.02\pi$ and $k_{x}=0.2\pi$. b) Values of $\tau$ at different values of $q$ and quasidisorder strength $\lambda$, for $k_{x}=0.02\pi$ and $k_{x}=0.2\pi$. In all cases, the IPR is averaged for the states within the energy range $E\in[0.05,1]$. #### IV.5.2 Aubry-André disorder Fig. 16(b) shows the values of $\tau(q)$ for $k_{x}=0.02\pi$ and $k_{x}=0.2\pi$, for several values of quasidisorder strength $\lambda$ for the size interval $L_{1}$. Unlike the previous case with Anderson disorder, we see that the results differ for each $k_{x}$, and that for some values of disorder strength $\tau(q)$ follows the line $q-1$ closely until some value of $q$ where the behaviour suddenly changes. In Fig. 16(b) we show, as before, values of $\tau(q)$ fitted for the considered size intervals $L_{1}$, $L_{2}$, $L_{3}$ and $L_{4}$. or lower values of $q$, $\tau$ remains at the values defined by the Eq. $\tau(q)=D(q)(q-1)$ with $D(q)=1$. However, at higher values of $q$, this behaviour changes. Contrary to the case with Anderson disorder, there is no clear tendency of $\tau(q)$ at increased system sizes, and the behaviour also depends on the value of $q$. This deviation from the $D(q)=1$ line is verified as soon as disorder is introduced, and suggests the system is in a multifractal regime. Accordingly, we see the appearance of critical bulk states in the system. From inspection of Fig. 15(b) and of the corresponding values of $\tau(q)$ at larger system sizes, we identify a transition to a localized phase around $\lambda\in[2.0,2,1]$. ## V Conclusions In this work we studied a two-dimensional topological superconductor in the presence of a magnetic field. We introduced disorder and quasidisorder in the system with the aim of studying the effects on topological and localization properties. Considering previous results on other systems such as insulators, semimetals and one-dimensional superconductors, and the results we found on the effect of quasidisorder in two-dimensional superconductors, we may expect that the results may be general considering gapless systems or topological systems (gapped or gapless), in which regions displaying similar topological and localization properties may be found. The system was first studied under a perpendicular magnetic field $B_{z}$. Four types of disorder were considered: Anderson disorder (two-dimensional), Anderson disorder (one-dimensional along $y$, uniform along $x$), Aubry-André disorder (one-dimensional along $y$, uniform along $x$) and Aubry-André disorder (two-dimensional). We observed that the response of the topological phases of the system differs depending on the type of disorder, and that quasidisorder induces topological phases in new regions of $B_{z}$, characterized by integer values of the Chern number $C$. The critical points at these phase boundaries were shown to become sharper as the system size increases, allowing us to conclude that the obtained phase diagrams apply to bigger system sizes. The real space system was also briefly studied when a parallel magnetic field is applied in the $y$ direction. We studied the cases of bulk disorder, with the same four different spatial configurations. For two-dimensional Anderson disorder, we found that the average IPR of the system increases with an exponential behaviour as a function of $\lambda$ for $\lambda>1.5$. When uniformity in the $x$ direction is imposed in the Anderson disorder potential, we found that the IPR shows a linear increase as a function of $\lambda$, for $\lambda>0.5$. For Aubry-André disorder, the behaviour of the average IPR of the system reveals the existence of two different regimes. In the first, the average IPR shows a slow increase with $\lambda$, and in the second the IPR greatly increases. The transition between the two regimes is located around $\lambda=2$. We then studied the system in a mixed ($k_{x},y$) space with an applied parallel magnetic field. The clean superconducting system is known to possess flat bands in the gapless regime. At the corresponding values of $k_{x}$ these have a winding number $\mathcal{W}$ of 1, which is defined from reducing the two dimensional system to an effective one dimension. We showed that these are also characterized by a $\pi-$quantized Berry phase at the same values of $k_{x}$. We showed that the introduction of quasidisorder induces new gapless phases in parameter regimes where they were absent in the clean case. For the $p$-wave system subject to a parallel magnetic field this leads to new regimes with Majorana flat bands. This is not only true for phases with a gapless bulk but also for gapped phases, where quasidisorder closes the bulk gap and Majorana flat bands appear. We then obtained the Berry phase with twisted boundary conditions and concluded the quasidisorder induced flat bands also have a quantized Berry phase of $\pi$. For the noncentrosymmetric superconductor with added $s$-wave superconducting pairing and Rashba spin orbit coupling, we showed that new regimes with unidirectional Majorana edge states appear. In particular, we showed that for a phase where right-moving unidirectional edge states were present in the system, the introduction of quasidisorder leads to the appearance of edge modes in the opposite moving direction, and for a certain quasidisorder range these modes coexist in the system. Figure 17: Phase diagrams of a system with 20x20 sites indexed by the Chern number $C$, for several values of Aubry-André disorder strength $\lambda$ and perpendicular magnetic field $B_{z}$, obtained for the average over 20 disorder configurations. The values of the parameters are $t=1$, $d=1/6$ and $\mu=4.5$ . Figure 18: Values of the Chern number $C$ vs. one-dimensional Aubry-André disorder strength $\lambda$ for the system sizes $20\times 20$, $30\times 30$ and $41\times 41$ and for a) $t=1$, $\mu=0$, $d=0.6$, $B_{z}=0.4$, b) $\mu=1$, $d=0.6$, $B_{z}=1.1$, c) $\mu=-3.5$, $d=1/6$, $B_{z}=0.3$. The results were averaged over 20 disorder configurations. Figure 19: Values of the Chern number $C$ vs. two-dimensional Aubry-André disorder strength $\lambda$ for the system sizes $20\times 20$, $30\times 30$ and $41\times 41$ and for a) $t=1$, $\mu=0$, $d=0.6$, $B_{z}=0.4$, b) $\mu=1$, $d=0.6$, $B_{z}=1.1$, c) $\mu=-3.5$, $d=1/6$, $B_{z}=0.3$. The results were averaged over 10 disorder configurations. The identification of the quasidisorder induced topologically non-trivial flat bands with a quantized Berry phase of $\pi$ allowed us to study in detail two topological transitions, for the $p$-wave superconductor with a parallel applied magnetic field $B_{y}$. The two critical points were identified and studied by obtaining the density of induced Majorana bound states in relation to $k_{x}$ points. We found that the values of the critical points show almost no variation with the system size for systems bigger than $76$ sites along $y$. The values of the dynamical critical exponents and correlation length critical exponents were obtained as $z=1.27\pm 0.04$ and $\nu=0.95\pm 0.05$ for the first critical point and $z=1.23\pm 0.03$, $\nu=1.00\pm 0.05$ for the second critical point, which puts these transitions in novel universality classes. We then investigated the multifractal nature of the wavefunctions by calculating the values of $\tau(q)$ from the IPR values at several values of disorder, at the same parameter values as those in which the topological transition was studied. From the behaviour as the thermodynamic limit is approached, we concluded that the introduction of quasidisorder induces multifractality in the system. A transition to a localized regime was identified for $\lambda_{AA}\in[2.0,2.1]$. The same analysis was made for the system with Anderson disorder. The behaviour of $\tau(q)$ as the system size tends to infinity suggests that the introduction of Anderson disorder will drive the system to a localized phase (in the thermodynamic limit). We have also shown that the average inverse participation ratio is not very sensitive to the magnitude of $B_{z}$ or $B_{y}$. Although $B_{y}$ leads to gapless behavior and $B_{z}$ in general leads to gapped behavior, and although each magnetic field direction leads to different topological properties and symmetries in the clean system, the localization properties are similar and the existence of critical states is also similar. It seems that a magnetic field in the $y$ direction, $B_{y}$, leads to a more localized behavior in the presence of a quasidisordered potential. For both magnetic field directions we found critical states and no mobility edges were found. We found a crossover as a function of $\lambda$ with a mixture of extended and critical states that grow in number as quasidisorder increases. In this context, Aubry-André along $1d$ in the two dimensional system or along $2d$ does not lead to qualitatively different results (in the sense that the crossover in localization is seen for both cases), besides the differences in induced topology. ###### Acknowledgements. We acknowledge partial support from FCT through the Grant UID/CTM/04540/2019. M.F.M. acknowledges partial support through the grant (1801P.01102.1.01) QMSP2021 - CEFEMA - IST-ID. ## Appendix A Additional details on the disordered model under a perpendicular magnetic field In Figs. 6(h) and 6(i), we observe that for low magnetic fields the increase of quasidisorder induces topological phases in regions for which the Chern number was zero. In the clean system, however, these regions correspond to a phase that is topological and characterized by a finite value of $I(k_{y})$. In Fig. 17 we show a phase diagram for the parameters $d=1/6$ and $\mu=4.5$. In this case, the topological invariant $I(k_{y}=0,\pi)$ is trivial for low values of magnetic field ($B_{z}<0.5$) when $C=0$. Contrarily to what is observed in Fig. 6(i), there is no induced topological region with $C=-1$. This thus suggests that the reentrant regions observed in Fig. 6(h) and Fig. 6(i) can possibly be related with the topological nature of the phases characterized by $I(k_{y})$ and with $C=0$. To see how the different critical values for quasiperiodic disorder scale with the system size, three transitions for the phase diagrams obtained with Aubry- André disorder along $y$ with uniformity along $x$ (third row in Fig. 6) and Aubry-André in two dimensions (fourth row in Fig. 6) were considered, at fixed values of $B_{z}$ and $\mu$. The results are presented in Figs. 18 and 19 for the system sizes $20\times 20$, $30\times 30$ and $41\times 41$. We found that within the considered system size range, the transitions become sharper as the size increases, thus suggesting that the obtained phase diagrams in Fig. 6 should apply to larger systems. Figure 20: Average participation ratio, APR, for a $p$-wave superconductor in the presence of Aubry-André quasidisorder for a) one-dimensional system and b) two-dimensional system, as a function of the superconducting pairing term $d$ and disorder strength $\lambda$. Figure 21: Average participation ratio, APR, for a $2d$ system with Anderson or Aubry-André (AA) quasidisorder, for $d=t/6,\mu=-3.5$, in a a) perpendicular or b) parallel magnetic field. ## Appendix B Participation ratio of $2d$ Aubry-André quasidisorder In Fig. 7 we considered the average inverse participation ratio regarding both disorder or quasidisorder along one spatial direction (the $y$ axis) and disorder or quasidisorder in the plane, comparing a set of values of parallel magnetic field. In this Appendix we carry out a more extensive analysis. We want to focus our attention on the regime of increasing disorder strength, from small values to larger values as localization takes place, in particular on the possible separation between extended, critical and localized regimes. The inverse participation ratio is particularly useful to study the transition to the localized regime, but is not as revealing in the extended-critical regimes. In this Appendix we will consider instead the participation ratio, which is given by the inverse of Eq. 26. It is of the order of one for extended states, and becomes of the order of the inverse of the system size in the localized regime. Figure 22: Average participation ratio as a function of $\lambda$ for perpendicular magnetic field, $B_{z}$, and a) $1d$ and b) $2d$ quasidisorder and parallel magnetic field, $B_{y}$, and c) $1d$ and $2d$ quasidisorder. Figure 23: Participation ratios of the eigenstates for $\lambda=1,d=t/6,\mu=-3.5$, as a function of energy for a) $B_{z}=2$ and b) $B_{y}=2$. In the case of a one-dimensional $p$-wave superconductor (Kitaev model) in the presence of an Aubry-André potential and with no magnetic field, it is known that the three regimes of extended, critical and localized states are present, as one changes the amplitude of the pairing, $d$, and the quasidisorder strength, $\lambda$ [42]. The average participation ratio (APR), considering a single disorder configuration, is shown in Fig. 20(a), where clear transitions are shown separating the various regimes (here we only consider a disorder configuration, but the results are characteristic of a larger set of disorder configurations). The average participation ratio has different plateaus as the parameters change. Considering a two-dimensional $p$-wave superconductor with two-dimensional quasidisorder, there is no clear transition between the regimes and one finds crossovers as $\lambda$ increases. The system sizes considered in Fig. 20(b) are small ($21\times 21$), but a smooth transition showing the decreasing of the average participation ratio seems to indicate no clear separation of a critical regime before the localized phase takes place. In Fig. 21 we consider the average participation ratio for the two-dimensional case in the presence of a perpendicular or parallel magnetic field, for the cases of Anderson disorder and quasidisorder. These results highlight the extended/critical regimes at lower values of $\lambda$. Anderson disorder behaves similarly for the two magnetic field directions, and quasidisorder leads to higher values of the average participation ratio as disorder increases, with a sharper transition to the localized regime, as shown in Fig. 7. In the case of the parallel magnetic field quasidisorder has a stronger localization effect. A difference with respect to the perpendicular magnetic field is the existence of gapless states, more sensitive to the long-range disorder associated with the Aubry-André potential. In the case of perpendicular magnetic field, the system remains gapped (except at the transitions between the various topological regimes) and therefore is expected to be less sensitive to the quasidisorder potential. The crossover behavior is clearly seen in Fig. 22, independently of the magnetic field direction ($B_{z}$ or $B_{y}$). Also, the consideration of quasidisorder along the $y$ direction and periodic along $x$ or quasidisorder that is fully two-dimensional leads to similar results. Some differences are visible for small magnetic fields or small values of $\lambda$. Except for these regions, the average participation ratio is quite independent of the amplitude of the magnetic field, but the effect of a parallel magnetic field is more significant, as discussed. Figure 24: Wave function amplitudes for the parameters of Fig. 23, as a function of space location in a $21\times 21$ system, for states with a) $B_{z}=2$ and participation ratio $PR=0.781$, b) $B_{z}=2$ and $PR=0.186$, c) $B_{y}=2$ and $PR=0.643$, d) $B_{y}=2$ and $PR=0.169$. In order to have a better understanding of the possible existence of critical states in the regime prior to the transition to localization, we show a typical case in Fig. 23 of the participation ratios of the various eigenstates, as a function of their energies, for perpendicular and parallel magnetic fields. The parameters are chosen so that we are in an intermediate regime, where the average participation ratio is in the crossover between fully extended and localized states. There are significant fluctuations between states with high participation ratios (characteristic of extended states) and low participation ratios (characteristic of intermediate, critical, states) but still larger than values that correspond to the localized regime. The results do not show a mobility edge, and the states mix throughout the energy range. Also, as $\lambda$ increases, we have found that the percentage of critical-like states increases, explaining the crossover behavior. The extended versus critical character nature of the states is shown in Fig. 24, where a few wave functions are shown (for the system with periodic boundary conditions in both directions), characteristic of extended and critical states coexisting in the same energy spectrum. ## Appendix C Energy spectra evolution for the noncentrosymmetric superconductor with Aubry-André and Anderson disorder in $(k_{x},y)$ space Figure 25: Energy spectra evolution with a) Anderson disorder and b) Aubry- André disorder, for $B_{y}=d$, $\alpha=0.2d$ and $\Delta_{s}=0.5d$. Here we present results for the evolution of the energy spectra for the noncentrosymmetric superconductor, with spin orbit coupling $\alpha$ and mixed $p$ and $s$-wave pairings, subject to Anderson and Aubry-André disorder in the $(k_{x},y)$ space. We take the same parameter values for $t$, $d$ and $\mu$ as in section IV.2, and consider two different cases: $B_{y}=d$, $\alpha=0.2d$, $\Delta_{s}=0.5d$ and $B_{y}=4d$, $\alpha=0.2d$ and $\Delta_{s}=0.3d$. In Fig. 25 we present the evolution of the energy spectrum for a) Anderson disorder and b) Aubry-André disorder, for $B_{y}=d$, $\alpha=0.2d$, $\Delta_{s}=0.5d$. As Anderson disorder is introduced in the system, the edge states are destroyed and the considered energy range gets filled with bulk states, but the tilt of the spectrum is preserved. As a result, the flat bands which were previously lifted due to the introduction of finite values of $\alpha$ and $\Delta_{s}$ do not collapse to zero energy. For high values of $\lambda$ the density of states exhibits two peaks which result from the inclination of the bulk energy spectrum (Fig. 27 a) ). Figure 26: Energy spectra evolution with a) Anderson disorder and b) Aubry- André disorder, for $B_{y}=4d$, $\alpha=0.2d$, and $\Delta_{s}=0.3d$. In Fig. 25(b), as Aubry-André disorder is introduced, we see an evolution that is similar to Fig. 9(b), but instead of new flat band regimes, new unidirectional edge states appear. Unlike what happens for Anderson disorder, at high values of $\lambda$ a gap opens for values of $k_{x}$ around $k_{x}=0$ (although the bulk as a whole remains gapless). This is reflected in the density of states, that drops around $E=0$ for higher disorder values. Similarly to what was observed for Anderson disorder, the tilt of the energy spectrum is preserved as disorder increases. Figure 27: Density of states evolution with a) Anderson disorder and b) Aubry-André (AA) disorder, for $B_{y}=d$, $\alpha=0.2d$ and $\Delta_{s}=0.5d$ and with c) Anderson disorder and d) Aubry-André disorder for $B_{y}=4d$, $\alpha=0.2d$, and $\Delta_{s}=0.3d$. In Fig. 26, the clean system is in the regime where unidirectional MESs appear. The values of the $p$-wave pairing and spin orbit term are kept constant in relation to the case of Fig. 25, but the $s$-wave pairing term is decreased from $\Delta_{s}=0.5d$ to $\Delta_{s}=0.3d$ and the magnetic field is increased from $B_{y}=d$ to $B_{y}=4d$. The spectrum acquires a tilt in the opposite direction if compared to the clean system in Fig. 25, as a result of the increased magnetic field. With Anderson disorder, Fig. 26(a), The unidirectional Majorana edge states are robust to small values of disorder strength but as disorder increases the structure of the band is lost, as bulk states fill the lower energy values. This differs from Fig. 25(a) where the tilt of the spectrum is preserved even at higher values of disorder. As disorder is increased, there is at first an increase in the value of the DOS at zero energy, which then decreases for higher values of disorder. For $\lambda>1.8$ the density of states becomes nearly constant in the considered range of $E\in[-0.4,0.4]$. In Fig. 26(b), when a certain value of Aubry-André disorder is reached, ”flipped” unidirectional states appear in the system. This is seen clearly in Fig. 26(b) for the values of $\lambda=1.2$ and $\lambda=1.4$, as a band with negative slope appears for values of $k_{x}$ around $k_{x}=0$. At $\lambda=1.2$ there is a coexistence of unidirectional ”flipped” left-moving edge modes (with negative slope) around $k_{x}=0$ and right-moving edge modes (with positive slope) for higher (absolute) values of $k_{x}$. A backflow current that balances the current on the edges is created on the bulk: extra right or left moving modes will appear depending on the net current on the edges. ## References * [1] C. W. J. Beenakker, Annu. Rev. Condens. Matter Phys. 4, 113 (2013). * [2] M. Z. Hasan and C. L. Kane, Rev. Mod. Phys. 82, 3045 (2010). * [3] X.-L. Qi and S.-C. Zhang, Rev. Mod. Phys. 83, 1057 (2011). * [4] J. Alicea, Rep. Prog. Phys. 75, 076501 (2012). * [5] M. Sato and Y. Ando, Rep. Prog. Phys. 80, 076501 (2017). * [6] Y. Tanaka, Y. Mizuno, T. Yokoyama, K. Yada, and M. Sato, Phys. Rev. Lett. 105, 097002 (2010). * [7] M. Sato, Y. Tanaka, K. Yada, and T. Yokoyama, Phys. Rev. B 83, 224511 (2011). * [8] C. L. M. Wong, J. Liu, K. T. Law, and P. A. Lee, Phys. Rev. B 88, 060504 (2013). * [9] A. P. Schnyder and S. Ryu, Phys. Rev. B 84, 060504 (2011). * [10] A. P. Schnyder, P. M. R. Brydon, and C. Timm, Phys. Rev. B 85, 024522 (2012). * [11] G. E. Volovik, Jetp Lett. 107, 516 (2018). * [12] S. Peotta and P. Törmaä, Nature Communications 6, 8944 (2015). * [13] S. Bose, A. M. García-García, M. M. Ugeda, J. D. Urbina, C. H. Michaelis, I. Bruhega, and K. Kern, LABEL:https://www.nature.com/articles/nmat2768Nat. Materials 9, 550 (2010). * [14] M. A. N. Araújo, A. M. García-García, and P. D. Sacramento, Phys. Rev. B 84, 172502 (2011). * [15] I. S. Burmistrov, I. V. Gornyi, and A. D. Mirlin, Phys. Rev. Lett. 108, 017002 (2012). * [16] J. Mayoh and A. M. García-García, Phys. Rev. B 90, 134513 (2014). * [17] B. Fan and A. M. García-García, Phys. Rev. B 102, 184507 (2020). * [18] C. Carbillet, V. Cherkez, M. A. Skvortsov, M. V. Feigelḿan, F. Debontridder, L. B. Ioffe, V. S. Stolyarov, K. Ilin, M. Siegel, D. Roditchev, T. Cren, and C. Brun, Phys. Rev. B 102, 024504 (2020). * [19] A. Sĺebarski, M. Fijalkowski, P. Zajdel, M.M. Maśka, J. Deniszczyk, M. Zubko, O. Pavlosiuk, K. Sasmal, and M. B. Maple, Phys. Rev. B 102, 054514 (2020). * [20] A. Sĺebarski, M. Fijalkowski, M. M. Maśka, J. Deniszczyk, P. Zajdel, B. Trump, and A. Yakovenko, Phys. Rev. B 103, 155133 (2021). * [21] M. V. Feigelḿan, L. B. Ioffe, V. E. Kravtsov, and E. A. Yuzbashyan, Phys. Rev. Lett. 98, 027001 (2007). * [22] M. V. Feigelḿan, L. B. Ioffe, V. E. Kravtsov, and E. Cuevas, Ann. Phys. 325, 1390 (2010). * [23] J. Mayoh, and A. M. García-García, Phys. Rev. B 92, 174526 (2015). * [24] K. Zhao, H. Lin, X. Xiao, W. Huang, W. Yao, M. Yan, Y. Xing, Q. Zhang, Z.-X. Li, S. Hoshino, J. Wang, S. Zhou, L. Gu, M. S. Bahramy, H. Yao, N. Nagaosa, Q.-K. Xue, K. T. Law, X. Chen, and S.-H. Ji, Nat. Phys. 15, 904 (2019). * [25] C. Rubio-Verdú, A. M. García-García, H. Ryu, D.-J. Choi, J. Zaldívar, S. Tang, B. Fan, Z.-X. Shen, S.-K. Mo, J. I. Pascual, and M. M. Ugeda, Nano Letters 20, 5111 (2020). * [26] B. Fan and A. M. García-García, Phys. Rev. B 101, 104509 (2020). * [27] Z. Fan, G.-W. Chern, and S.-Z. Lin, Phys. Rev. Research 3, 023195 (2021). * [28] M. Stosiek, F. Evers, and I. S. Burmistrov, Phys. Rev. Research 3, L042016 (2021). * [29] X. Zhang and M. Foster, Phys. Rev. B 106, L180503 (2022). * [30] K.-E. Huhtinen, J. Herzog-Arbeitman, A. Chew, B. A. Bernevig, and P. Törmä, Phys. Rev. B 106, 014518 (2022). * [31] Y. Hu, S. M. L. Teicher, B. R. Ortiz, Y. Luo, S. Peng, L. Huai, J. Ma, N. C. Plumb, S. D. Wilson, J. He, and M. Shi, Science Bulletin 67, 495 (2022). * [32] K. Jiang, T. Wu, J.-X. Yin, Z. Wang, M. Z. Hasan, S. D. Wilson, X. Chen, and J. Hu, National Science Review nwac199 (2022). * [33] X. Wu, T. Schwemmer, T. Müller, A. Consiglio, G. Sangiovanni, D. Di Sante, Y. Iqbal, W. Hanke, A. P. Schnyder, M. M. Denner, M. H. Fischer, T. Neupert, and R. Thomale, Phys. Rev. Lett. 127, 177001 (2021). * [34] B. R. Ortiz, S. M. L. Teicher, Y. Hu, J. L. Zuo, P. M. Sarte, E. C. Schueller, A. M. M. Abeykoon, M. J. Krogstad, S. Rosenkranz, R. Osborn, R. Seshadri, L. Balents, J. He, and S. D. Wilson, Phys. Rev. Lett. 125, 247002 (2020). * [35] C. Mielke, D. Das, J.-X. Yin, H. Liu, R. Gupta, Y.-X. Jiang, M. Medarde, X. Wu, H. C. Lei, J. Chang, P. Dai, Q. Si, H. Miao, R. Thomale, T. Neupert, Y. Shi, R. Khasanov, M. Z. Hasan, H. Luetkens, and Z. Guguchia, Nature 602, 245 (2022). * [36] J. Sutradhar, S. Mukerjee, R. Pandit, and S. Banerjee, Phys. Rev. B 99, 224204 (2019). * [37] M. Saha, B. P. Venkatesh, and B. K. Agarwalla, Phys. Rev. B 105, 224204 (2022). * [38] L.-J. Lang, X. Cai, and S. Chen, Phys. Rev. Lett. 108, 220401 (2012). * [39] M. Tezuka and N. Kawakami, Phys. Rev. B 85, 140508 (2012). * [40] W. DeGottardi, D. Sen, and S. Vishveshwara, Phys. Rev. Lett. 110, 146404 (2013). * [41] X. Cai, L.-J. Lang, S. Chen, and Y. Wang, Phys. Rev. Lett. 110, 176403 (2013). * [42] J. Wang, X.-J. Liu, G. Xianlong, and H. Hu, Phys. Rev. B 93, 104504 (2016). * [43] X. Cai, J. Phys.: Condens. Matter 26, 155701 (2014). * [44] Y. E. Kraus, Y. Lahini, Z. Ringel, M. Verbin, and O. Zilberberg, Phys. Rev. Lett. 109, 106402 (2012). * [45] S. Nakajima, N. Takei, K. Sakuma, Y. Kuno, P. Marra, and Y. Takahashi, Nat. Phys. 17, 844 (2021). * [46] S. Longhi, Opt. Lett. 45, 4036 (2020). * [47] O. Zilberberg, Opt. Mater. Express 11, 1143 (2021). * [48] T. Liu and H. Guo, Phys. Letters A 382, 3287 (2018). * [49] M. I. N. Rosa, Y. Guo, and M. Ruzzene, Appl. Phys. Lett. 118, 131901 (2021) * [50] M. Verbin, O. Zilberberg, Y. E. Kraus, Y. Lahini, and Y. Silberberg, Phys. Rev. Lett. 110, 076403 (2013). * [51] Y. E. Kraus and O. Zilberberg, Phys. Rev. Lett. 109, 116404 (2012). * [52] J. H. Pixley, J. H. Wilson, D. A. Huse, and S. Gopalakrishnan, Phys. Rev. Lett. 120, 207604 (2018). * [53] V. Goblot, A. Štrkalj, N. Pernet, J. L. Lado, C. Dorow, A. Lemaître, L. Le Gratiet, A. Harouri, I. Sagnes, S. Ravets, A. Amo, J. Bloch, and O. Zilberberg, Nat. Phys. 16, 832 (2020). * [54] T. Xiao, D. Xie, Z. Dong, T. Chen, W. Yi, and B. Yan, Science Bulletin 66, 2175 (2021). * [55] F. A. An, E. J. Meier, and B. Gadway, Phys. Rev. X 8, 031045 (2018). * [56] G. Roati, C. D’Errico, L. Fallani, M. Fattori, C. Fort, M. Zaccanti, G. Modugno, M. Modugno, and M. Inguscio, Nature 453, 895 (2008). * [57] L. Guidoni, C. Triché, P. Verkerk, and G. Grynberg, Phys. Rev. Lett. 79, 3363 (1997). * [58] H. P. Lüschen, S. Scherg, T. Kohlert, M. Schreiber, P. Bordia, X. Li, S. Das Sarma, and I. Bloch, Phys. Rev. Lett. 120, 160404 (2018). * [59] Y. Lahini, R. Pugatch, F. Pozzi, M. Sorel, R. Morandotti, N. Davidson, and Y. Silberberg, Phys. Rev. Lett. 103, 013901 (2009). * [60] S. Aubry and G. André, Ann. Israel Phys. Soc 3, 18 (1980). * [61] J. M. B. Lopes dos Santos, N. M. R. Peres, and A. H. Castro Neto, Phys. Rev. Lett. 99, 256802 (2007). * [62] R. Bistritzer and A. H. MacDonald, Proc. Natl. Acad. Sci. U.S.A. 108, 12233 (2011). * [63] J. M. B. Lopes dos Santos, N. M. R. Peres, and A. H. Castro Neto, Phys. Rev. B 86, 155449 (2012). * [64] D. Mao and T. Senthil, Phys. Rev. B 103, 115110 (2021). * [65] M. Gonçalves, H. Z. Olyaei, B. Amorim, R. Mondaini, P. Ribeiro, and E. V. Castro, 2D Mater. 9, 011001 (2022). * [66] J. Wang, X.-J. Liu, G. Xianlong, and H. Hu, Phys. Rev. B 93, 104504 (2016). * [67] J. Fraxanet, U. Bhattacharya, T. Grass, D. Rakshit, M. Lewenstein, and A. Dauphin, Phys. Rev. Research 3, 013148 (2021). * [68] X. Tong, Y.-M. Meng, X. Jiang, C. Lee, G. D. de M. Neto, and G. Xianlong, Phys. Rev. B 103, 104202 (2021). * [69] T. Lv, T.-C. Yi, L. Li, G. Sun, and W.-L. You, Phys. Rev. A 105, 013315 (2022). * [70] D. R. Hofstadter, Phys. Rev. B 14, 2239 (1976). * [71] S. Das Sarma, Song He, and X. C. Xie Phys. Rev. Lett. 61, 2144 (1988). * [72] S. Das Sarma, Song He, and X. C. Xie, Phys. Rev. B 41, 5544 (1990). * [73] J. Biddle and S. Das Sarma, Phys. Rev. Lett. 104, 070601 (2010). * [74] S. Ganesham, J. H. Pixley and S. Das Sarma, Phys. Rev. Lett. 114, 146601 (2015). * [75] S. Gopalakrishnan, Phys. Rev. B 96, 054202 (2017). * [76] A. Duthie, S. Roy, and D. E. Logan, Phys. Rev. B 103, L060201 (2021). * [77] T. Liu, X. Xia, S. Longhi, L. S.-Palencia, SciPost Phys. 12, 027 (2022). * [78] S. Ganeshan, K. Sun, and S. Das Sarma, Phys. Rev. Lett. 110, 180403 (2013). * [79] M. Gonçalves, B. Amorim, E. V. Castro and P. Ribeiro, SciPost Phys. 13, 046 (2022). * [80] A. P. Mackenzie and Y. Maeno, Rev. Mod. Phys. 75, 657 (2003). * [81] R. Joynt and L. Taillefer, Rev. Mod. Phys. 74, 235 (2002). * [82] K. Matano, M. Kriener, K. Segawa, Y. Ando, and G. Zheng, Nature Phys 12, 852 (2016). * [83] E. Bauer, G. Hilscher, H. Michor, Ch. Paul, E. W. Scheidt, A. Gribanov, Yu. Seropegin, H. Noël, M. Sigrist, and P. Rogl, Phys. Rev. Lett. 92, 027003 (2004). * [84] I. Sugitani, Y. Okuda, H. Shishido, T. Yamada, A. Thamizhavel, E. Yamamoto, T. D. Matsuda, Y. Haga, T. Takeuchi, R. Settai, and Y. Ōnuki, J. Phys. Soc. Jpn. 75, 043703 (2006). * [85] N. Kimura, K. Ito, K. Saitoh, Y. Umeda, H. Aoki, and T. Terashima, Phys. Rev. Lett. 95, 247004 (2005). * [86] S. Fujimoto, J. Phys. Soc. Jpn. 76, 051008 (2007). * [87] P. A. Frigeri, D. F. Agterberg, A. Koga, and M. Sigrist, Phys. Rev. Lett. 92, 097001 (2004). * [88] M. Sato and S. Fujimoto, Phys. Rev. B 79, 094504 (2009). * [89] A. Daido and Y. Yanase, Phys. Rev. B 95, 134507 (2017). * [90] A. V. Balatsky, I. Vekhter, and J.-X. Zhu, Rev. Mod. Phys. 78, 373 (2006). * [91] I. Reis, D. J. J. Marchand, and M. Franz, Phys. Rev. B 90, 085124 (2014). * [92] P. D. Sacramento, J. Phys.: Condens. Matter 27, 445702 (2015). * [93] K. Pöyhönen, I. Sahlberg, A. Westström, and T. Ojanen, Nat Commun 9, 2103 (2018). * [94] T. C̆adez̆, R. Mondaini, E. V. Castro, and P. D. Sacramento, Acta Phys. Pol. A 135, 1180 (2019). * [95] T. Devakul and D. A. Huse, Phys. Rev. B 96, 214201 (2017). * [96] J. T. Chalker and G. J. Daniell, Phys. Rev. Lett. 61, 593 (1988). * [97] J. T. Chalker, Physica A: Stat. Mech. Appl. 167, 253 (1990). * [98] B. Huang and W. V. Liu, Phys. Rev. B 100, 144202 (2019). * [99] J. H. Pixley, P. Goswami, and S. Das Sarma, Phys. Rev. Lett. 115, 076601 (2015). * [100] Y. Fu, J. H. Wilson and J. H. Pixley, Phys. Rev. B 104, L041106 (2021). * [101] M. Buchhold, S. Diehl, and A. Altland, Phys. Rev. B 98, 205134 (2018). * [102] M. Gonçalves, P. Ribeiro, E. V. Castro and M. A. N. Araújo, Phys. Rev. Lett. 124, 136405 (2020). * [103] Y. Fu, E. J. König, J. H. Wilson, Y.-Z. Chou, and J. H. Pixley, npj Quantum Materials (2020)5:71. * [104] W. A. Benalcazar, B. A. Bernevig, and T. L. Hughes, Phys. Rev. B 96, 245115 (2017). * [105] D. Xiao, M.-C. Chang, and Q. Niu, Rev. Mod. Phys. 82, 1959 (2010). * [106] R. Resta, J. Phys.: Condens. Matter 22, 123201 (2010). * [107] T. Fukui, Y. Hatsugai, and H. Suzuki, J. Phys. Soc. Jpn. 74, 1674 (2005). * [108] Y.-F. Zhang, Y.-Y. Yang, Y. Ju, L. Sheng, R. Shen, D.-N. Sheng, and D.-Y. Xing, Chinese Phys. B 22, 117312 (2013). * [109] Here we limit ourselves to the analysis of a system of size $41\times 41$, although draft results for a system of size $20\times 20$ followed the same behaviour as the one discussed. * [110] R. Queiroz and A. P. Schnyder, Phys. Rev. B 89, 054501 (2014). * [111] A. Mildenberger, F. Evers, A. D. Mirlin, and J. T. Chalker, Phys. Rev. B 75, 245321 (2007). * [112] N. Sedlmayr, J. M. Aguiar-Hualde, and C. Bena, Phys. Rev. B 91, 115415 (2015). * [113] K. Kobayashi, T. Ohtsuki, K.-I. Imura, and I. F. Herbut, Phys. Rev. Lett. 112, 016402 (2014). * [114] J. C. C. Cestari, A. Foerster, M. A. Gusmão, and M. Continentino, Phys. Rev. A 84, 055601 (2011). * [115] F. Wegner, Z Physik B 36, 209 (1980). * [116] F. Evers and A. D. Mirlin, Rev. Mod. Phys. 80, 1355 (2008).
§ INTRODUCTION The Neil Gehrels Swift Observatory (Swift hereafter; [Gehrels et al. , 2004]) was launched on 2004 November 20, with the primary goal of discovering and localising Gamma-Ray Bursts (GRBs) to arcsec accuracy within a few hundred seconds, and then following the X-ray and optical afterglows as they fade. While GRBs are still an important part of Swift's work, over the years the satellite has become the go-to mission for observations of all kinds of transients as they vary in the optical, ultraviolet (UV), X-ray and $\gamma$-ray regimes. In this paper, we focus on novae observed in the UV by the UV/Optical Telescope (UVOT; [Roming et al. , 2005]). Novae are thermonuclear explosions formed in accreting white dwarf (WD) systems. Material is transferred from the secondary, or companion, star onto the surface of the WD primary, until the pressure and temperature are sufficient to trigger a thermonuclear runaway (see [Bode & Evans , 2008, Woudt & Ribeiro , 2014] for review articles), flinging material outwards, and shrouding the WD surface from view. At this point, a new optical source – the nova – is formed, with the peak optical brightness occurring when the photosphere reaches maximum expansion[While most novae are first discovered at optical wavelengths, V959 Mon was initially detected in $\gamma$-rays by the Large Area Telescope onboard Fermi, when too close to the Sun for ground-based telescopes to observe [Cheung et al. , 2012, Cheung et al. , 2012]. Models of nova eruptions also predict there should be a brief, soft X-ray flash after hydrogen ignition, before the optical emission is detectable. This was finally observed for the first time in 2020 by eROSITA (extended Roentgen Survey with an Imaging Telescope Array) for the nova YZ Ret [König et al. , 2022].]. As the ejecta expand, they become optically thin, typically allowing the surface nuclear burning (and hence soft X-rays) to become visible; this is the so-called super-soft source (SSS) phase[It is, however, possible for the nuclear burning to cease before the ejecta fully clear, in which case the soft X-rays may fade away before they can be detected. In the case of V745 Sco [Page et al. , 2015], it was suggested that only the cooling, tail-end of the SSS emission was seen, with the actual hydrogen burning having ended very quickly, placing it close to this `unobservable region'.]. Once the hydrogen shell has been burnt, the system fades back to quiescence. There are two main groups of novae, termed Classical Novae (CNe) and Recurrent Novae (RNe), respectively. CNe – which make up the majority – have only been seen in eruption once. These systems have orbital periods of a few hours, and the secondary star is typically a late-type main sequence (MS) object <cit.>. RNe, on the other hand, have had multiple detected eruptions, with recurrence timescales of up to around 100 yr (though this is, of course, something of a selection effect, dependent on historical records; even the so-called CNe are expected to erupt more than once, with duty cycles of typically thousands of years or longer <cit.>). The orbital periods of RNe cover a wide range, from a few hours to a year or more [Schaefer , 2010, Anupama , 2008], with the longer periods corresponding to systems where the secondary star is an evolved giant; the short period systems are more similar to CNe, with an MS or subgiant companion star. A binary system with a higher-mass WD and higher accretion rate is expected to have a shorter recurrence time [Starrfield , 1989] (though T Pyx is unusual, in that it is thought to have a comparatively low mass WD; <cit.>). Given the shorter timescales involved, RNe accrete, and then eject, less material during each nova cycle <cit.>, which means the SSS phase tends to become visible more rapidly in these systems. The optical/UV light-curves in some RNe have been noted to show a flat, plateau phase, which is speculated to arise from the re-radiation of the SSS emission from an accretion disc [Schaefer , 2010, Hachisu et al. , 2006, Hachisu et al. , 2008]. There are ten confirmed Galactic RNe [Schaefer , 2010], with more known in other galaxies, such as M31 [Shafter et al. , 2015, Darnley , 2021]. Our sample contains some novae (classical and recurrent) which occurred in symbiotic binaries. These are systems where the secondary star is a late-type giant, with outflowing wind encompassing the WD; they are sometimes known as embedded novae, to differentiate them from the so-called `symbiotic novae', which are a different population of slowly-evolving eruptive variable stars [Mürset & Nussbaumer , 1994, Mikołajewska , 2007, Mikołajewska , 2008], and are not considered here. The first nova which Swift monitored in detail was the 2006 eruption of the RN RS Oph [Bode et al. , 2006, Osborne et al. , 2011]. This impressive dataset showed that Swift was well suited for monitoring such sources, and this subsequently led to many more novae being observed in detail with the observatory. Examples within our own Galaxy include V458 Vul [Ness et al. , 2009, Schwarz et al. , 2011], V2491 Cyg [Ibarra et al. , 2009, Page et al. , 2010], HV Cet [Beardmore et al. , 2012], V5668 Sgr [Gehrz et al. , 2018], V745 Sco [Page et al. , 2015], V407 Lup [Aydi et al. , 2018] and V3890 Sgr [Page et al. , 2020]. The best monitored Swift light-curves up until the end of 2017 were collated by [Page et al. , 2020], where both X-ray and UV results were presented. In this paper, we focus in more detail on the evolution of the UV emission in a sample of 12 novae. These were chosen to be the sources best monitored by Swift-UVOT over an extended period of time, in at least one UV filter, where the light-curves follow a series of power-law decays. While this paper considers only novae, this approach to fitting UV light-curves could be extended to other transients, such as supernovae or active galactic nuclei. § OBSERVATIONS The Swift-UVOT is one of two narrow field instruments onboard Swift; the X-ray Telescope (XRT; [Burrows et al. , 2005]) is the other. The UVOT has, among other observing capabilities, six broadband filters: three optical and three UV. Given that there are a large number of ground-based observers who frequently follow novae at optical wavelengths, Swift typically utilises one or more of the UV filters when monitoring novae: $uvw1$ (central wavelength, $\lambda_c$ = 2600 Å; Full Width at Half Maximum, FWMH = 693 Å), $uvm2$ ($\lambda_c$ = 2246 Å; FWHM = 498 Å) and $uvw2$ ($\lambda_c$ = 1928 Å; FWHM = 657 Å). More details on the instrument can be found in [Poole et al. , 2008, Breeveld et al. , 2011]. Novae can peak at very bright magnitudes, some even reaching naked eye visibility – RS Oph, for example, peaks around magnitude 4.5 in the visual band. While this is good news for the typical stargazer, sources brighter than around magnitude 10–11 lead to significant coincidence loss in the UVOT, which is not recovered when using the standard Swift photometric tools. A technique was developed by [Page et al. , 2013] whereby the read-out streak formed by bright sources can be used to obtain measurements up to 2.4 mag brighter than the previous limit, thus extending the useful dynamic range of the instrument. We have utilised this method where relevant in the following analysis. In 2020 September, a more accurate UVOT calibration file was released, to account for the loss of sensitivity over time. The light-curves presented here have been corrected for this degradation[See details at <https://www.swift.ac.uk/analysis/uvot/>.], leading to small differences between these results and those previously published for the older novae (up to $\sim$ 0.02 mag for the sources in this sample). Not only do novae show intrinsic variability as the source itself evolves, novae can be very different from each other, showing a variety of light-curve shapes <cit.> across all wavebands. Some show periodic modulations (e.g., HV Cet; [Beardmore et al. , 2012]), while others brighten and fade apparently randomly, sometimes with the UV in antiphase with the X-ray emission (e.g., V458 Vul; [Ness et al. , 2009, Schwarz et al. , 2011]); see Fig. <ref>. Many, however, do just (mainly) fade in the UV after peak brightness, and, in this paper, we concentrate on these novae. Specifically, we present a sample where detailed UVOT monitoring, starting promptly (within a few days) after the eruption[V959 Mon is the exception, with observations starting some time after the actual nova eruption, because of its location with respect to the Sun at outburst (see footnote 1).], was performed in at least one UV filter, over an extended period of time. Table <ref> lists the names of the novae, together with their eruption dates, orbital periods, interstellar reddening E(B$-$V), and (approximate) distances[Over the years, distance estimates to novae have changed. We use the most recent derivations by [Schaefer , 2022], using parallax data from the third Gaia data release (DR3) among other methods, for all our sample bar V407 Cyg, which was not included in this catalogue.]. Those novae which occurred in symbiotic systems are noted in the sixth column. Examples of light-curves where the UV emission is not simply a series of power-law decays. Left: HV Cet. This nova shows a strong periodic oscillation in both the UV and X-ray bands, as well as an underlying slower modulation. The inset shows a zoom-in of the 1.77-d period, plotted in linear time. Right: V458 Vul. This nova shows aperiodic UV modulation, which is approximately anti-correlated with the corresponding X-ray emission. In both cases the XRT count rate is over 0.3–10 keV. Names, eruption details, orbital periods, interstellar reddening E(B$-$V), and distances of the novae in this paper, ordered by eruption date; symbiotic systems are marked in the sixth column. The novae marked in bold are recurrent; the eruption date given is that for the data included in this article. $^a$ Date of $\gamma$-ray discovery by Fermi-LAT. $^b$ While [Joshi et al. , 2015] refer to V1534 Sco as a symbiotic system, [Munari et al. , 2017] note that there are some inconsistencies with the evidence for a cool giant. We will therefore not treat V1534 Sco as a symbiotic in this work. $^c$ There has been significant uncertainty in the distance to RS Oph over the years, with 1.6 kpc being assumed for a long time, although it now appears this larger estimate is preferred; see discussions in [Schaefer , 2022, Orio et al. , 2022] for example. $^d$ https://www.aavso.org/aavso-alert-notice-752 Nova Alternative name Eruption date UT P$_{\rm orb}$ E(B$-$V) Symbiotic? distance (kpc) References V2491 Cyg Nova Cyg 2008 No. 2 2008-04-10.73 2.3 h 0.23 4.8 [Nakano et al. , 2008, Baklanov et al. , 2008, Helton et al. , 2008, Rudy et al. , 2008, Schaefer , 2022] U Sco – 2010-01-28.44 1.23 d 0.20 6.3 [Schaefer , 2010, Schaefer et al. , 2010, Pagnotta et al. , 2015, Schaefer , 2022] V407 Cyg – 2010-03-10.80 43 y 0.5 Y 3.9 [Nishiyama et al. , 2010, Munari et al. , 1990, Hachisu & Kato , 2019, Shore et al. , 2011] T Pyx – 2011-04-14.24 1.83 h 0.25 3.6 [Schaefer , 2010, Waagan et al. , 2011, Shore et al. , 2011, Schaefer , 2022] V959 Mon Nova Mon 2012 2012-06-22$^{a}$ 7.1 h 0.38 2.9 [Fujikawa et al. , 2012, Cheung et al. , 2012, Cheung et al. , 2012, Page et al. , 2013, Shore et al. , 2013, Schaefer , 2022] V339 Del Nova Del 2012 2013-08-14.58 0.163 d 0.18 1.6 [Nakano , 2013, Shore et al. , 2016, Schaefer , 2022] V745 Sco – 2014-02-06.69 2440 d 1.00 Y 8.0 [Schaefer , 2010, Waagen & Pearce , 2014, Banerjee et al. , 2014, Schaefer , 2022] V1534 Sco Nova Sco 2014 2014-03-26.85 520 d 1.11 ?$^b$ 8.2 [Nishiyama & Kabashima , 2014, Joshi et al. , 2015, Munari et al. , 2017, Hachisu & Kato , 2019, Schaefer , 2022, Schaefer , 2022] V1535 Sco Nova Sco 2015 2015-02-11.84 50 d 0.80 Y 7.8 [Nakano , 2015, Walter , 2015, Nelson et al. , 2015, Linford et al. , 2017, Hachisu & Kato , 2019, Schaefer , 2022, Schaefer , 2022] V3890 Sgr – 2019-08-27.87 747.6 d 0.59 Y 8.5 [Schaefer , 2010, Pereira , 2019, Munari & Walter , 2019, Evans et al. , 2022, Schaefer , 2022, Mikołajewska et al. , 2021] V1674 Her Nova Her 2021 2021-06-12.19 0.153 d 0.50 3.2 [Patterson et al. , 2021, Drake et al. , 2021, Woodward et al. , 2021, Schaefer , 2022] RS Oph – 2021-08-09.54 453.6 d 0.73 Y 2.7$^c$ AAVSO$^d$; [Brandi et al. , 2009, Schaefer , 2010, Page et al. , 2022, Schaefer , 2022, Snijders , 1987] § ANALYSIS AND RESULTS Following [Page et al. , 2013, Page et al. , 2015, Page et al. , 2020], we parameterise the UV decay as a series of power-laws. That is, we consider the magnitudes to be proportional to log(time), which is equivalent to flux proportional to time: f $\propto$ (t/1 day)$^{-\alpha}$; T$_0$ is defined as the start of the nova eruption. While optical/UV light-curves are traditionally plotted in terms of magnitude versus linear time, a log scale allows us to infer useful information, with trends easily visible to the eye. A series of power-laws was thus fitted to each UVOT light-curve, using a $\chi^2$ minimisation routine for both slopes and break times; breaks were included if they were significant at the 3$\sigma$ level. The power-law slope $\alpha$ was then calculated as \begin{equation} \alpha = \frac{1}{2.5} . \frac{m_2 - m_1}{log\frac{t_2}{t_1}} \label{eqn} \end{equation} where m$_1$ and m$_2$ are the magnitudes from the fitted model at the break times of t$_1$ and t$_2$. For each nova, the different UV filter data were fitted independently, and all breaks are assumed to be instantaneous, with no smoothing applied. Figs. <ref>–<ref> show the X-ray and UV light-curves for each of the novae in the sample, plotted as days since eruption. The solid grey lines show the broken power-laws fitted to the UVOT data; the corresponding power-law indices for each segment are listed in Table <ref>. Where the value of $\alpha$ is negative, this indicates that the source has rebrightened. The error bars on the data points are mostly smaller than the marker size, and are usually $\sim$ 0.02–0.05 mag, occasionally up to $\sim$ 0.3 mag for some of the shorter (and fainter) observations. The XRT and UVOT light-curve of V2491 Cyg, with the best-fit broken power-law decline shown as the grey solid line. The UV light-curve shows a rebrightening `cusp' after day 12, so only the data after day 19 (marked with the vertical dashed line) were fitted. The X-ray data (top) are in count s$^{-1}$ over 0.3–10 keV. The XRT and UVOT light-curves of U Sco, with the best-fit broken power-law declines shown as the grey solid lines. The X-ray data (top) are in count s$^{-1}$ over 0.3–10 keV. The XRT and UVOT light-curves of V407 Cyg, with the best-fit broken power-law decline shown as the grey solid line. The X-ray data (top) are in count s$^{-1}$ over 0.3–10 keV. The XRT and UVOT light-curves of T Pyx, with the best-fit broken power-law decline shown as the grey solid line. Only data after the UV peak on day 70 (marked with the vertical dashed line) were considered. The observations of T Pyx are plotted as days since discovery; the optical peak occurred around 28 days later. The X-ray data (top) are in count s$^{-1}$ over 0.3–10 keV. The XRT and UVOT light-curves of V959 Mon, with the best-fit broken power-law declines shown as the grey solid lines. Late-time data beyond day 300 were excluded from the fits, because there are not enough bins to constrain the break times accurately. The observations of V959 Mon are plotted as days since Fermi discovery. The X-ray data (top) are in count s$^{-1}$ over 0.3–10 keV. The XRT and UVOT light-curves of V339 Del, with the best-fit broken power-law declines shown as the grey solid lines. The X-ray data (top) are in count s$^{-1}$ over 0.3–10 keV. The XRT and UVOT light-curves of V745 Sco, with the best-fit broken power-law declines shown as the grey solid lines. The X-ray data (top) are in count s$^{-1}$ over 0.3–10 keV. The XRT and UVOT light-curves of V1534 Sco, with the best-fit broken power-law declines shown as the grey solid lines. The X-ray data (top) are in count s$^{-1}$ over 0.3–10 keV. The XRT and UVOT light-curves of V1535 Sco, with the best-fit broken power-law decline shown as the grey solid line. Data were only collected with the $uvw1$ filter intermittently for the first 19 days, so have not been included. The X-ray data (top) are in count s$^{-1}$ over 0.3–10 keV. The XRT and UVOT light-curves of V3890 Sgr, with the best-fit broken power-law decline shown as the grey solid line. $uvw2$ data were only collected until day 25, so have not been included. Late-time data after day 100 have been excluded from the fits. The X-ray data (top) are in count s$^{-1}$ over 0.3–10 keV. The XRT and UVOT light-curves of V1674 Her, with the best-fit broken power-law declines shown as the grey solid lines. The earliest bin in the $uvm2$ light-curve has been excluded from the fit, since the break time cannot be constrained. The X-ray data (top) are in count s$^{-1}$ over 0.3–10 keV. The XRT and UVOT light-curves of RS Oph, with the best-fit broken power-law declines shown as the grey solid lines. Only the UVOT data before the first solar conjunction have been fitted. The X-ray data (top) are in count s$^{-1}$ over 0.3–10 keV. Parameterising the UV light-curves as a series of power-law decays. The uncertainties on the fitted break times are given at the 90% level, and the power-law slopes are calculated using Equation <ref>. The errors on the slopes have been estimated using the uncertainties of the magnitudes from the fits. Nova filter $\alpha_1$ T$_{\rm break,1}$ $\alpha_2$ T$_{\rm break,2}$ $\alpha_3$ T$_{\rm break,3}$ $\alpha_4$ T$_{\rm break,4}$ $\alpha_5$ T$_{\rm break,5}$ $\alpha_6$ (day) (day) (day) (day) (day) V2491 $uvw2$ 1.93$^{+0.02}_{-0.02}$ 40.2$^{+0.1}_{-0.1}$ 4.60$^{+0.01}_{-0.01}$ 55.1$^{+0.1}_{-0.1}$ 2.91$^{+0.09}_{-0.09}$ 124$^{+2}_{-3}$ 1.5$^{+0.1}_{-0.1}$ U Sco $uvw1$ 0.54$^{+0.05}_{-0.05}$ 4.4$^{+0.1}_{-0.3}$ 3.0$^{+0.1}_{-0.1}$ 12.52$^{+0.01}_{-0.01}$ $-$0.28$^{+0.01}_{-0.01}$ 24.19$^{+0.06}_{-0.06}$ 4.99$^{+0.01}_{-0.01}$ 44.51$^{+0.06}_{-0.14}$ $-$2.85$^{+0.06}_{-0.06}$ 52.28$^{+0.07}_{-0.03}$ 8.34$^{+0.02}_{-0.02}$ $uvw2$ 0.44$^{+0.02}_{-0.02}$ 3.9$^{+0.1}_{-0.1}$ 3.20$^{+0.02}_{-0.02}$ 12.53$^{+0.01}_{-0.09}$ 0.02$^{+0.01}_{-0.01}$ 25.15$^{+0.01}_{-0.06}$ 5.61$^{+0.02}_{-0.02}$ 41.8$^{+0.1}_{-0.2}$ $-$0.33$^{+0.04}_{-0.04}$ 52.34$^{+0.02}_{-0.07}$ 7.56$^{+0.02}_{-0.02}$ V407 $uvm2$ 0.35$^{+0.01}_{-0.01}$ 41.4$^{+0.3}_{-0.4}$ 4.39$^{+0.07}_{-0.07}$ 63$^{+1}_{-1}$ 2.23$^{+0.06}_{-0.06}$ T Pyx $uvm2$ 6.54$^{+0.01}_{-0.01}$ 113.90$^{+0.01}_{-0.01}$ 1.27$^{+0.01}_{-0.01}$ 142.59$^{+0.01}_{-0.01}$ 7.03$^{+0.05}_{-0.05}$ 150.2$^{+0.2}_{-0.3}$ 4.05$^{+0.01}_{-0.01}$ 268.7$^{+0.9}_{-0.8}$ 1.38$^{+0.02}_{-0.02}$ V959 $uvw1$ 2.02$^{+0.05}_{-0.05}$ 212$^{+1}_{-1}$ 5.33$^{+0.08}_{-0.08}$ Mon $uvm2$ 1.83$^{+0.01}_{-0.01}$ 205$^{+1}_{-1}$ 4.91$^{+0.07}_{-0.07}$ $uvw2$ 2.07$^{+0.04}_{-0.04}$ 205$^{+1}_{-1}$ 5.29$^{+0.06}_{-0.06}$ V339 $uvw1$ – – 1.1$^{+0.2}_{-0.2}$ 138$^{+16}_{-14}$ 2.4$^{+0.2}_{-0.2}$ 258$^{+12}_{-4}$ 1.5$^{+0.5}_{-0.5}$ Del $uvm2$ 3.0$^{+0.5}_{-0.5}$ 81$^{+15}_{-7}$ 0.8$^{+0.5}_{-0.2}$ 129$^{+2}_{-1}$ 2.17$^{+0.07}_{-0.07}$ 266$^{+11}_{-9}$ 1.7$^{+0.2}_{-0.2}$ $uvw2$ – – 0.94$^{+0.08}_{-0.08}$ 133$^{+5}_{-4}$ 2.96$^{+0.02}_{-0.02}$ 264$^{+2}_{-6}$ 2.1$^{+0.1}_{-0.1}$ V745 $uvw1$ 0.68$^{+0.01}_{-0.01}$ 6.3$^{+0.1}_{-0.1}$ 3.01$^{+0.05}_{-0.05}$ 15.7$^{+0.5}_{-0.4}$ 1.45$^{+0.03}_{-0.03}$ Sco $uvm2$ 0.53$^{+0.01}_{-0.01}$ 5.5$^{+0.1}_{-0.1}$ 2.4$^{+0.1}_{-0.1}$ 18$^{+1}_{-1}$ 1.40$^{+0.03}_{-0.03}$ $uvw2$ 0.73$^{+0.01}_{-0.01}$ 7.8$^{+0.1}_{-0.1}$ 3.62$^{+0.06}_{-0.06}$ 15.5$^{+0.4}_{-0.4}$ 1.46$^{+0.04}_{-0.04}$ V1534 $uvw1$ 0.75$^{+0.02}_{-0.02}$ 9.5$^{+0.4}_{-0.4}$ 3.65$^{+0.06}_{-0.06}$ 22$^{+1}_{-1}$ 1.1$^{+0.1}_{-0.1}$ Sco $uvm2$ 0.5$^{+0.1}_{-0.1}$ 4.9$^{+1.6}_{-0.6}$ 1.7$^{+0.2}_{-0.2}$ – – $uvw2$ 0.43$^{+0.09}_{-0.09}$ 9.4$^{+1.1}_{-0.7}$ 3.7$^{+0.3}_{-0.3}$ – – V1535 $uvm2$ 0.71$^{+0.01}_{-0.01}$ 16.2$^{+0.1}_{-0.1}$ 5.20$^{+0.01}_{-0.01}$ 19.0$^{+0.2}_{-0.1}$ 2.0$^{+0.1}_{-0.1}$ 31.4$^{+0.3}_{-0.8}$ 3.7 $^{+0.2}_{-0.2}$ 44.4$^{+0.6}_{-0.6}$ $-$1.32$^{+0.02}_{-0.02}$ 67$^{+1}_{-1}$ 5.00$^{+0.06}_{-0.06}$ Sco $uvw2$ 0.85$^{+0.05}_{-0.05}$ 16.0$^{+0.2}_{-0.2}$ 6.4$^{+0.4}_{-0.4}$ 18.2$^{+0.6}_{-0.4}$ 2.2$^{+0.1}_{-0.1}$ – – – – – – V3890 $uvm2$ 0.88$^{+0.06}_{-0.06}$ 13$^{+1}_{-1}$ 1.4$^{+0.1}_{-0.1}$ 21.6$^{+0.2}_{-0.2}$ 4.7$^{+0.1}_{-0.1}$ 30.2$^{+0.6}_{-0.6}$ 2.9$^{+0.3}_{-0.3}$ 41$^{+3}_{-3}$ 2.1$^{+0.1}_{-0.1}$ V1674 $uvw1$ – – 2.2$^{+0.2}_{-0.2}$ 30$^{+2}_{-4}$ 2.1$^{+0.1}_{-0.1}$ 77$^{+8}_{-3}$ 1.2$^{+0.2}_{-0.2}$ Her $uvm2$ 2.1$^{+0.2}_{-0.2}$ 12.3$^{+1.2}_{-0.3}$ 1.96$^{+0.01}_{-0.01}$ 32$^{+8}_{-5}$ 2.02$^{+0.01}_{-0.01}$ 86$^{+6}_{-5}$ 0.80$^{+0.06}_{-0.06}$ $uvw2$ 1.6$^{+0.2}_{-0.2}$ 15$^{+1}_{-1}$ 2.6$^{+0.1}_{-0.1}$ 30$^{+2}_{-3}$ 2.2 $^{+0.2}_{-0.2}$ – – RS uvm2 0.99$^{+0.08}_{-0.08}$ 27$^{+1}_{-1}$ 1.70$^{+0.03}_{-0.03}$ 76$^{+2}_{-3}$ 3.6$^{+0.6}_{-0.6}$ Oph uvw2 1.1$^{+0.2}_{-0.2}$ 40$^{+3}_{-2}$ 1.8$^{+0.6}_{-0.6}$ 68$^{+17}_{-9}$ 2.6$^{+0.10}_{-0.10}$ §.§ Light-curve descriptions Below we briefly summarise the fits for each nova. We note that the models applied may not be a unique description of the data. For example, in some cases consecutive slope changes are in the same direction – i.e., we see a gradual flattening across two break times, rather than a flattening followed by a steepening – which may suggest a smoother, slower change is occurring, rather than the instantaneous breaks we fit. The series of sharply-broken power-laws is, however, a simple and convenient characterisation of the decline in UV brightness. §.§.§ V2491 Cyg Data were collected in the $uvw2$ filter between days 7 and 236; the earliest magnitudes were estimated using the read-out streak method [Page et al. , 2013] (Fig. <ref>). The UV light-curve shows a rebrightening starting around day 12 (a `cusp'; [Strope, Schaefer & Henden , 2010]), so only the data after day 19 were fitted. The data before the start of the cusp align with an extrapolation of the earliest fitted power-law, however. During the post-cusp time interval, the rate of the UV decline steepened around day 40, before starting to flatten off at day 55 and flattening further after day 133. §.§.§ U Sco Following the 2010 eruption of the RN U Sco, the majority of the UVOT data were collected using the $uvw1$ and $uvw2$ filters, between days 0.6 and 63 post-eruption; the earliest magnitudes were estimated using the read-out streak method [Page et al. , 2013] (Fig. <ref>). U Sco is an eclipsing system <cit.>, with the 1.23 d period clearly visible in the UVOT data (Fig. <ref>). However, the underlying long-term evolution (including all the data within and outside the eclipse times) can be approximated by a series of power-law segments, showing both fading and flattening/brightening intervals. The fitted break times in the light-curves agree between the filters to within a few days. Zoom-in of part of the U Sco $uvw1$ light-curve, showing a series of eclipses. §.§.§ V407 Cyg Between days 5 and 141, observations were performed using the $uvm2$ filter (Fig. <ref>). During this time, the decay steepened on day 41, before flattening off again after day 63. §.§.§ T Pyx T Pyx was observed while the UV emission was still rising; these early data, starting on day 7, have been excluded, with the decay fitting performed after day 70, once the emission had peaked, and running through till day 369. The $uvm2$ light-curve (Fig. <ref>) showed a complex decay, with at least five separate segments, including a short, week-long interval between days 142 and 149, which showed a significant steepening. §.§.§ V959 Mon V959 Mon was initially monitored only using the $uvm2$ filter between days 58 and 150; after this time, the other two UV filters were also brought into play (Fig. <ref>). There was no notable change in decay rate at the earlier times, and all three filters show breaks from a flatter to steeper decay around day 205-212. Only data until day 256 are considered; after this time there were observations on days 423 (all three filters), 687 ($uvw1$) and 1002 ($uvm2$), which do not provide enough data points to constrain additional fitting. The measurements do suggest a further flattening of the decay between days 256 and 423, however. §.§.§ V339 Del The very earliest observations of V339 Del with Swift were performed with the UVOT blocked, because of the extreme brightness (it peaked at $V$ $\sim$ 4.3; [Munari et al. , 2015]); the UV data collection commenced on day 57 in the $uvm2$ filter, and day 60 in $uvw1$ and $uvw2$, continuing until day 291 in $uvw1$, and day 409 in both $uvm2$ and $uvw2$ (Fig. <ref>). Again, the read-out streak method [Page et al. , 2013] was required for data brighter than around magnitude 10–11. Since the observations were performed mainly with the high time resolution event mode, data were subsequently averaged over bins of $\sim$ 1 d in length, to improve the statistics. The $uvm2$ filter was the one used most frequently at the start, and this higher density of early data points is likely why an additional early-time break, from steep to flat can be constrained in this light-curve. All three UV light-curves subsequently steepen around day 130–140, before flattening between day 258–266. §.§.§ V745 Sco V745 Sco was observed with all three filters from 0.5 day after eruption until day 66. There was also an earlier $uvm2$ observation at day 0.16, and further observations out till day 173 in $uvw1$ (Fig. <ref>). All three light-curves steepen around day 5–8, and flatten off again between $\sim$ day 16–19. §.§.§ V1534 Sco While V1534 Sco was initially observed with all UV filters, beyond three weeks after eruption only the $uvw1$ observations were continued (Fig. <ref>). Considering the longest ($uvw1$) dataset, which runs between days 0.3 and 83, the decay broke twice, first steepening after day 9.5, then flattening again after day 22. The $uvm2$ and $uvw2$ light-curves only cover the time of the first break, with data collected between days 0.9–15.2 and 0.9–19.8, respectively. While the break time fitted to the $uvw2$ data, and the post-break slope, are in close agreement with $uvw1$, the initial slope is flatter. The break in $uvm2$ appears to occur days earlier. §.§.§ V1535 Sco V1535 Sco was observed between days 3.8 and 80.3 with the $uvm2$ filter, and 6.0 to 42.2 with $uvw2$ (Fig. <ref>). The $uvw1$ filter was only used for seven observations, so these data have not been included. Because of the longer duration of the $uvm2$ dataset, it is better fitted with a further three breaks, in addition to the two which are required for the $uvw2$ light-curve, around 16 and 18–19 days. After day 44, the $uvm2$ data show a brightening trend for around 23 days, before starting to fade again. §.§.§ V3890 Sgr The vast majority of the UVOT observations of V3890 Sgr were obtained using the $uvm2$ filter, so these are the data considered here (Fig. <ref>). Observations in this filter began on day 3, continuing until the nova entered the solar observing constraint for Swift on day 75. Further observations were taken after the source re-emerged from behind the Sun, until day 250, but these have not been included in the overall fit (though are shown in Fig. <ref>). The light-curve decline steepens twice and then flattens off again, in two stages. The UV brightness appears to have stayed constant during the observing constraint (or the source faded and rebrightened by about the same amount), before starting to fade again, with $\alpha$ $\sim$ 2.8. We note that this flat interval is much later than the X-ray SSS phase, so will not correspond to the plateau phase of [Hachisu et al. , 2008]. §.§.§ V1674 Her An early observation was obtained on day 1.3 using the $uvm2$ filter, with following observations on days 6.5, 7.6 and 10.3 in $uvw2$, $uvm2$ and $uvw1$ respectively, before subsequent observations used all three filters until day 86, after which filters were again alternated (Fig. <ref>). The last observations in $uvw1$, $uvm2$ and $uvw2$ were on days 126.5, 341.4 and 86.5, respectively. There is clearly a change in slope between the first two observations obtained with the $uvm2$ filter on days 1.3 and 7.6, but a break time cannot be sensibly estimated with no measurements in between. Therefore the initial $uvm2$ data point was not included in the analysis. While the $uvm2$ data show evidence for three breaks in the power-law decay, the $uvw1$ data do not require the earliest break, most likely since this filter dataset starts latest. Similarly, the $uvw2$ light-curve does not require the final change in slope because of the lack of data beyond day 86. Although $\alpha_2$ and $\alpha_3$ are very similar for the $uvm2$ data, the slight change is an improvement (the break is significant at the 3$\sigma$ level), and, given the requirement for a break around day 30 in the other two filters, it seems likely to be real. §.§.§ RS Oph The very earliest Swift observations of RS Oph in 2021 had the UVOT blocked, because of the optical brightness of the source. The first photometric observation was obtained 15.6 days after the optical peak, and ran through till the solar observing constraint began on day 87; observations began again on day 197.5 and continued until the following solar constraint, on day 448. The read-out streak method [Page et al. , 2013] was used for the first two months. The majority of observations were obtained with either the $uvm2$ or $uvw2$ filter (sometimes with grism exposures as well; [Azzollini et al. , 2022]), and these are shown in Fig. <ref>. Both filters show two steepening breaks, the first around day 30–40, the second, day 70–75. Following solar conjunction, the $uvm2$ light-curve shows signs of slow rebrightening; these data have not been included in the fits. §.§ Normalised break times In order to compare the different novae more directly, in Table <ref> the break times have been normalised to the end of the X-ray SSS phase as T$_{\rm break}$/T$_{\rm SSSend}$. Each SSS end point has been estimated from the X-ray light-curves (Figs. <ref>-<ref>; see also [Page et al. , 2020, Page et al. , 2020, Drake et al. , 2021]), and is defined to be the time at which the X-ray count rate starts to decrease steadily. There can obviously be some uncertainty in this time, given that there are gaps in the light-curve coverage, as well as short-term variability, but this is typically only a couple of days or less[The SSS phase of V339 Del ended during the solar observing constraint, when no observations could be performed. However, an extrapolation backwards of the decline once observations restarted suggests the X-ray fading began around the same time as the constraint.], as can be seen from the top panels of Figs. <ref>–<ref>. The nth break time normalised to the end time of the X-ray SSS: T$_{\rm break,n}$ (norm.) = T$_{\rm break,n}$/T$_{\rm SSSend}$. The power-law indices are the same as in Table <ref>, so have not been listed. $^{a}$ The SSS switched off during the Sun constraint, which ran from days 144–202; given the shape of the decay, it seems likely the SSS ended around the time V339 Del entered the observing constraint. $^b$ In V1534 Sco, the early brightening of the X-ray emission appears to be caused by the absorbing column declining, rather than a new soft component appearing. The X-ray count rate peaked on day 7, and this is the value used for the break time normalisation. Nova T$_{\rm SSSstart}$–T$_{\rm SSSend}$ filter T$_{\rm break,1}$ T$_{\rm break,2}$ T$_{\rm break,3}$ T$_{\rm break,4}$ T$_{\rm break,5}$ (day) (norm.) (norm.) (norm.) (norm.) (norm.) V2491 33–43 $uvw2$ 0.9 1.3 2.9 U Sco 12–33 $uvw1$ 0.1 0.4 0.7 1.4 1.6 $uvw2$ 0.1 0.4 0.8 1.3 1.6 V407 12–40 $uvm2$ 1.0 1.6 T Pyx 123–180 $uvm2$ 0.6 0.79 0.83 1.5 V959 150–200 $uvw1$ 1.1 Mon $uvm2$ 1.0 $uvw2$ 1.0 V339 60–144$^{a}$ $uvw1$ – 0.9 1.8 Del $uvm2$ 0.6 0.9 1.8 $uvw2$ – 0.9 1.8 V745 3–6 $uvw1$ 1.1 2.6 Sco $uvm2$ 0.9 3.2 $uvw2$ 1.3 2.6 V1534 7$^b$ $uvw1$ 1.4 3.2 Sco $uvm2$ 0.7 – $uvw2$ 1.2 – V1535 12–25 $uvm2$ 0.6 0.8 1.3 1.8 2.7 Sco $uvw2$ 0.6 0.7 – – – V3890 8–20 $uvm2$ 0.7 1.1 1.5 2.1 V1674 19–45 $uvw1$ – 0.7 1.7 Her $uvm2$ 0.3 0.7 1.9 $uvw2$ 0.3 0.7 – RS 21–62 uvm2 0.4 1.2 Oph uvw2 0.6 1.1 Each light-curve shows a break close in time to the end of the SSS phase (i.e., there is a T$_{\rm break}$ (norm) close to 1 – in the range 0.7–1.3 – in Table <ref>), with the possible exception of V1534 Sco, which does not break until a little later in the $uvw1$ band. However, as mentioned in the caption of Table <ref>, V1534 Sco did not show an obvious SSS phase, but rather became softer and brighter as the absorbing column decreased, so this discrepancy (and the mis-match between the break times in the different filters) can be disregarded. Fig. <ref> shows the normalised break times (using the best-sampled light-curve for each nova which was observed using more than one UVOT filter) plotted as a histogram. The inset shows only the break which is closest in time to the end of the SSS phase, since some of the light curves have multiple breaks around this interval. There is a clear peak in the number of breaks around T$_{\rm SSSend}$. The fact that some breaks occur before the apparent end of the SSS emission in the X-ray band will be returned to in Section <ref>. A histogram showing the normalised break times fitted in the UVOT light-curves for all novae in the sample; the best-sampled light-curve has been used in each case where observations were performed in more than one filter. The inset plots only the break for each nova light-curve which is closest in time to T$_{\rm break}$/T$_{\rm SSSend}$ = 1. § DISCUSSION The UV light-curves presented here were chosen to be well sampled, and it is obvious that a higher cadence of observations is more likely to highlight changes in slope, especially over shorter time intervals[U Sco was observed at a particularly high cadence when trying to map the 1.23-day eclipses.]. This should be borne in mind when comparing the fits to different novae. §.§ Individual novae Two thirds of the novae in this sample have multi-filter light-curves. While there are distinct similarities between the different filters, the best fits are not always identical. Considering the break times normalised with respect to the end of the SSS in Table <ref>, there is a slight difference in the third and fourth break times for U Sco (with T$_{\rm break,4}$ being around 2.7 d earlier in $uvw2$ from Table <ref>), causing different slopes in the plateau, fading and late re-brightening. V959 Mon shows a later break (by about a week) in $uvw1$, although the $uvw1$ and $uvw2$ slopes are very similar; $uvm2$ shows flatter declines both pre- and post-break. In V745 Sco, the first break happens in the order $uvm2$, $uvw1$, $uvw2$; however, the second break time is consistent for $uvw1$ and $uvw2$, with $uvm2$ apparently breaking a couple of days later. We note that the $uvw1$ and $uvw2$ filter transmission curves have extended red tails – the so-called `red leak'[<https://swift.gsfc.nasa.gov/analysis/uvot_digest/redleak.html>]. This might in general affect the break times in these two filters compared with $uvm2$ (which does not suffer from this red leak) in systems with a red companion. In the case of V1534 Sco, only the first break was covered by all three filters, and the break in $uvm2$ is noticeable earlier (around 4.5 d), leading to a flatter post-break decay slope. The longest dataset, $uvw1$, fades more steeply than the other two before this early-time break. V1535 Sco has very similar break times for both filters used for the early monitoring. The slope between the breaks is steeper in the $uvw2$ data, however. V1674 Her shows an earlier first break in $uvm2$ than $uvw2$ (no break required in $uvw1$ due to lack of early data). Around this break, the $uvm2$ light-curve flattens, whereas the $uvw2$ data steepen. The break around day 30 is consistent in all three filters, and the third break agrees between the two filters used at that time, though the error bars on the time are relatively large. For RS Oph, the first break occurs slightly earlier in $uvm2$ than $uvw2$, but the second break occurs a little later – although the second break time is consistent within the uncertainties. The initial decay slopes are similar for both filters, though the final decay is steeper in $uvm2$. In some of these situations, the differences in observation times (and sometimes cadences) between the UVOT filters may account for some of the discrepancies found. There do seem to be some intrinsic differences between the results at different wavelengths, but there are no obvious trends which are the same in each case: we don't always see that the bluer $uvw2$ data fade in a different manner from the redder $uvw1$, for example. As noted above, $uvw1$ and $uvw2$ are affected by a `red leak', which could be a more significant issue where the secondary star is redder. UV spectra of novae often show strong emission lines, which can vary over time. Such emission will affect the overall magnitudes in different filters by differing amounts, depending on the wavelengths of the lines, which may lead to changes in the apparent break times between filters. We recall that [Page et al. , 2015] extended the fitting to the optical and infrared (IR) bands for V745 Sco, finding that the optical curves changed more gradually than the UV, while the IR was more strongly affected by light from the red giant (RG) secondary. §.§ Comparing the sample as a whole If the pre-day-70 data for T Pyx, and the single early $uvm2$ bin for V1674 Her, are included, then, with the exception of the $uvm2$ data for V339 Del, all the light-curves in the sample show a first break from a flatter to a steeper decay. This initial break occurs before or at the time of the SSS switch-off in the X-ray band[V1534 Sco is a possible exception, as previously noted.]. Whether it is the earliest fitted break or not, all the novae (with obvious supersoft emission) show a change in slope around the end of the SSS phase (given the uncertainties in determining the exact SSS end time), as has been previously noted by, e.g., [Page et al. , 2010, Page et al. , 2013, Page et al. , 2015]; see also Fig. <ref>. The end of the SSS phase corresponds to the point at which nuclear burning ceases, thus leading to a drop in the ionisation of the RG wind in the symbiotic-like systems, and a general decline in any UV emission coming directly from the hot WD photosphere. All the breaks are from flatter to steeper at this time, with the exception of V1674 Her, for which $uvw1$ and $uvw2$ both show a flattening; the $uvm2$ curve shows a (slight) steepening, however. T Pyx shows two breaks in quick succession (only a week apart) close to 0.8 T$_{\rm SSSend}$, with the first showing a flat to steep transition. The V1535 Sco $uvm2$ light-curve has two breaks relatively close in time to the SSS end point (0.76$\times$ and 1.26$\times$ T$_{\rm SSSend}$), with the second of these times corresponding to a steepening, which may suggest that is the break related to the end of the SSS phase (and is therefore the normalised break time included in the inset of Fig. <ref>). Fig. <ref> shows that, for some novae, the break closest in time to the end of the SSS phase actually occurs a little before T$_{\rm SSSend}$ measured from the X-ray light-curves. While it is not immediately obvious why the UV emission would start to fade before the nuclear burning switches off, it could be that fitting the light-curves with sharp breaks affects the results, given that some breaks, at least, might in reality be a smoother transition between power-law slopes. In addition, changes in the continuum spectral shape, or the emergence (or variation) of emission lines could affect the rate at which the overall UV brightness changes. All the novae then show further breaks in their UV decline after the end of the SSS phase. There are limited later-time measurements for V959 Mon, but it is clear that the decay flattened again; there just isn't enough information to constrain the fit. Following the 100+ day gap caused by the solar conjunction (which started around 10–15 days after the SSS phase started to fade), the data for RS Oph show a brightening trend. These later changes in slope do not occur at fixed times with respect to the SSS turn-off, neither do different novae show similar numbers of breaks over given time intervals. The earliest fitted break occurs at only 10% of the SSS end time in U Sco, corresponding to around day 4 post-eruption (followed by a further early break at 40% of the SSS time – day 12.5 – corresponding to the start of the plateau interval, which then persists for around 12 days). The UV decay in V745 Sco also breaks early on, around day 6, though that already equates to the end of the SSS phase in this rapidly-evolving nova. We note that T Pyx shows a short `wiggle' at the end of its month-long plateau (days 114–143), fading steeply with $\alpha$ $\sim$ 7 for about a week, before continuing to decay with $\alpha$ $\sim$ 4 for more than 100 days. The different novae were followed for differing lengths of time overall, with respect to the end of the SSS phase (or, indeed, in general: usually out till at least 100 days post-eruption, though). The duration of the monitoring was generally set by the brightness of the fading nova (either X-ray or UV), though sometimes will also have been affected by Swift observing constraints. This means that we cannot sensibly compare the numbers of late-time (i.e., after the end of the SSS emission) breaks. There is a wide range of decay slopes seen, from the flat plateau phase in some RNe, to steep declines of $\alpha$ $>$ 5. Two of the novae (U Sco and V1535 Sco) show temporary rebrightenings at later times; in fact, both of these happen around day 40 post-eruption, though these equate to different normalised times with respect to the SSS end point. The UV emission of RS Oph also rebrightens at late times; this rebrightening has not yet ended with the observations collected to date. Besides the break in decay caused by the end of nuclear burning, the other observed changes in slope are likely to be related to physical processes in the expanding ejecta, such as declining density, or (before the end of the SSS phase) decreasing incident flux due to greater distance from the WD. In some novae, the accretion process may remain unsettled for quite some time after the eruption, rather than promptly relaxing back into a typical accretion disc [Mason et al. , 2021], and this could affect the rate at which the UV emission fades. §.§ Populations In Fig. <ref>, we show the power-law fits to the light-curves in this sample (one per nova, choosing the filter which covers the longest interval of time if more than one was used), transformed into absolute magnitude using the distances in Table <ref>, and plotted against the time normalised to the SSS end. Corrections for reddening have also been applied at the wavelength of the UV filter in use each time, based on values of E(B$-$V) taken from the literature (mainly from [Schaefer , 2022]; see Table <ref>), and assuming that A$_{\rm V}$ = 3.1 $\times$ E(B$-$V) for the Milky Way [Draine , 2003]. The early times for T Pyx and V2491 Cyg which were excluded from the power-law fitting are not shown, likewise the V959 Mon and RS Oph late time data. The sub-samples of CNe, RNe with short periods, RNe with long periods, and symbiotic-like systems (which have not been seen to recur) are plotted in black, red, magenta and blue, respectively. Different line styles highlight which filter is plotted. The symbiotic RNe with long periods are at the brighter absolute magnitude end of the sample, with the CNe mainly at a much lower luminosity (V1534 Sco being the exception, though there is a possibility that this source is actually a symbiotic system); there do not appear to be any other obvious differences between the populations. Power-law fits to the UVOT light-curves, converted to absolute magnitude, and plotted against time normalised to the end of the SSS. In the case where a nova was observed in multiple filters, the best sampled curve is shown. Classical novae are plotted in black, short-period recurrent nova in red, long-period recurrent novae with RG companions in magenta, and non-recurrent symbiotic-like systems (those with giant companions) in blue. Solid lines mark the light-curves collected using the $uvw1$ filter, dashed – $uvm2$ and dotted – $uvw2$. §.§.§ Classical novae The classical (non-symbiotic) novae in this current sample are V2491 Cyg, V959 Mon, V339 Del, V1674 Her and (probably) V1534 Sco. The UV light-curves of V959 Mon started later after eruption because of the proximity to the Sun delayed observations, so early-time breaks in the decay may have been missed. Despite this, these four objects show similar power-law decays and breaks. The spread of unreddened, absolute magnitudes is not large, either. §.§.§ Recurrent novae The recurrent novae in this paper consist of both short-period (U Sco and T Pyx), and long-period (with a giant companion) systems (V745 Sco, V3890 Sgr and RS Oph). In this sample of 12 novae, the long-period, symbiotic RNe are typically the overall brightest in terms of absolute magnitude (although we did not measure the peak for CN V339 Del, since the UVOT was operated in blocked mode for instrument safety). These measurements do, of course, rely on having accurate distance and extinction measurements. For example, if RS Oph were at 1.6 kpc (the previously favoured distance) instead of 2.7 kpc, it would be about one magnitude fainter in absolute terms. U Sco and T Pyx, the two short-period systems, both show early time breaks and the (oft-seen) plateau corresponding to the X-ray SSS interval. None of the long period, symbiotic systems in this sample shows a plateau in their UV decline. The earliest break in the V745 Sco light-curve corresponds to the end of the SSS phase, which occurred very early, running only between days 3–6 post-eruption; there might not have been time for an optical/UV plateau to form during this brief interval. V3890 Sgr showed a longer SSS duration of $\sim$ 12 d, though this is still shorter than in U Sco ($\sim$ 21 d) and T Pyx ($\sim$ 57 d) so, again, this might account for the lack of plateau. While RS Oph does not show an obvious plateau in these UV data, its optical light-curves do, as shown by [Strope, Schaefer & Henden , 2010, Schwarz et al. , 2011]. §.§.§ Symbiotic-like novae In this sample there are five systems which have cool RG companions: V407 Cyg, V1535 Sco, V745 Sco, V3890 Sgr and RS Oph. The first two are CNe, having only been seen in eruption once, while the other three are known RNe. V1535 Sco has an orbital period of 50 d, while the other four have even longer P$_{\rm orb}$ $>$ 1 yr. Following a nova explosion in these embedded systems, there will be shock interactions between the expanding ejecta and the RG wind. As well as producing hard X-rays <cit.>, these shocks also lead to UV line emission [Azzollini , 2021, Azzollini et al. , 2022]. The presence or otherwise of a giant companion and accompanying wind does not directly have an effect on the end time of the X-ray SSS emission. In a summary paper of Swift-XRT observations of SSS novae, [Schwarz et al. , 2011] also showed that there is no correlation between orbital period (typically longer for the symbiotic-like novae) and the SSS turn-off time. The SSS interval is generally accepted to be related to the WD mass, with an early onset and short duration being signs of a massive WD <cit.> – as expected for RNe. We do indeed see that the RNe U Sco, V745 Sco and V3890 Sgr show early and short SSS phases; T Pyx does not, but this system is known to be unusual, evolving more slowly than most other recurrents [Schaefer , 2010]. The start of the SSS phase in RS Oph is not quite as early as some, but has a well-defined beginning and end. For the symbiotic-like sample here, V745 Sco has a very short SSS phase (ending on day 6), while the other four switched off after 20–60 d. The SSS phases of the non-symbiotic systems end between days 33–200[As noted in Table <ref>, V1534 Sco does not show an obvious SSS phase, with day 7 actually corresponding to the time when the X-ray emission peaked due to the decrease of the absorption. This nova has therefore be omitted from this range.]. In the case of the symbiotic-like novae, there will be a RG wind `bubble' surrounding the system. The outer extent of this region depends on the recurrence time of the nova: the longer the time between eruptions (or if only one nova event has so far occurred), the further the RG wind can flow before disruption. Once the nova ejecta reach this limit (where the wind density becomes too low to be significant), the shock `breaks out' of the medium, which will likely lead to a drop in the UV flux as the shock-related emission decreases, and so a steepening in the light-curve decay. As an example, in the 2006 eruption of RS Oph (not included in this sample, since the majority of the UVOT observations were taken using the grism; [Azzollini , 2021, Azzollini et al. , 2022]), the shock breakout time was found to occur around day 80 post-outburst [Anupama , 2008]. There is no obvious difference between the UV light-curve evolution for the symbiotic-like systems and the others in this sample (Fig. <ref>) – although it is likely that one of the later breaks (possibly later than Swift has followed, and therefore not included in the plots above) in the UV decline in the symbiotic/embedded systems is related to the drop-off of the RG wind. §.§ Comparison with previous work Strope et al. [Strope, Schaefer & Henden , 2010] published almost 100 detailed nova optical light-curves, including some of the sources in our sample, categorising the results depending on the shapes of the light-curves. Those which follow a series of power-law declines in magnitude-log(time) space, as in the sample considered here, are classified as `S' for smooth (or stereotypical), and constitute almost 40% of their sample. As mentioned above, RNe often show a plateau in their optical/UV decay, coincident with the SSS phase seen in X-rays; these are included in the `P'-class by [Strope, Schaefer & Henden , 2010], meaning an `S'-type curve, but with a long, almost-flat interval a few magnitudes below peak. Considering the RNe in this sample, U Sco and T Pyx are indeed both marked as `P' by [Strope, Schaefer & Henden , 2010], and a flat interval is also seen in our fits (see Table <ref>, and Figs. <ref> and <ref>). RS Oph is also classified as as `P', and does indeed show a plateau in the optical light-curve [Strope, Schaefer & Henden , 2010], but no such flattening is obvious in the UV data (Fig. <ref>). RN V3890 Sgr does not show a plateau, and is marked as `S' in [Strope, Schaefer & Henden , 2010]; see also [Page et al. , 2020]. RN V745 Sco is not classified by [Strope, Schaefer & Henden , 2010], but does not show a plateau in the UVOT data presented here. V2491 Cyg is one of only a few novae which have been found to show a cusp-shaped secondary maximum (`C' in the Strope classification scheme: “power-law decline plus secondary maximum with steepening rise then steep decline”), which is clearly visible in the UVOT light-curve in Fig. <ref>. We do note that the specific RN eruptions considered by [Strope, Schaefer & Henden , 2010] are not the exact same events as those presented here, with our data corresponding to more recent outbursts. Optical/UV light-curves from recurrent eruptions are frequently very similar – possibly identical – for both Galactic and extragalactic novae, however <cit.>. Hachisu & Kato [Hachisu & Kato , 2006] proposed a `universal decline law' for (classical) novae, considering optical and IR light-curves, which shows a similar shape to the `S'-class. Modelling the eruption as a radiatively-driven wind, their template has a decay slope steepening from $\alpha$ $\sim$ 1.75 to $\sim$ 3.5, around 6 mag below peak; this break is taken to be caused by an abrupt decrease in the wind mass-loss rate. As the wind ceases, the decline flattens to $\alpha$ $\sim$ 3. However, [Strope, Schaefer & Henden , 2010] find a much larger scatter in their measurements than can be physically explained by this universal model by [Hachisu & Kato , 2006], and suggest this means there is an additional mechanism at work. Indeed, [Shen & Quataert , 2022] show how the initial mass loss can be driven by binary interaction rather than radiation pressure. § SUMMARY Nova UV light-curves show a variety of shapes, evolutions and variabilities following the initial eruption. In some cases, the UV emission varies in phase (or anti-phase) with the X-rays, in others the wavebands seem completely unrelated (at least during the super-soft X-ray phase). In this paper, we concentrated on this latter population, where the UV decline can typically be modelled by a series of (usually) declining power-law slopes, while the X-rays first brighten through the SSS phase, and then fade. We have presented the first uniform analysis of nova UV light-curves – including classical and recurrent, short-period and long – which were well monitored by the Neil Gehrels Swift Observatory. Within this sample we find that all sources show a break in the UV decay slope around the point the SSS phase ends in the X-ray band, while additional breaks also occur over a range of times, both earlier and later. While the light-curves using different filters for a given nova sometimes show very similar break times and slopes, at other times they disagree; there are no definitive trends between the different filter wavelengths in all cases. Overall, there are no strong population trends between the evolution of the UV light-curves of the classical, symbiotic-like, short-period recurrent and long-period recurrent novae, although it is likely that the symbiotic-like, embedded systems will show a late-time break when the ejecta shock reaches the outer edge of the wind bubble formed by the giant companion. There is also no large difference in the range of unreddened, absolute magnitudes for the four populations considered, although the three symbiotic, long-period RNe are at the brighter end of the sample, while the majority of the CNe are significantly less luminous. Conceptualization, K.L.P., N.P.M.K., J.P.O.; methodology, K.L.P., N.P.M.K., J.P.O.; validation, K.L.P., N.P.M.K., J.P.O.; formal analysis, K.L.P., J.P.O.; visualization, K.L.P.; writing — original draft preparation, K.L.P.; writing — review and editing, K.L.P., N.P.M.K., J.P.O. All authors have read and agreed to the published version of the manuscript. The Swift project in the UK is funded by the UK Space Agency. The data underlying this paper are available in the Swift archive at <https://www.swift.ac.uk/swift_live/> and the HEASARC Browse archive at The authors declare no conflict of interest. [Gehrels et al. , 2004] Gehrels, N.; Chincarini, G.; Giommi, P.; Mason, K.O.; Nousek, J.A.; Wells, A.A.; White, N.E.; Barthelmy, S.D.; Burrows, D.N.; Cominsky, L.R.; et al. The Swift Gamma-Ray Burst Mission. Astrophys. J. 2004, 611, 1005–1020. [Roming et al. , 2005] Roming, P.W.A.; Kennedy, T.E.; Mason, K.O.; Nousek, J.A.; Ahr, L.; Bingham, R.E., Broos, P.S.; Carter, M.J.; Hancock, B.K.; Huckle, H.E.; et al. The Swift Ultra-Violet/Optical Telescope. Space Sci. Rev. 2005, 120, 95–142. [Bode & Evans , 2008] Bode M.F.; Evans A. Classical Novae, 2nd ed.; Cambridge Astrophysics Series 43, Cambridge Univ. Press, UK, 2008 [Woudt & Ribeiro , 2014]Woudt P.A.; Ribeiro V.A.R.M. Stella Novae: Past and Future Decades; Astronomical Society of the Pacific Conference Series, Vol. 490, San Francisco: Astronomical Society of the Pacific, 2014 [Cheung et al. , 2012] Cheung, C.C.; Hays, E.; Venters, T.; Donato, D.; Corbet, R. H. D. Fermi LAT Detection of a New Gamma-ray Transient in the Galactic Plane: Fermi J0639+0548. Astron. Telegram 2012, 4224, 1–1. [Cheung et al. , 2012] Cheung, C.C.; Shore, S.N.; De Gennaro Aquino, I.; Charbonnel, S.; Edlin, J.; Hays, E.; Corbet, R.H.D.; Wood, D.L. Possible Association of the Gamma-ray Transient Fermi J0639+0548 with Nova Mon 2012. Astron. Telegram 2012b, 4310, 1–1. [König et al. , 2022] König, O.; Wilms, J.; Arcodia, R.; Dauser, T.; Dennerl, K.; Doroshenko, V.; Haberl, F.; Hämmerich, S.; Kirsch, C.; Kreykenbohm, I.; et al. X-ray detection of a nova in the fireball phase. Nature 2022, 605, 248–250. [Page et al. , 2015] Page, K.L.; Osborne, J.P.; Kuin, N.P.M.; Henze, M.; Walter, F.M.; Beardmore, A.P.; Bode, M.F.; Darnley, M.J.; Delgado, L.; Drake, J.J.; Hernanz, M.; Mukai, K.; Nelson, T.; Ness, J.-U.; Schwarz, G.J.; Shore, S.N.; Starrfield, S.; Woodward, C.E. Swift detection of the super-swift switch-on of the super-soft phase in nova V745 Sco (2014). Mon. Not. R. Astron. Soc. 2015, 454, 3108–3120. [Darnley et al. , 2012] Darnley, M.J.; Ribeiro, V.A.R.M.; Bode, M.F.; Hounsell, R.A.; Williams, R.P. On the Progenitors of Galactic Novae. em Astrophys. J. 2012, 746, 61:1–61:10. [Truran & Livio , 1986]Truran, J.W.; Livio, M. On the Frequency of Occurrence of Oxygen-Neon-Magnesium White Dwarfs in Classical Nova Systems. Astrophys. J. 1986, 308, 721–727. [Schaefer , 2010]Schaefer B.E. Comprehensive Photometric Histories of All Known Galactic Recurrent Novae. Astrophys. J. Supp. 2010a, 187, 275–373. [Anupama , 2008]Anupama, G.C. One Year of Monitoring of the 2006 Outburst in Optical and Low Radio Frequencies in RS Ophiuchi (2006) and the Recurrent Nova Phenomenon, ed. A. Evans et al., ASP Conf. Ser. 401, (San Francisco, CA: ASP), 2008, 251–255. [Starrfield , 1989]Starrfield S., Thermonuclear processes and the classical nova outburst in Classical Novae, eds. Bode M., Evans A., Wiley, NY, USA, 1989, 39–60. [Tofflemire et al. , 2013] Tofflemire, B.M.; Orio, M.; Page, K.L.; Osborne, J.P.; Ciroi, S.; Cracco, V.; Di Mille, F.; Maxwell, M. X-Ray Grating Observations of Recurrent Nova T Pyxidis during the 2011 Outburst. Astrophys. J. 2013, 779, 22:1–22:13. [Nelson et al. , 2014] Nelson, T.; Chomiuk, L.; Roy, N.; Sokoloski, J.L.; Mukai, K.; Krauss, M.I.; Mioduszewski, A.J.; Rupen, M.P.; Weston, J. The 2011 Outburst of Recurrent Nova T Pyx: Radio Observations Reveal the Ejecta Mass and Hint at Complex Mass Loss. Astrophys. J. 2014, 785, 78:1–78:13. [Chomiuk et al. , 2014] Chomiuk, L.; Nelson, T.; Mukai, K.; Sokoloski, J.L.; Rupen, M.P.; Page, K.L.; Osborne, J.P.; Kuulkers, E.; Mioduszewski, A.J.; Roy, N.; Weston, J.; Krauss, M.I. The 2011 Outburst of Recurrent Nova T Pyx: X-ray Observations Expose the White Dwarf Mass and Ejection Dynamics. Astrophys. J. 2014, 788, 130:1–130:13. [Wolf et al. , 2013]Wolf, W.M.; Bildsten, L.; Brooks, J.; Paxton, B. Hydrogen Burning on Accreting White Dwarfs: Stability, Recurrent Novae, and the Post-nova Supersoft Phase. Astrophys. J. 2013, 777, 136:1–136:15. [Hachisu et al. , 2006] Hachisu, I.; Kato, M.; Kiyota, S.; Kubotera, K.; Maehara, H.; Nakajima, K.; Ishii, Y.; Kamada, M.; Mizoguchi, S.; Nishiyama, S.; et al. The Hydrogen-Burning Turnoff of RS Ophiuchi (2006). Astrophys. J. Lett. 2006, 651, L141–144. [Hachisu et al. , 2008]Hachisu, I.; Kato, M.; Kiyota, S.; Kubotera, K.; Maehara, H.; Nakajima, K.; Ishii, Y.; Kamada, M.; Mizoguchi, S.; Nishiyama, S.; Sumitomo, N.; Tanaka, K.; Yamanaka, M.; Sadakane, K. Optical Light Curves of RS Oph (2006) and Hydrogen Burning Turnoff in RS Ophiuchi (2006) and the Recurrent Nova Phenomenon, ed. A. Evans et al., ASP Conf. Ser. 401, (San Francisco, CA: ASP), 2008, 206–209. [Shafter et al. , 2015] Shafter, A.W.; Henze, M.; Rector, T.A.; Schweizer, F.; Hornoch, K.; Orio, M.; Pietsch, W.; Darnley, M.J.; Williams, S.C.; Bode, M.F.; Bryan, J. Recurrent Novae in M31. Astrophys. J. Supp. 2015, 216, 34:1–34:35. [Darnley , 2021]Darnley, M.J. Accrete, Accrete, Accrete... Bang! (and repeat): The Remarkable Recurrent Novae. PoS 2021, 368, 44:0–44:18 [Mürset & Nussbaumer , 1994] Mürset, U.; Nussbaumer, H. Temperatures and luminosities of symbiotic novae. em Astron. Astrophs. 1994, 282, 586–604. [Mikołajewska , 2007]Mikołajewska, J. Symbiotic Stars: Continually Embarrassing Binaries. BaltA bf 2007, 16, 1–9. [Mikołajewska , 2008]Mikołajewska, J. The Place of Recurrent Novae Among the Symbiotic Stars, in RS Ophiuci (2006) and the Recurrent Nova Phenomenon, ed. A. Evans et al., San Francisco, CA, ASP Conf. Ser. 401, 2008, 42-51. [Bode et al. , 2006] Bode, M.F.; O'Brien, T.J.; Osborne, J.P.; Page, K.L.; Senziani, F.; Skinner, G.K.; Starrfield, S.; Ness, J.-U.; Drake, J.J.; Schwarz, G.; et al. Swift Observations of the 2006 Outburst of the Recurrent Nova RS Ophiuchi. I. Early X-Ray Emission from the Shocked Ejecta and Red Giant Wind. Astrophys. J. 2006, 652, 629–635. [Osborne et al. , 2011] Osborne, J.P.; Page, K.L.; Beardmore, A.P.; Bode, M.F.; Goad, M.R.; O'Brien, T.J.; Starrfield, S.; Rauch, T.; Ness, J.-U.; Krautter, J.; et al. The Supersoft X-ray Phase of Nova RS Ophiuchi 2006. Astrophys. J. 2011, 727, 124:1–124:10. [Ness et al. , 2009]Ness, J.-U.; Drake, J.J.; Beardmore, A.P.; Boyd, D.; Bode, M.F.; Brady, S.; Evans, P.A.; Gaensicke, B.T.; Kitamoto, S.; Knigge, C.; Miller, I.; Osborne, J.P.; Page, K.L.; Rodriguez-Gil, P.; Schwarz, G.; Staels, B.; Steeghs, D.; Takei, D.; Tsujimoto, M.; Wesson, R. ; Zijlstra, A. Swift X-Ray and Ultraviolet Monitoring of the Classical Nova V458 Vul (Nova Vul 2007). AJ 2009, 137, 4160–4168. [Schwarz et al. , 2011] Schwarz, G.J.; Ness, J.U.; Osborne, J.P.; Page, K.L.; Evans, P.A.; Beardmore, A.P.; Walter, F.M.; Helton, L.A.; Woodward, C.E.; Bode, M.; Starrfield, S.; Drake, J.J. Swift X-Ray Observations of Classical Novae. II. The Super Soft Source Sample. Astrophys. J. Supp. 2011, 197, 31:1–31:26. [Ibarra et al. , 2009]Ibarra, A.; Kuulkers, E.; Osborne, J.P.; Page, K.; Ness, J.U.; Saxton, R.D.; Baumgartner, W.; Beckmann, V.; Bode, M.F.; Hernanz, M.; Mukai, K.; Orio, M.; Sala, G.; Starrfield, S.; Wynn, G.A. Pre-nova X-ray observations of V2491 Cygni (Nova Cyg 2008b). Astron. Astrophys. 2009, 497, L5–L8. [Page et al. , 2010] Page, K.L.; Osborne, J.P.; Evans, P.A.; Wynn, G.A.; Beardmore, A.P.; Starling, R.L.C.; Bode, M.F.; Ibarra, A.; Kuulkers, E.; Ness, J.-U.; Schwarz, G.J. Swift observations of the X-ray and UV evolution of V2491 Cyg (Nova Cyg 2008 No. 2). Mon. Not. R. Astron. Soc. 2010, 401, 121–130. [Beardmore et al. , 2012] Beardmore, A.P.; Osborne, J.P.; Page, K.L.; Hakala, P.J.; Schwarz, G.J.; Rauch, T.; Balman; S.; Evans, P.A.; Goad, M.R.; Ness, J.-U; Starrfield, S.; Wagner, R.M. The outburst of Nova CSS 081007:030559+054715 (HV Ceti). Astron. Astrophys. 2012, 545, A116:1–A116:10. [Gehrz et al. , 2018] Gehrz, R.D.; Evans, A.; Woodward, C.E.; Helton, L.A.; Banerjee, D.P.K.; Srivastava, M.K.; Ashok, N.M.; Joshi, V.; Eyres, S.P.S.; Krautter, J.; Kuin, N.P.M.; Page, K.L.; Osborne, J.P.; Schwarz, G.J.; Shenoy, D.P.; Shore, S.N.; Starrfield, S.G.; Wagner, R. M. The Temporal Development of Dust Formation and Destruction in Nova Sagittarii 2015#2 (V5668 SGR): A Panchromatic Study. Astrophys. J. 2018, 858, 78:1–78:17. [Aydi et al. , 2018]Aydi, E.; Orio, M.; Beardmore, A.P.; Ness, J.-U.; Page, K.L.; Kuin, N.P.M.; Walter, F.M.; Buckley, D.A.H.; Mohamed, S.; Whitelock, P.; Osborne, J.P.; Strader, J.; Chomiuk, L.; Darnley, M.J.; Dobrotka, A.; Kniazev, A.; Miszalski, B.; Myers, G.; Ospina, N.; Henze, M.; Starrfield, S.; Woodward, C.E. Multiwavelength observations of V407 Lupi (ASASSN-16kt) - a very fast nova erupting in an intermediate polar. Mon. Not. R. Astron. Soc. 2018, 480, 572–609. [Page et al. , 2020] Page, K.L.; Kuin, N.P.M.; Beardmore, A.P.; Walter, F.M.; Osborne, J.P.; Markwardt, C.B.; Ness, J.-U.; Orio, M.; Sokolovsky, K.V. The 2019 eruption of recurrent nova V3890 Sgr: observations by Swift, NICER, and SMARTS. Mon. Not. R. Astron. Soc. 2020a, 499, 4814-4831. [Page et al. , 2020] Page, K.L.; Beardmore, A.P.; Osborne, J.P. Neil Gehrels Swift Observatory studies of supersoft novae. Ad. Space Res. 2020b, 66, 1169-1192. [Burrows et al. , 2005] Burrows, D.N.; Hill, J.E.; Nousek, J.A.; Kennea, J.A.; Wells, A.; Osborne, J.P.; Abbey, A.F.; Beardmore, A.; Mukerjee, K.; Short, A.D.T.; et al. The Swift X-Ray Telescope. Space Sci. Rev. 2005, 120, 165–195. [Poole et al. , 2008] Poole, T.S.; Breeveld, A.A.; Page, M.J.; Landsman, W.; Holland, S.T.; Roming, P.; Kuin, N.P.M.; Brown, P.J.; Gronwall, C.; Hunsberger, S.; et al. Photometric calibration of the Swift ultraviolet/optical telescope. Mon. Not. R. Astron. Soc. 2008, 383, 627–645. [Breeveld et al. , 2011] Breeveld, A.A.; Landsman, W.; Holland, S.T.; Roming, P.; Kuin, N.P.M.; Page, M.J. An Updated Ultraviolet Calibration for the Swift/UVOT in Gamma-Ray Bursts 2010), AIP Conf. Proc. 1358, 2011, 373–376. [Page et al. , 2013]Page, M.J.; Kuin, N.P.M.; Breeveld, A.A.; Hancock, B.; Holland, S.T.; Marshall, F.E.; Oates, S.; Roming, P.W.A.; Siegel, M.H.; Smith, P.J.; Carter, M.; De Pasquale, M.; Symeonidis, M.; Yershov, V.; Beardmore, A.P. The use and calibration of read-out streaks to increase the dynamic range of the Swift Ultraviolet/Optical Telescope. Mon. Not. R. Astron. Soc. 2013, 436, 1684-1693. [Schaefer , 2022]Scahefer, B.E. Comprehensive Catalogue of the Overall Best Distances and Properties of 402 Galactic Novae. Mon. Not. R. Astron. Soc. 2022a, 517, 6150–6169. [Nakano et al. , 2008] Nakano, S.; Beize, J.; Jin, Z.-W.; Gao, X.; Yamaoka, H.; Haseda, K.; Guido, E.; Sostero, G.; Klingenberg, G.; Kadota, K. V2491 Cygni. IAU Circ. 2008, 8934, 1–1. [Baklanov et al. , 2008]Baklanov, A.; Pavlenko, E.; Berezina, E. Discovery of the optical variability of Nova V2491 Cyg. Astron. Telegram 2008, 1514, 1–1. [Helton et al. , 2008] Helton, L.A.; Woodward, C.E.; Vanlandingham, K.; Schwarz, G.J. V2491 Cygni. IAU Circ. 2008, 1379, 1–1. [Rudy et al. , 2008]Rudy, R.J.; Lynch, D.K.; Russell, R.W.; Woodward, C.E.; Covey, K. V2491 Cygni. IAU Circ. 2008, 8938, 2–2. [Schaefer et al. , 2010]Schaefer, B.E.; Harris, B.G.; Dvorak, S.; Templeton, M.; Linnolt, M. U Scorpii. IAU Circ. 2010, 9111, 1–1. [Pagnotta et al. , 2015] Pagnotta, A.; Schaefer, B.E.; Clem, J.L.; Landolt, A.U.; Handler, G.; Page, K.L.; Osborne, J.P.; Schlegel, E.M.; Hoffman, D.I.; Kiyota, S.; Maehara, H. The 2010 Eruption of the Recurrent Nova U Scorpii: The Multi-wavelength Light Curve Astrophys. J. 2015, 811, 32:1–32:42. [Nishiyama et al. , 2010] Nishiyama, K.; Kabashima, F.; Kojima, T.; Sakaniwa, K.; Tago, A.; Schmeer, P.; Munari, U. V407 Cygni. IAU Circ. 2010, 9130, 1–1. [Munari et al. , 1990] Munari, U.; Margoni, R.; Stagni, R. The extreme, possible symbiotic Mira V407 CYG and its relevance to the OH/IR sources. Mon. Not. R. Astron. Soc. 1990, 242, 563–569. [Hachisu & Kato , 2019]Hachisu, I., Kato, M. A Light-curve Analysis of 32 Recent Galactic Novae: Distances and White Dwarf Masses. Astrophys. J. Supp. 2019, 242, 1–18. [Shore et al. , 2011] Shore, S.N.; Wahlgren, G.M.; Augusteijn, T.; Liimets, T.; Page, K.L.; Osborne, J.P.; Beardmore, A.P.; Koubsky, P.; Šlechta, M.; Votruba, V. The spectroscopic evolution of the symbiotic-like recurrent nova V407 Cygni during its 2010 outburst. I. The shock and its evolution. Astron. Astrophys. 2011a, 527, A98:1–A98:15. [Waagan et al. , 2011]Waagan, E.; Linnolt, M.; Bolzoni, S.; Amorim, A.; Plummer, A.; Williams, P. F.; Matsuyama, H.; Kerr, S.; Pearce, A. T Pyxidis. CBET 2011, 2700, 1–1. [Shore et al. , 2011] Shore, S.N.; Augusteijn, T.; Ederoclite, A.; Uthas, H. The spectroscopic evolution of the recurrent nova T Pyxidis during its 2011 outburst. I. The optically thick phase and the origin of moving lines in novae . Astron. Astrophys. 2011b, 533, L8:1–L8:5 [Fujikawa et al. , 2012] Fujikawa, S.; Yamaoka, H.; Nakano, S. Nova Monocerotis 2012 = PNV J06393874+0553520. CBET 2012, 3202, 1–1. [Page et al. , 2013] Page, K.L.; Osborne, J.P.; Wagner, R.M.; Beardmore, A.P.; Shore, S.N.; Starrfield, S.; Woodward, C.E. The 7.1 Hr X-Ray-Ultraviolet-Near-infrared Period of the $\gamma$-Ray Classical Nova Monocerotis 2012. Astrophys. J. 2013, 768, L26:1–L26:6. [Shore et al. , 2013] Shore, S.N.; De Gennaro Aquino, I.; Schwarz, G.J.; Augusteijn, T.; Cheung, C.C.; Walter, F.M.; Starrfield, S. The spectroscopic evolution of the $\gamma$-ray emitting classical nova Nova Mon 2012. I. Implications for the ONe subclass of classical novae. Astron. Astrophys. 2013, 553, A123:1–A123:17. [Nakano , 2013]Nakano, S. V339 Delphini = Nova Delphini 2013 = PNV J20233073+2046041. IAUC 9258, 1–1. [Shore et al. , 2016] Shore, S.N.; Mason, E.; Schwarz, G.J.; Teyssier, F.M.; Buil, C.; De Gennaro Aquino, I.; Page, K.L.; Osborne, J.P.; Scaringi, S.; Starrfield, S.; van Winckel, H.; Williams, R.E.; Woodward, C.E. The panchromatic spectroscopic evolution of the classical CO nova V339 Delphini (Nova Del 2013) until X-ray turnoff. Astron. Astrophys. 2016, 590, A123:1–A123:16. [Waagen & Pearce , 2014] Waagen, E.O.; Pearce, A. V745 Scorpii. CBET 2014, 3803, 1–1. [Banerjee et al. , 2014] Banerjee, D.P.K.; Joshi, V.; Venkataraman, V.; Ashok, N.M.; Marion, G.H.; Hsiao, E.Y.; Raj, A. Near-IR Studies of Recurrent Nova V745 Scorpii during its 2014 Outburst. Astrophys. J. 2014, 785 L11:1–L11:5. [Nishiyama & Kabashima , 2014] Nishiyama, K.; V1534 Scorpii = Nova Scorpii 2014 = TCP J17154683-3128303. IAU Circ. 2014, 9271, 3–3. [Joshi et al. , 2015] Joshi, Vishal; Banerjee, D.P.K.; Ashok, N.M.; Venkataraman, V.; Walter, F.M. Infrared studies of Nova Scorpii 2014: an outburst in a symbiotic system sans an accompanying blast wave. Mon. Not. R. Astron. Soc. 2015, 452, 3696-3703. [Munari et al. , 2017] Munari, U.; Hambsch, F.-J.; Frigo, A. Photometric evolution of seven recent novae and the double-component characterizing the light curve of those emitting in gamma rays. Mon. Not. R. Astron. Soc. 2017, 469, 4341-4358. [Schaefer , 2022] Schaefer, B.E. Comprehensive listing of 156 reliable orbital periods for novae, including 49 new periods. Mon. Not. R. Astron. Soc. 2022b, 517, 3640-3659. [Nakano , 2015]Nakano, S. Nova Scorpii 2015 = PNV J17032620-3504140. IAU Circ. 2015, 9274, 1–1. [Walter , 2015]Walter, F. Nova Sco 2015 = PNV J17032620-3504140. Astron. Telegram 2015, 7060, 1–1. [Nelson et al. , 2015] Nelson, T.; Linford, J.; Chomiuk, L.; Sokoloski, J.; Mukai, K.; Finzell, T.; Weston, J.; Rupen, M.; Mioduszewski, A. Early X-ray and radio observations of Nova Sco 2015 implicate strong shocks against a red giant wind. Astron. Telegram 2015, 7085, 1–1. [Linford et al. , 2017] Linford, J.D.; Chomiuk, L.; Nelson, T.; Finzell, T.; Walter, F.M.; Sokoloski, J.L.; Mukai, K.; Mioduszewski, A.J.; van der Horst, A.J.; Weston, J.H.S.; Rupen, M.P. The Peculiar Multiwavelength Evolution Of V1535 Sco. Astrophys. J. 2017, 842, 1–16. [Pereira , 2019]Pereira, A. V3890 Sgr outburst. vsnet-alert 2019 23505. [Munari & Walter , 2019]Munari, U., Walter, F.M. The reddening and distance to recurrent nova V3890 Sgr from interstellar features. Astron. Telegram 2019, 13069, 1–1. [Evans et al. , 2022] Evans, A.; Geballe, T.R.; Woodward, C.E.; Banerjee, D.P.K.; Gehrz, R.D.; Starrfield, S.; Shahbandeh, M. Infrared spectroscopy of the 2019 eruption of the recurrent nova V3890 Sgr: separation into equatorial and polar winds revealed. Mon. Not. R. Astron. Soc. 2022, 517, 6077–6090. [Mikołajewska et al. , 2021] Mikołajewska, J.; Iłkiewicz, K.; Gałan, C.; Monard, B.; Otulakowska-Hypka, M.; Shara, M.M.; Udalski, A. The symbiotic recurrent nova V3890 Sgr: binary parameters and pre-outburst activity. Mon. Not. R. Astron. Soc. 2021, 504, 2122-2132. [Patterson et al. , 2021]Patterson, J.; Epstein-Martin, M.; Vanmunster, T.; Kemp, J. The Orbital and Pulse Periods of V1674 Herculis. Astron. Telegram 2021, 14856, 1–1. [Drake et al. , 2021] Drake, J.J.; Ness, J.-U.; Page, K.L.; Luna, G.J.M.; Beardmore, A.P.; Orio, M.; Osborne, J.P.; Mròz, P.; Starrfield, S.; Banerjee, D.P.K.; et al. The Remarkable Spin-down and Ultrafast Outflows of the Highly Pulsed Supersoft Source of Nova Herculis 2021. Astrophys. J. 2021, 922, L42:1–L42:9. [Woodward et al. , 2021] Woodward, C.E.; Banerjee, D.P.K.; Geballe, T.R.; Page, K.L.; Starrfield, S.; Wagner, R. M. Near-infrared Studies of Nova V1674 Herculis: A Shocking Record Breaker. Astrophys. J. 2021, 922, L10:1–L10:7. [Brandi et al. , 2009] Brandi, E.; Quiroga, C.; Mikołajewska, J.; Ferrer, O.E.; García, L.G. Spectroscopic orbits and variations of RS Ophiuchi. Astron. Astrophys. 2009, 497, 815–825. [Page et al. , 2022] Page, K.L.; Beardmore, A.P.; Osborne, J.P.; Munari, U.; Ness, J.-U.; Evans, P.A.; Bode, M.F.; Darnley, M.J.; Drake, J.J.; Kuin, N.P.M.; O'Brien, T.J.; Orio, M.; Shore, S.N.; Starrfield, S.; Woodward, C.E. The 2021 outburst of the recurrent nova RS Ophiuchi observed in X-rays by the Neil Gehrels Swift Observatory: a comparative study. Mon. Not. R. Astron. Soc. 2022, 514, 1557–1574. [Snijders , 1987]Snijders, M.A.J. Multi-frequency observations of the 1985 outburst of RS Ophiuchi. Astrophys. Space Science 1987, 130, 243–254. [Orio et al. , 2022] Orio, M.; Behar, E.; Luna, G.J.M.; Drake, J.J.; Gallagher, J.; Nichols, J.S.; Ness, J.U.; Dobrotka, A.; Mikolajewska, J.; Della Valle, M.; Ignace, R.; Rahin, R. Shocks in the Outflow of the RS Oph 2021 Eruption Observed with X-Ray Gratings. Astrophys. J 2022, 938, 34:1–34:14. [Strope, Schaefer & Henden , 2010]Strope, R.J.; Schaefer, B.E.; Henden, A.A. Catalog of 93 Nova Light Curves: Classification and Properties. Astron. J. 2010, 140, 34–62. [Schaefer et al. , 2010]Schaefer, B.E.; Pagnotta, A.; Xiao, L.; Darnley, M.J.; Bode, M.F.; Harris, B.G.; Dvorak, S.; Menke, J.; Linnolt, M.; Templeton, M.; et al. Discovery of the Predicted 2010 Eruption and the Pre-eruption Light Curve for Recurrent Nova U Scorpii. Astron. J 2010c, 140, 925-932. [Ness et al. , 2012] Ness, J.-U.; Schaefer, B.E.; Dobrotka, A.; Sadowski, A.; Drake, J.J.; Barnard, R.; Talavera, A.; Gonzalez-Riestra, R.; Page, K.L.; Hernanz, M.; Sala, G.; Starrfield, S. From X-Ray Dips to Eclipse: Witnessing Disk Reformation in the Recurrent Nova U Sco. Astrophys. J. 2012, 745, 43:1–43:16 [Munari et al. , 2015] Munari, U.; Maitan, A.; Moretti, S.; Tomaselli, S. 500 days of Stromgren b, y and narrow-band [OIII], H$\alpha$ photometric evolution of gamma-ray Nova Del 2013 (=V339 Del) . New Astronomy 2015, 40, 28–40. [Mason et al. , 2021] Mason, E.; Shore, S.N.; Drake, J.; Howell, S.B.; Kuin, P.; Magaudda, E. The aftermath of nova Centauri 2013 (V1369 Centauri). Astron. Astrophys. 2021, 649, A28:1–A28:15. [Draine , 2003]Draine, B.T. Interstellar Dust Grains. Ann. Rev. Astron. Astrophys. 2003, 41, 241-289. [Azzollini , 2021]Azzollini, A. Multi-wavelength Spectroscopic Study of Shock Phenomena Driven by Explosive Outbursts in Symbiotic-like Recurrent Novae. Master thesis, Univ. Pisa, Pisa, Italy, 2021. [Azzollini et al. , 2022]Azzollini, A.; Shore, S.N.; Kuin, N.P.M.; Page, K.L. Multiwavelength spectroscopic study of shock driven phenomena in explosive outbursts in symbiotic-like recurrent novae with emphasis on RS Ophiuchi. Astron. Astrophys. 2022, submitted. [Kato , 1997]Kato, M. Optically Thick Winds from Degenerate Dwarfs. I. Classical Novae of Populations I and II. Astrophys. J. Supp. 1997, 113, 121–129. [Yaron et al. , 2005] Yaron, O.; Prialnik, D.; Shara, M.M.; Kovetz, A. An Extended Grid of Nova Models. II. The Parameter Space of Nova Outbursts. Astrophys. J. 2005, 623, 398–410. [Henze et al. , 2011] Henze, M.; Pietsch, W.; Haberl, F.; Hernanz, M.; Sala, G.; Hatzidimitriou, D.; Della Valle, M.; Rau, A.; Hartmann, D.H.; Burwitz, V. X-ray monitoring of classical novae in the central region of M 31. II. Autumn and winter 2007/2008 and 2008/2009. Astron. Astrophys. 2011, 533, A52:1–A52:30. [Henze et al. , 2018] Henze, M.; Darnley, M.J.; Williams, S.C.; Kato, M.; Hachisu, I.; Anupama, G.C.; Arai, A.; Boyd, D.; Burke, D.; Ciardullo, R.; et al. Breaking the Habit: The Peculiar 2016 Eruption of the Unique Recurrent Nova M31N 2008-12a. Astrophys. J. 2018, 857, 68:1–68:29. [Hachisu & Kato , 2006] Hachisu, I.; Kato, M. A Universal Decline Law of Classical Novae. Astrophys. J. Supp. 2006, 167, 59–80. [Shen & Quataert , 2022]Shen, K.J.; Quataert, E. Binary Interaction Dominates Mass Ejection in Classical Novae. Astrophys. J. 2022, 938, 31:1–31:13.
email<EMAIL_ADDRESS>email<EMAIL_ADDRESS> # Assessing quantum thermalization in physical and configuration spaces via many-body weak values Carlos F. Destefani1 [ Xavier Oriols1 [ 1 Department of Electronic Engineering, Universitat Autònoma de Barcelona, 08193 Bellaterra, Barcelona, Spain ###### Abstract We explore the origin of the arrow of time in an isolated quantum system described by the Schrödinger equation. We provide an explanation from weak values in the configuration space, which are understood as operational properties obtained in the laboratory following a well-defined protocol. We show that quantum systems satisfying the eigenstate thermalization hypothesis can simultaneously provide thermalized ensemble expectation values and nonthermalized weak values of the momentum, both from the same operational probability distribution. The reason why weak values of the momentum may escape from the eigenstate thermalization hypothesis is because they are linked only to off-diagonal elements of the density matrix in the energy representation. For indistinguishable particles, however, operational properties can not be defined in the configuration space. Therefore, we state that the origin of the arrow of time in isolated quantum systems described by the Schrödinger equation comes from dealing with properties obtained by averaging (tracing out) some degrees of freedom of the configuration space. We then argue that thermalization does not occur in the properties defined in the configuration space, and our argument is compatible with defending that thermalization is a real phenomenon in the properties defined in the physical space. All of these conclusions are testable in the laboratory through many- body weak values. ## I Introduction The arrow of time has always been a topic of lively debate [1, 2, 3, 4, 5, 6, 7, 8]. It appears in many disciplines as, for example, a cosmological arrow of time pointing in the direction of the Universe expansion [2]. A related arrow of time appears in the time evolution of time-irreversible macroscopic systems governed by the second law of thermodynamics, where entropy always increases with time [1, 7]. The puzzle implicit in such irreversibility is that most fundamental microscopic laws have no arrow of time. They are time-reversible laws, in the sense that time appears as a variable, just like position, without any privileged direction. But, if macroscopic laws emerge from microscopic laws, what can make them so different? A possible explanation is that such fundamental laws are in fact not time-reversal. For example, it has been argued that the time-reversible Schrödinger equation is not the true law at a fundamental level, and that it should be substituted by laws from spontaneous collapse theories which, by construction, are time-irreversible [3]. Another explanation argues that real systems are never perfectly isolated, so that time-reversible fundamental laws, despite being the true laws, are not directly applicable [4]. And yet another argumentation claims that the evolution of a real system depends, apart from the true time- reversible microscopic laws, on the initial conditions which provide an irreversible time evolution [6]. In this paper we explore a different path, via weak values in the configuration space, to understand the physical origins of the arrow of time in perfectly isolated nonrelativistic systems described by the Schrödinger equation, assumed as a true reversible law where initial conditions are not relevant to explain the observed time-irreversibility in our results. Our goal is to link the evolution of microscopic (or macroscopic) properties of a model system with the configuration (or physical) space where such properties are defined. The wave function solution of the Schrödinger equation is defined in a $3N$-dimensional configuration space, while typical properties where time- irreversibility is observed, are defined in smaller spaces where some degrees of freedom of the configuration space are integrated. Thus, the question that motivates us is whether the presence or absence of an arrow of time in the time evolution of the properties of a quantum system is a consequence of defining them in the full configuration space or in a smaller space. For our goal, the renewed interest in statistical mechanics of closed quantum systems [9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24] provides the perfect scenario. Such interest has been generated by the successful experimental ability to isolate and manipulate bosonic [25, 26, 27, 28, 29] and fermionic [30, 31, 32, 33, 34] many-body systems built on ultra- cold atomic gases subjected to optical lattices. Such quantum systems can be described by the many-particle Schrödinger equation; an arrow of time appears because, despite these systems being expected to present unitary evolution, some of their initial nonequilibrium nonthermalized properties may later thermalize. A preliminary consideration is that it is not at all obvious whether the configuration space is more or less fundamental than the ordinary physical space. The nonrelativistic Schrödinger equation can be seen as a sort of approximation to the relativistic quantum field theory [35]. For indistinguishable particles, quantum field theory does not require knowledge of the exact position of each particle in the configuration space, but only of how many particles are present in a position of the physical space and, as such, configuration space seems less fundamental than physical space. Another consideration is that thermalization is typically reported in properties mensurable in a laboratory, so that one needs to discuss, within similar empirical protocols, nonthermalized properties also mensurable in a laboratory. But making conclusions testable in a laboratory opens new difficulties since, strictly speaking, a measured closed system is no longer a closed system because of its interaction with the measuring apparatus [36]. Such measurement produces a collapse of the quantum state of the isolated system, which is at the origin of quantum randomness and backaction. Within the orthodox theory, such a collapse requires a time-asymmetric law, different from the time-symmetric Schrödinger equation, so that an “orthodox quantum- mechanical arrow of time” seems to enter into play [37]. We will see how the weak values protocol provides operational properties of the quantum system without backaction or quantum randomness (avoiding the role of the collapse in our discussion). It is important to differentiate between operational (empirical) and ontological (real) properties of a system. Operational properties are those whose definition comes exclusively from operations done on the laboratory over the system (with or without ontological meaning for that property), which are defined independently on any quantum theory. Ontological properties, on the other hand, are those that a specific quantum theory postulates to exist. Therefore, it is possible that a given operational property coincides with an ontological property in one theory, but not in another. A typical example is the velocity of a particle, known to be an operational property computed from weak values, independently on any quantum theory; such operational property coincides with an ontological property in the Bohmian theory (the velocity itself), but it is not an ontological property in the Orthodox theory. It is far from our scope to discuss whether or not thermalization is an ontological phenomenon or not occurring in the configuration space or in the physical space, since this depends on which quantum theory is invoked. Our less controversial focus here is to approach thermalization in closed systems from operational properties testable in laboratory. The structure of the paper is the following. Sect. II presents the many-body generalization of the single-particle weak values, stating them as operational properties in configuration space without explicit dependence on quantum randomness and backaction, for both distinguishable and indistinguishable particles. Sect. III discusses thermalization and equilibration concepts as found in the literature for closed quantum systems, and address the eigenstate thermalization hypothesis [38, 39]. Sect. IV defines our model system and its nonequilibrium dynamics, and summarizes our results for both expectation values and weak values from the Schrödinger equation dynamics. In Sect. V we conclude. ## II Operational properties in the configuration space To simplify notation, along the paper we use natural units and consider a $1$-dimensional physical space with degree of freedom $x$, so that $\mathbf{x}=\\{x_{1},...,x_{N}\\}$ is the position in the $N$-dimensional configuration space; the extension to a $3N$-dimensional space should be straightforward. As already mentioned, a measured closed system is no longer a closed system, and a strong measurement, for example, of the momentum operator $\hat{p}$ yields the eigenvalue $p$, and produces the initial state to be converted into the momentum eigenstate. On the other hand, expectation values and weak values yield operational information of the system without backaction and quantum randomness. ### II.1 Expectation values We define the expectation value of the momentum $\langle p(t)\rangle$ as an operational property of the system in the sense that it is linked to a well defined protocol in the laboratory, $\langle p(t)\rangle=\int dp\;p\;\mathbb{P}(p,t),$ (1) where the probability distribution $\mathbb{P}(p,t)$ can be obtained as follows: i) many identical initial states $|\Psi(t)\rangle$ are prepared at time $t$; ii) for each initial state, a (weak or strong) measurement of momentum is done, yielding the value $p$ at time $t$; iii) $\mathbb{P}(p,t)$ is constructed by counting how many $p$ occurs when repeating ii) on ensemble i). When then applying Born law to predict the value of $\mathbb{P}(p,t)$ one can easily identifies $\langle p(t)\rangle=\int dp\;p\;\mathbb{P}(p,t)=\langle\Psi(t)|\hat{p}|\Psi(t)\rangle.$ (2) Notice that the right hand side of (2) depends on the state of the system $|\Psi(t)\rangle$ before a measurement is done, without neither randomness nor backaction. In fact, $\langle p(t)\rangle$ is a typical property used to analyze when an isolated quantum system thermalizes. We are here interested in discussing thermalization from operational properties of the isolated quantum system requiring the measurement of both momentum and position simultaneously. Let us then start by discussing weak values in physical space. ### II.2 Weak values in the physical space It has recently been shown that weak values [40] are able to yield dynamic information on two noncommuting operators at a single time avoiding backaction [41, 42] and quantum randomness. Weak values have attracted a lot of theoretical [41, 42, 43, 44, 45, 46] and experimental [47, 48, 49] interests in many research fields. At the laboratory the single-particle weak value of momentum $p_{W}(x,t)$ is given by $p_{W}(x,t)=\frac{\int dp\;p\;\mathbb{P}(p,x,t)}{\int dp\;\mathbb{P}(p,x,t)},$ (3) computed from the probability distribution $\mathbb{P}(p,x,t)$ via the following procedure: i) many identical initial states $|\Psi(t)\rangle$ are prepared at time $t$; ii) for each initial state, a weak measurement of the momentum is done, yielding the value $p$ at time $t$; iii) subsequently, a strong measurement of the position is done, yielding the value $x$ at time $t$; iv) $\mathbb{P}(p,x,t)$ is constructed by counting how many $p$ and $x$ occurs when repeating ii) and iii) on ensemble i). Since $x$ is post-selected in (3), but not integrated out, one gets information on how the expectation value of the momentum is distributed in the physical space. Again, when applying Born law to predict the value of $\mathbb{P}(p,x,t)$ one can easily identifies $p_{W}(x,t)=\frac{\int dp\;p\;\mathbb{P}(p,x,t)}{\int dp\;\mathbb{P}(p,x,t)}=\text{Real}\left(\frac{\langle x|\hat{p}|\Psi(t)\rangle}{\langle x|\Psi(t)\rangle}\right).$ (4) Similarly to (2), the ensemble-over-identical-experiments in the right hand side in (4) eliminates the undesired backaction and quantum randomness induced by the first measuring apparatus [40, 42, 46, 50, 51], so that $p_{W}(x,t)$ in (4) depends only on the initial state before the measurements took place. Expression (4) allows us to give the weak value a very simple interpretation. In a single-particle system, from $\mathbb{I}=\int dx|x\rangle\langle x|$, one can rewrite the expectation value of the momentum in (2) (see Appendix A) as $\langle p\rangle(t)=\int dx\langle\Psi(t)|x\rangle\langle x|\hat{p}|\Psi(t)\rangle=\int dx|\Psi(x,t)|^{2}p_{W}(x,t),$ (5) so that the same probability distribution $\mathbb{P}(p,x,t)$ used to compute $p_{W}(x,t)$ in (4) can be employed to obtain $\mathbb{P}(p,t)$ used to compute $\langle p(t)\rangle$ in (2), since $\mathbb{P}(p,t)=\int dx\;\mathbb{P}(p,x,t).$ (6) Using the mathematics (and not necessarily the ontology) of Bohmian mechanics, one can also re-interpret $p_{W}(x,t)$ as the (operational) velocity of the particle at position $x$ and time $t$ [43, 44, 45, 46, 52, 53, 54, 55, 56, 57], $p_{W}(x,t)=\text{Imag}\left(\frac{1}{\Psi(x,t)}\frac{\partial\Psi(x,t)}{\partial x}\right)=\frac{J(x,t)}{|\Psi(x,t)|^{2}},$ (7) with $J(x,t)=\text{Imag}(\Psi^{*}(x,t)\partial\Psi(x,t)/\partial x)$ the current density (see Appendix B). ### II.3 Weak values in the configuration space for distinguishable particles Notice that $p_{W}(x,t)$ is an operational property in the ordinary physical space, but we now need to define an operational property in the configuration space for dealing with an isolated quantum system with $N$ particles. Therefore, in this paper, we extend the original single-particle weak values in (4) to $N$-particle scenarios for both distinguishable and indistinguishable cases. For the former case, the weak values for the $j$-particle is $p_{W}^{j}(\mathbf{x},t)=\frac{\int dp_{j}\;p_{j}\;\mathbb{P}(p_{j},\mathbf{x},t)}{\int dp_{j}\;\mathbb{P}(p_{j},\mathbf{x},t)},$ (8) where now the probability distribution $\mathbb{P}(p_{j},\mathbf{x},t)$ is obtained as follows: i) many identical initial states $|\Psi(t)\rangle$ are prepared at time $t$; ii) for each initial state, a weak measurement of the momentum of the $j$-particle is done, yielding the value $p_{j}$ at time $t$; iii) subsequently, a strong measurement of the positions of particles $1$,$2$,…,$N$ yielding respectively the values $x_{1}$,$x_{2}$,…,$x_{N}$ is done; iv) $\mathbb{P}(p_{j},\mathbf{x},t)$ is constructed by counting how many $p_{j}$,$x_{1}$,$x_{2}$,…,$x_{N}$ occurs when repeating ii) and iii) on ensemble i). Our definition of distinguishable particles above is operational in the sense that the measuring apparatus is somehow able to distinguish particles, for example, by measuring their masses but, to avoid unnecessary notation, we have not indicated this extra measurement in the above protocol. Again Born law allows us to rewrite (8) as $p_{W}^{j}(\mathbf{x},t)=\frac{\int dp_{j}\;p_{j}\;\mathbb{P}(p_{j},\mathbf{x},t)}{\int dp_{j}\;\mathbb{P}(p_{j},\mathbf{x},t)}=\text{Real}\left(\frac{\langle\mathbf{x}|\hat{p}_{j}|\Psi(t)\rangle}{\langle\mathbf{x}|\Psi(t)\rangle}\right),$ (9) where once more from the mathematics (and not necessarily from the ontology) of Bohmian mechanics, one can also re-interpret $p_{W}^{j}(\mathbf{x},t)$ as the (operational) velocity of the $j$-particle at the position $\mathbf{x}$ in the configuration space and time $t$ [43, 44, 45, 46, 52, 53, 54, 55, 56, 57], $p_{W}^{j}(\mathbf{x},t)=\frac{J^{j}(\mathbf{x},t)}{|\Psi(\mathbf{x},t)|^{2}},$ (10) with $J^{j}(\mathbf{x},t)=\text{Imag}(\Psi^{*}(\mathbf{x},t)\partial\Psi(\mathbf{x},t)/\partial x_{j})$ the current density in the $x_{j}$ direction. ### II.4 Weak values for indistinguishable particles The most common situation in the laboratory however relates to identical particles, for which a proper many-body wave function should implicitly include the exchange symmetry among the particles. Equation (9) then becomes inaccessible in a laboratory because it is no longer possible to know, operationally, which position belongs to each particle. To deal with indistinguishable particles, one needs to construct a many-body weak value defined in physical space coordinate $x$ by averaging (integrating) all degrees of freedom (see Appendix A). By doing so one obtains $\tilde{p}_{W}(x,t)=\frac{\int dp\;p\;\mathbb{\tilde{P}}(p,x,t)}{\int dp\;\mathbb{\tilde{P}}(p,x,t)},$ (11) where the probability distribution $\mathbb{\tilde{P}}(p,x,t)$ is now obtained as follows: i) many identical initial states $|\Psi(t)\rangle$ are prepared at time $t$; ii) for each initial state, a weak measurement of the momentum of one nonidentified particle is done, yielding the value $p$ at time $t$; iii) subsequently, a strong measurement of the position of the same or another nonidentified particle is done, yielding the value $x$ at time $t$; iv) $\mathbb{\tilde{P}}(p,x,t)$ is constructed by counting how many $p$ and $x$ occurs when repeating ii) and iii) on ensemble i). Born law again allows us to rewrite (11) as (see Appendix A) $\tilde{p}_{W}(x,t)=\frac{\int dp\;p\;\mathbb{\tilde{P}}(p,x,t)}{\int dp\;\mathbb{\tilde{P}}(p,x,t)}=\frac{1}{N^{2}}\sum_{j=1}^{N}\sum_{k=1}^{N}p_{W}^{j,k}(x,t),$ (12) with $p_{W}^{j,k}(x,t)=\frac{\int dx_{1}...\int dx_{k-1}\int dx_{k+1}...\int dx_{N}\;p_{W}^{j}(..,x_{k-1},x,x_{k+1},..,t)|\Psi(..,x_{k-1},x,x_{k+1},..,t)|^{2}}{\int dx_{1}...\int dx_{k-1}\int dx_{k+1}...\int dx_{N}|\Psi(..,x_{k-1},x,x_{k+1},..,t)|^{2}}.$ (13) Notice that $\tilde{p}_{W}(x,t)$ is, in fact, the local velocity as used in quantum hydrodynamic models [52, 53, 54, 55], being empirically accessible in both distinguishable and indistinguishable scenarios. The operational protocol for computing $\mathbb{\tilde{P}}(p,x,t)$ for indistinguishable particles is related to $\mathbb{P}(p_{k},\mathbf{x},t)$ for distinguishable particles as $\mathbb{\tilde{P}}(p,x,t)=\frac{1}{N^{2}}\sum_{j=1}^{N}\sum_{k=1}^{N}\int dx_{1}...\int dx_{k-1}\int dx_{k+1}...\int dx_{N}\;\mathbb{P}(p_{j},..,x_{k-1},x,x_{k+1},..,t).$ (14) ## III Thermalization in isolated systems from expectation values Our main contribution in the study of quantum thermalization, as detailed in next section, is the inclusion of many-body weak values in the configuration space as operational properties. However, such a study has usually been done in the literature in terms of expectation values as in (2), and as such we summarize in this section the role of expectation values to characterize thermalization. For an initial nonequilibrium pure state $|\Psi(0)\rangle$, the Schrödinger equation provides its unitary evolution as $|\Psi(t)\rangle=\sum_{n}c_{n}e^{-iE_{n}t}|n\rangle$, where $|n\rangle$ is an energy eigenstate with eigenvalue $E_{n}$, and $c_{n}=\langle n|\Psi(0)\rangle$ is defined by the initial conditions. The expectation value of some observable $\hat{A}$ is given by $\langle A\rangle(t)=\sum_{n}\rho_{n,n}A_{n,n}+\sum_{n,m\neq n}\rho_{n,m}(t)A_{m,n},$ (15) with the time-dependent off-diagonal elements of the density matrix in the energy representation defined as $\rho_{n,m}(t)=c_{m}^{*}c_{n}e^{i(E_{m}-E_{n})t},$ (16) and the time-independent diagonal elements as $\rho_{n,n}(t)=c_{n}^{*}c_{n}=\rho_{n,n}(0),$ (17) with the operator $\hat{A}$ in the energy representation being $A_{m,n}=\langle m|\hat{A}|n\rangle.$ (18) A system is said to equilibrate if, after some time $t_{eq}$ enough for full dephasing between different energy eigenstates, the off-diagonal terms (coherences) cancel out so that (15) can be computed solely from the time- independent diagonal terms (populations), that is, $\langle A\rangle(t)\approx\sum_{n}\rho_{n,n}A_{n,n}$ for most times $t>t_{eq}$ (except for some recurrence times). The properties of the system after equilibration are fully determined by the initial conditions $\rho_{n,n}(0)=|c_{n}|^{2}$, since the density matrix populations are time-independent. The closed system is then said to thermalize when $\langle A\rangle(t)$ becomes roughly equal to the expectation value as computed from the classical density matrix in the microcanonical ensemble, $\rho_{cl}$, in which equal probabilities are attached to each microstate within an energy window defined by the initial conditions; that is, a subset $N_{act}$ of the energy eigenstates are initially activated at $t=0$, and they will remain as the only states dictating the system dynamics at any $t$. It is important to notice that $N_{act}$ refers to a given number of relevant elements in a basis set, but it has no relation with the number of particles $N$ of the system. In other words, one can also expect thermalization even in few-particle systems, as detailed in next section; in fact, thermalization has also been studied in laboratories in small systems with as little as 6 [27], 5 [14], or 2-4 bosons [58, 59], 3 qubits [60], and even single-particle systems [61, 62, 63]. The eigenstate thermalization hypothesis (ETH) [38, 39] has become the standard theory dealing with quantum thermalization in closed systems. It states that the dephasing above mentioned is typical to nondegenerate and chaotic many-body nonintegrable systems, where the off-diagonals $A_{m,n}$ in (18) become exponentially smaller than $A_{n,n}$. In recent years a large amount of numerical experiments has successfully tested such a hypothesis by directly diagonalizing some sort of short range many-body lattice Hamiltonian, like Fermi- or Bose-Hubbard [26, 31, 32, 64, 65, 66], and XXZ- or XYZ- Heisenberg [14, 23, 27, 67, 68, 69, 70], in the search of chaotic signatures in the statistics of their spectra, as in general induced by local impurities, without the need to explicitly evolve $|\Psi(0)\rangle$. The ETH states that nonintegrable systems (where total energy may be the only conserved quantity), after a quench (which may create a nonequilibrium initial state by activating a subset $N_{act}$ of excited eigenstates), can present a ‘chaotic’ spectrum ruled by the Wigner-Dyson statistics (which contains level repulsion), and that the long time average of the expectation value of some observable roughly equals its thermal equilibrium value in an (microcanonical) ensemble. Such a hypothesis claims that thermalization is indeed hidden in the chaotic initial nature of the Hamiltonian eigenstates themselves. Lattice models typically handle $\approx 24$ sites with $\approx 1/3$ filling, and the above local impurities are added to break their otherwise integrable character, as the ETH overall claims that integrable systems are not expected to thermalize. On the other hand, our time evolution dwells in true configuration space with an antisymmetrized wave function and full long range electron-electron interaction; due to the tensorial nature of our problem, and since we need to employ a grid with $M\approx 10^{3}$ points per degree of freedom for decent position and momentum resolutions, we can realistically only deal with $N\lesssim 4$ particles ($N=3$ already implies $M\approx 10^{9}$ grid points at each time step). ## IV Numerical results for expectation values and weak values We now apply the many-body weak values machinery for the analysis of quantum thermalization in a model with $N$ spinless electrons [20, 71, 72, 73, 74, 75, 76, 77], typical of condensates in harmonic oscillator traps under a speckle field. Such a field translates to a ‘chaotic’ random disorder potential, which can yield a chaotic energy spectrum as requested by the ETH and so induce thermalization even in systems with small $N$. Our model can attach a disorder to each grid point, and an initial velocity at $t=0$ is given to the electrons as to simulate the initial quench of the confining potential, for each of the considered $N=1,2,3$ systems. In fact, most of the thermalization literature dealing with identical particles employs some lattice-based model, since such models avoid the need of explicit knowledge of the exact position of each particle in a point of the configuration space; instead, they only need to know how many particles are present in each site of the physical space. From a computational point of view lattice models have unquestionable advantages but are inappropriate for our goal of discussing whether or not thermalization occurs in configuration space, a goal that forces us to directly solve the time-evolution of the few-body Schrödinger equation in configuration space, and to analyze thermalization by monitoring the time-evolution of both expectation values and weak values. ### IV.1 Initial state The pure initial $N$-electron nonequilibrium antisymmetric state is $\langle\mathbf{x}|\Psi(0)\rangle=\frac{1}{\mathcal{C}}\sum_{n=1}^{N!}\text{sign}(\vec{p}(n))\prod_{j=1}^{N}\psi_{j}(x_{p(n)_{j}},0),$ (19) with $\mathcal{C}$ a normalization constant and $\text{sign}(\vec{p}(n))$ the sign of the permutation $\vec{p}(n)=\\{p(n)_{1},..,P(n)_{N}\\}$. Each initial Gaussian state in (19) is $\psi_{j}(x)=\exp\left[-\frac{(x-x_{0j})^{2}}{2\sigma_{j}^{2}}\right]\exp\left[ip_{0j}(x-x_{0j})\right],$ (20) with spatial dispersion $\sigma_{j}$, central position $x_{0j}$, and central velocity $p_{0j}$. The dynamical evolution of $|\Psi(t)\rangle$ is determined by the Schrödinger equation, $i\partial|\Psi(t)\rangle/\partial t=\hat{H}|\Psi(t)\rangle$, where the Hamiltonian $\hat{H}$ is described in next section. ### IV.2 Full Hamiltonian The $N$-electron Hamiltonian of our model system is $\hat{H}=\hat{H}_{0}+\hat{D}$, with $\hat{H}_{0}=\sum_{j=1}^{N}\left[\hat{k}_{j}+\hat{v}_{j}+\sum_{k<j}^{N}\hat{e}_{k,j}\right],\;\;\;\;\;\;\hat{D}=\sum_{j=1}^{N}\hat{d}_{j}.$ (21) In $\hat{H}_{0}$, $\langle x_{j}|\hat{e}_{k,j}|x_{k}\rangle=1/\sqrt{(x_{j}-x_{k})^{2}+\alpha^{2}}$ takes care of the Coulomb repulsion with a smooth parameter $\alpha$, $\langle x_{j}|\hat{k}_{j}|x_{j}\rangle=-\partial^{2}/(2\partial x_{j}^{2})$ stands for the kinetic energy, and $\langle x_{j}|\hat{v}_{j}|x_{j}\rangle=\omega^{2}x_{j}^{2}/2$ is the harmonic trap potential. On the other hand, $\hat{D}$ introduces random disorder at every grid point, where $\langle x_{j}|\hat{d}_{j}|x_{j}\rangle=\gamma_{D}\sum_{k=1}^{M}a_{k}\exp[-4(x_{j}-g_{k})^{2}/\sigma_{D}^{2}]$, with $\gamma_{D}$ its strength and $\sigma_{D}$ its spatial dispersion, while $g_{k}$ runs through $M$ grid points; the set of random numbers $a_{k}$ satisfies $\langle a_{k}\rangle=0$ and $\langle a_{k}^{2}\rangle=1$, and the disorder potential is normalized via $\int\langle x_{j}|\hat{d}_{j}|x_{j}\rangle^{2}dx_{j}=\gamma_{D}^{2}$. Such random disorder can be mapped onto speckle field potentials typical in some optical lattice experiments. All simulation parameters are found in [78]. Figure 1: (a): Zoom of a disordered harmonic potential for $N=1$, $\langle x_{1}|\hat{v}_{1}+\hat{d}_{1}|x_{1}\rangle$, from a larger simulation box. (b): Successive energy splittings $\Delta E=E_{n}-E_{n-1}$ from a grid diagonalization of the respective Hamiltonian $\hat{H}$. (d): zoom around the peak of the corresponding density matrix modulus $|\rho_{n,m}(t=0)|$, which remains the same at any time $t$ in an unitary evolution. (c): initial positions $x_{0j}$ for $N=1,2,3$, and initial velocity $p_{0j}$ which is the same in every case. Figure 1(a) exemplifies, for $N=1$, a typical shape of the disordered harmonic potential, $\langle x_{1}|\hat{v}_{1}+\hat{d}_{1}|x_{1}\rangle$, while figure 1(b) shows the corresponding successive energy splittings $\Delta E=E_{n}-E_{n-1}$, as obtained from a direct diagonalization of $\hat{H}$, which oscillate around the pure harmonic oscillator value of $1$. Figure 1(d) shows the related shape of the density matrix modulus $|\rho_{n,m}(t=0)|$, from where one identifies $N_{act}\approx 70$ (counting every level above $10\%$ of peak value), while figure 1(c) shows initial positions $x_{0j}$, $j=1...N$, and the initial velocity $p_{0}$, the same for any $j$ and $N$, for the $N=1,2,3$ systems. Notice that $p_{0}$ not only takes the role of simulating the initial quench of the trap and so to initiate the nonequilibrium dynamics, but it is also responsible for determining the size of the energy window and so the value of $N_{act}$. The initial energy $E_{0}$ of the wave packet defines, in the language of the ETH, the center of the microcanonical energy window; since $E_{0}\approx\omega+p_{0}^{2}/2=201$, the peak is at $n=m=201$. As mentioned in (15), the diagonal terms are time- independent and, although real and imaginary parts of the off-diagonals terms oscillate in time, their modulus remain constant so that, in an unitary evolution, the modulus seen in figure 1(d) remains the same at any $t$ (see Appendix B). Both initial wave packet and Hamiltonian in (19)-(21) have many adjustable parameters in [78] upon which the value of $t_{eq}$ depends on: (i) Coulomb correlation by varying $\omega$ or $\alpha$; (ii) initial Gaussian by varying $\sigma_{j}$, $x_{0j}$, or $p_{0j}$; (iii) disorder potential by varying $\gamma_{D}$ or $\sigma_{D}$. All of them play a role in determining the $\Delta E$ splittings of the involved $N_{act}$ activated eigenstates, which is what drives the nonequilibrium dynamics of both expectation values and weak values. It is not our goal to fully characterize the equilibration process as a function of all those parameters. Neither to address the topic of many-body localization [79, 80, 81], which should work against thermalization, nor to go deeper in the issue of quantum-to-classical transition at $t\gg t_{eq}$; these two latter issues are beyond the scope of our work and are focus of extensive studies elsewhere. The main goal of our paper is to present a distinct perspective in the understanding of quantum thermalization by looking at the many-body weak values of the momentum in the configuration space. Each numerical experiment corresponds to a static realization of a disorder pattern; $t_{eq}$ hardly changes among different runs, given that all other model parameters remain unchanged. The disorder amplitude should not be too strong, to avoid localization due to all small wells created on the top of the trap, neither too weak, to avoid too long simulation times until reaching $t_{eq}$. We emphasize that even in the absence of electron-electron collision, disorder collision is able to create correlation among distinct degrees of freedom in configuration space when $N>1$. ### IV.3 Time evolution Figure 2: Wave packet dynamics. Upper panels for $N=3$: initial $|\Psi(x_{1},x_{2},[x_{3}],t=0)|^{2}$ (a) and final $|\Psi(x_{1},x_{2},[x_{3}],t=150)|^{2}$ (b) shapes. Middle panels for $N=2$: initial $|\Psi(x_{1},x_{2},t=0)|^{2}$ (c) and final $|\Psi(x_{1},x_{2},t=150)|^{2}$ (d) shapes. Lower panels: (e) for $N=1$: initial (dotted) $|\Psi(x_{1},t=0)|^{2}$ and final (solid) $|\Psi(x_{1},t=150)|^{2}$ shapes; (f): $1$D-view for the three systems: $|\Psi(x_{1},[x_{2}],[x_{3}],t)|^{2}$ for $N=3$ (red), $|\Psi(x_{1},[x_{2}],t)|^{2}$ for $N=2$ (blue), and $|\Psi(x_{1},t)|^{2}$ for $N=1$ (green), with solid (dotted) lines for $t=150$ ($t=0$); also shown the classical microcanonical harmonic oscillator distribution $\rho_{cl}(x_{1})$ (black dashed line). All plots are in log-scale, and the horizontal axis in (a),(c) ((b),(d)) is the same as in (e) ((f)). Figure 2 plots in configuration space the time evolution of the $N$-electron wave function for the three $N=1,2,3$ systems from the initial nonequilibrium state. We show $|\Psi(x_{1},x_{2},[x_{3}],t)|^{2}$ for $N=3$ and $|\Psi(x_{1},x_{2},t)|^{2}$ for $N=2$ at initial ($t=0$, panels (a),(c)) and final ($t=150$, panels (b),(d)) simulation times, while panel (e) shows $|\Psi(x_{1},t)|^{2}$ for $N=1$ for both initial ($t=0$, dotted) and final ($t=150$, solid) simulation times; the notation $[x_{j}]$ means that the degree $j$ is integrated out, with results independent on chosen $j$ due to the antisymmetry of the problem. The initially localized wave packets fully spread out due to random scattering generated by both disorder potential and Coulomb repulsion, with such spreading becoming more homogeneous as $N$ increases. In panel (f) we show the respective $1$D plots of $|\Psi(x_{1},[x_{2}],[x_{3}],t)|^{2}$ ($N=3$, red), $|\Psi(x_{1},[x_{2}],t)|^{2}$ ($N=2$, blue), and $|\Psi(x_{1},t)|^{2}$ ($N=1$, green), with initial (final) results at dotted (solid) lines; the probabilities at large $t$, as one expects to have crossed $t_{eq}$, approach the microcanonical distribution of a classical harmonic oscillator (dashed black line), $\rho_{cl}(x)=[\pi l_{0}\sqrt{p_{0}^{2}-x^{2}/l_{0}^{2}}]^{-1}$, so that the dynamics develops in between the classical turning points at $x_{TP}=\pm 20$ ($=p_{TP}$ since $\omega=1/l_{0}^{2}=1$); the deviation from $\rho_{cl}(x)$ decreases as $N$ increases. In Appendix C we analyze the same nonequilibrium dynamics in momentum representation. ### IV.4 Thermalized expectation values Figures 3 and 4 show time evolution and thermalization of some typical expectation values $\langle A\rangle(t)=\langle\Psi(t)|\hat{A}|\Psi(t)\rangle$, for $N=1,2,3$ respectively in panels (a), (b), (c). On one hand, figure 3 focus on energy terms normalized by $N$: kinetic $\langle K\rangle(t)$, with $\hat{K}=\sum_{j}\hat{k}_{j}$, potential $\langle V\rangle(t)=\langle V_{HO}\rangle(t)+\langle V_{D}\rangle(t)+\langle V_{Cou}\rangle(t)$, with $\hat{V}_{HO}=\sum_{j}\hat{v}_{j}$, $\hat{V}_{D}=\sum_{j}\hat{d}_{j}$, and $\hat{V}_{Cou}=\sum_{j,k<j}\hat{e}_{k,j}$, and half of total energy $\langle E\rangle(t)=\langle K\rangle(t)+\langle V\rangle(t)$. On other hand, figure 4 focus on position $\langle x_{j}\rangle(t)$ and momentum $\langle p_{j}\rangle(t)$ for the $j$-electron, and on their RMS values $z_{j,RMS}(t)=\sqrt{\langle z^{2}_{j}\rangle(t)-\langle z_{j}\rangle^{2}(t)}$, with $z=x,p$, and results independing on chosen $j$. _Without_ random disorder, such expectation values would only exhibit harmonic oscillations with period $2\pi/\omega$: while $\langle x_{j}\rangle(t)$ and $\langle p_{j}\rangle(t)$ would respectively oscillate within position $\pm x_{TP}=\pm 20$ and momentum $\pm p_{TP}=\pm 20$ turning points, $\langle V\rangle(t)/N$ and $\langle K\rangle(t)/N$ would respectively oscillate within $0$ and $x_{TP}^{2}/2=200$ and within 0 and $p_{TP}^{2}/2=200$; these latter values increase a little upon $N$ due to Coulomb repulsion, whose isolated contribution is shown ($100\times$-magnified) in figure 3. Figure 3: Energy expectation values from the dynamics in figure 2. Panels (a), (b), (c) respectively for $N=1,2,3$. Kinetic $\langle K\rangle(t)$, potential $\langle V\rangle(t)$, half of total energy $\langle E\rangle(t)=\langle K\rangle(t)+\langle V\rangle(t)$, and isolated contribution of $\langle V_{Cou}\rangle(t)$ ($100\times$-magnified) are all normalized by $N$. Inset for $N=2$ shows a longer propagation time, $t=[150-300]$. Legend in (a) and horizontal axis in (c) apply to all panels. The _presence_ of random disorder, even though $\langle V_{D}\rangle(t)\approx 0$ at any $t$, brings the initial nonequilibrium state into a final _equilibrium_ state after a relaxation time $t_{eq}$. We know from the discussions in (15) and in figure 1(d) that thermalization of an observable $\hat{A}$ is determined by the diagonal populations of the density matrix, while its off-diagonal coherences should dephase and only yield small fluctuations around the relaxed value (see Appendix B). So one may estimate the value of $t_{eq}$ either from figure 3, when the virial theorem $\langle K\rangle\approx\langle V\rangle\approx\langle E\rangle/2$ is roughly satisfied (since $\langle V_{Cou}\rangle\ll\langle V_{HO}\rangle$) [74], or from figure 4, when $\langle p_{j}\rangle\approx\langle x_{j}\rangle\approx 0$ seemingly indicating a frozen dynamics after thermalization. From this latter result the RMS values become $z_{j,RMS}(t)\approx\sqrt{\langle z^{2}_{j}\rangle(t)}$, becoming also constant in figure 4 at $t>t_{eq}$ (from $\approx 14.2$ for $N=1$ to $\approx 14.4$ for $N=3$), so that the values of $p^{2}_{j,RMS}/2=\langle p^{2}_{j}\rangle/2$ and $x^{2}_{j,RMS}/2=\langle x^{2}_{j}\rangle/2$ roughly yield the respective values of $\langle K\rangle/N$ and $\langle V\rangle/N$ at $t>t_{eq}$ in figure 3. As expected for an unitary evolution, $\langle E\rangle(t)/N$ remains conserved at any $t$, from $\approx 201$ for $N=1$ to $\approx 207$ for $N=3$. Figure 4: Position $\langle x_{j}\rangle(t)$ and momentum $\langle p_{j}\rangle(t)$ expectation values from the dynamics in figure 2. Panels (a), (b), (c) respectively for $N=1,2,3$. Their respective RMS values, as defined in the text, are also shown, where results do not depend on chosen $j$. Inset for $N=2$ shows a longer propagation time, $t=[150-300]$. Legend in (a) and horizontal axis in (c) apply to all panels. The value of $t_{eq}$ depends on all parameters [78] in (19)-(21), e.g., the smaller is $p_{0j}$ or the higher is $\gamma_{D}$ the smaller is $t_{eq}$. By increasing the influence of $\langle V_{Cou}\rangle$ in comparison to $\langle K\rangle$ (by decreasing $\omega$ or $\alpha$), $t_{eq}$ increases since the oscillation period and so $x_{TP}$ increases. The plots of $\langle V_{Cou}\rangle(t)$ in figure 3 show that the Coulomb repulsion is more effective at the turning points for $t\ll t_{eq}$, where electrons spend more time reversing their movements, while the disorder potential overall acts through a whole oscillation, although one may take it as more effective at the origin. Coulomb correlation has a striking influence on the thermalization process: in configuration space the only scattering mechanism for $N=1$ is due to disorder, while for $N>1$ Coulomb scattering also makes more difficult for the system to relax. This is seen by the slightly increasing values of $t_{eq}$ as one moves in figure 3 from (a) ($t_{eq}\approx 70$) to (b) ($t_{eq}\approx 80$) to (c) ($t_{eq}\approx 90$). The vanishing of $\langle p\rangle(t)$ in figure 4 seems more effective as one moves from (a) to (b) to (c) but, however, it does not necessarily imply that electrons have achieved a stationary-state null velocity at $t\gg t_{eq}$, as our following analysis on weak values of the momentum will clarify (see also ‘phase-space’ in Appendix C). ### IV.5 Nonthermalized weak values We can at last elaborate on how the many-body weak values of the momentum may improve our understanding on thermalization. In table 1 we summarize the five types of operational properties accessible for the three different types of quantum systems considered in this section: single-particle, distinguishable particles, indistinguishable particles. $N=1$ | $\langle p(t)\rangle$ (2) | $p_{W}(x,t)$ (4) | | | ---|---|---|---|---|--- $N>1$ (Dis) | $\langle p(t)\rangle$ (2) | | $p_{W}^{j}(\mathbf{x},t)$ (9) | $\tilde{p}_{W}(x,t)$ (12) | $p_{W}^{j,k}(x,t)$ (13) $N>1$ (Ind) | $\langle p(t)\rangle$ (2) | | | $\tilde{p}_{W}(x,t)$ (12) | Table 1: Five operational properties accessible in the laboratory for each of the three quantum systems considered in our work: (i) with $N=1$ particles, (ii) with $N>1$ distinguishable particles, and (iii) with $N>1$ indistinguishable particles. The plot in figure 5(a) corresponds to the quantum system $N=1$ in table 1. Although the expectation value $\langle p_{j}\rangle(t)$ seems to indicate that the quantum behavior at $t\gg t_{eq}$ roughly equals the behavior of a diagonal density matrix in (17), the weak values $p_{W}^{1,1}(0,t)$ (which obviously corresponds to $p_{W}(0,t)$ in (4)) certifies that the off-diagonal terms in (16) do not vanish after thermalization. For $N=1$, the fact that expectation values thermalize is just a result that positive and negative off- diagonals elements can roughly compensate each other, but they certainly do not disappear as indicated by $p_{W}^{1,1}(x,t)$. We remind that $\langle p\rangle(t)=\int dx\int dp\;p\;\mathbb{P}(p,x,t)$ and $p_{W}(x,t)$ can, both, be computed from the same empirical probability $\mathbb{P}(p,x,t)$. In other words, $\mathbb{P}(p,x,t)$ provides simultaneous thermalized and nonthermalized results depending on how it is treated. Figure 5: Local-in-position many-body weak values of the momentum from the dynamics in figure 2 for $N=1$ in (a), and for distinguishable $N=2$ particles in (b),(c) where Coulomb/exchange terms are disconsidered. Panels (a), (b) show $p_{W}^{j,j}(x,t)$ from (13), which does not depend on $j$. Panel (c) shows both $p_{W}^{j,k}(x,t)$ from (13) and $\tilde{p}_{W}(x,t)$ from (12). Values of $x$ are the respective initial $x_{0j}$ values. The expectation value of the momentum $\langle p_{j}\rangle(t)$ is also shown. Horizontal axis in (a),(b) the same as in (c). The plots in figures 5(b) and 5(c) correspond to the quantum system $N>1(\text{Dis})$ in table 1, because neither exchange nor Coulomb interaction among the particles are included; that is, the many-body wave function here could be written as $\Psi(\mathbf{x},t)=\psi_{1}(x_{1},t)\psi_{2}(x_{2},t)$. The weak values $p_{W}^{1,1}(0,t)$ and $p_{W}^{1,1}(-4,t)$ in panel (b) confirm the nonthermalized operational properties even at $t\gg t_{eq}$. In panel (c), one notices that $p_{W}^{1,2}(0,t)$, which corresponds to the weak measurement of the momentum of particle 1 and strong measurement of the position of particle 2, shows a thermalized behavior as it overlaps $\langle p_{j}\rangle(t)$; but this is because here, as the particles have no correlations among them, one gets $p_{W}^{j,k}(x,t)=\langle p_{j}\rangle(t)$ when $j\neq k$, as discussed in (32). Also in panel (c) $\tilde{p}_{W}(0,t)$, from (12), which here is just a particle-average of the oscillating term $p_{W}^{1,1}(0,t)$ and the non-oscillating term $p_{W}^{1,2}(0,t)$, presents less oscillations but still clearly showing a non-thermalized behavior of these operational properties. Figure 6: Local-in-position many-body weak values of the momentum from the dynamics in figure 2 for indistinguishable particles with $N=2$ in (a),(b) and $N=3$ in (c). Panels (a), (c) show $p_{W}^{j,j}(x,t)$ from (13), which does not depend on $j$, for the respective initial $x_{0j}$ values. Panel (b) shows both $p_{W}^{j,k}(x,t)$ from (13) and $\tilde{p}_{W}(x,t)$ from (12). The expectation value of the momentum $\langle p_{j}\rangle(t)$ is also shown. Inset in (a) shows a longer propagation time, $t=[150-300]$. Horizontal axis in (a),(b) the same as in (c). Figure 6 corresponds to the (most common) quantum system $N>1(\text{Ind})$ in table 1, in which one only has operational access in the laboratory to the expectation value $\langle p\rangle(t)$ in (2) and to the weak value $\tilde{p}_{W}(x,t)$ in (12). Panels (a) and (b) for $N=2$; panel (c) for $N=3$. Since we also have mathematical (not operational) access in our simulations (directly from configuration space) to the weak value $p_{W}^{j,k}(x,t)$, we have also plotted it to help us to understand the behavior of the operational weak value $\tilde{p}_{W}(x,t)$, which is a particle average over different $p_{W}^{j,k}(x,t)$. We see in Fig. 6(a) that $p_{W}^{1,1}(x,t)$ has smaller oscillations than in Fig. 5(b), but it still does not thermalize, while $\langle p_{j}\rangle(t)$ effectivelly thermalize (this later result is independent on $j$); the larger time window in the inset so confirms. In figure 6(b) one notices that both $p_{W}^{1,2}(0,t)$ and $p_{W}^{1,1}(0,t)$ have a similar non-thermalizing behavior. From these two latter values we obtain $\tilde{p}_{W}(0,t)=(p_{W}^{1,1}(0,t)+p_{W}^{1,2}(0,t))/2$ (thanks to the properties in (31)), whose plot in panel (b) shows that this operational parameter for identical particles does not fully thermalize. The $N=3$ case in panel (c) starts to show the trend that, at higher $N$, $p_{W}^{1,1}(x,t)$ will approach the expectation value $\langle p_{1}\rangle(t)$ and so will also thermalize, which is seem for all used $x_{0j}$ values. This is understood from the fact that, for identical particles, $p_{W}^{j,k}(x,t)$ (and $p_{W}^{j,j}(x,t))$ contains $N-1$ spatial integrals, while $\langle p_{j}\rangle(t)$ contains $N$ and so, as one increases $N$, one gets $p_{W}^{j,k}(x,t)\approx\langle p_{j}\rangle(t)$ because $N-1\approx N$. The fact that the weak value of the momentum does not thermalize in the configuration space can be understood when $p_{W}^{j}(\mathbf{x},t)$ in (9) is mathematically interpreted as a Bohmian velocity, satisfying all of its mathematical properties without any ontologic implications. Without external perturbation, the initial state $|\Psi(0)\rangle$ of a closed system cannot change with time its condition of being or not an energy eigenstate [36]. In other words, the only energy eigenstates are the ones that are so at all times. We know that for a closed system with $\Psi(\mathbf{x},t)$ vanishing at the boundaries, the only Bohmian velocities in (9) that are zero are those linked to an energy eigenstate [56]. From (10) we see that the weak value of the momentum depends on the current density, and a time-dependent $J^{j}(\mathbf{x},t)$ (strictly different from zero) is required for a time- dependent probability presence in such a space, due to the continuity equation in the configuration space implicit in the Schrödinger equation. Thus, since our initial nonequilibrium state is not an energy eigenstate, from all previous arguments, we conclude that $p_{W}^{j}(\mathbf{x},t)$ will never vanish in the configuration space no matter the value of $N$, independently on the thermalization or not of the expectation value of the momentum $\langle p_{j}\rangle(t)$; that is, $p_{W}^{j}(\mathbf{x},t)$ will never thermalize when evaluated at a point $\mathbf{x}$ of the configuration space (the same conclusions apply to the presence probability in such space). Most of the developments done in this paper come from the fact that, in most cases, the weak value $p_{W}^{j}(\mathbf{x},t)$ in the configuration space is not empirically accessible in the laboratory, and as such it is not an operational property for indistinguishable particles; in such cases, though, the weak value $\tilde{p}_{W}(x,t)$ in the physical space tends to thermalize as $N$ increases. ## V Conclusions We have seen how, from the empirical knowledge of a unique distribution probability $\mathbb{P}(p,\textbf{x},t)$, it is possible to get, simultaneously, both thermalized (expectation values) and nonthermalized (weak values) operational properties of a closed quantum system described by the Schrödinger equation. A possible origin of the arrow of time in such systems comes from dealing with operational properties obtained by averaging (tracing out) some of the degrees of freedom defined in configuration space. As such there is no contradiction in that there are properties defined in the configuration space that do not thermalize, while some properties defined in the physical space (by averaging or tracing out some degrees of freedom) have their own time irreversible equations of motion. Such conclusion can be tested in the laboratory through the many-body weak values of the momentum for distinguishable particles where, at least conceptually, it is possible to get information of the position of all particles in the laboratory after their strong measurement. However, for indistinguishable particles, a position measurement at $x$ cannot be linked to a specific particle since there is no experimental protocol to tag identical particles and, as such, instead of $\mathbb{P}(p,\textbf{x},t)$ in configuration space, one has operational access to $\mathbb{P}(p,x,t)$ in physical space. But we have shown that $\mathbb{P}(p,x,t)$ can be understood as an averaging of $\mathbb{P}(p,\textbf{x},t)$ over degrees of freedom in the configuration space. For the simpler single-particle case, where obviously the distinction between distinguishable/indistinguishable particles and between configuration/physical spaces makes no sense, we have also seen that expectation values can thermalize while weak values may not. The simple explanation is that the expectation value has one integration over the single valid coordinate while the weak value has not. Even for a system with $N=2$ identical particles, we see different behaviors for the expectation value $\langle p\rangle(t)$ and the weak value $\tilde{p}_{W}(x,t)$ in the physical space. In general, for identical particles, $\tilde{p}_{W}(x,t)$ contains $N-1$ spatial integrals, while $\langle p_{j}\rangle(t)$ contains $N$ and so, as one increases $N$, one gets $p_{W}^{j,k}(x,t)\approx\langle p_{j}\rangle(t)$ because $N-1\approx N$; that is, the former thermalizes when the latter also does. But such thermalization is seen in the physical space, not in the configuration space. So does thermalization occur in configuration space? No thermalization occurs in the most common operational properties of a quantum system when evaluated in a point $\mathbf{x}$ of the configuration space, even when other properties defined in simpler spaces are thermalized. Notice that we have avoided along the paper to discuss ontologic properties. Instead, our operational approach has the advantage that the conclusions do not depend on which quantum theory is invoked, but it also forbids the discussion whether or not thermalization is a real (ontologic) phenomenon occurring in physical space. This latter discussion depends on the ontology of each quantum theory; there are quantum theories where the configuration space is not the fundamental space. In summary, as mentioned in our Intro [1, 2, 3, 4, 5, 6, 7, 8], there are many arrows of time which could require different explanations. We here only discuss the origin of the arrow of time in the non-relativistic many-body Schrödinger equation in closed quantum systems: a non-thermalized property in the configuration space, when some of its degrees of freedom are averaged, leads to a thermalized property in the physical space. We have also developed the operational many-body weak values of the momentum to make such a conclusion testable in the laboratory. ###### Acknowledgements. This research was funded by Spain’s Ministerio de Ciencia, Innovación y Universidades under Grant No. RTI2018-097876-B-C21 (MCIU/AEI/FEDER, UE), Grant PID2021-127840NB-I00 (MICINN/AEI/FEDER, UE), the “Generalitat de Catalunya” and FEDER for the project 001-P-001644 (QUANTUMCAT), the European Union’s Horizon 2020 research and innovation programme under Grant No. 881603 GrapheneCore3 and under the Marie Skłodowska-Curie Grant No. 765426 TeraApps. ## Appendix A Weak values equations in many-body systems In this appendix we develop the main weak values equations in the paper. ### A.1 Development of equations (4) and (9) in the paper For a single-particle system described by the wave function $\psi(x,t)=\langle x|\psi(t)\rangle$, the expression for the weak values of the momentum, $p_{W}(x,t)$, can be obtained from the position distribution of the mean momentum $\langle p\rangle(t)$ of the single-particle operator, $\hat{p}=p|p\rangle\langle p|$, as $\langle p\rangle(t)=\langle\psi(t)|\left(\int dx|x\rangle\langle x\right)|\hat{p}|\psi(t)\rangle=\int dx|\psi(x,t)|^{2}\frac{\langle x|\hat{p}|\psi(t)\rangle}{\langle x|\psi(t)\rangle}=\int dx|\psi(x,t)|^{2}p_{W}(x,t),$ (22) where we have defined the weak values as $p_{W}(x,t)=\text{Real}\left(\frac{\langle x|\hat{p}|\psi(t)\rangle}{\langle x|\psi(t)\rangle}\right)$. Since $\langle p\rangle(t)$ is real, as $\hat{p}$ is an hermitian operator, we have $\int dx|\psi(x,t)|^{2}\text{Imag}\left(\frac{\langle x|\hat{p}|\psi(t)\rangle}{\langle x|\psi(t)\rangle}\right)=0$. Thus, only the real part is considered for defining the weak values and so we reproduce equation (4) in the paper. From a similar development we can find the weak values for an $N$-particle system, with $\Psi(\mathbf{x},t)=\langle\mathbf{x}|\psi(t)\rangle$. The mean momentum $\langle p_{j}\rangle(t)$ of degree of freedom $j$ belonging to operator $\hat{P}_{j}\equiv\hat{1}\otimes...\otimes\hat{p}_{j}\otimes...\otimes\hat{1}$ with $\hat{p}_{j}=p_{j}|p_{j}\rangle\langle p_{j}|$ is $\displaystyle\langle p_{j}\rangle(t)$ $\displaystyle=$ $\displaystyle\langle\Psi(t)|\hat{P}_{j}|\Psi(t)\rangle=\int d\mathbf{x}\langle\Psi(t)|\mathbf{x}\rangle\langle\mathbf{x}|\hat{P}_{j}|\Psi(t)\rangle=\int d\mathbf{x}\langle\Psi(t)|\mathbf{x}\rangle\langle x_{1}|\otimes...\otimes\left(\langle x_{j}|\hat{p}_{j}\right)\otimes...\otimes\langle x_{N}|\Psi(t)\rangle$ (23) $\displaystyle=$ $\displaystyle\int d\mathbf{x}\Psi^{*}(\mathbf{x},t)(-i)\frac{\partial\Psi(\mathbf{x},t)}{\partial x_{j}}=\int d\mathbf{x}|\Psi(\mathbf{x},t)|^{2}\text{Imag}\left(\frac{\frac{\partial\Psi(\mathbf{x},t)}{\partial x_{j}}}{\Psi(\mathbf{x},t)}\right)=\int d\mathbf{x}|\Psi(\mathbf{x},t)|^{2}p_{W}^{j}(\mathbf{x},t),$ where we have used $\int d\mathbf{x}=\int dx_{1}...\int dx_{N}$, $|\mathbf{x}\rangle=|x_{1}\rangle\otimes...\otimes|x_{N}\rangle$, and $\left(\langle x_{j}|\hat{p}_{j}\right)=\int dx_{j}^{\prime}\langle x_{j}|\hat{p}_{j}|x_{j}^{\prime}\rangle\langle x_{j}^{\prime}|=\langle x_{j}|(-i)\frac{\partial}{\partial x_{j}}$. This result shows that $\langle p_{j}\rangle(t)$ can be decomposed into different components along the positions $\mathbf{x}$ on the configuration space. Each component $p_{W}^{j}(\mathbf{x},t)$ is the many-body weak values of the momentum of the $j$-th particle, $p_{W}^{j}(\mathbf{x},t)=\text{Real}\left(\frac{\langle\mathbf{x}|\hat{P}_{j}|\Psi(t)\rangle}{\langle\mathbf{x}|\Psi(t)\rangle}\right)=\text{Imag}\left(\frac{\frac{\partial\Psi(\mathbf{x},t)}{\partial x_{j}}}{\Psi(\mathbf{x},t)}\right)=\frac{J^{j}(\mathbf{x},t)}{|\Psi(\mathbf{x},t)|^{2}},$ (24) with $|\Psi(\mathbf{x},t)|^{2}$ the probability of finding a particle at the given configuration position $\mathbf{x}$. Expression (24) so reproduces equation (9) in the paper. The last identity in (24) shows that the weak values of the momentum is just the Bohmian velocity of the $j$-th particle in the configuration position $\mathbf{x}$. If one deals with few particles well- separated in the physical space, then the measurement of the many-body weak values of each particle in the laboratory is unproblematic. ### A.2 Development of equation (13) in the paper The problem with the weak values in (24) appears in the laboratory when we consider $N$ particles in the same region of the physical space, since $p_{W}^{j}(\mathbf{x},t)$ depends on all positions of the $N$ particles. Then, it seems impossible for practical purposes to develop a measurement protocol identifying the $N$ positions of the particles simultaneously. Thus, we want to rewrite (23) in a way that it only depends on one position of one of the $N$ particles. We are interested in an expression for computing $\langle p_{j}\rangle(t)$ as a product of a probability in the physical space, $\mathbb{P}^{k}(x,t)$, by a weak values which is also local in the physical space, $p_{W}^{j,k}(x,t)$, which goes like $\langle p_{j}\rangle(t)=\int dx\mathbb{P}^{k}(x,t)\;p_{W}^{j,k}(x,t),$ (25) where $\mathbb{P}^{k}(x,t)=\int dx_{1}...\int dx_{k-1}\int dx_{k+1}...\int dx_{N}\;|\Psi(\mathbf{x},t)|^{2}.$ (26) From (23), (24), and (26), one exactly gets equation (13) in the paper, $p_{W}^{j,k}(x,t)=\frac{\int dx_{1}...\int dx_{k-1}\int dx_{k+1}...\int dx_{N}\;p_{W}^{j}(..,x_{k-1},x,x_{k+1},..,t)|\Psi(..,x_{k-1},x,x_{k+1},..,t)|^{2}}{\int dx_{1}...\int dx_{k-1}\int dx_{k+1}...\int dx_{N}|\Psi(..,x_{k-1},x,x_{k+1},..,t)|^{2}}.$ (27) It is straightforward to show that $p_{W}^{j,k}(x,t)$ in (27) pondered by $\mathbb{P}^{k}(x,t)$ in (26) exactly gives $\langle p_{j}\rangle(t)$. ### A.3 Development of equation (12) in the paper The problem with the weak values in (27) is that it seems very difficult to identify on which $j$-th particle the momentum is (weakly) measured, and on which $k$-particle the position is (strongly) measured. Even worst, it seems not possible to repeat the experiment and get the position and momentum of the same two particles as in the previous measurement. In fact, the evaluation of $\langle p_{j}\rangle(t)$ is independent on which $k$-th particle one does the position measurement. If there are $N$ particles in the system we can write $\langle p_{j}\rangle(t)=\frac{1}{N}\sum_{k=1}^{N}\int dx\;\mathbb{P}^{k}(x,t)p_{W}^{j,k}(x,t)=\int dx\;\mathbb{P}(x,t)\frac{1}{N}\sum_{k=1}^{N}p_{W}^{j,k}(x,t),$ (28) since $\mathbb{P}^{k}(x,t)=\mathbb{P}(x,t)$ for identical particles. Finally, if we assume that we will not identify the $j$-th particle for the momentum measurement, then instead of trying to get $\langle p_{j}\rangle(t)$ we will get an average over the $N$ particles, $\displaystyle\langle p\rangle(t)\equiv\frac{1}{N}\sum_{j=1}^{N}\langle p_{j}\rangle(t)=\frac{1}{N}\frac{1}{N}\sum_{j=1}^{N}\sum_{k=1}^{N}\int dx\;\mathbb{P}(x,t)p_{W}^{j,k}(x,t)=\int dx\;\mathbb{P}(x,t)\frac{1}{N^{2}}\sum_{j=1}^{N}\sum_{k=1}^{N}p_{W}^{j,k}(x,t).$ (29) As such, we arrive to the weak values of the momentum as in equation (12) in the paper, $\tilde{p}_{W}(x,t)=\frac{1}{N^{2}}\sum_{j=1}^{N}\sum_{k=1}^{N}p_{W}^{j,k}(x,t),$ (30) which satisfies $\langle p\rangle(t)=\int dx\;\mathbb{P}(x,t)\;\tilde{p}_{W}(x,t)$. For identical particles, either fermions or bosons, we have $\mathbb{P}^{k}(x,t)=\mathbb{P}^{j}(x,t)\equiv\mathbb{P}(x,t)$ for all $j,k$ since $|\Psi(..,x_{k},..,x_{j},..,t)|^{2}=|\Psi(..,x_{j},..,x_{k},..,t)|^{2}$. We also have $p_{W}^{j}(..,x_{l},..,x_{j},..,t)=p_{W}^{l}(..,x_{j},..,x_{l},..,t)$ when $j,l\neq k$ because $J^{j}(..,x_{l},..,x_{j},..,t)=J^{l}(..,x_{j},..,x_{l},..,t)$ when $j,l\neq k$. As such, we obtain $\displaystyle p_{W}^{j,k}(x,t)$ $\displaystyle=$ $\displaystyle p_{W}^{l,k}(x,t)\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\text{for all }j,l\neq k,$ $\displaystyle p_{W}^{j,k}(x,t)$ $\displaystyle=$ $\displaystyle p_{W}^{j,l}(x,t)\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\text{for all }k,l\neq j,$ $\displaystyle p_{W}^{j,k}(x,t)$ $\displaystyle=$ $\displaystyle p_{W}^{k,j}(x,t)\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\text{for all }j\neq k,$ $\displaystyle p_{W}^{j,j}(x,t)$ $\displaystyle=$ $\displaystyle p_{W}^{k,k}(x,t)\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\text{for all }j,k.$ (31) For a separable wave function, $\Psi(\mathbf{x},t)=\psi_{1}(x_{1},t)...\psi_{N}(x_{N},t)$, we obtain $\displaystyle p_{W}^{j,k}(x,t)$ $\displaystyle=$ $\displaystyle\frac{\int dx\;p_{W}^{j}(x,t)|\psi_{j}(x,t)|^{2}}{\int dx|\psi_{j}(x,t)|^{2}}=\int dx\;J^{j}(x,t)=\langle p_{j}\rangle(t)\;\;\;\;\;\;\;\text{for all }j\neq k,$ (32) $\displaystyle p_{W}^{j,j}(x,t)$ $\displaystyle=$ $\displaystyle p_{W}^{j}(x,t)=\frac{J^{j}(x,t)}{|\psi_{j}(x,t)|^{2}}\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\text{for all }j,$ (33) where $J^{j}(x,t)$ and $p_{W}^{j}(x,t)$ are the current density and the weak values, respectively, linked to the single-particle $\psi_{j}(x_{j},t)$. So, for separable systems, the terms $p_{W}^{j,k}(x,t)$ are spatially uniform, while $p^{j}_{W}(x,t)$ depend strongly on the position. For nonseparable systems such differences are not true. Thus, $p_{W}^{j,k}(x,t)$ also provides a procedure to quantify the interaction between distinct particles. ### A.4 Time-averaging of weak values Figure 7: Time-averaged many-body weak values of the momentum, $\bar{f}_{W}(\\{x_{0j}\\},t,T)=\bar{p}_{W}^{j,k}(\\{x_{0j}\\},t,T)$, from the weak values presented in figure 6 in the paper. Only $j=k=1$ is plotted, which is the same as $j=k=2$ for $N=2$ and $j=k=2,3$ for $N=3$. The values of $\\{x_{0j}\\}$ are taken as the set of initial values of the respective wave packets for $N=1$ in (a),(b), $N=2$ in (c),(d), and $N=3$ in (e),(f). Left and right panels consider respectively a time $t=20<t_{eq}$ and $t=120>t_{eq}$, while the integration period $T/2$ in all panels runs from $0$ to $20$. We have seen in figure 6 in the paper that, contrary to the momentum expectation value in figure 4 in the paper, the weak values of the momentum only approach $0$ at higher $N$. For $N=2$ they remain within a finite range of values after $t_{eq}$, while for $N=1$ such range is much larger. We have also discussed how the weak values (as well as the density matrix coherences) have a random nature for $t>t_{eq}$. Thus we compute the time-average of such weak values on a period of time $T$, $\bar{f}_{W}(x,t,T)=\frac{1}{T}\int_{t-T/2}^{t+T/2}dt^{\prime}\;f_{W}(x,t^{\prime}),$ (34) where $f_{W}(x,t)$ can be any of the $4$ different types of weak values expressed in the paper, that is, $p_{W}(x,t)$, $p_{W}^{j}(x,t)$, $p_{W}^{j,k}(x,t)$, and $\tilde{p}_{W}(x,t)$. We consider in figure 7 the case $f_{W}(x,t)=p_{W}^{j,k}(x,t)$, and then plot $\bar{f}_{W}(x,t,T)=\bar{p}_{W}^{j,k}(x,t,T)$ from the weak values presented in figure 6 in the paper, as a function of $T$. Upper, middle, lower panels respectively for $N=1$, $N=2$, $N=3$; panels in left and right column respectively related to a time before ($t=20<t_{eq}$) and after ($t=120>t_{eq}$) thermalization. For $N>1$, the time-average of the weak values always vanish at $t>t_{eq}$, even at small $T$ ((d),(f)); for $t<t_{eq}$ this may only happen at much higher $T$ ((c),(e)). The $N=1$ case is, once more, much less smooth due to the many nodes of the wave function. ## Appendix B Diagonal and off-diagonal elements of the density matrix in the energy representation and its connection with weak values of the momentum The dynamics presented in our paper is based on an initial nonequilibrium _pure_ state evolving in a closed system, while being affected by some ‘chaotic’ disorder potential. Without disorder, such a pure state simply evolves periodically in the underneath harmonic oscillator potential. With disorder, we have seen its role in bringing the nonequilibrium state into an equilibrium regime characterized e.g. by the behaviour of the expectation values shown in figures 3 and 4 in the paper. But what is the role of disorder on the density matrix in the energy representation? ### B.1 Density matrix populations and coherences We consider a pure state as a superposition of (single-particle or many- particle) energy eigenstates, $\langle x_{1}|\otimes...\otimes\langle x_{N}|\Psi(t)\rangle=\langle\mathbf{x}|\Psi(t)\rangle=\sum_{n}c_{n}\langle\mathbf{x}|n\rangle e^{-iE_{n}t}=\sum_{n}\;c_{n}R_{n}(\mathbf{x})e^{-iE_{n}t},$ (35) being $|n\rangle$ an energy eigenstate with eigenvalue $E_{n}$, $c_{n}=\langle n|\Psi(0)\rangle$, and $\langle\mathbf{x}|n\rangle=R_{n}(\mathbf{x})$ a real function. The $n$-sum runs within a set $N_{act}$ of activated states, which is defined by some quench responsible for creating the nonequilibrium initial state; in our model, the quench is a sudden shift of the harmonic trap translated to an initial velocity for the electrons at $t=0$. The pure state density matrix, $\hat{\rho}(t)=|\Psi(t)\rangle\langle\Psi(t)|$, in the energy basis becomes $\rho_{n,m}(t)=\langle m|\hat{\rho}(t)|n\rangle=\langle m|\Psi(t)\rangle\langle\Psi(t)|n\rangle=c_{n}\;c_{m}^{*}\;e^{i(E_{m}-E_{n})t},$ (36) where the diagonal elements, $\rho_{n,n}$ (_populations_), are clearly time- independent, while the off-diagonal terms (_coherences_) are not. When time- averaged, within a time interval $T\to\infty$, the off-diagonal elements vanish and the density matrix becomes diagonal, $\lim_{T\to\infty}\frac{1}{T}\int_{-T/2}^{T/2}\;\rho_{n,m}(t)\;dt=c_{n}\;c_{m}^{*}\;\lim_{T\to\infty}\frac{1}{T}\int_{-T/2}^{T/2}e^{i(E_{m}-E_{n})t}dt=c_{n}\;c_{m}^{*}2\pi\delta_{n,m},$ (37) with $\delta_{n,m}$ the Kronecker delta. This does not imply that the density matrix in (36) (without time averaging) becomes diagonal as $t\to\infty$, but only that the off-diagonal elements oscillate around zero. In figure 8 we show the evolution of real and imaginary parts of a few off- diagonal elements $\rho_{n,m}(t)$ ($N=1$), which can be written from (36) by employing $c_{l}=|c_{l}|e^{i\theta_{l}}$ as $\rho_{n,m}(t)=|c_{n}|\;|c_{m}|\text{cos}(\theta_{n}-\theta_{m}+(E_{m}-E_{n})t)+i|c_{n}|\;|c_{m}|\text{sin}(\theta_{n}-\theta_{m}+(E_{m}-E_{n})t),$ (38) such that the modulus $|\rho_{n,m}(t)|=|c_{n}||c_{m}|$ is also time- independent, as seen in panel (a). The oscillation period depends inversely on their level splittings, $\Delta E=E_{m}-E_{n}$, as shown in panels (b)-(d), which relate to the same level $n_{p}=201$ at the peak of $\rho_{n,m}(t)$: as one increases the splitting, by considering a matrix element with $1$ (b), $3$ (c), and $6$ (d) levels apart, the oscillation period decreases. Also, if considering both $n,m$ values away from the peak at $n_{p}=201$, the respective matrix elements have exponentially smaller amplitudes. In a pure harmonic oscillator all level splittings are simply proportional to the difference in number of levels, but the disorder potential creates _random_ splittings. The full density matrix is given in figure 1(d) in the paper, which shows $N_{act}\approx 70$ activated states at $t=0$, which will remain as the main states determining the system unitary evolution at any $t$. From these results we conclude that the thermalized system keeps memory of its initial state through the propagation of its off-diagonal coherences (in particular by keeping the information $c_{l}=|c_{l}|e^{i\theta_{l}}$), which never disappear. Most importantly for the current operator as discussed in the paper, since its diagonal populations are zero by construction. Figure 8: Density matrix coherences for $N=1$. From the grid-diagonalization as depicted in the inset (b) of figure 1 in the paper, the time-evolution of a few of the off-diagonal coherences $\rho_{n,m}(t)$ is shown with respect to the peak at $n_{p}=201$, with real (imaginary) part in blue (red). As the energy splitting, $\Delta E=E_{m}-E_{n}$, increases from (b) to (c) to (d), the respective oscillation period decreases. In (a) one verifies that the respective moduli $|\rho_{n,m}(t)|^{2}$ remain constant. ### B.2 Connection between density matrix coherences and weak values First, we need the presence probability density $|\Psi(\mathbf{x},t)|^{2}$, given by $\displaystyle|\Psi(\mathbf{x},t)|^{2}$ $\displaystyle=$ $\displaystyle\text{trace}\left[\hat{\rho}(|x_{1}\rangle\otimes...\otimes|x_{N}\rangle\langle x_{1}|\otimes...\otimes\langle x_{N}|)\right]$ (39) $\displaystyle=$ $\displaystyle\sum_{n}\sum_{m}c_{n}c_{m}^{*}\;e^{i(E_{m}-E_{n})\;t}R_{n}(\mathbf{x})R^{*}_{m}(\mathbf{x}),$ and that can be decomposed into time-independent populations and time- dependent coherences as $|\Psi(\mathbf{x},t)|^{2}=|\Psi_{dia}(\mathbf{x},t)|^{2}+|\Psi_{off}(\mathbf{x},t)|^{2},$ (40) with $\displaystyle|\Psi_{dia}(\mathbf{x},t)|^{2}$ $\displaystyle=$ $\displaystyle\sum_{n}\rho_{n,n}|R_{n}(\mathbf{x})|^{2},$ $\displaystyle|\Psi_{off}(\mathbf{x},t)|^{2}$ $\displaystyle=$ $\displaystyle\sum_{n}\sum_{m\neq n}\rho_{n,m}(t)R_{n}(\mathbf{x})R^{*}_{m}(\mathbf{x}).$ (41) We also need the current density $J^{j}(\mathbf{x},t)$ due to the $j$-th particle at one given position in the configuration space, $\displaystyle J^{j}(\mathbf{x},t)$ $\displaystyle=$ $\displaystyle\text{trace}\left[\hat{\rho}\left(\hat{p}_{j}|x_{1}\rangle\otimes...\otimes|x_{N}\rangle\langle x_{1}|\otimes...\otimes\langle x_{N}|+|x_{1}\rangle\otimes...\otimes|x_{N}\rangle\langle x_{1}|\otimes...\otimes\langle x_{N}|\hat{p}_{j}\right)\right]$ (42) $\displaystyle=$ $\displaystyle\frac{i}{2}\sum_{n}\sum_{m}c_{n}c_{m}^{*}\;e^{i(E_{m}-E_{n})\;t}\left[R_{n}(\mathbf{x},t)\;\frac{\partial R_{m}(\mathbf{x},t)}{\partial x_{j}}-R_{m}(\mathbf{x},t)\;\frac{\partial R_{n}(\mathbf{x},t)}{\partial x_{j}}\right],$ that can also be decomposed in terms of diagonal and off-diagonal components as $J^{j}(\mathbf{x},t)=J^{j}_{dia}(\mathbf{x},t)+J^{j}_{off}(\mathbf{x},t),$ (43) with $\displaystyle J^{j}_{dia}(\mathbf{x},t)$ $\displaystyle=$ $\displaystyle\frac{i}{2}\sum_{n}\rho_{n,n}\left[R_{n}(\mathbf{x})\;\frac{\partial R_{n}(\mathbf{x})}{\partial x_{j}}-R_{n}(\mathbf{x})\;\frac{\partial R_{n}(\mathbf{x})}{\partial x_{j}}\right]=\sum_{n}\rho_{n,n}(t)\;J^{j}_{n,n}(\mathbf{x}),$ $\displaystyle J^{j}_{off}(\mathbf{x},t)$ $\displaystyle=$ $\displaystyle\frac{i}{2}\sum_{n}\sum_{m\neq n}\rho_{n,m}(t)\;\left[R_{n}(\mathbf{x})\;\frac{\partial R_{m}(\mathbf{x})}{\partial x_{j}}-R_{m}(\mathbf{x})\;\frac{\partial R_{n}(\mathbf{x})}{\partial x_{j}}\right]=\sum_{n}\sum_{m\neq n}\rho_{n,m}(t)\;J^{j}_{m,n}(\mathbf{x}),$ (44) where $J^{j}_{m,n}(\mathbf{x})=\frac{i}{2}\left[R_{n}(\mathbf{x})\;\frac{\partial R_{m}(\mathbf{x})}{\partial x_{j}}-R_{m}(\mathbf{x})\;\frac{\partial R_{n}(\mathbf{x})}{\partial x_{j}}\right]$. The relevant point is that only off-diagonal elements $\rho_{n,m}(t)\;J^{j}_{m,n}(\mathbf{x})$ for $n\neq m$ provide current densities, since the contribution of diagonal elements vanish, $J^{j}_{dia}(\mathbf{x},t)=0$, in closed systems with wave function vanishing at the boundaries. The diagonal terms $\rho_{n,n}(t)\;J^{j}_{n,n}(\mathbf{x})$ do not contribute to the total current because energy eigenstates are pure real (or pure imaginary), so that their current $J^{j}_{n,n}(\mathbf{x})=0$ in first equation in (44). Both results from (40) and (43) are in agreement with the well-known continuity equation, $\displaystyle 0$ $\displaystyle=$ $\displaystyle\frac{\partial|\Psi(\mathbf{x},t)|^{2}}{\partial t}+\sum_{j=1}^{N}\frac{\partial J^{j}(\mathbf{x},t)}{\partial x_{j}}$ (45) $\displaystyle=$ $\displaystyle\frac{\partial|\Psi_{dia}(\mathbf{x},t)|^{2}}{\partial t}+\sum_{j=1}^{N}\frac{\partial J^{j}_{dia}(\mathbf{x},t)}{\partial x_{j}}+\frac{\partial|\Psi_{off}(\mathbf{x},t)|^{2}}{\partial t}+\sum_{j=1}^{N}\frac{\partial J^{j}_{off}(\mathbf{x},t)}{\partial x_{j}}$ $\displaystyle=$ $\displaystyle\frac{\partial|\Psi_{off}(\mathbf{x},t)|^{2}}{\partial t}+\sum_{j=1}^{N}\frac{\partial J^{j}_{off}(\mathbf{x},t)}{\partial x_{j}},$ where we have used the trivial results $\frac{\partial|\Psi_{dia}(\mathbf{x},t)|^{2}}{\partial t}=0$ (because $|\Psi_{dia}(\mathbf{x},t)|$ is time-independent) and $\sum_{j=1}^{N}\frac{\partial J^{j}_{dia}(\mathbf{x},t)}{\partial x_{j}}=0$ (because $J^{j}_{dia}(\mathbf{x},t)=0$). Since we know from (36) and (38) that the coherences never vanish ($\rho_{n,m}(t)\neq 0$ for $n\neq m$) and always oscillate, we conclude that $\frac{\partial|\Psi_{off}(\mathbf{x},t)|^{2}}{\partial t}\neq 0$, so that (45) means that the off-diagonal probability presence $|\Psi_{off}(\mathbf{x},t)|^{2}$ and the off-diagonal current density $J^{j}_{off}(\mathbf{x},t)$ are dynamically changing during the whole simulation, before and after $t_{eq}$. We notice now that $|\Psi(\mathbf{x},t)|^{2}$ and $J^{j}_{off}(\mathbf{x},t)$ are the elements that define the weak values in equation (3) in the paper, $p_{W}^{j}(\mathbf{x},t)=\frac{J^{j}(\mathbf{x},t)}{|\Psi(\mathbf{x},t)|^{2}}=\frac{J^{j}_{off}(\mathbf{x},t)}{|\Psi(\mathbf{x},t)|^{2}}=\frac{1}{|\Psi(\mathbf{x},t)|^{2}}\left(\sum_{n}\sum_{m\neq n}\rho_{n,m}(t)\;J^{j}_{m,n}(\mathbf{x})\right).$ (46) This provides the required connection between weak values and coherences. ## Appendix C Momentum representation and ‘phase-space’ In this appendix we provide the dynamics evolution in momentum representation and construct a pseudo phase-space of the system. Figure 9: Wave packet dynamics in momentum representation. This figure is similar to figure 2 in the paper, which employed position representation, but instead it considers momentum representation. Upper panels: wave packet $|\Psi(p_{1},p_{2},[p_{3}],t)|^{2}$ for $N=3$ at $t=0$ in (a) and $t=150$ in (b). Middle panels: $|\Psi(p_{1},p_{2},t)|^{2}$ for $N=2$ at $t=0$ in (c) and $t=150$ in (d). Panel (e): wave packet $|\Psi(p_{1},t)|^{2}$ for $N=1$ at $t=0$ (dotted) and at $t=150$ (solid). Panel (f): $1$D-view of $|\Psi(p_{1},[p_{2}],[p_{3}],t)|^{2}$ for $N=3$ (red), $|\Psi(p_{1},[p_{2}],t)|^{2}$ for $N=2$ (blue), and $|\Psi(p_{1},t)|^{2}$ for $N=1$ (green), with solid (dotted) lines for the wave packet at $t=150$ ($t=0$), where all $t=0$ plots overlap at $p_{0j}=20$. All plots are in log- scale. Horizontal axis in (a),(c) ((b),(d)) is the same as in (e) ((f)). Figure 10: ‘Phase-space’ analysis. From position $\langle x_{1}\rangle$(t) and momentum $\langle p_{1}\rangle(t)$ expectation values in figure 4 in the paper, we compile a respective ‘phase-space’ for $N=1$ in (a), $N=2$ in (c), and $N=3$ in (e). Panels (b),(d),(f) are a zoom at the origin for the respective $N$. The initial value of $\langle x_{1}\rangle(t=0)$ for each $N$ is the average of the set of respective initial positions $\\{x_{0j}\\}$, while $\langle p_{1}\rangle(t=0)=20$ for any $N$. As one approaches the final simulation time, $t=150>t_{eq}$, the ‘phase-space’ looks noisier for $N=1,2$ as a consequence of the recurrences as seen in the respective expectation values in the paper. Horizontal axis in (a),(c) ((b),(d)) is the same as in (e) ((f)). ### C.1 Dynamics in momentum representation Figure 2 in the paper summarizes the dynamics for the systems with $N=1,2,3$ by showing the initial and final snapshots of the respective wave functions in _position_ representation. Since our algorithm for propagating the Schrödinger equation uses a split-operator method where the kinetic energy is handled in momentum representation, where it is diagonal, and then Fourier-transformed back to position representation (where the potential terms are diagonal), for consistency, we show in figure 9 exactly the same plots as in figure 2 in the paper, but in _momentum_ representation. That is, we plot in the upper panels $|\Psi(p_{1},p_{2},[p_{3}],t)|^{2}$, with $[p_{3}]$ integrated out, for $N=3$ at $t=0$ in (a) and $t=150$ in (b) (results would be the same if integrating on $[p_{1}]$ or $[p_{2}]$). The middle panels show $|\Psi(p_{1},p_{2},t)|^{2}$ for $N=2$ at $t=0$ in (c) and $t=150$ in (d). Panel (e) shows $|\Psi(p_{1},t)|^{2}$ for $N=1$ at $t=0$ ($t=150$) in dotted (solid) lines, while panel (f) compiles all respective $1$D plots of $|\Psi(p_{1},[p_{2}],[p_{3}],t)|^{2}$ for $N=3$, $|\Psi(p_{1},[p_{2}],t)|^{2}$ for $N=2$, and $|\Psi(p_{1},t)|^{2}$ for $N=1$, with solid (dotted) lines for $t=150$ ($t=0$). Since the initial velocity is the same and centered at $p_{0j}=20$ for each degree of freedom and for every $N$, all $1$D plots at $t=0$ overlap. One also notices in momentum representation the same full spread of the wave function after thermalization, at $t\gg t_{eq}$, which then remains in between the ’turning points’ $p_{TP}=\pm 20$ (since $\omega=1$). As in position representation, the higher the $N$ the more homogeneous is the wave packet spread. ### C.2 Pseudo phase-space We have shown in the paper that thermalization provides time-independent expectation values, so that another useful plot is a pseudo ‘phase-space’ as if that could be simply compiled from the expectation values of $\langle p_{j}\rangle(t)$ and $\langle x_{j}\rangle(t)$ in figure 4 in the paper. Figure 10 shows the ‘phase space’ for $N=1,2,3$ respectively in upper, middle, lower panels; left column the full range, right column a zoom at the origin $(0,0)$. The figures look the same no matter which $j$-electron is considered at each $N$. For any $N$ the curves start at $\langle p_{1}\rangle(0)=20$ and $\langle x_{1}\rangle(0)=0$, only exception being $N=2$ which starts at $\langle x_{1}\rangle(0)=-2$. The curves after thermalization (right column) are about the same at any $N$, with $\langle p_{1}\rangle(t)\approx\langle x_{1}\rangle(t)\approx 0$; they look visually different since some recurrences may happen causing a noisier ‘phase space’ as time keeps rolling after $t_{eq}$; however, once more, the behaviour is more smooth as $N$ increases. ## References * [1] R. Penrose, The Emperor’s New Mind: Concerning computers, Brains and the Laws of Physics (Oxford Univ. Press, Oxford 1989). * [2] S. W. Hawking, Phys. Rev. D 32, 2489 (1985). * [3] G. C. Ghirardi, A. Rimini, and T. Weber, Phys. Rev. D 34, 470 (1986). * [4] L. Sklar, Physics and Change: Philosophical Issues in the Foundations of Statistical Mechanics (Cambridge Univ. Press, Cambridge 1993). * [5] D. Z. Albert, Time and Change (MA: Harvard Univ. Press, Cambridge 2000). * [6] J. Bricmont, D. Durr, M. Galavotti, F. Petruccione, and N. Zanghi, Change in Physics: Foundations and Perspectives (Springer, Berlin 2001). * [7] R. Penrose, J. Stat. Phys. 77, 217 (1994). * [8] H. D. Zeh, The Physical Basis of the Direction of Time (5th ed., Springer, Berlin 2007). * [9] M. Ueda, Nat. Rev. Phys. 2, 669 (2020). * [10] C. Gogolin and J. Eisert, Rep. Prog. Phys. 79, 056001 (2016). * [11] J. Eisert, M. Friesdorf, and C. Gogolin, Nat. Phys. 11, 124 (2015). * [12] V. I. Yukalov, Laser Phys. Lett. 8, 485 (2011). * [13] P. Reimann, New J. Phys. 21, 053014 (2019). * [14] M. Rigol, V. Dunjko, and M. Olshanii, Nature 452, 854 (2008). * [15] R. Nandkishore and D. A. Huse, Annu. Rev. Condens. Matter Phys. 6, 15 (2015). * [16] L. D’Alessio, Y. Kafri, A. Polkovnikov, and M. Rigol, Adv. Phys. 65, 239 (2016). * [17] J. M. Deutsch, Rep. Prog. Phys. 81, 082001 (2018). * [18] T. Langen, R. Geiger, and J. Schmiedmayer, Annu. Rev. Condens. Matter Phys. 6, 201 (2015). * [19] M. Lewenstein, A. Sanpera, V. Ahufinger, B. Damski, A. Sen(De), and U. Sen, Adv. Phys. 56, 243 (2007). * [20] L. Sanchez-Palencia, D. Clément, P. Lugan, P. Bouyer, and A. Aspect, New J. Phys. 10, 045019 (2008). * [21] A. Polkovnikov, K. Sengupta, A. Silva, and M. Vengalattore, Rev. Mod. Phys. 83, 863 (2011). * [22] P. Reimann, New J. Phys. 17, 055025 (2015). * [23] T. R. Oliveira, C. Charalambous, D. Jonathan, M. Lewenstein, and A. Riera, New J. Phys. 20, 033032 (2018). * [24] H. Kim, T. N. Ikeda, and D. A. Huse, Phys. Rev. E 90, 052105 (2014). * [25] T. Kinoshita, T. Wenger, and D. Weiss, Nature 440, 900 (2006). * [26] S. Trotzky, Y-A. Chen, A. Flesch, I. P. McCulloch, U. Schollwöck, J. Eisert, and I. Bloch, Nat. Phys. 8, 325 (2012). * [27] A. M. Kaufman, M. E. Tai, A. Lukin, M. Rispoli, R. Schittko, P. M. Preiss, and M. Greiner, Science 353, 794 (2016). * [28] J. P. Ronzheimer, M. Schreiber, S. Braun, S. S. Hodgman, S. Langer, I. P. McCulloch, F. Heidrich-Meisner, I. Bloch, and U. Schneider, Phys. Rev. Lett. 110, 205301 (2013). * [29] D. Dries, S. E. Pollack, J. M. Hitchcock, and R. G. Hulet, Phys. Rev. A 82, 033603 (2010). * [30] S. Will, D. Iyer, and M. Rigol, Nat. Commun. 6, 6009 (2015). * [31] J. Kajala, F. Massel, and P. Torma, Phys. Rev. Lett. 106, 206401 (2011). * [32] U. Schneider, L. Hackermuller, S. Will, T. Best, I. Bloch, T. A. Costi, R. W. Helmes, D. Rasch, and A. Rosch, Science 322, 1520 (2008). * [33] B. Nagler, K. Jagering, A. Sheikhan, S. Barbosa, J. Koch, S. Eggert, I. Schneider, and A. Widera, Phys. Rev. A 101, 053633 (2020). * [34] U. Schneider, L. Hackermuller, J. P. Ronzheimer, S. Will, S. Braun, T. Best, I. Bloch, E. Demler, S. Mandt, D. Rasch, and A. Rosch, Nat. Phys. 8, 213 (2012). * [35] W. C. Myrvold, Synthese 192, 3247 (2015). * [36] H-P. Breuer and F. Petruccione, The Theory of Open Quantum Systems (Oxford Univ. Press, Oxford, 2002). * [37] In other quantum theories, like Bohmian mechanics or many-worlds, the collapse is just an ‘effective’ phenomenon whose consequences can still be treated from a bigger unitary Schrödinger equation that includes the measuring apparatus. * [38] M. Srednicki, Phys. Rev. E 50, 888 (1994). * [39] J. M. Deutsch, Phys. Rev. A 43, 2046 (1991). * [40] Y. Aharonov, D. Z. Albert, and L. Vaidman, Phys. Rev. Lett. 60, 1351 (1988). * [41] B. E. Y. Svensson, Quanta 2, 18 (2013). * [42] D. Pandey, R. Sampaio, T. Ala-Nissila, G. Albareda, and X. Oriols, Phys. Rev. A 103, 052219 (2021). * [43] H. M. Wiseman, New J. Physics 9, 165 (2007). * [44] D. Dürr, S. Goldstein, and N. Zanghì, J. Stat. Phys. 134, 1023 (2009). * [45] D. Marian, N. Zanghi, and X. Oriols, Phys. Rev. Lett. 116, 110404 (2016). * [46] F. L. Traversa, G. Albareda, M. Di Ventra, and X. Oriols, Phys. Rev. A 87, 052124 (2013). * [47] S. Kocsis, B. Braverman, S. Ravets, M. J. Stevens, R. P. Mirin, L. K. Shalm, and A. M. Steinberg, Science 332, 1170 (2011). * [48] A. Hariri, D. Curic, L. Giner, and J. S. Lundeen, Phys. Rev. A 100, 032119 (2019). * [49] R. Ramos, D. Spierings, I. Racicot, and A. M. Steinberg, Nature 583, 529 (2020). * [50] A.G. Kofman, S. Ashhab, and F. Nori, Physics Reports 520, 43 (2012). * [51] J. Dressel, M. Malik, F. M. Miatto, A. N. Jordan, and R. W. Boyd, Rev. Mod. Phys. 86, 307 (2014). * [52] K. Renziehausen and I. Barth, Found. Phys. 50, 772 (2020). * [53] O. A. Castro-Alvaredo, B. Doyon, and T. Yoshimura, Phys. Rev. X 6, 041065 (2016). * [54] V. Alba, B. Bertini, M. Fagotti, L. Piroli, and P. Ruggiero, J. Stat. Mech., 114004 (2021). * [55] I. Bouchole and J. Dubail, J. Stat. Mech., 014003 (2022). * [56] X. Oriols and J. Mompart, Applied Bohmian Mechanics: From Nanoscale Systems to Cosmology (Pan Stanford, Singapore, 2012). * [57] A. O. T. Pang, H. Ferretti, N. Lupu-Gladstein, W.-K. Tham, A. Brodutch, K. Bonsma-Fisher, J. E. Sipe, and A. M. Steinberg, Quantum 4, 365 (2020). * [58] T. Fogarty, M. Á. García-March, L. F. Santos, and N. L. Harshman, Quantum 5, 486 (2021). * [59] G. Zisling, L. F. Santos, and Y. B. Lev, SciPost. Phys. 10, 088 (2021). * [60] C. Neill, P. Roushan, M. Fang, Y. Chen, M. Kolodrubetz, Z. Chen, A. Megrant, R. Barends, B. Campbell, B. Chiaro, A. Dunsworth, E. Jeffrey, J. Kelly, J. Mutus, P. J. J. O’Malley, C. Quintana, D. Sank, A. Vainsencher, J. Wenner, T. C. White, A. Polkovnikov, and J. M. Martinis, Nat. Phys. 12, 1037 (2016). * [61] Md. M. Ali, W.-M. Huang, and W.-M. Zhang, Sci. Rep. 10, 13500 (2020). * [62] J.-Q. Liao, H. Dong, and C. P. Sun, Phys. Rev. A 81, 052121 (2010). * [63] P. Lydzba, Y. Zhang, M. Rigol, and L. Vidmar, Phys. Rev. B 104, 214203 (2021). * [64] C. Nation and D. Porras, Phys. Rev. E 102, 042115 (2020). * [65] M. Rigol, Phys. Rev. A 80, 053607 (2009). * [66] L. F. Santos and M. Rigol, Phys. Rev. E 81, 036206 (2010). * [67] Y. Tang, W. Kao, K.-Y. Li, S. Seo, K. Mallayya, M. Rigol, S. Gopalakrishnan, and B. L. Lev, Phys. Rev. X 8, 021030 (2018). * [68] C. Gogolin, M. P. Muller, and J. Eisert, Phys. Rev. Lett. 106, 040401 (2011). * [69] M. Brenes, T. LeBlond, J. Goold, and M. Rigol, Phys. Rev. Lett. 125, 070605 (2020). * [70] M. Brenes, J. Goold, and M. Rigol, Phys. Rev. B 102, 075127 (2020). * [71] Y.-W. Hsueh, C.-H. Hsueh, and W.-C. Wu, Entropy 22, 855 (2020). * [72] L. Pezze, B. Hambrecht, and L. Sanchez-Palencia, EPL 88, 30009 (2009). * [73] L. Pezze, and L. Sanchez-Palencia, Phys. Rev. Lett. 106, 040601 (2011). * [74] C.-H. Hsueh, R. Ong, J.-F. Tseng, M. Tsubota, and W.-C. Wu, Phys. Rev. A 98, 063613 (2018). * [75] R. Ong, C.-H. Hsueh, and W.-C. Wu, Phys. Rev. A 100, 053619 (2019). * [76] D. Clément, A. F. Varón, J. A. Retter, L. Sanchez-Palencia, A. Aspect, and P. Bouyer, New J. Phys. 8, 165 (2006). * [77] J. Billy, V. Josse, Z. Zuo, A. Bernard, B. Hambrecht, P. Lugan, D. Clément, L. Sanchez-Palencia, P. Bouyer, and Alain Aspect, Nature 453, 891 (2008). * [78] All figures in the paper consider $\alpha=0.1$, $\omega=1.0$, $\sigma_{j}=l_{0}=1/\sqrt{\omega}=1.0$; initial velocities are $p_{0j}=20$ for any $j$ and $N$; initial positions are $x_{01}=0$ for $N=1$, $(x_{01},x_{02})=(-4,0)$ for $N=2$, $(x_{01},x_{02},x_{03})=(-4,0,4)$ for $N=3$, having in mind that the initial wave packet is antisymmetrized such that the label $j$ in $x_{0j}$ is redundant; for the disorder, $\gamma_{D}=25$ and $\sigma_{D}\approx\Delta x$, with $\Delta x$ being the position resolution of the respective grid, defined in a way that its momentum resolution is $\Delta p\approx\Delta x$, since our propagator uses both position and momentum representations. Initial time is $t_{0}=0$, final time is $t_{f}=150$ (for $N=2$ $t_{f}=300$), with time step $\Delta t=0.001$. Simulation boxes are roughly twice the extension of the turning points, and $\approx 10^{3}$ points per dimension are used; so e.g for $N=3$ our configuration space has $\approx 10^{9}$ points. * [79] A. Maksymov, P. Sierant, and J. Zakrzewski, Phys. Rev. B 102, 134205 (2020). * [80] J. Smith, A. Lee, P. Richerme, B. Neyenhuis, P. W. Hess, P. Hauke, M. Heyl, D. A. Huse, and C. Monroe, Nat. Phys. 12, 907 (2016). * [81] L. F. Santos and M. Rigol, Phys. Rev. E 82, 031130 (2010).
# Randomized Milstein algorithm for approximation of solutions of jump- diffusion SDEs Paweł Przybyłowicz Verena Schwarz Michaela Szölgyenyi ###### Abstract We investigate the error of the randomized Milstein algorithm for solving scalar jump-diffusion stochastic differential equations. We provide a complete error analysis under substantially weaker assumptions than known in the literature. In case the jump-commutativity condition is satisfied, we prove optimality of the randomized Milstein algorithm by proving a matching lower bound. Moreover, we give some insight into the multidimensional case by investigating the optimal convergence rate for the approximation of jump- diffusion type Lévys’ areas. Finally, we report numerical experiments that support our theoretical findings. Keywords: jump-diffusion SDEs, randomized Milstein algorithm, Lévy’s area, $n$-th minimal error, optimality of algorithms, information-based complexity MSC (2020): 68Q25, 65C30, 60H10 ## 1 Introduction Consider the following jump-diffusion stochastic differential equation (SDE) $\displaystyle\mathop{}\\!\mathrm{d}X(t)=\mu(t,X(t))\mathop{}\\!\mathrm{d}t+\sigma(t,X(t))\mathop{}\\!\mathrm{d}W(t)+\rho(t,X(t-))\mathop{}\\!\mathrm{d}N(t),\quad t\in[0,T],\quad X(0)=X_{0},$ (1) where $\mu,\sigma,\rho:[0,T]\times{\mathbb{R}}\to{\mathbb{R}}$ are (at least) measurable functions, $T\in(0,\infty)$, $W=(W(t))_{t\in[0,T]}$ is a standard Wiener process, and $N=(N(t))_{t\in[0,T]}$ is a homogeneous Poisson process with intensity $\lambda>0$ on a filtered probability space $(\Omega,\mathcal{F},(\mathcal{F}_{t})_{t\geq 0},{\mathbb{P}})$ with a filtration $(\mathcal{F}_{t})_{t\geq 0}$ that satisfies the usual conditions. Furthermore, we assume $p\in[2,\infty)$ and $X_{0}$ to be an $\mathcal{F}_{0}$-measurable random variable with ${\mathbb{E}}[|X_{0}|^{2p}]<\infty$. Due to their numerous applications in mathematical finance, control theory, and the modelling of energy markets, cf. [15, 20, 21, 23], jump-diffusion SDEs continue to gain scientific interest. Only in very special cases exact solutions are available. It is therefore important to develop efficient (or even in some sense optimal) numerical algorithms. In this paper, we define the randomized Milstein algorithm, which is a Milstein-type scheme that uses randomization of the drift coefficient in time. We prove $L^{p}$-error and optimality of the scheme applied to SDE (1). We provide appropriate upper and lower error estimates in the multidimensional case, extending the findings from [2]. In the case that only a finite number of evaluations of $W$ and $N$ are allowed, this enables us to address the problem of the optimal approximation of jump-diffusion type Lévys’ areas. As these Lévys’ areas are naturally present in approximation schemes for jump- diffusion SDEs, our result implies lower error bounds. Analysis of the lower bounds and optimality is provided in the Information-Based Complexity (IBC) framework, see [24]. This setting is widely used for investigating optimal algorithms for approximation of solutions of SDEs, see, for example, [5, 6, 11], [12, 13, 17, 18, 19, 22, 10]. The approximate the solutions of SDEs using randomized algorithms is, for example, studied in [11, 12, 17, 18, 22], where the authors consider the randomized Euler–Maruyama scheme for SDEs in the jump-free case, and provide error bounds and optimality results. The articles [4] and [5] discuss the properties of randomized quadrature rules used for approximating stochastic Itô integrals. Error bounds and optimality results of the randomized Milstein scheme for SDEs without jumps were investigated in [13] and [8]. The latter constructed a two-stage version of the randomized Milstein scheme and examined its error. In this paper we extend the results from [13] and [2] to provide results for jump-diffusion SDEs. In the scalar case, we consider SDEs (1) with coefficients that are Hölder continuous in time and Lipschitz continuous and differentiable with Lipschitz continuous derivative in space. Under these assumptions we provide upper bounds for the error of the randomized Milstein algorithm. Our assumptions are significantly weaker than any other in the literature, where it is usually assumed that the coefficients are at least twice continuously differentiable in space, cf. [13, 15]. In addition to that, in case jump-commutativity condition (JCC) is satisfied, we prove optimality of the randomized Milstein algorithm among those randomized algorithms that use finitely many evaluations of the driving processes. It turns out that randomization of the drift coefficient in time improves the convergence rate, see Remark 4.2 and Theorem 4.1. In the multidimensional case we establish upper and lower bounds for the approximation of jump-diffusion Lévys’ areas using the trapezoidal rule. These error bounds imply lower error bounds for any class of coefficients of multidimensional SDEs for which the two-dimensional SDE generating the Levy’s area is a subproblem. In particular, it implies optimality of the multidimensional Euler–Maruyama algorithm in the class of algorithms that use only finitly many evaluations of $W$ and $N$. Therefore, the scalar and multidimensional case can differ a lot from a point of view of optimality of algorithms under certain admissible information about $W$, $N$, and under certain regularity assumptions. Our numerical experiments match the theoretical results on the convergence of the randomized Milstein algorithm. Most interestingly our experiments suggest that for the simulation of jump- diffusion SDEs the $L^{p}$-convergence rate is indeed dependent on $p$. The main contributions of the paper are: * • We perform rigorous error analysis for the randomized Milstein algorithm for scalar jump-diffusion SDEs (1) under relatively mild assumptions on the coefficients (Theorem 3.2). * • We investigate lower error bounds in the worst-case setting in the scalar (Theorem 4.1) and multidimensional case (Theorem 4.3). This essentially allows us to establish optimality of the randomized Milstein algorithm in the scalar case with $p=2$. * • We show that numerical experiments match our theoretical results. The paper is organized as follows. Section 2 states the assumptions under which we perform error analysis for the randomized Milstein algorithm. Section 3 is devoted to error analysis of the randomized Milstein process. Lower bounds and optimality analysis in the IBC framework are given in Section 4. In Section 5 we show the results of the numerical experiments. Finally, some auxiliary results used in the proofs can be found in the Appendix. ## 2 Preliminaries For a random variable $X\colon\Omega\to\mathbb{R}$ we denote by $\|X\|_{L^{p}(\Omega)}=(\mathbb{E}[|X|]^{p})^{1/p}$, where $p\in[2,\infty)$. We take $\displaystyle{\mathcal{F}_{\infty}=\sigma\Big{(}\bigcup_{t\geq 0}\mathcal{F}_{t}\Big{)}}$. Moreover, for $Z\in\\{W,N\\}$ we define $\displaystyle{\mathcal{F}_{\infty}^{Z}=\sigma\Big{(}\bigcup_{t\geq 0}\mathcal{F}_{t}^{Z}\Big{)}}$ where $\displaystyle{\mathcal{F}_{t}^{Z}=\sigma\Big{(}\bigcup_{0\leq s\leq t}\sigma\big{(}Z(s)\big{)}\Big{)}}$. The processes $W$ and $N$ are independent, i.e. $\mathcal{F}_{\infty}^{N}\mathchoice{\mathrel{\hbox to0.0pt{$\displaystyle\perp$\hss}\mkern 2.0mu{\displaystyle\perp}}}{\mathrel{\hbox to0.0pt{$\textstyle\perp$\hss}\mkern 2.0mu{\textstyle\perp}}}{\mathrel{\hbox to0.0pt{$\scriptstyle\perp$\hss}\mkern 2.0mu{\scriptstyle\perp}}}{\mathrel{\hbox to0.0pt{$\scriptscriptstyle\perp$\hss}\mkern 2.0mu{\scriptscriptstyle\perp}}}\mathcal{F}_{\infty}^{W}$, cf. [23, p. 64, Theorem 97]. We denote for all functions $f\in C^{0,1}([0,T]\times{\mathbb{R}};{\mathbb{R}})$ the partial derivative of $f$ with respect to $y$ by $f^{\prime}_{y}$. Further we define for all functions $f\in C^{0,1}([0,T]\times{\mathbb{R}};{\mathbb{R}})$, $L_{1}f(t,y)=\sigma(t,y)f^{\prime}_{y}(t,y)$ and $L_{-1}f(t,y)=f(t,y+\rho(t,y))-f(t,y)$ for all $t\in[0,T]$, $y\in{\mathbb{R}}$. We impose the following assumptions on the coefficient functions. ###### Assumption 2.1. For the functions $\mu,\sigma,\rho\colon[0,T]\times{\mathbb{R}}\to{\mathbb{R}}$ and for $p\in[2,\infty)$ we assume that there exist constants $\varrho_{1},\varrho_{2},\varrho_{3}\in(0,1]$ such that: * (i) For all $f\in\\{\mu,\sigma,\rho\\}$, $f\in C^{0,1}([0,T]\times{\mathbb{R}};{\mathbb{R}})$. * (ii) There exists a constant $K_{1}>0$ such that for all $t,s\in[0,T]$, $y,z\in{\mathbb{R}},$ and all $f\in\\{\mu,\sigma,\rho\\}$ it holds that $\displaystyle|f(t,y)-f(t,z)|\leq K_{1}|y-z|,$ (2) $|f^{\prime}_{y}(t,y)-f^{\prime}_{y}(t,z)|\leq K_{1}|y-z|,$ (3) $|f(t,y)-f(s,y)|\leq K_{1}(1+|y|)|t-s|^{\varrho_{f}},$ (4) where $(\varrho_{f},f)\in\\{(\varrho_{1},\mu),(\varrho_{2},\sigma),(\varrho_{3},\rho)\\}$. * (iii) There exists a constant $K_{2}>0$ such that for all $y\in{\mathbb{R}}$, $t,s\in[0,T]$, $|\mu^{\prime}_{y}(t,y)-\mu^{\prime}_{y}(s,y)|\leq K_{2}(1+|y|)|t-s|^{\varrho_{1}}.$ (5) * (iv) There exists a constant $K_{3}>0$ such that for all $t\in[0,T]$, $y,z\in{\mathbb{R}}$, and all $f\in\\{\sigma,\rho\\}$ it holds that $\displaystyle|L_{1}f(t,y)-L_{1}f(t,z)|\leq K_{3}|y-z|.$ (6) * (v) For the initial value $X_{0}$ we assume that it is an $\mathcal{F}_{0}$-measurable random variable and that $\|X_{0}\|_{L^{2p}(\Omega)}<\infty.$ (7) By the Lipschitz assumption (2) we obtain that for all $(t,y)\in[0,T]\times{\mathbb{R}}$ and $f\in\\{\mu,\sigma,\rho\\}$ we have $\displaystyle|f(t,y)|\leq K_{4}(1+|y|),$ (8) with $K_{4}=\max\limits_{i=1,2,3}\\{\max\\{|f(0,0)|,K_{1}\\}+K_{1}T^{\varrho_{i}}\\}$, and by (i) and (2), $\displaystyle|f^{\prime}_{y}(t,y)|\leq K_{1}.$ (9) Furthermore, we know that for $f\in\\{\mu,\sigma,\rho\\}$ it holds that for all $t\in[0,T]$ the first partial derivative $f^{\prime}_{y}(t,\cdot)$ is absolutely continuous, since it is Lipschitz continuous. Hence, for all $t\in[0,T]$ the second partial derivative $f^{\prime\prime}_{yy}(t,\cdot)$ exists almost everywhere on ${\mathbb{R}}$. For all $t\in[0,T]$ denote by $S_{f}(t)$ the set of Lebesgue measure $0$ for which the second partial derivative $f^{\prime\prime}_{yy}(t,\cdot)$ does not exist. Then for $f\in\\{\mu,\sigma,\rho\\}$, $t\in[0,T]$, and all $y\in{\mathbb{R}}\setminus S_{f}(t)$ it holds that $\displaystyle|f^{\prime\prime}_{yy}(t,y)|\leq K_{1}.$ (10) On $S_{f}(t)$ we define $f^{\prime\prime}_{yy}(t,\cdot)\equiv 0$. At this point, we like to emphasise that the choice of the values of $f^{\prime\prime}_{yy}(t,\cdot)$ on $S_{f}(t)$ does not influence the proof of the main result, since by using the local time theory we see that the suitable bounds we compute are not dependent on these values. Morever, for $f\in\\{\sigma,\rho\\}$ we have that there exists a constant $K_{5}\in(0,\infty)$ such that for all $(t,y)\in[0,T]\times{\mathbb{R}}$, $\max\\{|L_{1}f(t,y)|,|L_{-1}f(t,y)|\\}\leq K_{5}(1+|y|).$ (11) Under Assumption 2.1 the existence and uniqueness of a strong solution to the SDE (1) is well-known, see, for example [16, p. 255, Theorem 6]. Since $\mathbb{E}[|X_{0}|^{2p}]<\infty$, there exists $K_{6}\in(0,\infty)$ such that $\displaystyle{\mathbb{E}}\Big{[}\sup_{0\leq t\leq T}|X(t)|^{2p}\Big{]}\leq K_{6},$ (12) and for all $s,t\in[0,T]$, $\displaystyle{\mathbb{E}}\Big{[}|X(t)-X(s)|^{p}\Big{]}\leq K_{6}|t-s|,$ (13) see [22, Lemma 1]. The estimate (13) can be improved if $\rho\equiv 0$. Moreover, under the Assumption 2.1 (i), (ii), for $f\in\\{\mu,\sigma,\rho\\}$ and fixed $t\in[v_{1},v_{2}]\subset[0,T]$, it is possible to apply the Meyer- Itô formula [16, p. 221, Theorem 71] to the function ${\mathbb{R}}\ni y\mapsto f(t,y)\in\mathbb{R}$ and to the solution process $(X(s))_{s\in[v_{1},v_{2}]}$. This gives the following parametric version of the Meyer-Itô formula: For all $s,t\in[v_{1},v_{2}]$ it holds that $f(t,X(s))=f(t,X(v_{1}))+\int\limits_{v_{1}}^{s}\alpha(f,t,u)\mathop{}\\!\mathrm{d}u+\int\limits_{v_{1}}^{s}\beta(f,t,u)\mathop{}\\!\mathrm{d}W(u)+\int\limits_{v_{1}}^{s}\gamma(f,t,u)\mathop{}\\!\mathrm{d}N(u),$ (14) where $\displaystyle\alpha(f,t,u)$ $\displaystyle=\alpha_{1}(f,t,u)+\alpha_{2}(f,t,u),$ (15) $\displaystyle\alpha_{1}(f,t,u)$ $\displaystyle=f^{\prime}_{y}(t,X(u))\mu(u,X(u)),$ $\displaystyle\alpha_{2}(f,t,u)$ $\displaystyle=\frac{1}{2}f^{\prime\prime}_{yy}(t,X(u))\sigma^{2}(u,X(u)),$ $\displaystyle\beta(f,t,u)$ $\displaystyle=f^{\prime}_{y}(t,X(u))\sigma(u,X(u)),$ $\displaystyle\gamma(f,t,u)$ $\displaystyle=f(t,X(u-)+\rho(u,X(u-)))-f(t,X(u-)).$ Lemma A.1 states basic estimates for the functions above. For approximating the solution of SDE (1) we use the randomized Milstein algorithm. For $n\in{\mathbb{N}}$ we set $\delta=T/n$ and let $t_{i}=i\delta$ for $i\in\\{0,\ldots,n\\}$. Moreover, we use the notation $\Delta Y_{i}=Y(t_{i+1})-Y(t_{i})$ for $i\in\\{0,1,\ldots,n-1\\}$, and $\displaystyle{I_{s,t}(Y,Z)=\int\limits_{s}^{t}\int\limits_{s}^{u-}\mathop{}\\!\mathrm{d}Y(v)\mathop{}\\!\mathrm{d}Z(u)}$ for $Y,Z\in\\{W,N\\}$ and $s,t\in[0,T]$. Note that $I_{s,t}(N,W)+I_{s,t}(W,N)=(W(t)-W(s))(N(t)-N(s)),$ (16) and the $\sigma$-fields $\sigma(I_{s,t}(Y,Z))$ and $\mathcal{F}_{s}$ are independent, cf. [7, Fact B.28 (ii)]. Let $\\{\xi_{i}\\}_{i=0}^{n-1}$ be independent random variables on the probability space $(\Omega,\mathcal{F},\mathbb{P})$, such that the $\sigma$-fileds $\sigma(\xi_{0},\xi_{1},\ldots,\xi_{n-1})$ and $\mathcal{F}_{\infty}$ are independent, with $\xi_{i}$ being uniformly distributed on $[t_{i},t_{i+1}]$. Then the randomized Milstein algorithm $X^{(\delta)}$ is defined recursively through $\displaystyle X^{(\delta)}(t_{0})$ $\displaystyle=X_{0},$ (17) $\displaystyle X^{(\delta)}(t_{i+1})$ $\displaystyle=X^{(\delta)}(t_{i})+\mu(\xi_{i},X^{(\delta)}(t_{i}))\delta+\sigma(t_{i},X^{(\delta)}(t_{i}))\Delta W_{i}+\rho(t_{i},X^{(\delta)}(t_{i}))\Delta N_{i}$ $\displaystyle\quad+L_{1}\sigma(t_{i},X^{(\delta)}(t_{i}))I_{t_{i},t_{i+1}}(W,W)+L_{-1}\rho(t_{i},X^{(\delta)}(t_{i}))I_{t_{i},t_{i+1}}(N,N)$ $\displaystyle\quad+L_{-1}\sigma(t_{i},X^{(\delta)}(t_{i}))I_{t_{i},t_{i+1}}(N,W)+L_{1}\rho(t_{i},X^{(\delta)}(t_{i}))I_{t_{i},t_{i+1}}(W,N),$ $\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad\quad i\in\\{0,\ldots,n-1\\}.$ In order to analyse the error of the randomized Milstein algorithm we use the so-called time-continuous Milstein approximation $(X^{(\delta)}_{c}(t))_{t\in[0,T]}$, called the randomized Milstein process. It is defined as follows $\displaystyle X^{(\delta)}_{c}(t_{0})$ $\displaystyle=X_{0},$ (18) $\displaystyle X^{(\delta)}_{c}(t)$ $\displaystyle=X^{(\delta)}_{c}(t_{i})+\mu(\xi_{i},X^{(\delta)}_{c}(t_{i}))(t-t_{i})+\sigma(t_{i},X^{(\delta)}_{c}(t_{i}))(W(t)-W(t_{i}))$ $\displaystyle\quad+\rho(t_{i},X^{(\delta)}_{c}(t_{i}))(N(t)-N(t_{i}))$ $\displaystyle\quad+L_{1}\sigma(t_{i},X^{(\delta)}_{c}(t_{i}))I_{t_{i},t}(W,W)+L_{-1}\rho(t_{i},X^{(\delta)}_{c}(t_{i}))I_{t_{i},t}(N,N)$ $\displaystyle\quad+L_{-1}\sigma(t_{i},X^{(\delta)}_{c}(t_{i}))I_{t_{i},t}(N,W)+L_{1}\rho(t_{i},X^{(\delta)}_{c}(t_{i}))I_{t_{i},t}(W,N),$ for $t\in(t_{i},t_{i+1}]$, $i\in\\{0,\ldots,n-1\\}$. This implies that for all $i\in\\{0,\ldots,n\\}$, $X^{(\delta)}(t_{i})=X^{(\delta)}_{c}(t_{i})$. Now, analog to [13], we extend the filtration $(\mathcal{F}_{t})_{t\geq 0}$ in the following way: we take $\mathcal{\bar{F}}^{n}_{t}=\sigma(\mathcal{F}_{t}\cup\mathcal{G}^{n})$, where $\mathcal{G}^{n}=\sigma(\xi_{0},\ldots,\xi_{n-1})$. Since $\mathcal{G}^{n}$ and $\mathcal{F}_{\infty}$ are independent, $W$ and $N$ are still Wiener and Poisson processes with respect to $(\mathcal{\bar{F}}^{n}_{t})_{t\geq 0}$, respectively. Since in the paper we are integrating * • $(\mathcal{\bar{F}}^{n}_{t})_{t\geq 0}$-progressively measurable processes with respect to the continuous $(\mathcal{\bar{F}}^{n}_{t})_{t\geq 0}$-semi- martingales $(t)_{t\in[0,T]}$, $(W(t))_{t\in[0,T]}$, * • $(\mathcal{\bar{F}}^{n}_{t})_{t\geq 0}$-adapted càglàd processes with respect to the càdlàg $(\mathcal{\bar{F}}^{n}_{t})_{t\geq 0}$-semimartingale$(N(t))_{t\in[0,T]}$, the (stochastic) integrals are well-defined, see, for example, [16]. Moreover, the randomized Milstein process is $(\mathcal{\bar{F}}^{n}_{t})_{t\geq 0}$-progressively measurable, since it is càdlàg and adapted. Note that the randomised Milstein process is not an implementable algorithm since it uses all values of $W$ and $N$ and these are not accessible. However we will use it as an auxiliary scheme for our proof that the randomized Milstein algorithm has convergence orders $\delta^{\min\\{\frac{2}{p},\varrho_{1}+\frac{1}{p},\varrho_{2},\varrho_{3}\\}}$. ## 3 Error analysis for the randomized Milstein process Let for all $i\in\\{1,\ldots,n\\}$, $U_{i}=(t_{i},X^{(\delta)}(t_{i})),\quad V_{i}=(\xi_{i},X^{(\delta)}(t_{i})).$ (19) The processes $X$ and $X^{(\delta)}_{c}$ can be written for all $t\in[0,T]$ as $\displaystyle X(t)=X(0)+A(t)+B(t)+C(t),$ (20) $\displaystyle X^{(\delta)}_{c}(t)=X(0)+A^{(\delta)}(t)+B^{(\delta)}(t)+C^{(\delta)}(t),$ (21) where $\displaystyle A(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\mu(s,X(s))\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s,$ (22) $\displaystyle B(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\sigma(s,X(s))\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}W(s),$ $\displaystyle C(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\rho(s,X(s-))\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}N(s),$ $\displaystyle A^{(\delta)}(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\mu(V_{i})\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s,$ (23) $\displaystyle B^{(\delta)}(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Big{(}\sigma(U_{i})+\int\limits_{t_{i}}^{s}L_{1}\sigma(U_{i})\mathop{}\\!\mathrm{d}W(u)+\int\limits_{t_{i}}^{s}L_{-1}\sigma(U_{i})\mathop{}\\!\mathrm{d}N(u)\Big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}W(s),$ $\displaystyle C^{(\delta)}(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Big{(}\rho(U_{i})+\int\limits_{t_{i}}^{s}L_{1}\rho(U_{i})\mathop{}\\!\mathrm{d}W(u)+\int\limits_{t_{i}}^{s-}L_{-1}\rho(U_{i})\mathop{}\\!\mathrm{d}N(u)\Big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}N(s).$ ###### Lemma 3.1. Under the Assumption 2.1 it holds that there exists a constant $K_{8}\in(0,\infty)$ such that for all $n\in{\mathbb{N}}$ it holds that $\displaystyle\sup_{0\leq t\leq T}{\mathbb{E}}\big{[}|X^{(\delta)}_{c}(t)|^{p}\big{]}\leq K_{8}.$ (24) ###### Proof. By induction and the fact that $\mathbb{E}[|X_{0}|^{p}]<\infty$ we get that $\max\limits_{0\leq i\leq n}{\mathbb{E}}\big{[}|X^{(\delta)}(t_{i})|^{p}\big{]}<\infty.$ (25) Moreover, by (25) and (18) we obtain that for all $n\in{\mathbb{N}}$ there exists a constant $c_{1}\in(0,\infty)$ such that $\sup\limits_{0\leq t\leq T}\mathbb{E}\big{[}|X^{\delta}_{c}(t)|^{p}\big{]}\leq c_{1}(1+\max\limits_{0\leq i\leq n-1}\mathbb{E}\big{[}|X^{(\delta)}(t_{i})|^{p}\big{]})<\infty.$ (26) Now, we denote for all $t\in[0,T]$, $X_{c}^{(\delta)}(t)=X(0)+\int\limits_{0}^{t}\Psi_{1,n}(s)\mathop{}\\!\mathrm{d}s+\int\limits_{0}^{t}\Psi_{2,n}(s)\mathop{}\\!\mathrm{d}W(s)+\int\limits_{0}^{t}\Psi_{3,n}(s)\mathop{}\\!\mathrm{d}N(s),$ (27) where $\Psi_{1,n}(s)=\sum_{i=0}^{n-1}\mu(V_{i})\mathds{1}_{(t_{i},t_{i+1}]}(s),$ (28) $\Psi_{2,n}(s)=\sum_{i=0}^{n-1}\Bigg{(}\sigma(U_{i})+\int\limits_{t_{i}}^{s}L_{1}\sigma(U_{i})\mathop{}\\!\mathrm{d}W(u)+\int\limits_{t_{i}}^{s}L_{-1}\sigma(U_{i})\mathop{}\\!\mathrm{d}N(u)\Bigg{)}\mathds{1}_{(t_{i},t_{i+1}]}(s),$ (29) $\Psi_{3,n}(s)=\sum_{i=0}^{n-1}\Bigg{(}\rho(U_{i})+\int\limits_{t_{i}}^{s}L_{1}\rho(U_{i})\mathop{}\\!\mathrm{d}W(u)+\int\limits_{t_{i}}^{s-}L_{-1}\rho(U_{i})\mathop{}\\!\mathrm{d}N(u)\Bigg{)}\mathds{1}_{(t_{i},t_{i+1}]}(s).$ (30) By Lemma A.6 we have for all $(k,Z)\in\\{(1,s),(2,W),(3,N)\\}$ that $\mathbb{E}\Bigg{[}\Biggl{|}\int\limits_{0}^{t}\Psi_{k,n}(s)dZ(s)\Biggl{|}^{p}\Bigg{]}\leq\hat{c}\int\limits_{0}^{t}\mathbb{E}\big{[}|\Psi_{k,n}(s)|^{p}\big{]}ds.$ (31) From (8) we get that there exist constants $c_{2},c_{3}\in(0,\infty)$ such that $\mathbb{E}\big{[}|\Psi_{1,n}(s)|^{p}\big{]}\leq K_{4}^{p}\sum\limits_{i=0}^{n-1}\mathbb{E}\big{[}(1+|X^{(\delta)}(t_{i})|)^{p}\big{]}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\leq c_{2}+c_{3}\sum\limits_{i=0}^{n-1}\mathbb{E}\big{[}|X^{(\delta)}(t_{i})|^{p}\big{]}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s).$ (32) By (8), (11), and Lemma A.6 we obtain that there exist constants $c_{4},c_{5}\in(0,\infty)$ such that for all $(k,f)\in\\{(2,\sigma),(3,\rho)\\}$ it holds that $\displaystyle\int\limits_{0}^{t}\mathbb{E}\big{[}|\Psi_{k,n}(s)|^{p}\big{]}\mathop{}\\!\mathrm{d}s$ $\displaystyle\leq\hat{c}\,{\mathbb{E}}\Bigg{[}\sum_{i=0}^{n-1}\int\limits_{0}^{t}|f(U_{i})|^{p}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s\Bigg{]}$ (33) $\displaystyle\quad+\hat{c}\,{\mathbb{E}}\Bigg{[}\sum_{i=0}^{n-1}\int\limits_{0}^{t}\Bigg{|}\int\limits_{t_{i}}^{s}L_{1}f(U_{i})\mathop{}\\!\mathrm{d}W(u)\Bigg{|}^{p}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s\Bigg{]}$ $\displaystyle\quad+\hat{c}\,{\mathbb{E}}\Bigg{[}\sum_{i=0}^{n-1}\int\limits_{0}^{t}\Bigg{|}\int\limits_{t_{i}}^{s}L_{-1}f(U_{i})\mathop{}\\!\mathrm{d}N(u)\Bigg{|}^{p}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s\Bigg{]}$ $\displaystyle\leq c_{4}+c_{5}\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\mathbb{E}\big{[}|X^{(\delta)}(t_{i})|^{p}\big{]}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s.$ Here we used that $\displaystyle{\int\limits_{t_{i}}^{s}L_{-1}f(U_{i})\mathop{}\\!\mathrm{d}N(u)}$ and $\displaystyle{\int\limits_{t_{i}}^{s-}L_{-1}f(U_{i})\mathop{}\\!\mathrm{d}N(u)}$ differ only at finitely many points. Combining (27), (32), and (33) we obtain that there exist constants $c_{6},c_{7},c_{8}\in(0,\infty)$ such that $\displaystyle{\mathbb{E}}\Big{[}\big{|}X_{c}^{(\delta)}(t)\big{|}^{p}\Big{]}\leq c_{6}\Bigg{(}{\mathbb{E}}\Big{[}|X(0)|^{p}\Big{]}+\sum_{k=1}^{3}\int\limits_{0}^{t}{\mathbb{E}}\Big{[}|\Psi_{k,n}(s)|^{p}\Big{]}\mathop{}\\!\mathrm{d}s\Bigg{)}$ (34) $\displaystyle\leq c_{7}\Big{(}{\mathbb{E}}\Big{[}|X(0)|^{p}\Big{]}+1\Big{)}+c_{8}\int\limits_{0}^{t}\sup_{0\leq u\leq s}{\mathbb{E}}\Big{[}\big{|}X_{c}^{(\delta)}(u)\big{|}^{p}\Big{]}\mathop{}\\!\mathrm{d}s.$ Hence, $\displaystyle\sup_{0\leq s\leq t}{\mathbb{E}}\Big{[}\big{|}X_{c}^{(\delta)}(s)\big{|}^{p}\Big{]}\leq c_{7}\Big{(}{\mathbb{E}}\Big{[}|X(0)|^{p}\Big{]}+1\Big{)}+c_{8}\int\limits_{0}^{t}\sup_{0\leq u\leq s}{\mathbb{E}}\Big{[}\big{|}X_{c}^{(\delta)}(u)\big{|}^{p}\Big{]}\mathop{}\\!\mathrm{d}s.$ (35) The mapping $t\mapsto\sup_{0\leq s\leq t}{\mathbb{E}}\Big{[}\big{|}X_{c}^{(\delta)}(s)\big{|}^{p}\Big{]}$ is monotone and hence Borel measurable. Moreover, by (26) it is bounded. Hence, applying Gronwall’s lemma proves the claim. ∎ Next we prove the convergence rate of the randomized Milstein algorithm. ###### Theorem 3.2. Let Assumption 2.1 hold. Then there exists $C\in(0,\infty)$ such that for all $n\in{\mathbb{N}}$ it holds that $\displaystyle\sup_{0\leq t\leq T}\|X(t)-X^{(\delta)}_{c}(t)\|_{L^{p}(\Omega)}\leq C\delta^{\min\\{\frac{2}{p},\varrho_{1}+\frac{1}{p},\varrho_{2},\varrho_{3}\\}}.$ (36) ###### Proof. For all $t\in[0,T]$ it holds that $\displaystyle X(t)-X^{(\delta)}_{c}(t)=\big{(}A(t)-A^{(\delta)}(t)\big{)}+\big{(}B(t)-B^{(\delta)}(t)\big{)}+\big{(}C(t)-C^{(\delta)}(t)\big{)}.$ (37) We first rewrite each summand of the right hand side of equation (37). We obtain $\displaystyle A(t)-A^{(\delta)}(t)$ $\displaystyle=\tilde{A}^{(\delta)}_{1}(t)+\tilde{A}^{(\delta)}_{2}(t)+\tilde{A}^{(\delta)}_{3}(t),$ (38) where $\displaystyle\tilde{A}^{(\delta)}_{1}(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{(}\mu(s,X(s))-\mu(s,X(t_{i}))\big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s,$ (39) $\displaystyle\tilde{A}^{(\delta)}_{2}(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{(}\mu(s,X(t_{i}))-\mu(\xi_{i},X(t_{i}))\big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s,$ $\displaystyle\tilde{A}^{(\delta)}_{3}(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{(}\mu(\xi_{i},X(t_{i}))-\mu(\xi_{i},X^{(\delta)}_{c}(t_{i}))\big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s.$ For $\tilde{A}^{(\delta)}_{1}(t)$ we apply the parametric version of the Meyer-Itô formula (14), that is $\mu(s,X(s))-\mu(s,X(t_{i}))=\int\limits_{t_{i}}^{s}\alpha(\mu,s,u)\mathop{}\\!\mathrm{d}u+\int\limits_{t_{i}}^{s}\beta(\mu,s,u)\mathop{}\\!\mathrm{d}W(u)+\int\limits_{t_{i}}^{s}\gamma(\mu,s,u)\mathop{}\\!\mathrm{d}N(u).$ (40) Hence, $\tilde{A}_{1}^{(\delta)}(t)=\sum\limits_{j=1}^{3}\tilde{M}_{j}^{(\delta)}(t),$ (41) where $\displaystyle\tilde{M}_{1}^{(\delta)}(t)=\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\Bigg{(}\int\limits_{t_{i}}^{s}\alpha(\mu,s,u)\mathop{}\\!\mathrm{d}u\Bigg{)}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s,$ (42) $\displaystyle\tilde{M}_{2}^{(\delta)}(t)=\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\Bigg{(}\int\limits_{t_{i}}^{s}\beta(\mu,s,u)\mathop{}\\!\mathrm{d}W(u)\Bigg{)}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s,$ $\displaystyle\tilde{M}_{3}^{(\delta)}(t)=\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\Biggl{(}\int\limits_{t_{i}}^{s}\gamma(\mu,s,u)\mathop{}\\!\mathrm{d}N(u)\Biggr{)}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s.$ Next we obtain for the second summand of (37), $\displaystyle B(t)-B^{(\delta)}(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{(}\sigma(s,X(s))-\sigma(t_{i},X(s))\big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}W(s)$ (43) $\displaystyle\quad+\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Bigg{(}\sigma(t_{i},X(s))-\sigma(t_{i},X(t_{i}))$ $\displaystyle\quad\quad\quad\quad\quad\quad-\int\limits_{t_{i}}^{s}L_{1}\sigma(U_{i})\mathop{}\\!\mathrm{d}W(u)-\int\limits_{t_{i}}^{s}L_{-1}\sigma(U_{i})\mathop{}\\!\mathrm{d}N(u)\Bigg{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}W(s)$ $\displaystyle\quad+\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{(}\sigma(t_{i},X(t_{i}))-\sigma(U_{i})\big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}W(s).$ Again we apply the parametric version of the Meyer-Itô formula (14) and obtain $\displaystyle\sigma(t_{i},X(s))-\sigma(t_{i},X(t_{i}))-\int\limits_{t_{i}}^{s}L_{1}\sigma(U_{i})\mathop{}\\!\mathrm{d}W(u)-\int\limits_{t_{i}}^{s}L_{-1}\sigma(U_{i})\mathop{}\\!\mathrm{d}N(u)$ (44) $\displaystyle=\int\limits_{t_{i}}^{s}\alpha(\sigma,t_{i},u)\mathop{}\\!\mathrm{d}u+\int\limits_{t_{i}}^{s}\Bigl{(}\beta(\sigma,t_{i},u)-L_{1}\sigma(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}W(u)+\int\limits_{t_{i}}^{s}\Bigl{(}\gamma(\sigma,t_{i},u)-L_{-1}\sigma(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}N(u).$ Hence, $\displaystyle B(t)-B^{(\delta)}(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Bigg{(}\int\limits\limits_{t_{i}}^{s}\alpha(\sigma,t_{i},u)\mathop{}\\!\mathrm{d}u\Bigg{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}W(s)$ (45) $\displaystyle\quad+\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Biggl{(}\int\limits\limits_{t_{i}}^{s}\Bigl{(}\beta(\sigma,t_{i},u)-L_{1}\sigma(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}W(u)\Biggr{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}W(s)$ $\displaystyle\quad+\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Biggl{(}\int\limits_{t_{i}}^{s}\Bigl{(}\gamma(\sigma,t_{i},u)-L_{-1}\sigma(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}N(u)\Biggr{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}W(s)$ $\displaystyle\quad+\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{(}\sigma(s,X(s))-\sigma(t_{i},X(s))\big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}W(s)$ $\displaystyle\quad+\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{(}\sigma(t_{i},X(t_{i}))-\sigma(U_{i})\big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}W(s).$ For the third summand of (37) we get $\displaystyle C(t)-C^{(\delta)}(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{(}\rho(s,X(s-))-\rho(t_{i},X(s-))\big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}N(s)$ (46) $\displaystyle\quad+\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Bigg{(}\rho(t_{i},X(s-))-\rho(t_{i},X(t_{i}))$ $\displaystyle\quad\quad\quad\quad\quad\quad-\int\limits_{t_{i}}^{s}L_{1}\rho(U_{i})\mathop{}\\!\mathrm{d}W(u)-\int\limits_{t_{i}}^{s-}L_{-1}\sigma(U_{i})\mathop{}\\!\mathrm{d}N(u)\Bigg{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}N(s)$ $\displaystyle\quad+\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{(}\rho(t_{i},X(t_{i}))-\rho(U_{i})\big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}N(s).$ Again we apply the parametric version of the Meyer-Itô formula (14) to obtain $\displaystyle\rho(t_{i},X(s-))-\rho(t_{i},X(t_{i}))-\int\limits_{t_{i}}^{s}L_{1}\rho(U_{i})\mathop{}\\!\mathrm{d}W(u)-\int\limits_{t_{i}}^{s-}L_{-1}\rho(U_{i})\mathop{}\\!\mathrm{d}N(u)$ (47) $\displaystyle=\int\limits_{t_{i}}^{s}\alpha(\rho,t_{i},u)\mathop{}\\!\mathrm{d}u+\int\limits_{t_{i}}^{s}\Bigl{(}\beta(\rho,t_{i},u)-L_{1}\rho(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}W(u)+\int\limits_{t_{i}}^{s-}\Bigl{(}\gamma(\rho,t_{i},u)-L_{-1}\rho(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}N(u);$ due to the continuity of the processes $\int\limits_{t_{i}}^{s}\alpha(\rho,t_{i},u)\mathop{}\\!\mathrm{d}u=\int\limits_{t_{i}}^{s-}\alpha(\rho,t_{i},u)\mathop{}\\!\mathrm{d}u,\quad\int\limits_{t_{i}}^{s}\beta(\rho,t_{i},u)\mathop{}\\!\mathrm{d}W(u)=\int\limits_{t_{i}}^{s-}\beta(\rho,t_{i},u)\mathop{}\\!\mathrm{d}W(u).$ (48) Therefore, for all $t\in[0,T]$, $\displaystyle C(t)-C^{(\delta)}(t)$ $\displaystyle=\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Bigg{(}\int\limits_{t_{i}}^{s}\alpha(\rho,t_{i},u)\mathop{}\\!\mathrm{d}u\Bigg{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}N(s)$ (49) $\displaystyle\quad+\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Biggl{(}\int\limits_{t_{i}}^{s}\Bigl{(}\beta(\rho,t_{i},u)-L_{1}\rho(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}W(u)\Biggr{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}N(s)$ $\displaystyle\quad+\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Biggl{(}\int\limits_{t_{i}}^{s-}\Bigl{(}\gamma(\rho,t_{i},u)-L_{-1}\rho(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}N(u)\Biggr{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}N(s)$ $\displaystyle\quad+\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{(}\rho(s,X(s-))-\rho(t_{i},X(s-))\big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}N(s)$ $\displaystyle\quad+\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{(}\rho(t_{i},X(t_{i}))-\rho(U_{i})\big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}N(s).$ Now we estimate all terms in (38), (41), (45), and (49). We apply Lemma A.6 and Assumption 2.1 (i) for $(f,v,Z)\in\\{(\mu,\xi_{i},\operatorname{Id}),(\sigma,t_{i},W),(\rho,t_{i},N)\\}$ . This shows that there exists a constant $c_{1}\in(0,\infty)$ such that $\displaystyle{\mathbb{E}}\Bigg{[}\Bigg{|}\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{(}f(v,X(t_{i}))-f(v,X^{(\delta)}_{c}(t_{i}))\big{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}Z(s)\Bigg{|}^{p}\Bigg{]}$ (50) $\displaystyle\leq\hat{c}\,{\mathbb{E}}\Bigg{[}\int\limits_{0}^{t}\sum_{i=0}^{n-1}\big{|}f(v,X(t_{i}))-f(v,X^{(\delta)}_{c}(t_{i}))\big{|}^{p}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s\Bigg{]}$ $\displaystyle\leq c_{1}\int\limits_{0}^{t}\sum_{i=0}^{n-1}{\mathbb{E}}\Big{[}\big{|}X(t_{i})-X^{(\delta)}_{c}(t_{i})\big{|}^{p}\Big{]}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s.$ Moreover, we obtain that there exists $c_{2}\in(0,\infty)$ such that for $(f,Z)\in\\{(\sigma,W),(\rho,N)\\}$ and for all $t\in[0,T]$ it holds that $\displaystyle\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\Bigl{(}f(s,X(s-))-f(t_{i},X(s-))\Bigr{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}Z(s)\Bigg{|}^{p}\Bigg{]}$ (51) $\displaystyle\leq\hat{c}\sum\limits_{i=0}^{n-1}\mathbb{E}\Bigg{[}\int\limits_{t_{i}}^{t_{i+1}}|f(s,X(s-))-f(t_{i},X(s-))|^{p}\mathop{}\\!\mathrm{d}s\Bigg{]}$ $\displaystyle\leq\hat{c}K_{1}^{p}\sum\limits_{i=0}^{n-1}\mathbb{E}\Bigg{[}\int\limits_{t_{i}}^{t_{i+1}}(1+|X(s-)|)^{p}\cdot(s-t_{i})^{p\varrho_{f}}\mathop{}\\!\mathrm{d}s\Bigg{]}$ $\displaystyle\leq\hat{c}K_{1}^{p}\delta^{p\varrho_{f}}\sum\limits_{i=0}^{n-1}\mathbb{E}\Bigg{[}\int\limits_{t_{i}}^{t_{i+1}}(1+|X(s)|)^{p}\mathop{}\\!\mathrm{d}s\Bigg{]}\leq 2^{p-1}\hat{c}K_{1}^{p}\delta^{p\varrho_{f}}\Bigl{(}1+\mathbb{E}\Big{[}\sup\limits_{0\leq t\leq T}|X(t)|^{p}\Big{]}\Bigr{)}\leq c_{2}\delta^{p\varrho_{f}}.$ By Lemma A.6 we get that there exist constants $c_{3},c_{4}\in(0,\infty)$ such that for $(f,v,Z)\in\\{(\mu,s,\operatorname{Id}),(\sigma,t_{i},W),(\rho,t_{i},N)\\}$ $\displaystyle{\mathbb{E}}\Bigg{[}\Bigg{|}\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Bigg{(}\int\limits_{t_{i}}^{s}\alpha(f,v,u)\mathop{}\\!\mathrm{d}u\Bigg{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}Z(s)\Bigg{|}^{p}\Bigg{]}$ (52) $\displaystyle\leq c_{3}\,\sum_{i=0}^{n-1}\int\limits_{t_{i}}^{t_{i+1}}{\mathbb{E}}\Bigg{[}\Bigg{(}\int\limits_{t_{i}}^{s}|\alpha_{1}(f,v,u)|\mathop{}\\!\mathrm{d}u\Bigg{)}^{p}\Bigg{]}\mathop{}\\!\mathrm{d}s+c_{4}\,\sum_{i=0}^{n-1}\int\limits_{t_{i}}^{t_{i+1}}{\mathbb{E}}\Bigg{[}\Bigg{(}\int\limits_{t_{i}}^{s}|\alpha_{2}(f,v,u)|\mathop{}\\!\mathrm{d}u\Bigg{)}^{p}\Bigg{]}\mathop{}\\!\mathrm{d}s.$ Next we estimate the expectations in equation (LABEL:Meq11) separately. For the first term we use (8), (9), and (12) to obtain that there exists a constant $c_{5}\in(0,\infty)$ such that for all $s\in[t_{i},t_{i+1}]$, $v\in\\{s,t_{i}\\}$, $\displaystyle{\mathbb{E}}\Bigg{[}\Bigg{(}\int\limits_{t_{i}}^{s}|\alpha_{1}(f,v,u)|\mathop{}\\!\mathrm{d}u\Bigg{)}^{p}\Bigg{]}={\mathbb{E}}\Bigg{[}\Bigg{(}\int\limits_{t_{i}}^{s}\big{|}f^{\prime}_{y}(v,X(u))\big{|}\cdot\big{|}\mu(u,X(u))\big{|}\mathop{}\\!\mathrm{d}u\Bigg{)}^{p}\Bigg{]}$ (53) $\displaystyle\leq(K_{1}K_{4})^{p}\,{\mathbb{E}}\Bigg{[}\Bigg{(}\int\limits_{t_{i}}^{s}\big{(}1+|X(u)|\big{)}\mathop{}\\!\mathrm{d}u\Bigg{)}^{p}\Bigg{]}\leq(K_{1}K_{4})^{p}\delta^{p}\,{\mathbb{E}}\Big{[}\big{(}1+\sup_{0\leq t\leq T}|X(t)|\big{)}^{p}\Big{]}\leq c_{5}\,\delta^{p}.$ For the second term we obtain that there exist constants $c_{6},c_{7},c_{8}\in(0,\infty)$ such that for all $s\in[t_{i},t_{i+1}]$, $v\in\\{s,t_{i}\\}$, $\displaystyle{\mathbb{E}}\Bigg{[}\Bigg{(}\int\limits_{t_{i}}^{s}|\alpha_{2}(f,v,u)|\mathop{}\\!\mathrm{d}u\Bigg{)}^{p}\Bigg{]}={\mathbb{E}}\Bigg{[}\Bigg{(}\frac{1}{2}\int\limits_{t_{i}}^{s}\big{|}f^{\prime\prime}_{yy}(v,X(u))\big{|}\cdot\big{|}\sigma^{2}(u,X(u))\big{|}\mathop{}\\!\mathrm{d}u\Bigg{)}^{p}\Bigg{]}$ (54) $\displaystyle\leq\Big{(}\frac{K_{1}}{2}\Big{)}^{p}{\mathbb{E}}\Bigg{[}\Bigg{(}\int\limits_{t_{i}}^{s}\sigma^{2}(u,X(u))\mathop{}\\!\mathrm{d}u\Bigg{)}^{p}\Bigg{]}\leq c_{6}\,{\mathbb{E}}\Bigg{[}\Bigg{(}\int\limits_{t_{i}}^{t_{i+1}}\big{(}1+|X(u)|^{2}\big{)}\mathop{}\\!\mathrm{d}u\Bigg{)}^{p}\Bigg{]}$ $\displaystyle\leq c_{7}\,\delta^{p}\,{\mathbb{E}}\Big{[}\big{(}1+\sup_{0\leq t\leq T}|X(t)|^{2}\big{)}^{p}\Big{]}\leq c_{8}\,\delta^{p},$ since $|f_{yy}^{\prime\prime}(t,y)|\leq K_{1}$ for all $t\in[0,T]$ and $y\in{\mathbb{R}}$. Hence, combining equations (LABEL:Meq11), (LABEL:Meq12), and (LABEL:Meq13) we obtain that there exists a constant $c_{9}\in(0,\infty)$ such that $\displaystyle{\mathbb{E}}\Bigg{[}\Bigg{|}\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Bigg{(}\int\limits_{t_{i}}^{s}\alpha(f,v,u)\mathop{}\\!\mathrm{d}u\Bigg{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}Z(s)\Bigg{|}^{p}\Bigg{]}\leq c_{9}\,\delta^{p}.$ (55) For $(f,Z)\in\\{(\sigma,W),(\rho,N)\\}$ and $t\in[0,T]$ we get $\displaystyle\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\Biggl{(}\int\limits_{t_{i}}^{s}\Bigl{(}\beta(f,t_{i},u)-L_{1}f(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}W(u)\Biggr{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}Z(s)\Bigg{|}^{p}\Bigg{]}$ (56) $\displaystyle\leq\hat{c}\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{t_{i}}^{s}\Bigl{(}\beta(f,t_{i},u)-L_{1}f(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}W(u)\Bigg{|}^{p}\Bigg{]}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s.$ Further, there exists a constant $c_{10}\in(0,\infty)$ such that for all $s\in[t_{i},t_{i+1}]$, $\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{t_{i}}^{s}\Bigl{(}\beta(f,t_{i},u)-L_{1}f(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}W(u)\Bigg{|}^{p}\Bigg{]}\leq c_{10}(s-t_{i})^{\frac{p}{2}-1}\cdot\mathbb{E}\Bigg{[}\int\limits_{t_{i}}^{s}\big{|}\beta(f,t_{i},u)-L_{1}f(U_{i})\big{|}^{p}\mathop{}\\!\mathrm{d}s\Bigg{]}.$ (57) Moreover, for $u\in[t_{i},t_{i+1}]$, $\displaystyle|\beta(f,t_{i},u)-L_{1}f(U_{i})|\leq|\beta(f,t_{i},u)-L_{1}f(t_{i},X(u))|+|L_{1}f(t_{i},X(u))-L_{1}f(t_{i},X^{(\delta)}(t_{i}))|$ (58) $\displaystyle\leq K_{3}|X(u)-X(t_{i})|+K_{3}|X(t_{i})-X^{(\delta)}(t_{i})|+K_{1}^{2}(1+|X(u)|)\cdot|u-t_{i}|^{\varrho_{2}},$ and by (12), (13) we have that there exist constants $c_{11},c_{12},c_{13}\in(0,\infty)$ such that $\mathbb{E}\big{[}|\beta(f,t_{i},u)-L_{1}f(U_{i})|^{p}\Big{]}\leq c_{11}(u-t_{i})+c_{12}(u-t_{i})^{p\varrho_{2}}+c_{13}\mathbb{E}\big{[}|X(t_{i})-X^{(\delta)}(t_{i})|^{p}\big{]}.$ (59) Hence, there exist constants $c_{14},c_{15},c_{16}\in(0,\infty)$ such that $\displaystyle\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\Biggl{(}\int\limits_{t_{i}}^{s}\Bigl{(}\beta(f,t_{i},u)-L_{1}f(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}W(u)\Biggr{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}Z(s)\Bigg{|}^{p}\Bigg{]}$ (60) $\displaystyle\leq c_{14}\delta^{\frac{p}{2}+1}+c_{15}\delta^{p(\varrho_{2}+\frac{1}{2})}+c_{16}\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\mathbb{E}\Big{[}\big{|}X(t_{i})-X^{(\delta)}(t_{i})\big{|}^{p}\Big{]}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s.$ Moreover, for $(f,Z)\in\\{(\sigma,W),(\rho,N)\\}$ and $t\in[0,T]$ we obtain $\displaystyle\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Biggl{(}\int\limits_{t_{i}}^{s-}\Bigl{(}\gamma(f,t_{i},u)-L_{-1}f(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}N(u)\Biggr{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}Z(s)\Bigg{|}^{p}\Bigg{]}$ (61) $\displaystyle\leq\hat{c}\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{t_{i}}^{s}\Bigl{(}\gamma(f,t_{i},u)-L_{-1}f(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}N(u)\Bigg{|}^{p}\Bigg{]}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s,$ where $\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{t_{i}}^{s}\Bigl{(}\gamma(f,t_{i},u)-L_{-1}f(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}N(u)\Bigg{|}^{p}\Bigg{]}\leq\hat{c}\,\mathbb{E}\Bigg{[}\int\limits_{t_{i}}^{s}|\gamma(f,t_{i},u)-L_{-1}f(U_{i})|^{p}\mathop{}\\!\mathrm{d}s\Bigg{]}.$ (62) Further, there exist constants $c_{17},c_{18},c_{19}\in(0,\infty)$ such that for all $s\in[t_{i},t_{i+1}]$, $\displaystyle\int\limits_{t_{i}}^{s}|\gamma(f,t_{i},u)-L_{-1}f(U_{i})|^{p}\mathop{}\\!\mathrm{d}u$ (63) $\displaystyle\leq c_{17}\int\limits_{t_{i}}^{s}|X(u)-X(t_{i})|^{p}\mathop{}\\!\mathrm{d}u+c_{18}\delta|X(t_{i})-X^{(\delta)}(t_{i})|^{p}+c_{19}\int\limits_{t_{i}}^{s}(1+|X(u)|)^{p}\cdot|u-t_{i}|^{p\varrho_{3}}\mathop{}\\!\mathrm{d}u,$ hence, by (12) and (13) there exist constants $c_{20},c_{21},c_{22}\in(0,\infty)$ such that $\displaystyle\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{0}^{t}\sum_{i=0}^{n-1}\Biggl{(}\int\limits_{t_{i}}^{s-}\Bigl{(}\gamma(f,t_{i},u)-L_{-1}f(U_{i})\Bigr{)}\mathop{}\\!\mathrm{d}N(u)\Biggr{)}\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}Z(s)\Bigg{|}^{p}\Bigg{]}$ (64) $\displaystyle\leq c_{20}\delta^{2}+c_{21}\delta^{p(\varrho_{3}+\frac{1}{p})}+c_{22}\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\mathbb{E}\Big{[}|X(t_{i})-X^{(\delta)}(t_{i})|^{p}\Big{]}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s.$ With this preliminary calculations we can estimate the error of the randomized Milstein algorithm as follows. It holds that there exists a constant $c_{23}\in(0,\infty)$ such that $\displaystyle{\mathbb{E}}\Big{[}|X(t)-X^{(\delta)}_{c}(t)|^{p}\Big{]}$ (65) $\displaystyle\leq c_{23}\Big{(}{\mathbb{E}}\Big{[}\big{|}A(t)-A^{(\delta)}(t)\big{|}^{p}\Big{]}+{\mathbb{E}}\Big{[}\big{|}B(t)-B^{(\delta)}(t)\big{|}^{p}\Big{]}+{\mathbb{E}}\Big{[}\big{|}C(t)-C^{(\delta)}(t)\big{|}^{p}\Big{]}\Big{)}.$ Combining (45) resp. (49) with (LABEL:Meq10), (LABEL:Meq10a), (LABEL:Meq13a), (LABEL:Meq13b), and (LABEL:Meq13c), we obtain that there exist constants $c_{24},c_{25},c_{26},c_{27}\in(0,\infty)$ such that for all $t\in[0,T]$, $\mathbb{E}\Big{[}\big{|}B(t)-B^{(\delta)}(t)\big{|}^{p}\Big{]}\leq c_{24}\delta^{p\min\\{\frac{2}{p},\varrho_{2},\varrho_{3}+\frac{1}{p}\\}}+c_{25}\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\mathbb{E}\Big{[}\big{|}X(t_{i})-X^{(\delta)}(t_{i})\big{|}^{p}\Big{]}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s,$ (66) $\mathbb{E}\Big{[}\big{|}C(t)-C^{(\delta)}(t)\big{|}^{p}\Big{]}\leq c_{26}\delta^{p\min\\{\frac{2}{p},\varrho_{3},\varrho_{2}+\frac{1}{2}\\}}+c_{27}\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\mathbb{E}\Big{[}\big{|}X(t_{i})-X^{(\delta)}(t_{i})\big{|}^{p}\Big{]}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s.$ (67) Next we estimate the remaining terms in (38) and (41). The estimation of $\mathbb{E}[|\tilde{M}_{1}^{(\delta)}(t)|^{p}]$ is already included in (LABEL:Meq13a). Analog to the steps in [13, pages 8–10] and by applying Lemma A.1 we obtain that there exists a constant $c_{28}\in(0,\infty)$ such that for all $t\in[0,T]$, $\mathbb{E}\Big{[}\big{|}\tilde{M}^{(\delta)}_{2}(t)\big{|}^{p}\Big{]}\leq c_{28}\delta^{p\min\\{\frac{1}{2}+\varrho_{1},1\\}}.$ (68) We now show the upper bound for $\mathbb{E}[|\tilde{M}^{(\delta)}_{3}(t)|^{p}]$. There exists a constant $c_{29}\in(0,\infty)$ such that for all $t\in[0,T]$ there exists $\ell\in\\{0,1,\ldots,n-1\\}$ with $t\in[t_{\ell},t_{\ell+1}]$ and $\displaystyle\mathbb{E}\Big{[}\big{|}\tilde{M}^{(\delta)}_{3}(t)\big{|}^{p}\Big{]}\leq c_{29}\Bigg{(}\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{0}^{t_{\ell}}\sum\limits_{i=0}^{n-1}\Bigg{(}\int\limits_{t_{i}}^{s}(\gamma(\mu,s,u)-\gamma(\mu,t_{i},u))\mathop{}\\!\mathrm{d}N(u)\Bigg{)}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s\Bigg{|}^{p}\Bigg{]}$ (69) $\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad\quad+\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{0}^{t_{\ell}}\sum\limits_{i=0}^{n-1}\Bigg{(}\int\limits_{t_{i}}^{s}\gamma(\mu,t_{i},u)\mathop{}\\!\mathrm{d}\tilde{N}(u)\Bigg{)}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s\Bigg{|}^{p}\Bigg{]}$ $\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad\quad+\lambda^{p}\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{0}^{t_{\ell}}\sum\limits_{i=0}^{n-1}\Bigg{(}\int\limits_{t_{i}}^{s}\gamma(\mu,t_{i},u)\mathop{}\\!\mathrm{d}u\bigg{)}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s\Bigg{|}^{p}\Bigg{]}$ $\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad\quad+\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{t_{\ell}}^{t}\Bigg{(}\int\limits_{t_{\ell}}^{s}\gamma(\mu,s,u)\mathop{}\\!\mathrm{d}N(u)\Bigg{)}\mathop{}\\!\mathrm{d}s\Bigg{|}^{p}\Bigg{]}\Big{)}.$ By Lemmas A.6 and A.1 we obtain that there exists a constant $c_{30}\in(0,\infty)$ such that $\displaystyle\mathbb{E}\Bigg{[}\Bigg{|}\int\limits_{0}^{t_{\ell}}\sum\limits_{i=0}^{n-1}\Bigg{(}\int\limits_{t_{i}}^{s}(\gamma(\mu,s,u)-\gamma(\mu,t_{i},u))\mathop{}\\!\mathrm{d}N(u)\Bigg{)}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s\Bigg{|}^{p}\Bigg{]}$ (70) $\displaystyle\leq\hat{c}\sum\limits_{i=0}^{n-1}\int\limits_{t_{i}}^{t_{i+1}}\mathbb{E}\Bigg{[}\int\limits_{t_{i}}^{s}|\gamma(\mu,s,u)-\gamma(\mu,t_{i},u)|^{p}\mathop{}\\!\mathrm{d}u\Bigg{]}\mathop{}\\!\mathrm{d}s$ $\displaystyle\leq\hat{c}K_{7}\sum\limits_{i=0}^{n-1}\int\limits_{t_{i}}^{t_{i+1}}\mathbb{E}\Bigg{[}\int\limits_{t_{i}}^{s}(1+|X(u-)|)^{p}\cdot(s-t_{i})^{p\varrho_{1}}\mathop{}\\!\mathrm{d}u\Bigg{]}\mathop{}\\!\mathrm{d}s\leq c_{30}\delta^{p(\varrho_{1}+\frac{1}{p})}.$ Similar as above, by the Hölder inequality, (12), and Lemma A.1 we obtain that there exists a constant $c_{31}\in(0,\infty)$ such that for all $t\in[t_{\ell},t_{\ell+1}]$, $\displaystyle{\mathbb{E}}\Bigg{[}\Bigg{|}\int\limits_{t_{\ell}}^{t}\Bigg{(}\int\limits_{t_{\ell}}^{s}\gamma(\mu,s,u)\mathop{}\\!\mathrm{d}N(u)\Bigg{)}\mathop{}\\!\mathrm{d}s\Bigg{|}^{p}\Bigg{]}\leq{\mathbb{E}}\Bigg{[}\Bigg{(}\int\limits_{t_{\ell}}^{t}\Bigg{|}\int\limits_{t_{\ell}}^{s}\gamma(\mu,s,u)\mathop{}\\!\mathrm{d}N(u)\Bigg{|}\mathop{}\\!\mathrm{d}s\Bigg{)}^{p}\Bigg{]}$ (71) $\displaystyle\leq\delta^{p-1}\,\int\limits_{t_{\ell}}^{t_{\ell+1}}{\mathbb{E}}\Bigg{[}\Bigg{|}\int\limits_{t_{\ell}}^{s}\gamma(\mu,s,u)\mathop{}\\!\mathrm{d}N(u)\Bigg{|}^{p}\Bigg{]}\mathop{}\\!\mathrm{d}s\leq\hat{c}\delta^{p-1}\,\int\limits_{t_{\ell}}^{t_{\ell+1}}{\mathbb{E}}\Bigg{[}\int\limits_{t_{\ell}}^{s}|\gamma(\mu,s,u)|^{p}\mathop{}\\!\mathrm{d}u\Bigg{]}\mathop{}\\!\mathrm{d}s\leq c_{31}\delta^{p+1}.$ Further we obtain that there exists a constant $c_{32}\in(0,\infty)$ such that $\displaystyle{\mathbb{E}}\Bigg{[}\Bigg{|}\int\limits_{0}^{t_{\ell}}\sum\limits_{i=0}^{n-1}\Bigg{(}\int\limits_{t_{i}}^{s}\gamma(\mu,t_{i},u)\mathop{}\\!\mathrm{d}u\Bigg{)}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s\Bigg{|}^{p}\Bigg{]}$ (72) $\displaystyle\leq T^{p-1}\sum\limits_{i=0}^{n-1}\int\limits_{t_{i}}^{t_{i+1}}{\mathbb{E}}\Biggl{[}\Biggl{|}\int\limits_{t_{i}}^{s}\gamma(\mu,t_{i},u)\mathop{}\\!\mathrm{d}u\Biggl{|}^{p}\Biggr{]}\mathop{}\\!\mathrm{d}s\leq T^{p-1}\delta^{p-1}\sum\limits_{i=0}^{n-1}\int\limits_{t_{i}}^{t_{i+1}}{\mathbb{E}}\Biggl{[}\int\limits_{t_{i}}^{s}|\gamma(\mu,t_{i},u)|^{p}\mathop{}\\!\mathrm{d}u\Biggr{]}\mathop{}\\!\mathrm{d}s\leq c_{32}\delta^{p}.$ Moreover, we have $\mathbb{E}\Bigg{[}\Biggl{|}\int\limits_{0}^{t_{\ell}}\sum\limits_{i=0}^{n-1}\Bigg{(}\int\limits_{t_{i}}^{s}\gamma(\mu,t_{i},u)\mathop{}\\!\mathrm{d}\tilde{N}(u)\Bigg{)}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s\Bigg{|}^{p}\Bigg{]}={\mathbb{E}}\Big{[}\big{|}\tilde{Z}_{\ell-1}\big{|}^{p}\Big{]},$ (73) where $\tilde{Z}_{k}=\sum\limits_{i=0}^{k}\tilde{Y}_{i},\quad k\in\\{0,1,\ldots,n-1\\},$ (74) with $Z_{-1}=0$, and $\tilde{Y}_{i}=\int\limits_{t_{i}}^{t_{i+1}}\Bigl{(}\int\limits_{t_{i}}^{s}\gamma(\mu,t_{i},u)\mathop{}\\!\mathrm{d}\widetilde{N}(u)\Bigr{)}\mathop{}\\!\mathrm{d}s.$ (75) Let $\mathcal{G}_{k}:=\mathcal{F}_{t_{k+1}}$. Then it holds that $\\{\tilde{Z}_{k},\mathcal{G}_{k}\\}_{k=0,1,\ldots,n-1}$ is a discrete-time martingale, since $\widetilde{Z}_{k}$ is adapted to $\mathcal{G}_{k}$ for $k\in\\{0,\ldots,n-1\\}$ and from Fubini’s theorem for conditional expectations (see [1]) we have $\displaystyle{\mathbb{E}}\big{[}\widetilde{Z}_{k+1}-\widetilde{Z}_{k}\big{|}\mathcal{G}_{k}\big{]}$ $\displaystyle={\mathbb{E}}\Bigg{[}\int\limits_{t_{k+1}}^{t_{k+2}}\Bigg{(}\int\limits_{t_{k+1}}^{s}\gamma(\mu,t_{k+1},u)\mathop{}\\!\mathrm{d}\widetilde{N}(u)\Bigg{)}\mathop{}\\!\mathrm{d}s\Bigg{|}\mathcal{F}_{t_{k+1}}\Bigg{]}$ (76) $\displaystyle=\int\limits_{t_{k+1}}^{t_{k+2}}{\mathbb{E}}\Bigg{[}\int\limits_{t_{k+1}}^{s}\gamma(\mu,t_{k+1},u)\mathop{}\\!\mathrm{d}\widetilde{N}(u)\Bigg{|}\mathcal{F}_{t_{k+1}}\Bigg{]}\mathop{}\\!\mathrm{d}s=0.$ Hence, by the discrete version of the Burkholder-Davis-Gundy inequality and Jensen’s inequality we obtain that there exist constants $c_{33},c_{34}\in(0,\infty)$ such that ${\mathbb{E}}\Big{[}\big{|}\tilde{Z}_{k}\big{|}^{p}\Big{]}\leq c_{33}{\mathbb{E}}\Bigl{[}\Bigl{(}\sum\limits_{i=0}^{k}|\tilde{Y}_{i}|^{2}\Bigr{)}^{p/2}\Bigr{]}\leq c_{33}n^{\frac{p}{2}-1}\sum\limits_{i=0}^{n-1}{\mathbb{E}}\Big{[}\big{|}\tilde{Y}_{i}\big{|}^{p}\Big{]}\leq c_{34}\delta^{\frac{p}{2}+1},$ (77) for $k\in\\{0,1,\ldots,n-1\\}$. Combining (LABEL:mainM3p1), (LABEL:mainM3p2), (LABEL:mainM3p3), (LABEL:mainM3p4), (73), and (77) we get that there exists a constant $c_{35}\in(0,\infty)$ such that for all $t\in[0,T]$, ${\mathbb{E}}\Big{[}\big{|}\tilde{M}_{3}(t)\big{|}^{p}\Big{]}\leq c_{35}\delta^{p\min\\{\varrho_{1}+\frac{1}{p},\frac{1}{2}+\frac{1}{p},1\\}}.$ (78) By (LABEL:Meq13a), (68), and (78) we get that there exists a constant $c_{36}\in(0,\infty)$ such that ${\mathbb{E}}\Big{[}\big{|}\tilde{A}_{1}^{(\delta)}(t)\big{|}^{p}\Big{]}\leq c_{36}\delta^{p\min\\{\varrho_{1}+\frac{1}{p},\frac{1}{2}+\frac{1}{p},1\\}}.$ (79) Analog to the proof of [13, equation 33] we get that there exists a constant $c_{37}\in(0,\infty)$ such that for all $t\in[0,T]$ it holds that $\mathbb{E}\Big{[}\big{|}\tilde{A}^{(\delta)}_{2}(t)\big{|}^{p}\Big{]}\leq c_{37}\delta^{p(\varrho_{1}+\frac{1}{2})}.$ (80) Moreover, we can estimate $\mathbb{E}[|\tilde{A}_{3}^{(\delta)}(t)|^{p}]$ by using (LABEL:Meq10). Therefore, by (LABEL:Meq10), (79), and (80) we have that there exist constants $c_{38},c_{39}\in(0,\infty)$ such that for all $t\in[0,T]$ it holds that ${\mathbb{E}}\Big{[}\big{|}A(t)-A^{(\delta)}(t)\big{|}^{p}\Big{]}\leq c_{38}\delta^{p\min\\{\varrho_{1}+\frac{1}{p},\frac{1}{2}+\frac{1}{p},1\\}}+c_{39}\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\mathbb{E}\Big{[}\big{|}X(t_{i})-X^{(\delta)}(t_{i})\big{|}^{p}\Big{]}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s.$ (81) Using (LABEL:part_est_XXd), (66), (67), and (81) we have that there exist constants $c_{40},c_{41}\in(0,\infty)$ such that for all $t\in[0,T]$ it holds that $\displaystyle\mathbb{E}\Big{[}\big{|}X(t)-X^{(\delta)}_{c}(t)\big{|}^{p}\Big{]}\leq c_{40}\delta^{p\min\\{\frac{2}{p},\varrho_{1}+\frac{1}{p},\varrho_{2},\varrho_{3}\\}}+c_{41}\int\limits_{0}^{t}\sum\limits_{i=0}^{n-1}\mathbb{E}\Big{[}\big{|}X(t_{i})-X^{(\delta)}(t_{i})\big{|}^{p}\Big{]}\cdot\mathds{1}_{(t_{i},t_{i+1}]}(s)\mathop{}\\!\mathrm{d}s$ (82) $\displaystyle\leq c_{40}\delta^{p\min\\{\frac{2}{p},\varrho_{1}+\frac{1}{p},\varrho_{2},\varrho_{3}\\}}+c_{41}\int\limits_{0}^{t}\sup\limits_{0\leq u\leq s}\mathbb{E}\Big{[}\big{|}X(u)-X^{(\delta)}_{c}(u)\big{|}^{p}\Big{]}\mathop{}\\!\mathrm{d}s,$ and hence, $\sup\limits_{0\leq u\leq t}\mathbb{E}\Big{[}\big{|}X(u)-X^{(\delta)}_{c}(u)\big{|}^{p}\Big{]}\leq c_{40}\delta^{p\min\\{\frac{2}{p},\varrho_{1}+\frac{1}{p},\varrho_{2},\varrho_{3}\\}}+c_{41}\int\limits_{0}^{t}\sup\limits_{0\leq u\leq s}\mathbb{E}\Big{[}\big{|}X(u)-X^{(\delta)}_{c}(u)\big{|}^{p}\Big{]}\mathop{}\\!\mathrm{d}s.$ (83) The function $\displaystyle{[0,T]\ni t\mapsto\sup\limits_{0\leq u\leq t}{\mathbb{E}}\Big{[}\big{|}X(u)-X^{(\delta)}_{c}(u)\big{|}^{p}\Big{]}\in[0,\infty)}$ is Borel measurable, since it is non-decreasing and bounded due to (12) and (24). Therefore, by applying Grownall’s lemma to (83) we obtain that there exists a constant $C\in(0,\infty)$ such that for all $t\in[0,T]$, $\displaystyle\sup_{0\leq u\leq t}{\mathbb{E}}\Big{[}\big{|}X(u)-X^{(\delta)}_{c}(u)\big{|}^{p}\Big{]}\leq C\delta^{p\min\\{\frac{2}{p},\varrho_{1}+\frac{1}{p},\varrho_{2},\varrho_{3}\\}}.$ (84) ∎ ###### Remark 3.3. Note that for the classical Milstein scheme $\bar{X}^{(\delta)}$, which can be seen as a randomized Milstein process with $\xi_{i}$ fixed to $t_{i}$ for all $i\in\\{0,1,\ldots,n-1\\}$, we have that there exists a constant $K_{9}\in(0,\infty)$ such that for all $n\in{\mathbb{N}}$, $\displaystyle\sup_{0\leq t\leq T}\|X(t)-\bar{X}^{(\delta)}_{c}(t)\|_{L^{p}(\Omega)}\leq K_{9}\delta^{\min\\{\frac{2}{p},\varrho_{1},\varrho_{2},\varrho_{3}\\}}.$ (85) This follows from a straightforward modification of the proof of Theorem 3.2. ###### Remark 3.4. In the jump-free case ($\rho=0$) we get from the proof of Theorem 3.2 that there exists $K_{10}\in(0,\infty)$ such that for all $n\in{\mathbb{N}}$, $\displaystyle\sup_{0\leq t\leq T}\|X(t)-X^{(\delta)}_{c}(t)\|_{L^{p}(\Omega)}\leq K_{10}\delta^{\min\\{\varrho_{1}+\frac{1}{2},\varrho_{2}\\}}.$ (86) Hence our result implies the same upper error bound for the randomized Milstein process as in [13, Proposition 1] but under slightly weaker assumptions on $\mu$ and $\sigma$. Moreover, for $\varrho_{2}=\min\\{\frac{1}{2}+\varrho_{1},1\\}$ we recover the upper error bound from [8], which therein is obtained for a two-stage randomized Milstein scheme. ## 4 Lower bounds and optimality In this section we first consider the scalar case under the JCC and then we study the multidimensional case. In both cases we set $p=2$ and assume availability only of standard information, given by values of $W$ and $N$ at a finite number of points. We investigate lower bounds and optimality in the IBC framework, cf. [24]. ### 4.1 Scalar case and optimality of the randomized Milstein algorithm We investigate lower bound and optimality of the randomized Milstein algorithm in the case where $p=2$ and when the JCC is satisfied, that is $L_{-1}\sigma(t,y)=L_{1}\rho(t,y),\ (t,y)\in[0,T]\times\mathbb{R},$ (87) see [15]. It follows that under condition (87) the randomized Milstein algorithm uses only standard discrete information about $W,N$, i.e., the values $W(t_{1}),\ldots,W(t_{n}),N(t_{1}),\ldots,N(t_{n})$, since, by (16) and (17), in that case it has the form $\displaystyle X^{(\delta)}(t_{0})=X_{0},$ (88) $\displaystyle X^{(\delta)}(t_{i+1})=X^{(\delta)}(t_{i})+\mu(\xi_{i},X^{(\delta)}(t_{i}))\delta+\sigma(t_{i},X^{(\delta)}(t_{i}))\Delta W_{i}+\rho(t_{i},X^{(\delta)}(t_{i}))\Delta N_{i}$ $\displaystyle\quad\quad\quad\quad\quad\quad+L_{1}\sigma(t_{i},X^{(\delta)}(t_{i}))I_{t_{i},t_{i+1}}(W,W)+L_{-1}\rho(t_{i},X^{(\delta)}(t_{i}))I_{t_{i},t_{i+1}}(N,N)$ $\displaystyle\quad\quad\quad\quad\quad\quad+L_{-1}\sigma(t_{i},X^{(\delta)}(t_{i}))\Delta W_{i}\Delta N_{i},\quad\quad\hbox{for }i\in\\{0,\ldots,n-1\\}.$ Hence, if the JCC holds, the scheme is implementable. In order to provide worst-case error and optimality analysis we define the following function classes. For $K\in(0,\infty)$, and $\gamma\in(0,1]$, a function $f:[0,T]\times\mathbb{R}\to\mathbb{R}$ belongs to the function class $F_{K}^{\gamma}$ if and only if for all $t,s\in[0,T]$ and all $y,z\in\mathbb{R}$ it satisfies * (i) $f\in C^{0,1}\left([0,T]\times\mathbb{R}\right)$, * (ii) $|f(0,0)|\leq K$, * (iii) $|f(t,y)-f(t,z)|\leq K|y-z|$, * (iv) $|f(t,y)-f(s,y)|\leq K(1+|y|)|t-s|^{\gamma}$, * (v) $\left|\frac{\partial f}{\partial y}(t,y)-\frac{\partial f}{\partial y}(t,z)\right|\leq K|y-z|$. In this paper we consider drift coefficients $\mu$ from the class $\mathcal{M}^{\varrho_{1}}_{K}=\Biggl{\\{}\mu\in F^{\varrho_{1}}_{K}\colon\left|\frac{\partial\mu}{\partial y}(t,y)-\frac{\partial\mu}{\partial y}(s,y)\right|\leq K(1+|y|)|t-s|^{\varrho_{1}}\ \hbox{for all}\ t,s\in[0,T],y\in\mathbb{R}\Biggr{\\}},$ while we assume that the diffusion and jump coefficients $(\sigma,\rho)$ are from the class $\displaystyle\mathcal{B}^{\varrho_{2},\varrho_{3}}_{K}=\Bigl{\\{}(\sigma,\rho)\in F^{\varrho_{2}}_{K}\times F^{\varrho_{3}}_{K}\colon|L_{1}\sigma(t,y)-L_{1}\sigma(t,z)|\leq K|y-z|,$ (89) $\displaystyle|L_{1}\rho(t,y)-L_{1}\rho(t,z)|\leq K|y-z|,\ L_{-1}\sigma(t,y)=L_{1}\rho(t,y),\ \ \hbox{for all}\ t\in[0,T],y,z\in\mathbb{R}\Bigr{\\}},$ where we recall that $L_{1}f(t,y)=\sigma(t,y)f^{\prime}_{y}(t,y)$ and $L_{-1}f(t,y)=f(t,y+\rho(t,y))-f(t,y)$. Moreover, for all $p\in[2,\infty)$, $\mathcal{J}^{p}_{K}=\\{X_{0}\colon\Omega\to\mathbb{R}\colon X_{0}\ \hbox{is}\ \mathcal{F}_{0}-\hbox{measurable},\mathbb{E}[|X_{0}|^{2p}]\leq K\\}.$ The class of input data $(\mu,\sigma,\rho,X_{0})$ is defined by $\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},p,K)=\mathcal{M}^{\varrho_{1}}_{K}\times\mathcal{B}^{\varrho_{2},\varrho_{3}}_{K}\times\mathcal{J}^{p}_{K}.$ (90) We call $\varrho_{1},\varrho_{2},\varrho_{3},p,K,T$ the parameters of the class $\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},p,K)$. We now describe the model of computation. An information vector has the form $\displaystyle\mathcal{N}(\mu,\sigma,\rho,X_{0},W,N)=$ $\displaystyle[\mu(\xi_{0},y_{0}),\ldots,\mu(\xi_{k_{1}-1},y_{k_{1}-1}),\sigma(t_{0},y_{0}),\ldots,\sigma(t_{k_{1}-1},y_{k_{1}-1}),$ (91) $\displaystyle\ \ \rho(t_{0},y_{0}),\ldots,\rho(t_{k_{1}-1},y_{k_{1}-1}),\frac{\partial\sigma}{\partial y}(t_{0},y_{0}),\ldots,\frac{\partial\sigma}{\partial y}(t_{k_{1}-1},y_{k_{1}-1}),$ $\displaystyle\ \ \sigma(t_{0},z_{0}),\ldots,\sigma(t_{k_{1}-1},z_{k_{1}-1}),\rho(t_{0},v_{0}),\ldots,\rho(t_{k_{1}-1},v_{k_{1}-1}),$ $\displaystyle\ \ W(s_{0}),\ldots,W(s_{k_{2}-1}),N(q_{0}),\ldots,N(q_{k_{3}-1}),X_{0}],$ where $k_{1}$, $k_{2}$, $k_{3}\in{\mathbb{N}}$ and $[\xi_{0},\xi_{1},\ldots,\xi_{k_{1}-1}]$ is a random vector on $(\Omega,\mathcal{F},\mathbb{P})$ with values in $[0,T]^{k_{1}}$. We assume that the $\sigma$-fields $\sigma(\xi_{0},\xi_{1},\ldots,\xi_{k_{1}-1})$ and $\mathcal{F}_{\infty}$ are independent. Moreover, $t_{0},t_{1},\ldots,t_{k_{1}-1}\in[0,T]$, $s_{0},s_{1},\ldots,s_{k_{2}-1}\in[0,T]$, and $q_{0},q_{1},\ldots,q_{k_{3}-1}\in[0,T]$ are given time points. We assume that $s_{i}\neq s_{j}$, $q_{i}\neq q_{j}$ for all $i\neq j$. The evaluation points $y_{j},z_{j},v_{j}$ for the spatial variables of $\mu,\sigma$, $\partial\sigma/\partial y$, and $\rho$ can be given in an adaptive way with respect to $(\mu,\sigma,\rho,X_{0})$ and $[W,N]$. This means that for some measurable mappings $\psi_{j}$, $j\in\\{0,1,\ldots,k_{1}-1\\}$, it holds that $(y_{0},z_{0},v_{0})=\psi_{0}(W(s_{0}),\ldots,W(s_{k_{2}-1}),N(q_{0}),\ldots,N(q_{k_{3}-1}),X_{0})$ (92) and $\displaystyle(y_{j},z_{j},v_{j})=\psi_{j}(\mu(\xi_{0},y_{0}),\ldots,\mu(\xi_{j-1},y_{j-1}),\sigma(t_{0},y_{0}),\ldots,\sigma(t_{j-1},y_{j-1}),$ (93) $\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad\rho(t_{0},y_{0}),\ldots,\rho(t_{j-1},y_{j-1}),\frac{\partial\sigma}{\partial y}(t_{0},y_{0}),\ldots,\frac{\partial\sigma}{\partial y}(t_{j-1},y_{j-1}),$ $\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad\sigma(t_{0},z_{0}),\ldots,\sigma(t_{j-1},z_{j-1}),\rho(t_{0},v_{0}),\ldots,\rho(t_{j-1},v_{j-1}),$ $\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad W(s_{0}),\ldots,W(s_{k_{2}-1}),N(q_{0}),\ldots,N(q_{k_{3}-1}),X_{0}).$ The total number of evaluations of $\mu,\sigma,\rho$, $W$, and $N$ is $l=6k_{1}+k_{2}+k_{3}$. Any algorithm $\mathcal{A}$ that uses the information $\mathcal{N}(\mu,\sigma,\rho,X_{0},W,N)$ and computes the approximation to $X(T)$, is of the form $\mathcal{A}(\mu,\sigma,\rho,X_{0},W,N)=\varphi(\mathcal{N}(\mu,\sigma,\rho,X_{0},W,N)),$ (94) where $\varphi:\mathbb{R}^{3k_{1}+k_{2}+k_{3}+1}\to\mathbb{R}$ is a Borel measurable function. For a fixed $n\in\mathbb{N}$ we denote by $\Phi_{n}$ the class of all algorithms (94) with total number of evaluations $l\leq n$. For $(\mu,\sigma,\rho,X_{0})\in\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},p,K)$ we define the error of $\mathcal{A}\in\Phi_{n}$ as $e^{(p)}(\mathcal{A},\mu,\sigma,\rho,X_{0},W,N)=\|\mathcal{A}(\mu,\sigma,\rho,X_{0},W,N)-X(\mu,\sigma,\rho,X_{0})(T)\|_{p}.$ (95) The worst-case error of $\mathcal{A}$ in a subclass $\mathcal{G}$ of $\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},p,K)$ is defined by $e^{(p)}(\mathcal{A},\mathcal{G},W,N)=\sup\limits_{(\mu,\sigma,\rho,X_{0})\in\mathcal{G}}e^{(p)}(\mathcal{A},\mu,\sigma,\rho,X_{0},W,N),$ (96) while the $n$-th minimal error in $\mathcal{G}$ is $e^{(p)}_{n}(\mathcal{G},W,N)=\inf\limits_{\mathcal{A}\in\Phi_{n}}e^{(p)}(\mathcal{A},\mathcal{G},W,N).$ (97) The aim is to find sharp bounds for $e^{(p)}_{n}(\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},p,K),W,N)$, i.e. lower and upper error bounds which match up to constants. The randomized Milstein algorithm can be written as $\mathcal{A}^{RM}_{n}(\mu,\sigma,\rho,X_{0},W,N)=X^{(\delta)}(T),$ (98) where $X^{(\delta)}(T)$ is defined in (17), and we have that $\mathcal{A}^{RM}_{n}\in\Phi_{8n}$. ###### Theorem 4.1. It holds that $e^{(2)}_{n}(\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},2,K),W,N)=\Theta(n^{-\min\\{\varrho_{1}+\frac{1}{2},\varrho_{2},\varrho_{3}\\}})$ (99) as $n\to+\infty$. ###### Proof. The upper bound $O(n^{-\min\\{\varrho_{1}+\frac{1}{2},\varrho_{2},\varrho_{3}\\}})$ on $e^{(2)}_{n}(\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},2,K),W,N)$ follows from Theorem 3.2 and the fact that $e^{(2)}_{n}(\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},2,K),W,N)\leq e^{(2)}(\mathcal{A}^{RM}_{n},\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},2,K),W,N)$. We now turn to the lower bound. Let $\mathcal{A}$ be any algorithm from $\Phi_{n}$ that uses at most $n$ evaluations of $(\mu,\sigma,\rho)$, $W$, and $N$. We consider the following subclasses of $\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},2,K)$: $\mathcal{G}_{1}(\varrho_{1},1,1,2,K)=\mathcal{\bar{M}}_{K}^{\varrho_{1}}\times\\{(0,0)\\}\times\\{0\\},$ where $\mathcal{\bar{M}}_{K}^{\varrho_{1}}=\\{\mu\in\mathcal{M}_{K}^{\varrho_{1}}\ |\ \mu(t,x)=\mu(t,0)\text{ for all }t\in[0,T],x\in\mathbb{R}\\},$ and $\mathcal{G}_{2}(1,\varrho_{2},1,2,K)=\\{0\\}\times\mathcal{\bar{B}}^{\varrho_{2},1}_{K}\times\\{0\\},$ where $\mathcal{\bar{B}}^{\varrho_{2},1}_{K}=\Bigl{\\{}(\sigma,0)\in\mathcal{B}^{\varrho_{2},1}_{K}\,\Bigl{|}\,\sigma(t,y)=\sigma(t,0)\ \hbox{for all}\ t\in[0,T],y\in\mathbb{R}\Bigr{\\}},$ and $\mathcal{G}_{3}(1,1,\varrho_{3},2,K)=\\{0\\}\times\mathcal{\bar{B}}^{1,\varrho_{3}}_{K}\times\\{0\\},$ where $\mathcal{\bar{B}}^{1,\varrho_{3}}_{K}=\Bigl{\\{}(0,\rho)\in\mathcal{B}^{1,\varrho_{3}}_{K}\,\Bigl{|}\,\rho(t,y)=\rho(t,0)\ \hbox{for all}\ t\in[0,T],y\in\mathbb{R}\Bigr{\\}}.$ For $(\mu,\sigma,\rho,X_{0})\in\mathcal{G}_{1}(\varrho_{1},1,1,2,K)$ we have $\displaystyle{X(\mu,\sigma,\rho,X_{0})(T)=\int_{0}^{T}\mu(t,0)\mathop{}\\!\mathrm{d}t}$. Since $k_{1}=O(n)$ by [14, Section 2.2.9, Proposition 2] we obtain that $e(\mathcal{A},\mathcal{G}_{1}(\varrho_{1},1,1,2,K))=\Omega(n^{-(\varrho_{1}+\frac{1}{2})}).$ (100) Next, for $(\mu,\sigma,\rho,X_{0})\in\mathcal{G}_{2}(1,\varrho_{2},1,2,K)$ we have $\displaystyle{X(\mu,\sigma,\rho,X_{0})(T)=\int_{0}^{T}\sigma(t,0)\mathop{}\\!\mathrm{d}W(t)}$. Since $k_{2}=O(n)$, [11, Proposition 5.1(i)] gives $e(\mathcal{A},\mathcal{G}_{2}(1,\varrho_{2},1,2,K))=\Omega(n^{-\varrho_{2}}).$ (101) Finally, for $(\mu,\sigma,\rho,X_{0})\in\mathcal{G}_{3}(1,1,\varrho_{3},2,K)$ we have $\displaystyle{X(\mu,\sigma,\rho,X_{0})(T)=\int_{0}^{T}\rho(t,0)\mathop{}\\!\mathrm{d}N(t)}$. Since $k_{3}=O(n)$, [22, Lemma 6] yields $e(\mathcal{A},\mathcal{G}_{3}(1,1,\varrho_{3},2,K))=\Omega(n^{-\varrho_{3}}).$ (102) Due to the fact that $\mathcal{G}_{1}(\varrho_{1},1,1,2,K)\cup\mathcal{G}_{2}(1,\varrho_{2},1,2,K)\cup\mathcal{G}_{3}(1,1,\varrho_{3},2,K)\subset\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},2,K)$, it holds that $\displaystyle e(\mathcal{A},\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},2,K))$ (103) $\displaystyle\geq\max\\{e(\mathcal{A},\mathcal{G}_{1}(\varrho_{1},1,1,2,K)),e(\mathcal{A},\mathcal{G}_{2}(1,\varrho_{2},1,2,K)),e(\mathcal{A},\mathcal{G}_{3}(1,1,\varrho_{3},2,K))\\}$ $\displaystyle=\Omega(n^{-\min\\{\varrho_{1}+\frac{1}{2},\varrho_{2},\varrho_{3}\\}}).$ This, together with the upper bound proves the claim. ∎ ###### Remark 4.2. For $\varrho_{2}=\varrho_{3}=1$ and $\varrho_{1}\in(1/2,1)$ we compare the worst case errors for the classical Euler–Maruyama algorithm $\mathcal{A}^{E}_{n}$, randomized Euler–Maruyama algorithm $\mathcal{A}^{RE}_{n}$, classical Milstein algorithm $\mathcal{A}^{M}_{n}$, and randomized Milstein algorithm $\mathcal{A}^{RE}_{n}$ in the class $\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},2,K)$: $\displaystyle e^{(2)}(\mathcal{A}^{E}_{n},\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},2,K),W,N)=O(n^{-1/2}),\ e^{(2)}(\mathcal{A}^{RE}_{n},\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},2,K),W,N)=O(n^{-1/2}),$ (104) $\displaystyle e^{(2)}(\mathcal{A}^{M}_{n},\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},2,K),W,N)=O(n^{-\varrho_{1}}),\ e^{(2)}(\mathcal{A}^{RM}_{n},\mathcal{F}(\varrho_{1},\varrho_{2},\varrho_{3},2,K),W,N)=O(n^{-1}).$ We observe that in the considered case the randomized Milstein algorithm outperforms the other (classical) algorithms. ### 4.2 Multidimensional case and optimality of the Euler–Maruyama algorithm In this section we show results that give insight into lower error bounds for the problem of approximation of solutions of systems of jump-diffusion SDEs in the case when only standard information about $W$ and $N$ is available. For this we extend results from [2] and consider the following jump-diffusion Lévy’s area $\displaystyle J(N,W)=I_{0,T}(N,W)=\int\limits_{0}^{T}\int\limits_{0}^{t-}\mathop{}\\!\mathrm{d}N(s)\mathop{}\\!\mathrm{d}W(t)=\int\limits_{0}^{T}N(t-)\mathop{}\\!\mathrm{d}W(t)=\int\limits_{0}^{T}N(t)\mathop{}\\!\mathrm{d}W(t).$ (105) The last equality holds since $W$ is continuous while $N(t)$ and $N(t-)$ differ in an at most finite number of time points almost surely. Note that $J(N,W)=X(T)$ where $X$ is the solution of the two-dimensional SDE $\displaystyle dY(t)=\mathop{}\\!\mathrm{d}N(t),$ (106) $\displaystyle dX(t)=Y(t)\mathop{}\\!\mathrm{d}W(t),\ t\in[0,T].$ In order to approximate (LABEL:LA1) we consider an arbitrary algorithm of the form $\displaystyle\mathcal{A}_{n}(N,W)=\varphi_{n}(\mathcal{N}_{n}(N,W))$ (107) for some Borel-measurable function $\varphi_{n}\colon{\mathbb{R}}^{2n}\to{\mathbb{R}}$, where $\displaystyle\mathcal{N}_{n}(N,W)=[N(t_{1}),\ldots,N(t_{n}),W(t_{1}),\ldots,W(t_{n})]$ (108) and $0=t_{0}<t_{1}<\ldots<t_{n}=T$ (109) is a fixed discretization of $[0,T]$. In particular, we consider the trapezoidal method $\mathcal{A}_{n}^{T}(N,W)$ based on the mesh (109), which is defined as $\mathcal{A}_{n}^{T}(N,W)=\sum_{i=0}^{n-1}\frac{1}{2}\big{(}W(t_{i+1})-W(t_{i})\big{)}\big{(}N(t_{i+1})+N(t_{i})\big{)}.$ (110) ###### Theorem 4.3. For the trapezoidal method (110) based on the equidistant mesh $t_{i}=iT/n$, $i\in\\{0,1,\ldots,n\\}$, it holds that $\lim_{n\to\infty}n^{1/2}\cdot\|J(N,W)-\mathcal{A}_{n}^{T}(N,W)\|_{2}=\lim\limits_{n\to\infty}n^{1/2}\cdot\inf\limits_{\mathcal{A}_{n}}\|J(N,W)-\mathcal{A}_{n}(N,W)\|_{2}=\frac{\lambda^{1/2}T}{2},$ (111) and therefore it is the optimal method among all methods of the form (107). ###### Proof. From the projection property for the conditional expectation we get for any algorithm (107) that $\displaystyle{\mathbb{E}}\Big{[}\big{|}J(N,W)-\mathcal{A}_{n}(N,W)\big{|}^{2}\Big{]}\geq{\mathbb{E}}\Big{[}\big{|}J(N,W)-{\mathbb{E}}\big{[}J(N,W)\big{|}\mathcal{N}_{n}(N,W)\big{]}\big{|}^{2}\Big{]},$ (112) since $\mathcal{A}_{n}(N,W)$ is $\sigma(\mathcal{N}_{n}(N,W))$-measurable. Therefore, we also have $\inf_{\mathcal{A}_{n}}{\mathbb{E}}\Big{[}\big{|}J(N,W)-\mathcal{A}_{n}(N,W)\big{|}^{2}\Big{]}\geq\inf\limits_{0=t_{0}<t_{1}\ldots<t_{n}=T}{\mathbb{E}}\Big{[}\big{|}J(N,W)-{\mathbb{E}}\big{[}J(N,W)\big{|}\mathcal{N}_{n}(N,W)\big{]}\big{|}^{2}\Big{]}.$ (113) Hence, we need to compute $\displaystyle{\mathbb{E}}\big{[}J(N,W)\big{|}\mathcal{N}_{n}(N,W)\big{]}=\sum_{i=0}^{n-1}{\mathbb{E}}\Bigg{[}\int\limits_{t_{i}}^{t_{i+1}}N(t)\mathop{}\\!\mathrm{d}W(t)\Big{|}\mathcal{N}_{n}(N,W)\Bigg{]}.$ (114) For all $i\in\\{0,\ldots,n-1\\}$ we define $J_{t_{i},t_{i+1}}(N,W)=\int\limits_{t_{i}}^{t_{i+1}}N(t)\mathop{}\\!\mathrm{d}W(t).$ (115) From the definition of the Itô integral we get for all $i\in\\{0,\ldots,n-1\\}$ that $\displaystyle J_{t_{i},t_{i+1}}(N,W)=\lim_{m\to\infty}J_{m}^{i}(N,W)\text{ in }L^{2}({\mathbb{R}}),$ (116) where $\displaystyle J_{m}^{i}(N,W)=\sum_{j=0}^{m-1}N(s_{j}^{i})(W(s_{j+1}^{i})-W(s_{j}^{i})),$ (117) with $s_{j}^{i}=t_{i}+j(t_{i+1}-t_{i})/m$ for all $j\in\\{0,\ldots,m\\}$. Further, we denote $\Delta W_{j}^{i}=W(s_{j+1}^{i})-W(s_{j}^{i})$ for all $i\in\\{1,\ldots,n-1\\}$ and $j\in\\{1,\ldots,m-1\\}$. Then $\displaystyle{\mathbb{E}}\big{[}J_{m}^{i}(N,W)\big{|}\mathcal{N}_{n}(N,W)\big{]}=\sum_{j=0}^{m-1}{\mathbb{E}}\big{[}N(s_{j}^{i})\Delta W_{j}^{i}\big{|}\mathcal{N}_{n}(N,W)\big{]}.$ (118) Since by Proposition A.5 the processes $N$ and $W$ are conditionally independent given the $\sigma$-algebra $\sigma(\mathcal{N}_{n}(N,W))$, we have that $\displaystyle{\mathbb{E}}\big{[}N(s_{j}^{i})\Delta W_{j}^{i}\big{|}\mathcal{N}_{n}(N,W)\big{]}={\mathbb{E}}\big{[}N(s_{j}^{i})\big{|}\mathcal{N}_{n}(N)\big{]}\cdot{\mathbb{E}}\big{[}\Delta W_{j}^{i}\big{|}\mathcal{N}_{n}(W)\big{]}.$ (119) Moreover, by [6, Lemma 8] and [19, Lemma 3.1], we have for all $s\in[t_{i},t_{i+1}]$ $\displaystyle{\mathbb{E}}\big{[}N(s)\big{|}\mathcal{N}_{n}(N)\big{]}=\frac{N(t_{i+1})(s-t_{i})+N(t_{i})(t_{i+1}-s)}{t_{i+1}-t_{i}}$ (120) and ${\mathbb{E}}\big{[}\Delta W_{j}^{i}\big{|}\mathcal{N}_{n}(W)\big{]}=\frac{\big{(}W(t_{i+1})-W(t_{i})\big{)}(s_{j+1}^{i}-s_{j}^{i})}{t_{i+1}-t_{i}}.$ (121) Plugging (119),(LABEL:LA14), and (121) into (118), we obtain $\displaystyle{\mathbb{E}}\big{[}J_{m}^{i}(N,W)\big{|}\mathcal{N}_{n}(N,W)\big{]}=\sum_{j=0}^{m-1}{\mathbb{E}}\big{[}N(s_{j}^{i})\big{|}\mathcal{N}_{n}(N)\big{]}\cdot{\mathbb{E}}\big{[}\Delta W_{j}^{i}\big{|}\mathcal{N}_{n}(W)\big{]}$ (122) $\displaystyle=\frac{W(t_{i+1})-W(t_{i})}{t_{i+1}-t_{i}}\sum_{j=0}^{m-1}{\mathbb{E}}\big{[}N(s_{j}^{i})\big{|}\mathcal{N}_{n}(N)\big{]}\cdot(s_{j+1}^{i}-s_{j}^{i}).$ Since $\displaystyle{\sum_{j=0}^{m-1}{\mathbb{E}}\big{[}N(s_{j}^{i})\big{|}\mathcal{N}_{n}(N)\big{]}\cdot(s_{j+1}^{i}-s_{j}^{i})}$ is a (pathwise) Riemann sum for the stochastic process $\big{(}{\mathbb{E}}\big{[}N(s)\big{|}\mathcal{N}_{n}(N)\big{]}\big{)}_{s\in[t_{i},t_{i+1}]}$ with continuous sample paths, it holds for all $i\in\\{0,\ldots,m-1\\}$ almost surely that $\displaystyle\lim_{m\to\infty}{\mathbb{E}}\big{[}J_{m}^{i}(N,W)\big{|}\mathcal{N}_{n}(N,W)\big{]}=\frac{W(t_{i+1})-W(t_{i})}{t_{i+1}-t_{i}}\int\limits_{t_{i}}^{t_{i+1}}{\mathbb{E}}\big{[}N(t)\big{|}\mathcal{N}_{n}(N)\big{]}\mathop{}\\!\mathrm{d}t.$ (123) Moreover, from (117) and by Jensen’s inequality for the conditional expectation we obtain $\displaystyle{\mathbb{E}}\Big{[}\big{|}{\mathbb{E}}[J_{t_{i},t_{i+1}}(N,W)|\mathcal{N}_{n}(N,W)]-{\mathbb{E}}[J_{m}^{i}(N,W)|\mathcal{N}_{n}(N,W)]\big{|}^{2}\Big{]}$ (124) $\displaystyle\leq{\mathbb{E}}\Big{[}\big{|}J_{t_{i},t_{i+1}}(N,W)-J_{m}^{i}(N,W)\big{|}^{2}\Big{]}\to 0\text{ as }m\to\infty.$ Hence, by (LABEL:LA18) $\displaystyle{\mathbb{E}}\big{[}J_{m}^{i}(N,W)\big{|}\mathcal{N}_{n}(N,W)\big{]}\to{\mathbb{E}}\big{[}J_{t_{i},t_{i+1}}(N,W)\big{|}\mathcal{N}_{n}(N,W)\big{]}\text{ as }m\to\infty\text{ in }L^{2}(\Omega)$ (125) and by (123) $\displaystyle{\mathbb{E}}\big{[}J_{m}^{i}(N,W)\big{|}\mathcal{N}_{n}(N,W)\big{]}\to\frac{W(t_{i+1})-W(t_{i})}{t_{i+1}-t_{i}}\int\limits_{t_{i}}^{t_{i+1}}{\mathbb{E}}\big{[}N(t)\big{|}\mathcal{N}_{n}(N)\big{]}\mathop{}\\!\mathrm{d}t\text{ as }m\to\infty\text{ a.s.}$ (126) Convergence in $L^{2}(\Omega)$ as well as almost sure convergence imply convergence in probability. Moreover by the uniqueness of the limit in probability we get that for all $i\in\\{0,\ldots,m-1\\}$, $\displaystyle{\mathbb{E}}\big{[}J_{t_{i},t_{i+1}}(N,W)\big{|}\mathcal{N}_{n}(N,W)\big{]}=\frac{W(t_{i+1})-W(t_{i})}{t_{i+1}-t_{i}}\int\limits_{t_{i}}^{t_{i+1}}{\mathbb{E}}\big{[}N(t)\big{|}\mathcal{N}_{n}(N)\big{]}\mathop{}\\!\mathrm{d}t\text{ a.s.}$ (127) Moreover, we have $\int\limits_{t_{i}}^{t_{i+1}}{\mathbb{E}}\big{[}N(t)\big{|}\mathcal{N}_{n}(N)\big{]}\mathop{}\\!\mathrm{d}t=\frac{1}{2}\big{(}N(t_{i+1})+N(t_{i})\big{)}(t_{i+1}-t_{i}).$ (128) Hence by plugging (128) into (LABEL:LA21) we get for all $i\in\\{0,\ldots,m-1\\}$ that $\displaystyle{\mathbb{E}}\big{[}J_{t_{i},t_{i+1}}(N,W)\big{|}\mathcal{N}_{n}(N,W)\big{]}=\frac{1}{2}(W(t_{i+1})-W(t_{i}))(N(t_{i+1})+N(t_{i})).$ (129) Combining (129) and (114) we get that ${\mathbb{E}}[J(N,W)|\mathcal{N}_{n}(N,W)]=\mathcal{A}^{T}_{n}(N,W)$ corresponds to the trapezoidal method. Now we investigate its error in order to get the minimal possible error among all methods of the form (107). To do so, let us consider the step process given for all $t\in[0,T]$ by $\displaystyle\hat{N}_{n}(t):=\sum_{i=0}^{n-1}\mathds{1}_{(t_{i},t_{i+1}]}(t)\frac{N(t_{i})+N(t_{i+1})}{2}.$ (130) The process $(\hat{N}_{n}(t))_{t\in[0,T]}$ is not adapted to the initial filtration $(\mathcal{F}_{t})_{t\geq 0}$. However, it is adapted to $\mathcal{\widetilde{F}}_{t}:=\sigma(\mathcal{F}_{t}^{W}\cup\mathcal{F}_{\infty}^{N})$, $t\geq 0$. Moreover, $(W(t))_{t\geq 0}$ is still a one-dimensional Wiener process with respect to the filtration $(\mathcal{\widetilde{F}}_{t})_{t\geq 0}$, since $\mathcal{F}_{\infty}^{N}$ and $\mathcal{F}_{\infty}^{W}$ are independent. Hence $J(\hat{N}_{n},W)=\int\limits_{0}^{T}\hat{N}_{n}(t)\mathop{}\\!\mathrm{d}W(t)$ (131) is a well-defined Itô integral of the $(\mathcal{\widetilde{F}}_{t})_{t\geq 0}$-simple process $(\hat{N}_{n}(t))_{t\in[0,T]}$ and therefore $\displaystyle J(\hat{N}_{n},W)=\sum_{i=0}^{n-1}\frac{N(t_{i})+N(t_{i+1})}{2}(W(t_{i+1})-W(t_{i}))=\mathcal{A}_{n}^{T}(N,W).$ (132) Since $(N(t)-\hat{N}_{n}(t))_{t\in[0,T]}$ is a $(\mathcal{\widetilde{F}}_{t})_{t\geq 0}$-progressively measurable process, by the Itô isometry and Jensen’s inequality we get $\displaystyle{\mathbb{E}}\Big{[}\big{|}J(N,W)-\mathcal{A}_{n}^{T}(N,W)\big{|}^{2}\Big{]}=\sum_{i=0}^{n-1}\int\limits_{t_{i}}^{t_{i+1}}{\mathbb{E}}\Big{[}\big{|}N(t)-\hat{N}_{n}(t)\big{|}^{2}\Big{]}\mathop{}\\!\mathrm{d}t$ (133) $\displaystyle=\frac{1}{4}\sum_{i=0}^{n-1}\int\limits_{t_{i}}^{t_{i+1}}\Big{(}{\mathbb{E}}\Big{[}\big{(}N(t)-N(t_{i})\big{)}^{2}\Big{]}-2{\mathbb{E}}\Big{[}N(t)-N(t_{i})\Big{]}\cdot{\mathbb{E}}\Big{[}N(t_{i+1})-N(t)\Big{]}$ $\displaystyle\quad\quad\quad\quad+{\mathbb{E}}\Big{[}\big{(}N(t_{i+1})-N(t)\big{)}^{2}\Big{]}\Big{)}\mathop{}\\!\mathrm{d}t$ $\displaystyle=\frac{\lambda}{4}\sum_{i=0}^{n-1}(t_{i+1}-t_{i})^{2}+\frac{\lambda^{2}}{12}\sum_{i=0}^{n-1}(t_{i+1}-t_{i})^{3}\geq\frac{\lambda}{4n}\Big{(}\sum_{i=0}^{n-1}(t_{i+1}-t_{i})\Big{)}^{2}+\frac{1}{n^{2}}\,\frac{\lambda^{2}}{12}\Big{(}\sum_{i=0}^{n-1}(t_{i+1}-t_{i})\Big{)}^{3}$ $\displaystyle\geq\frac{\lambda T^{2}}{4n}+\frac{\lambda^{2}T^{3}}{12n^{2}}.$ Therefore, $\inf\limits_{0=t_{0}<t_{1}<\ldots t_{n}=T}{\mathbb{E}}\Big{[}\big{|}J(N,W)-\mathcal{A}_{n}^{T}(N,W)\big{|}^{2}\Big{]}\geq\frac{\lambda T^{2}}{4n}+\frac{\lambda^{2}T^{3}}{12n^{2}}.$ (134) Hence, by (113) we conclude $\displaystyle n^{1/2}\cdot\inf_{\mathcal{A}_{n}}\|J(N,W)-\mathcal{A}_{n}(N,W)\|_{L^{2}(\Omega)}\geq\sqrt{\frac{\lambda T^{2}}{4}+\frac{\lambda^{2}T^{3}}{12n}}.$ (135) Moreover, for the trapezoidal method $\mathcal{A}^{T}_{n}(N,W)$ based on the equidistant mesh $t_{i}=iT/n$, $i\in\\{0,1,\ldots,n\\}$, we get by (LABEL:LA29) that ${\mathbb{E}}\Big{[}\big{|}J(N,W)-\mathcal{A}_{n}^{T}(N,W)\big{|}^{2}\Big{]}=\frac{\lambda T^{2}}{4n}+\frac{\lambda^{2}T^{3}}{12n^{2}},$ (136) and hence $n^{1/2}\cdot\inf\limits_{\mathcal{A}_{n}}\|J(N,W)-\mathcal{A}_{n}(N,W)\|_{2}\leq n^{1/2}\cdot\|J(N,W)-\mathcal{A}^{T}_{n}(N,W)\|_{2}=\sqrt{\frac{\lambda T^{2}}{4}+\frac{\lambda^{2}T^{3}}{12n}}.$ (137) From (135) and (137) we arrive at $\displaystyle n^{1/2}\cdot\inf\limits_{\mathcal{A}_{n}}\|J(N,W)-\mathcal{A}_{n}(N,W)\|_{2}=\sqrt{\frac{\lambda T^{2}}{4}+\frac{\lambda^{2}T^{3}}{12n}},$ (138) which together with (136) imply the thesis. ∎ ###### Remark 4.4. Consider any class of coefficients of multidimensional SDEs for which (LABEL:2dim_sde) is a subproblem. Then by Theorem 4.3, in the worst case setting with respect to the coefficients, the error cannot be smaller than $\Omega(n^{-1/2})$. Therefore, no matter if the JCC (87) is satisfied or not, we can apply the Euler–Maruyama (or randomized Euler–Maruyama) scheme in order to achieve the optimal error bound $O(n^{-1/2})$ if the error is measured in the $L^{2}(\Omega)$-norm, see, for example, [11], [12]. ###### Remark 4.5. In Theorems 4.1 and 4.3 we have considered only the $L^{2}$-error. Matching upper and lower bounds that depend on $p$ remain an open problem. Our numerical experiments in Section 5 suggest that for jump-diffusion SDEs the error indeed depends on $p$. ## 5 Numerical experiments We implement111The program code is available as ancillary file from the arXiv page of this paper. the randomized Milstein algorithm for the SDE $\displaystyle\left\\{\begin{array}[]{lll}\mathop{}\\!\mathrm{d}X(t)=\sin(M\cdot X(t)(1+t)^{\varrho_{1}})\mathop{}\\!\mathrm{d}t+\cos(M\cdot X(t)\cdot(1+t)^{\varrho_{2}})\mathop{}\\!\mathrm{d}W(t)&\\\ \qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad+\Big{(}-X(t)+\frac{\pi}{2M\cdot(1+t)^{\varrho_{2}}}\Big{)}\mathop{}\\!\mathrm{d}N(t),&t\in[0,1],\\\ X(0)=1.&\\\ \end{array}\right.$ (139) that has already been considered in the jump-free case in [13]. The jump coefficient is chosen so that the JCC is satisfied. The verification of Assumption 2.1 is straight forward, for the JCC (87), see Remark 5.2. In our simulations we set $\lambda=100$, $M=100$, $\varrho_{1}=0.1$, and $\varrho_{2}=0.6$. We estimate the $L^{p}$-error similar as in [20, p. 14] by $\displaystyle\operatorname{error}(k)=\operatorname{mean}\big{(}\big{|}X^{(k)}(T)-X^{(k-1)}(T)\big{|}^{p}\big{)}^{\frac{1}{p}}.$ (140) Here $X^{(k)}(T)$ is the approximation of $X(T)$ with step size $\delta^{(k)}$, where $\delta^{(k)}=2^{-k}$ for $k\in{\mathbb{N}}$. The $\operatorname{mean}$ is taken over $2^{16}$ sample paths. ###### Remark 5.1. In the implementation the interesting part is how the values $\xi_{i}$ are computed. For the randomization we first simulate independent uniformly distributed random variables $\xi_{i}$ on the corresponding intervals for the finest discretization grid. Iteratively we compute the values for the discretization grid with doubled step size as follows: One time interval in the larger grid consists of two time intervals of equal length in the finer grid. For those two intervals we have simulated two values $\xi_{i}$. Now we simulate an independent Bernoulli$(0.5)$ random variable that determines which of the values $\xi_{i}$ we take. This choice is then uniformly distributed on the interval of the large grid and hence consistent with the randomized Milstein algorithm. For $p\in[2,\infty)$ we obtain by Theorem 3.2 the theoretical convergence rate $\displaystyle{\min\Big{\\{}\frac{2}{p},\varrho_{1}+\frac{1}{p},\varrho_{2},\varrho_{2}\Big{\\}}=\min\Big{\\{}\frac{2}{p},0.1+\frac{1}{p},0.6\Big{\\}}}.$ For $p=1$ we take as theoretical convergence rate the same rate as for $p=2$, because the $L^{1}$-error can be estimated by the $L^{2}$-error using the Cauchy-Schwarz inequality. In Figure 1 we plot the $\log_{2}(\operatorname{error}(k))$ over $\log_{2}(\delta^{(k)}))$ for $p\in\\{1,2,3,4\\}$ and the corresponding theoretical convergence orders. Figure 1: Error estimates and theoretical convergence order for $p\in\\{1,2,3,4\\}$ We see that the observed convergence order is decreasing with increasing $p$. Further we notice that for $p=1$ the convergence of the simulation is higher than the theoretical convergence rate. This is reasonable because we took the rate of the $L^{2}$-error. For $p=2$ we observe that the simulation confirms the theoretical results; the slope of the simulation matches the convergence rate, which we proved to be optimal. Also for $p=3$ and $p=4$ the simulations confirm the theoretical results, since the simulation converges at least as fast as the theoretically obtained upper bound; we have not proven any lower bound. Next, we regress the slope of the simulated $\log_{2}(\operatorname{error}(k))$ in dependence of the corresponding $\log_{2}(\delta^{(k)}))$ for all $p\in\\{1,\ldots,8\\}$ and compare it to the theoretical upper bounds on the convergence rates we have proven, see Figure 2. We observe that for the simulations the convergence order is dependent on $p$, which confirms also this theoretical finding. Figure 2: Slopes of the simulation (estimated by linear regression) in comparison to theoretical convergence rates ###### Remark 5.2. Let us assume that the diffusion coefficient is of the form $\sigma(t,y)=F(\alpha(t)y+\beta(t))$ while the jump coefficient $\rho(t,y)=-y+\gamma(t)$ for some functions $F:\mathbb{R}\to\mathbb{R}$ and $\alpha,\beta,\gamma:[0,T]\to\mathbb{R}$. Moreover, let us assume that there exists $x_{0}\in\mathbb{R}$ such that * • $F(x_{0})=0$, * • $\alpha(t)\cdot\gamma(t)+\beta(t)=x_{0}$ for all $t\in[0,T]$. Then the JCC (87) is satisfied for the pair $(\sigma,\rho)$. This provides a new class of functions $(\sigma,\rho)$ satisfying the JCC which may, in contrast to the class considered in [15], be nonlinear. ## Appendix A Appendix The proof of the following lemma is straightforward and will be omitted. ###### Lemma A.1. Under Assumption 2.1 there exists a constant $K_{7}\in(0,\infty)$ such that for $f\in\\{\mu,\sigma,\rho\\}$ and for all $t_{1},t_{2},t,u\in[0,T]$, $\displaystyle|\alpha_{1}(f,t,u)|$ $\displaystyle\leq K_{7}(1+|X(u)|),$ (141) $\displaystyle|\beta(f,t,u)|$ $\displaystyle\leq K_{7}(1+|X(u)|),$ $\displaystyle|\beta(\mu,t_{1},u)-\beta(\mu,t_{2},u)|$ $\displaystyle\leq K_{7}(1+|X(u)|^{2})\cdot|t_{1}-t_{2}|^{\varrho_{1}},$ $\displaystyle|\gamma(f,t,u)|$ $\displaystyle\leq K_{7}(1+|X(u-)|),$ $\displaystyle|\gamma(\mu,t_{1},u)-\gamma(\mu,t_{2},u)|$ $\displaystyle\leq K_{7}(1+|X(u-)|)|t_{1}-t_{2}|^{\varrho_{1}}.$ Following [3] and [7] we recall the notion of conditional independence and some of its useful consequences. ###### Definition A.2 ([3, p. 36-II, Definition 43]). Let $(\Omega,\mathcal{F},{\mathbb{P}})$ be a probability space and let $\mathcal{F}_{1}$, $\mathcal{F}_{2}$, and $\mathcal{F}_{3}$ be three sub-$\sigma$-fields of $\mathcal{F}$. $\mathcal{F}_{1}$ and $\mathcal{F}_{3}$ are called conditionally independent given $\mathcal{F}_{2}$, if for all positive random variables $Y_{1}$ and $Y_{3}$, which are measurable with respect to $\mathcal{F}_{1}$ respectively $\mathcal{F}_{3}$ it holds that $\displaystyle{\mathbb{E}}[Y_{1}Y_{3}|\mathcal{F}_{2}]={\mathbb{E}}[Y_{1}|\mathcal{F}_{2}]\cdot{\mathbb{E}}[Y_{3}|\mathcal{F}_{2}]\text{ a.s.}$ (142) ###### Theorem A.3 ([3, p. 36-II, Theorem 45]). Let $(\Omega,\mathcal{F},{\mathbb{P}})$ be a probability space and let $\mathcal{F}_{1}$, $\mathcal{F}_{2}$, and $\mathcal{F}_{3}$ be three sub-$\sigma$-fields of $\mathcal{F}$. Further let $\mathcal{F}_{12}$ be the $\sigma$-field generated by $\mathcal{F}_{1}$ and $\mathcal{F}_{2}$. Then $\mathcal{F}_{1}$ and $\mathcal{F}_{3}$ are called conditionally independent given $\mathcal{F}_{2}$, if and only if for all $\mathcal{F}_{3}$-measurable and integrable random variables $Y_{3}$ it holds that $\displaystyle{\mathbb{E}}[Y_{3}|\mathcal{F}_{12}]={\mathbb{E}}[Y_{3}|\mathcal{F}_{2}]\text{ a.s.}$ (143) ###### Proposition A.4 ([7, Proposition A.23]). Let $(\Omega,\mathcal{F},{\mathbb{P}})$ be a probability space and let $\mathcal{F}_{1}$, $\mathcal{F}_{2}$, and $\mathcal{F}_{3}$ be three sub-$\sigma$-fields of $\mathcal{F}$. Further let $Y_{1},Y_{3}\colon\Omega\to{\mathbb{R}}$ be integrable random variables such that ${\mathbb{E}}[|Y_{1}Y_{3}|]<\infty$. Assume that $\sigma(Y_{1})\subset\mathcal{F}_{1}$ and $\sigma(Y_{3})\subset\mathcal{F}_{3}$. Further assume that $\mathcal{F}_{1}$ and $\mathcal{F}_{3}$ are conditionally independent given $\mathcal{F}_{2}$. Then it holds that $\displaystyle{\mathbb{E}}[Y_{1}Y_{3}|\mathcal{F}_{2}]={\mathbb{E}}[Y_{1}|\mathcal{F}_{2}]\cdot{\mathbb{E}}[Y_{3}|\mathcal{F}_{2}]\text{ a.s.}$ (144) ###### Proposition A.5 ([7, Lemma B.18]). Let $(\Omega,\mathcal{F},{\mathbb{P}})$ be a probability space and let $X,Y\colon\Omega\times[0,\infty)\to{\mathbb{R}}$ be stochastic processes, which are both $\mathcal{F}\otimes\mathcal{B}([0,\infty))|\mathcal{B}({\mathbb{R}})$-measurable and independent, i.e. $\mathcal{F}^{X}_{\infty}\mathchoice{\mathrel{\hbox to0.0pt{$\displaystyle\perp$\hss}\mkern 2.0mu{\displaystyle\perp}}}{\mathrel{\hbox to0.0pt{$\textstyle\perp$\hss}\mkern 2.0mu{\textstyle\perp}}}{\mathrel{\hbox to0.0pt{$\scriptstyle\perp$\hss}\mkern 2.0mu{\scriptstyle\perp}}}{\mathrel{\hbox to0.0pt{$\scriptscriptstyle\perp$\hss}\mkern 2.0mu{\scriptscriptstyle\perp}}}\mathcal{F}^{Y}_{\infty}$, where $\mathcal{F}^{Z}_{\infty}=\sigma(\bigcup_{t\geq 0}\sigma(Z(t)))$ for both $Z\in\\{X,Y\\}$. Further, assume that ${\mathbb{E}}[|X(t)|]<\infty\text{ and }{\mathbb{E}}[|Y(t)|]<\infty\text{ for all }t\geq 0$. Additionally, let $m,n\in{\mathbb{N}}$, $t_{i}^{X},t_{j}^{Y}\in[0,\infty)$ for all $i\in\\{1,\ldots,n\\}$ and $j\in\\{1,\ldots,m\\}$ be such that $0\leq t_{1}^{X}<t_{2}^{X}<\ldots<t_{n}^{X}$, $0\leq t_{1}^{Y}<t_{2}^{Y}<\ldots<t_{m}^{Y}$. Then $\mathcal{F}_{\infty}^{X}$ and $\mathcal{F}_{\infty}^{Y}$ are conditionally independent given $\sigma(X(t_{1}^{X}),\ldots,X(t_{n}^{X}),Y(t_{1}^{Y}),\ldots,Y(t_{m}^{Y}))$. The following estimate is a direct consequence of the Hölder, the Burkholder- Davis-Gundy, and the Kunita inequalitiy, see [9]. ###### Lemma A.6. Let $q\in[2,\infty)$, $a,b\in[0,T]$ with $a<b$, $Z\in\\{\operatorname{Id},W,N\\}$, $Y=(Y(t))_{t\in[a,b]}$ is a predictable stochastic process such that ${\mathbb{E}}\Big{[}\int\limits_{a}^{b}|Y(t)|^{q}\mathop{}\\!\mathrm{d}t\Big{]}<\infty$ (145) Then there exists a constant $\hat{c}\in(0,\infty)$ such that for all $t\in[a,b]$ it holds that $\displaystyle{\mathbb{E}}\Biggl{[}\sup_{s\in[a,t]}\Biggl{|}\int\limits_{a}^{s}Y(u)\mathop{}\\!\mathrm{d}Z(u)\Biggl{|}^{q}\Biggr{]}\leq\hat{c}\int\limits_{a}^{t}{\mathbb{E}}[|Y(u)|^{q}]\mathop{}\\!\mathrm{d}u.$ (146) ## Acknowledgements V. Schwarz and M. Szölgyenyi are supported by the Austrian Science Fund (FWF): DOC 78. ## References * Brooks [1972] R. A. Brooks. Conditional expectations associated with stochastic processes. _Pacific Journal of Mathematics_ , 41:33–42, 1972. * Clark and Cameron [1980] J. M. C. Clark and R. J. Cameron. The maximum rate of convergence of discrete approximations for stochastic differential equations, in: Stochastic differential systems filtering and control. _Stochastic Differential Systems Filtering and Control_ , pages 162–171, 1980. * Dellacherie and Meyer [1978] C. Dellacherie and P. Meyer. _Probabilities and Potential_. North-Holland mathematic studies. Elsevier North-Holland, inc., 1978. * Eisenmann and Kruse [2018] M. Eisenmann and R. Kruse. Two quadrature rules for stochastic Itô-integrals with fractional Sobolev regularity. _Communications in Mathematical Sciences_ , 16:2125–2146, 2018. * Heinrich [2019] S. Heinrich. Complexity of stochastic integration in Sobolev classes. _Journal of Mathematical Analysis and Applications_ , 476:177–195, 2019. * Hertling [2001] P. Hertling. Nonlinear Lebesgue and Itô integration problems of high complexity. _Journal of Complexity_ , 17:366–387, 2001. * Kałuża [2020] A. Kałuża. _Optimal algorithms for solving stochastic initial-value problems with jumps_. PhD thesis, AGH University of Science and Technology, Kraków. https://winntbg.bg.agh.edu.pl/rozprawy2/11743/full11743.pdf, 2020. * Kruse and Wu [2019] R. Kruse and Y. Wu. A randomized Milstein method for stochastic differential equations with non-differentiable drift coefficients. _Discrete and Continuous Dynamical Systems - Series B_ , 24:3475–3502, 2019. * Kunita [2004] H. Kunita. Stochastic differential equations based on Lévy processes and stochastic flows of diffeomorphisms. In _Real and stochastic analysis_ , pages 305–373. Springer, 2004\. * Müller-Gronbach and Yaroslavtseva [2020] T. Müller-Gronbach and L. Yaroslavtseva. Sharp lower error bounds for strong approximation of sdes with discontinuous drift coefficient by coupling if noise. 2020\. arXiv:2010.00915. * Morkisz and Przybyłowicz [2014] P. Morkisz and P. Przybyłowicz. Strong approximation of solutions of stochastic differential equations with time-irregular coefficients via randomized Euler algorithm. _Applied Numerical Mathematics_ , 78:80–94, 2014. * Morkisz and Przybyłowicz [2017] P. M. Morkisz and P. Przybyłowicz. Optimal pointwise approximation of SDE’s from inexact information. _Journal of Computational and Applied Mathematics_ , 324:85–100, 2017. * Morkisz and Przybyłowicz [2021] P. M. Morkisz and P. Przybyłowicz. Randomized derivative-free Milstein algorithm for efficient approximation of solutions of SDEs under noisy information. _Journal of Computational and Applied Mathematics_ , 383, 2021. 113112\. * Novak [1988] E. Novak. _Deterministic and Stochastic Error Bounds in Numerical Analysis_. Lecture Notes in Mathematics, vol. 1349. Springer, 1988. * Platen and Bruti-Liberati [2010] E. Platen and N. Bruti-Liberati. _Numerical Solution of Stochastic Differential Equations with Jumps in Finance_. Springer Verlag, Berlin, Heidelberg, 2010. * Protter [2005] P. Protter. _Stochastic Integration and Differential Equations_. Stochastic Modelling and Applied Probability. Springer, Berlin-Heidelberg, 2005. * Przybyłowicz [2015a] P. Przybyłowicz. Minimal asymptotic error for one-point approximation of sdes with time-irregular coefficients. _Journal of Computational and Applied Mathematics_ , 282:98–110, 2015a. * Przybyłowicz [2015b] P. Przybyłowicz. Optimal global approximation of SDEs with time-irregular coefficients in asymptotic setting. _Applied Mathematics and Computation_ , 270:441–457, 2015b. * Przybyłowicz [2016] P. Przybyłowicz. Optimal global approximation of stochastic differential equations with additive poisson noise. _Numerical Algorithms_ , 73:323–348, 2016. * Przybyłowicz and Szölgyenyi [2021] P. Przybyłowicz and M. Szölgyenyi. Existence, uniqueness, and approximation of solutions of jump-diffusion SDEs with discontinuous drift. _Applied Mathematics and Computation_ , 403:126191, 2021\. * Przybyłowicz et al. [2021] P. Przybyłowicz, M. Szölgyenyi, and F. Xu. Existence and uniqueness of solutions of SDEs with discontinuous drift and finite activity jumps. _Statistics and Probability Letters_ , 174:109072, 2021\. * Przybyłowicz et al. [2022] P. Przybyłowicz, M. Sobieraj, and Ł. Stȩpień. Efficient approximtion of SDEs driven by countably dimensional Wiener process and Poisson random measure. _SIAM Journal of Numerical Analysis_ , 60:824–855, 2022\. * Situ [2005] R. Situ. _Theory of Stochastic Differential Equations with Jumps and Applications_. Mathematical and Analytical Techniques with Applications to Engineering. Springer, 2005. * Traub et al. [1988] J. F. Traub, G. W. Wasilkowski, and H. Woźniakowski. _Information-Based Complexity_. Academic Press, New York, 1988. Paweł Przybyłowicz Faculty of Applied Mathematics, AGH University of Science and Technology, Al. Mickiewicza 30, 30-059 Krakow, Poland <EMAIL_ADDRESS> Michaela Szölgyenyi Department of Statistics, University of Klagenfurt, Universitätsstraße 65-67, 9020 Klagenfurt, Austria <EMAIL_ADDRESS> Verena Schwarz 🖂 Department of Statistics, University of Klagenfurt, Universitätsstraße 65-67, 9020 Klagenfurt, Austria <EMAIL_ADDRESS>
# The interior Backus problem: local resolution in Hölder spaces Toru Kan Department of Mathematics, Osaka Metropolitan University, 1-1 Gakuen-cho, Naka-ku, Sakai, 599-8531, Japan<EMAIL_ADDRESS>, Rolando Magnanini Dipartimento di Matematica ed Informatica “U. Dini”, Università di Firenze, viale Morgagni 67/A, 50134 Firenze, Italy<EMAIL_ADDRESS>http://web.math.unifi.it/users/magnanin and Michiaki Onodera Department of Mathematics, Tokyo Institute of Technology, 2-12-1 Ookayama, Meguro-ku, Tokyo 152-8551, Japan<EMAIL_ADDRESS> ###### Abstract. We prove an existence result for the interior Backus problem in the Euclidean ball. The problem consists in determining a harmonic function in the ball from the knowledge of the modulus of its gradient on the boundary. The problem is severely nonlinear. From a physical point of view, the problem can be interpreted as the determination of the velocity potential of an incompressible and irrotational fluid inside the ball from measurements of the velocity field’s modulus on the boundary. The linearized problem is an irregular oblique derivative problem, for which a phenomenon of loss of derivatives occurs. As a consequence, a solution by linearization of the Backus problem becomes problematic. Here, we linearize the problem around the vertical height solution and show that the loss of derivatives does not occur for solutions which are either (vertically) axially symmetric or oddly symmetric in the vertical direction. A standard fixed point argument is then feasible, based on ad hoc weighted estimates in Hölder spaces. ###### Key words and phrases: Backus problem, fully nonlinear boundary conditions, irregular oblique derivative problem ###### 2010 Mathematics Subject Classification: 35J65, 35C15, 35B07 ## 1\. Introduction Let $\Omega$ be a bounded domain in the Euclidean space $\mathbb{R}^{N}$, $N\geq 2$, with boundary $\Gamma$. Let $g$ be a positive continuous function on $\Gamma$. The interior Backus problem consists in determining a function $u\in C^{1}(\overline{\Omega})\cap C^{2}(\Omega)$ such that (1.1) $\Delta u=0\ \mbox{ in }\ \Omega,\quad|\nabla u|=g\ \mbox{ on }\ \Gamma.$ This problem was first considered and completely solved in [2] (see also [12]), for $N=2$. There, Backus was motivated by a problem in geophysics, which entails the reconstruction of the gravitational or geomagnetic terrestrial field from measurements of its intensity on the Earth’s surface ${\mathcal{S}}$. In fact, if one models the Earth as the unit ball $B$, then the relevant geophysical problem amounts to determine solutions $u$ of (1.1), with $\Omega=\mathbb{R}^{N}\setminus\overline{B}$ and $\Gamma={\mathcal{S}}$, such that (1.2) $u(x)\to 0\ \mbox{ as }\ |x|\to\infty.$ This is what we call the exterior Backus problem. When $N=2$, by the Riemann mapping theorem, we know that any simply connected (seen as a domain on the Riemann sphere) proper subdomain of the complex plane $\mathbb{C}$ is conformally equivalent to the unit disk. Moreover, the harmonicity of functions is preserved by conformal mappings, while the modulus of their gradients changes just by a positive factor (which depends on the derivative of the conformal map). So, that is one reason why problem (1.1) has some interest. Another physical motivation, which genuinely pertains the interior problem setting, has to do with the study of incompressible and irrotational fluid flows. Let $\vec{V}$ be the velocity field of a fluid and let $\rho$ be its density. Any fluid flow obeys the continuity equation: $\mathop{\mathrm{div}}(\rho\,\vec{V})+\rho_{t}=0.$ If a fluid is incompressible, its density is constant. In particular, we have that $\rho_{t}=0$, and hence the continuity equation reduces to $\mathop{\mathrm{div}}(\vec{V})=0$. If the fluid is irrotational, then $\mathop{\mathrm{curl}}\,(\vec{V})=0$, and hence there exists a harmonic velocity potential $u$ such that $\nabla u=\vec{V}$. Thus, solving (1.1) can be interpreted as the determination of the velocity of the fluid inside the domain $\Omega$ from measurements of its modulus on the boundary. It is worthwhile to clarify from this point of view the results obtained in [2] for planar domains. As already mentioned, in this case we can always assume that $\Omega$ is the unit disk. We shall describe the situation in the simplest case in which $g$ is assumed to be constant, say $g\equiv 1$. As shown in [2], or by simply invoking Weierstrass factorization theorem (see [16]), in the complex variable $z=x+i\,y$, when $g\equiv 1$, the complex gradient $u_{x}-i\,u_{y}$ of $u$ is uniquely determined by the Blaschke product $u_{x}-i\,u_{y}=e^{i\,\alpha}\prod_{j=1}^{n}\frac{z-z_{j}}{1-\overline{z}_{j}\,z},$ where $\alpha\in\mathbb{R}$ and $z_{1},\dots,z_{n}\in\Omega$ are given parameters. In fact, each factor in the product has unitary modulus on the unit circle. Thus, up to a rotation of an angle $\alpha$, the velocity field of the fluid can be uniquely determined from its modulus on the boundary if we know the position and the nature (the multiplicity, so as to speak) of its stagnation points $z_{1},\dots,z_{n}$. In order to conclude our motivations, it may be of interest to mention [6], in which it can be found a possible application to encephalography by magnetic means. When $N\geq 3$, we can still use the Kelvin transformation to map the exterior of the ball to its interior (or an exterior domain to a bounded one) by preserving the harmonicity of functions. However, the boundary condition in (1.1) changes quite considerably. In fact, if ${\mathcal{K}}w(y)=|y|^{2-N}\,w\left(\frac{y}{|y|^{2}}\right),\ \ y\neq 0,$ is the standard Kelvin transformation of a function $w:\mathbb{R}^{N}\setminus\\{0\\}\to\mathbb{R}$, we have that $\nabla{\mathcal{K}}w(y)=|x|^{N}\left\\{\nabla w(x)-\Bigl{[}2\,x\cdot\nabla w(x)+(N-2)\,w(x)\Bigr{]}\frac{x}{|x|^{2}}\right\\},$ with $y=x/|x|^{2}$. In particular, if we apply the transformation to the exterior of unit ball $B$ centered at the origin, we obtain that the condition $|\nabla u|=g$ on ${\mathcal{S}}$ changes into $|\nabla U+(N-2)\,U\,\nu|=g\ \mbox{ on }\ {\mathcal{S}},$ with $U={\mathcal{K}}u$. Here, $\nu$ denotes the exterior unit normal vector field on ${\mathcal{S}}$. Another feature for which the interior and exterior problems differ from one another is that the latter admits solutions $u$ whose normal derivative $u_{\nu}$ does not change sign on ${\mathcal{S}}$ (e.g. the fundamental solution or, more in general, the capacity potential of a bounded domain), while in the former the divergence theorem tells us that the mean value on ${\mathcal{S}}$ of $u_{\nu}$ must be equal to zero. The positivity of the normal derivative of the solution has been useful in [8] to obtain the local resolution of the exterior gravitational Backus problem for the Earth near the so-called monopole $\Phi(x)=1/|x|$, based on a fixed-point argument. In other words, when $N=3$ and $\Omega=\mathbb{R}^{3}\setminus\overline{B}$, the existence and uniqueness of a solution $u$ of (1.1)-(1.2) is obtained as the perturbation $u=\Phi+v,$ where $v$ is harmonic in $\Omega$ and decays to zero at infinity. This result holds if the data $g$ is sufficiently close to $1$ in a Hölder norm. (Notice that $1$ is the modulus of the gradient of the monopole on ${\mathcal{S}}$.) We also mention that the positivity of $u_{\nu}$ is also used in [4] to construct a comparison principle for suitably defined viscosity solutions for the exterior Backus problem and hence develop a nonlinear approach to the problem. When the positivity property is not available, the only existence result up to date is given in [10]. There, we consider, in physical dimesion $N=3$, the local resolution of the exterior Backus problem (1.1)-(1.2) near the so-called dipole: $d(x)=\frac{x_{3}}{|x|^{3}}.$ The gradient $\nabla d$ models the terrestrial geomagnetic field. The problem of finding solutions of (1.1)-(1.2) of the type $u=d+v,$ with $v$ harmonic in $\Omega$, which decays to zero at infinity, has another level of difficulty, though. We can see this if we linearize (1.1)-(1.2) near $d$. Indeed, we obtain the boundary value problem: (1.3) $\Delta v=0\ \mbox{ in }\ \Omega,\quad\nabla d\cdot\nabla v=\varphi\ \mbox{ on }\ {\mathcal{S}},\quad v(x)\to 0\ \mbox{ as }\ |x|\to\infty.$ This can be classified as an irregular oblique derivative problem. In fact, in contrast with the monopole case, in which the vector field $\nabla\Phi$ governing the linearized problem is nothing else than the normal field on ${\mathcal{S}}$, in (1.3), instead, the field $\nabla d$ points outward to the Earth’s surface on the southern hemisphere, becomes tangential on the equator $\mathcal{E}=\\{x\in{\mathcal{S}}:x_{N}=0\\}$, and points inward on the northern hemisphere. For this reason, (1.3) suffers of two drawbacks. The first one is a severe lack of uniqueness, since its solutions can be uniquely determined only up to prescribing Dirichlet boundary values on $\mathcal{E}$. The second one is a loss of regularity: the expected solution $v$ does not gain the desired regularity. In other words, the regularity of $v$ does not improve that of the data $\varphi$ by one order — it can be seen that it falls short of $1/2$. This inconvenience makes the perturbation approach more complicated, because the iterative scheme on which a fixed-point argument is based upon loses derivatives at each step. In presence of a loss of derivatives, the Nash-Moser implicit function theorem has worked in other contexts (see the pioneering works [15, 13, 14, 7], for instance). Unfortunately, in attacking the Backus problem, this plan is so far out of reach. In fact, one would need sufficiently precise estimates for the relevant oblique derivative problems. Namely, it is necessary to have an accurate control not only for the solution of (1.3), but also for those of a class of oblique derivative problems obtained by perturbing $\nabla d$. Nevertheless, in [10] we showed that (1.1)-(1.2) is solvable near $d$ for solutions which are axially symmetric around the Earth’s axis $\mathcal{A}=\\{\lambda\,(0,0,1):\lambda\in\mathbb{R}\\}$. In fact, we show that the solutions of the linearized problem (1.3) with this symmetry no longer lose derivatives in an appropriate scale of fractional Sobolev spaces on ${\mathcal{S}}$. This result is made possible by the use of spherical harmonics on ${\mathcal{S}}$. As a consequence, the relevant fixed-point scheme can be mended and the existence of a solution of (1.1)-(1.2) is obtained if $g$ is sufficiently close to $|\nabla d|$ in some fractional Sobolev norm. In the present paper, for $N\geq 3$, we turn our attention to the local resolution of the interior Backus problem (1.1) in the framework of Hölder spaces. This framework is that used in [8] for the gravitational case. The simplest instance in this case is to consider solutions of (1.1) in the form: $u(x)=x_{N}+v(x),$ where $v$ is harmonic in $\Omega$. If we place the $x_{N}$-axis horizontally, in the fluidmechanical framework mentioned above, we want to determine the velocity of the fluid inside a domain from measurements of its modulus on the boundary as a perturbation of that of an horizontal laminar flow with potential $f(x)=x_{N}.$ In this case, the associated linearized problem is simply: (1.4) $\Delta v=0\ \mbox{ in }\ \Omega,\quad\partial_{x_{N}}v=\varphi\ \mbox{ on }\ {\mathcal{S}},$ where $\varphi$ is a given function. It is clear that the vector field governing (1.4) is $e_{N}=(0,\dots,0,1)$, that shows similar qualitative features to those of $\nabla d$, in the sense that $-e_{N}$ points inward on the northern hemisphere, is tangential on $\mathcal{E}$, and points outward on the southern hemisphere. Also in this case, though, a loss of derivatives occurs for problem (1.4). Nevertheless, we shall see that the somewhat easier oblique derivative condition in (1.4) allows a treatment of (1.1) in the framework of Hölder spaces, in the cases where solutions are oddly symmetric with respect to the hyperplane $x_{N}=0$ or axially symmetric around $x_{N}$-axis. In order to describe the main result of this paper, we need to introduce some notation. For $k=0,1,2,\dots$ and $\alpha\in(0,1)$, we define the function spaces: $\displaystyle C_{\rm even}^{k+\alpha}(\overline{B})=\\{\varphi\in C^{k+\alpha}(\overline{B}):\varphi(x^{\prime},x_{N})=\varphi(x^{\prime},-x_{N})\\},$ $\displaystyle C_{\rm odd}^{k+\alpha}(\overline{B})=\\{\varphi\in C^{k+\alpha}(\overline{B}):\varphi(x^{\prime},x_{N})=-\varphi(x^{\prime},-x_{N})\\},$ $\displaystyle C_{\rm ax}^{k+\alpha}(\overline{B})=\\{\varphi\in C^{k+\alpha}(\overline{B}):\varphi(x^{\prime},x_{N})=\varphi(|x^{\prime}|e_{1}^{\prime},x_{N})\\}.$ Here $e_{1}^{\prime}=(1,0,\ldots,0)\in\mathbb{R}^{N-1}$. We note that these are closed subspaces of the space $C^{k+\alpha}(\overline{B})$ of $k$-differentiable functions whose derivatives up to the order $k$ are $\alpha$-Hölder continuous. The usual Hölder seminorm and norm on $C^{k+\alpha}(\overline{B})$ are denoted by $[\cdot]_{\alpha,\overline{B}}$ and $|\cdot|_{k+\alpha,\overline{B}}$, respectively, and defined as: $\displaystyle[u]_{\alpha,\overline{B}}=\sup_{x,y\in\overline{B},\,x\neq y}\frac{|u(x)-u(y)|}{|x-y|^{\alpha}},$ $\displaystyle|u|_{k+\alpha,\overline{B}}=\sum_{|\gamma|\leq k}|D^{\gamma}u|_{0,\overline{B}}+\sum_{|\gamma|=k}[D^{\gamma}u]_{\alpha,\overline{B}}.$ Here, $|\cdot|_{0,\overline{B}}$ stands for the standard (uniform) maximum norm. With these premises, we can state the main result of this paper as follows. ###### Theorem 1.1. Let $\alpha\in(0,1)$ and set $\Omega=B$. Then there exist positive constants $\delta_{0}$ and $C$ with the following properties. * (i) If $g\in C^{1+\alpha}_{\rm even}(\overline{B})$ and $|g-1|_{1+\alpha,\overline{B}}\leq\delta_{0},$ then problem (1.1) has a solution $u\in C^{2}(B)\cap C^{1+\alpha}_{\rm odd}(\overline{B})$ satisfying $|u-f|_{1+\alpha,\overline{B}}\leq C\,|g-1|_{1+\alpha,\overline{B}}.$ * (ii) If $g\in C^{1+\alpha}_{\rm ax}(\overline{B})$, $h\in\mathbb{R}$ and $|g-1|_{1+\alpha,\overline{B}}\leq\delta_{0},$ then problem (1.1) has a solution $u\in C^{2}(B)\cap C^{1+\alpha}_{\rm ax}(\overline{B})$ satisfying $u=h\ \mbox{ on }\ \mathcal{E},$ and $|u-h-f|_{1+\alpha,\overline{B}}\leq C\,\left(|g-1|_{1+\alpha,\overline{B}}\right).$ Note that the solution $u$ obtained in (i) of Theorem 1.1 automatically satisfies the condition $u=0$ on $\mathcal{E}$, since it is odd in the variable $x_{N}$; while in (ii) we need to impose the additional boundary condition on $\mathcal{E}$, due to the fact that (1.1) is invariant under the addition of constants. It is worthwhile noticing at this point that, even if the height $f$ and the dipole $d$ are the Kelvin trasformations of one another, ${\mathcal{K}}$ maps the linearized exterior problem (1.3) into the interior oblique derivative problem: $\Delta v=0\ \mbox{ in }\ B,\quad\partial_{x_{N}}v+(N-2)\,f\,\partial_{\nu}v+(N-1)(N-2)\,f\,v=\varphi\ \mbox{ on }\ {\mathcal{S}}.$ This problem is more difficult to treat, even if the relevant vector field governing it has the same qualitative properties of $e_{N}$. A study of this problem will be the theme of future work. The proof of Theorem 1.1 hinges on some a priori estimates for the solution of the linearized problem (1.4) subject to the Dirichlet-type condition (1.5) $v=\psi\ \mbox{ on }\ \mathcal{E}.$ Thus, in Section 2, we will first derive an explicit representation formula for the solutions of (1.4)-(1.5). We will also adapt a couple of lemmas in [1] to our purposes, by making explicit the dependence of the relevant norms of $v$ on those of the data $\varphi$. Then, in Section 3, we shall derive crucial a priori estimates for the linearized problem (1.4)-(1.5) (see Theorem 3.1). Finally, in Section 4, based on these estimates, we will carry out the proof of Theorem 1.1. ## 2\. An explicit integral representation formula for the oblique derivative problem In this section, we carry out explicit computations which lead to the construction of an integral representation formula for the linearized problem (1.4)-(1.5). We follow the scheme introduced in [1]. ### 2.1. Uniqueness for problem (1.4)-(1.5) Let a function space $C^{1}_{N}(\overline{B})$ be defined by $C^{1}_{N}(\overline{B})=\\{v\in C(\overline{B}):\partial_{x_{N}}v\mbox{ exists and }\partial_{x_{N}}v\in C(\overline{B})\\}.$ Then $C^{1}_{N}(\overline{B})\cap C^{2}(B)$ is one of the natural spaces for solutions of (1.4)-(1.5). We show that a uniqueness result holds in this space. The argument follows the lines of one used in [9] and is based on Hopf’s boundary lemma. ###### Proposition 2.1 (Uniqueness). For any given $\varphi\in C({\mathcal{S}})$ and $\psi\in C(\mathcal{E})$, the problem (1.4)-(1.5) has at most one solution of class $C^{1}_{N}(\overline{B})\cap C^{2}(B)$. ###### Proof. Let $v_{1}$ and $v_{2}$ be two solutions of class $C^{1}_{N}(\overline{B})\cap C^{2}(B)$ of (1.4)-(1.5), and set $v=v_{1}-v_{2}$. Then $v$ solves (1.4)-(1.5) with $\varphi\equiv 0$ and $\psi\equiv 0$. If $v(x)$ is a (positive) maximum for $v$ on $\overline{B}$, then we have that $x\in{\mathcal{S}}$, by the maximum principle, and $x\notin\mathcal{E}$, being as $v(x)>0$. Now, if $x$ were in the upper hemisphere of ${\mathcal{S}}$, then we would have that $\partial_{x_{N}}v(x)>0$, by Hopf’s boundary lemma, since $e_{N}$ is an outward direction on that hemisphere. This is a contradiction. By a similar argument, we infer that $x$ cannot belong to the lower hemisphere of ${\mathcal{S}}$. Thus, we conclude that $v\leq 0$ on $\overline{B}$. By considering the minimum of $v$, we can infer that $v\geq 0$ on $\overline{B}$, and hence $v\equiv 0$ on $\overline{B}$. ∎ We can easily use this proposition to infer that the solution $v$ of (1.4)-(1.5) inherits possible symmetries of the data $\varphi$ and $\psi$. For instance, if $\varphi$ and $\psi$ are axially symmetric around $x_{N}$-axis, then $v$ is so. ### 2.2. Estimates for the Dirichlet problem for the Laplace equation In the next result, we derive estimates for solutions of the Dirichlet problem in $B$: (2.1) $\Delta w=0\ \mbox{ in }\ B,\quad w=\varphi\ \mbox{ on }\ {\mathcal{S}}.$ It is well-known that if $\varphi\in C({\mathcal{S}})$, this problem has the unique solution $w\in C(\overline{B})\cap C^{2}(B)$ given by the Poisson integral formula: (2.2) $w(x)=\int_{\mathcal{S}}P_{B}(x;y)\,\varphi(y)\,dS_{y}.$ Here, $P_{\Omega}$ stands for the Poisson kernel for a bounded domain $\Omega$. In particular, $P_{B}$ is explicitly given by $P_{B}(x;y)=\frac{1}{N\omega_{N}}\frac{1-|x|^{2}}{|x-y|^{N}}\ \mbox{ for }\ x\in B,\ y\in{\mathcal{S}},$ where $\omega_{N}$ is the volume of $B$. For our aims, we need the following refinement of [1, Lemma 2.2]. ###### Proposition 2.2. Suppose that $\varphi\in C^{k+\alpha}(\overline{B})$ for some non-negative integer $k$ and $\alpha\in[0,1)$. Let $\beta$ be a multi-index with $|\beta|>k+\alpha$. If $w$ is the solution of (2.1), then there exists a positive constant $C=C(N,\beta,k,\alpha)$ such that (2.3) $|D^{\beta}w(x)|\leq C\,|\varphi|_{k+\alpha,\overline{B}}\,(1-|x|^{2})^{-|\beta|+k+\alpha}$ for all $x\in B$. To prove Proposition 2.2, we first recall the following a priori estimate for the Laplace equation. For the proofs, see for instance [5, Theorem 6.6, Problem 6.2]. ###### Lemma 2.3. Suppose that $\varphi$ is of class $C^{k+\alpha}(\overline{B})$ for some integer $k\geq 2$ and $\alpha\in(0,1)$. Then the solution $w$ of (2.1) satisfies the inequality $|w|_{k+\alpha,\overline{B}}\leq C|\varphi|_{k+\alpha,\overline{B}}$ for some positive constant $C=C(N,k,\alpha)$. Next, we derive a pointwise estimate for the Poisson kernel. ###### Lemma 2.4. For any multi-index $\beta$, it holds that $D^{\beta}_{x}P_{B}(x;y)=\frac{a_{\beta}(x,y)}{|x-y|^{|\beta|+N-1}}\ \mbox{ for }\ x\in B,\ y\in{\mathcal{S}},$ where $|a_{\beta}(x,y)|\leq C_{*}$ for some positive constant $C_{*}$ which only depends on $N$ and $\beta$. ###### Proof. We have that $P_{B}(x;y)=\frac{1}{N\omega_{N}}\frac{1-|x|^{2}}{|x-y|^{N}}=\frac{1}{N\omega_{N}}\frac{|y|^{2}-|x|^{2}}{|x-y|^{N}}=\frac{1}{N\omega_{N}}\frac{(x-y)\cdot(x+y)}{|x-y|^{N}}.$ Hence, if we set $z=x-y$, we obtain that $P_{B}(z+y;y)=\frac{1}{N\omega_{N}}\frac{1}{|z|^{N-2}}+\frac{2}{N\omega_{N}}\,\frac{y\cdot z}{|z|^{N}}.$ This function of $z$ is the sum of a $(2-N)$-homogeneous and a $(1-N)$-homogeneous function. Thus, we infer that $D^{\beta}_{z}P_{B}(z+y;y)=A(z;y)+B(z;y),$ where $A(z;y)$ and $B(z;y)$ are homogeneous of degree $2-N-|\beta|$ and $1-N-|\beta|$ in $z$. As a consequence, we get: $A(z;y)+B(z;y)=|z|^{1-N-|\beta|}\bigl{[}|z|\,A(z/|z|;y)+B(z/|z|;y)\bigr{]}.$ The function in the brackets is bounded since both $z/|z|$ and $y$ have a unitary norm and $z\in 2B$. Therefore, we conclude by setting $a_{\beta}(x,y)=|x-y|\,A\left(\frac{x-y}{|x-y|};y\right)+B\left(\frac{x-y}{|x-y|};y\right),$ for $x\in B$ and $y\in{\mathcal{S}}$. ∎ We also need the following bound. ###### Lemma 2.5. Let a multi-index $\beta$ and a nonnegative number $\kappa$ satisfy $|\beta|>\kappa$. Then, there exists a positive constant $C=C(N,\beta,\kappa)$ such that $\int_{{\mathcal{S}}}|D^{\beta}_{x}P_{B}(x;y)|\,|y-x_{0}|^{\kappa}\,dS_{y}\leq C(1-|x|)^{-|\beta|+\kappa},$ for all $x\in B\setminus\\{0\\}$, where $x_{0}=x/|x|\in{\mathcal{S}}$. When $x=0$, we can choose $x_{0}$ to be any point in ${\mathcal{S}}$. ###### Proof. When $x=0$, the bound easily follows from Lemma 2.4. Let $x\in B\setminus\\{0\\}$, $x_{0}=x/|x|$, and set $r=1-|x|$. Then, for every $y\in{\mathcal{S}}$, we have that $|y-x|=\frac{2}{3}|y-x|+\frac{1}{3}|(y-x_{0})-(x-x_{0})|\geq\\\ \frac{2}{3}(|y|-|x|)+\frac{1}{3}(|y-x_{0}|-|x-x_{0}|)=\frac{1}{3}r+\frac{1}{3}|y-x_{0}|.$ This with Lemma 2.4 shows that $\int_{{\mathcal{S}}}|D^{\beta}_{x}P_{B}(x;y)|\,|y-x_{0}|^{\kappa}\,dS_{y}=\int_{{\mathcal{S}}}\frac{|a_{\beta}(x,y)||y-x_{0}|^{\kappa}}{|y-x|^{|\beta|+N-1}}\,dS_{y}\leq\\\ 3^{1-N-|\beta|}C_{*}\int_{\mathcal{S}}\frac{|y-x_{0}|^{\kappa}}{(r+|y-x_{0}|)^{|\beta|+N-1}}\,dS_{y}=\\\ 3^{1-N-|\beta|}C_{*}r^{-|\beta|+\kappa}\int_{|rz+e_{N}|=1}\frac{|z|^{\kappa}}{(1+|z|)^{|\beta|+N-1}}\,dS_{z},$ where we have used the change of variables $y=x_{0}+r{\mathcal{R}}z$ with the orthogonal matrix ${\mathcal{R}}$ satisfying ${\mathcal{R}}^{-1}x_{0}=e_{N}=(0,\ldots,0,1)$. The last integral on the right-hand side of the above inequality is bounded with respect to $r$, because as $r\to 0$ it converges to the integral $\int_{\mathbb{R}^{N-1}}\frac{|z|^{\kappa}}{(1+|z|)^{|\beta|+N-1}}\,dS_{z},$ which is finite, being as $|\beta|>\kappa$. We thus obtain the desired inequality. ∎ ###### Proof of Proposition 2.2. As usual, in this proof $C$ will denote a generic constant possibly depending on $N,\beta,k$, and $\alpha$. Pick any point $x_{0}\in{\mathcal{S}}$. Since $\varphi\in C^{k+\alpha}(\overline{B})$, we can write the following standard Taylor expansion for $\varphi$: $\varphi(y)=\sum_{j=0}^{k}\sum_{|\gamma|=j}\frac{D^{\gamma}\varphi(x_{0})}{\gamma!}\,(y-x_{0})^{\gamma}+\sum_{|\gamma|=k}\frac{D^{\gamma}\varphi(x_{0}+\theta\,(y-x_{0}))-D^{\gamma}\varphi(x_{0})}{\gamma!}\,(y-x_{0})^{\gamma}.$ Here, we use the standard conventions on the multi-index notation. Thus, integrating $\varphi(y)$ for $y\in{\mathcal{S}}$ against $D^{\beta}_{x}P_{B}(x;y)$ gives that (2.4) $D^{\beta}w(x)=\sum_{|\gamma|\leq k}\frac{D^{\gamma}\varphi(x_{0})}{\gamma!}\,D^{\beta}h_{\gamma}(x)+R_{k}(x),$ where $h_{\gamma}$ is the solution of (2.1) with $\varphi=(\cdot- x_{0})^{\gamma}$ and $R_{k}(x)=\sum_{|\gamma|=k}\int_{\mathcal{S}}D^{\beta}_{x}P_{B}(x;y)\,\frac{D^{\gamma}\varphi(x_{0}+\theta\,(y-x_{0}))-D^{\gamma}\varphi(x_{0})}{\gamma!}\,(y-x_{0})^{\gamma}\,dS_{y}.$ Now, let $x_{*}$ be any point in $B\setminus\\{0\\}$ such that $x_{*}=|x_{*}|x_{0}$. Then, Lemma 2.3 gives that (2.5) $|D^{\beta}h_{\gamma}(x_{*})|\leq|D^{\beta}h_{\gamma}|_{0,\overline{B}}\leq C.$ Moreover, if $\alpha\in(0,1)$, Lemma 2.5 shows that (2.6) $|R_{k}(x_{*})|\leq\sum_{|\gamma|=k}\frac{[D^{\gamma}\varphi]_{\alpha,\overline{B}}}{\gamma!}\int_{\mathcal{S}}|D^{\beta}_{x}P_{B}(x_{*};y)|\,|y-x_{0}|^{k+\alpha}\,dS_{y}\leq\\\ C\sum_{|\gamma|=k}[D^{\gamma}\varphi]_{\alpha,\overline{B}}\,(1-|x_{*}|)^{-|\beta|+k+\alpha}.$ This inequality is also valid for $\alpha=0$, if $[D^{\gamma}\varphi]_{\alpha,\overline{B}}$ is replaced by $|D^{\gamma}\varphi|_{0,\overline{B}}$. Plugging (2.5) and (2.6) into (2.4), then gives that $|D^{\beta}w(x_{*})|\leq C\,|\varphi|_{k+\alpha,\overline{B}}\,(1-|x_{*}|)^{-|\beta|+k+\alpha}.$ which yields (2.3), after an update of the constant $C$. ∎ ### 2.3. Representation formulas for problem (1.4)-(1.5) In order to obtain a representation formula, we consider the Dirichlet problem (2.7) $-\Delta_{x^{\prime}}Z(x^{\prime})=\partial_{x_{N}}w(x^{\prime},0)\ \mbox{ in }\ D,\quad Z=\psi\ \mbox{ on }\ \partial D,$ where $w$ is the solution of (2.1) and $D=\\{x^{\prime}\in\mathbb{R}^{N-1}:|x^{\prime}|<1\\}$. We identify $D$ and $\partial D$ with the equatorial ball $\\{x=(x^{\prime},x_{N})\in B:x_{N}=0\\}$ and the equator $\mathcal{E}$, respectively. From Lemma 2.5, we see that $w$ satisfies $|\partial_{x_{N}}w(x^{\prime},0)|\leq|\varphi|_{0,{\mathcal{S}}}\int_{{\mathcal{S}}}|\partial_{x_{N}}P_{B}(x^{\prime},0;y)|\,dS_{y}\leq C\,|\varphi|_{0,{\mathcal{S}}}\,(1-|x^{\prime}|)^{-1}$ for some constant $C$. Therefore, for any $\varphi\in C({\mathcal{S}})$ and $\psi\in C(\mathcal{E})$, the existence and uniqueness of solutions of (2.7) in $C(\overline{D})\cap C^{2}(D)$ are guaranteed by [5, Theorem 4.9]. ###### Proposition 2.6 (Existence and representation formula). Suppose that $\varphi\in C({\mathcal{S}})$ and $\psi\in C(\mathcal{E})$. Let $w$ and $Z$ be the solutions of (2.1) and (2.7), respectively, and set (2.8) $W(x)=\int_{0}^{x_{N}}w(x^{\prime},t)\,dt,\quad x=(x^{\prime},x_{N})\in\overline{B},$ where $w$ is defined in (2.2). Then the unique solution $v$ of class $C^{1}_{N}(\overline{B})\cap C^{2}(B)$ of (1.4)-(1.5) is given by (2.9) $v(x)=W(x)+Z(x^{\prime}),\quad x=(x^{\prime},x_{N})\in\overline{B}.$ ###### Proof. Let $v$ be defined by (2.9). Then $v=W+Z\in C(\overline{B})\cap C^{2}(B)$, since we know that $w\in C(\overline{B})\cap C^{2}(B)$ and $Z\in C(\overline{D})\cap C^{2}(D)$. Moreover, we have that $\partial_{x_{N}}v=w\in C(\overline{B})$, and therefore $v\in C^{1}_{N}(\overline{B})\cap C^{2}(B)$. Since $W(x^{\prime},0)=0$, we see that $\partial_{x_{N}}v=w=\varphi\ \mbox{ on }\ {\mathcal{S}},\quad v=Z=\psi\ \mbox{ on }\ \mathcal{E}.$ Hence the assertion follows if we show that $v$ is harmonic in $B$. By a direct calculation, we have that $\Delta W(x)=\int_{0}^{x_{N}}\Delta_{x^{\prime}}w(x^{\prime},t)\,dt+\partial_{x_{N}}w(x^{\prime},x_{N})=\\\ -\int_{0}^{x_{N}}\partial_{x_{N}x_{N}}^{2}w(x^{\prime},t)\,dt+\partial_{x_{N}}w(x^{\prime},x_{N})=\partial_{x_{N}}w(x^{\prime},0).$ We thus infer that $\Delta v(x)=\Delta W(x)+\Delta_{x^{\prime}}Z(x^{\prime})=\partial_{x_{N}}w(x^{\prime},0)-\partial_{x_{N}}w(x^{\prime},0)=0,$ as desired. ∎ Even if it is not needed in the proof of Theorem 1.1, we also derive for future reference an explicit integral representation formula. The formula may be helpful for numerical approximations. To derive the formula, we recall that the fundamental solution $\Gamma_{d}$ of the Laplace equation in a $d$-dimensional Euclidean space ($d\geq 2$) is given by $\Gamma_{2}(x)=\frac{1}{2\pi}\,\log\frac{1}{|x|},\qquad\Gamma_{d}(x)=\frac{1}{d(d-2)\,\omega_{d}}|x|^{2-d}\ \mbox{ if }\ d\geq 3.$ Then, the Green’s function for $D$ is written as $G_{D}(x^{\prime};y^{\prime})=\Gamma_{N-1}(x^{\prime}-y^{\prime})-\Gamma_{N-1}\left(|x^{\prime}|({\mathcal{I}}(x^{\prime})-y^{\prime})\right),$ where ${\mathcal{I}}$ denotes the inversion ${\mathcal{I}}(x^{\prime})=x^{\prime}/|x^{\prime}|^{2}$ for $x^{\prime}\neq 0$. If we now define the kernel $K(x;y)=\int_{0}^{x_{N}}P_{B}(x^{\prime},t;y)\,dt+\int_{D}G_{D}(x^{\prime},z^{\prime})\,\partial_{x_{N}}P_{B}(z^{\prime},0;y)\,dz^{\prime},$ for $x=(x^{\prime},x_{N})\in B$ and $y\in{\mathcal{S}}$, the representation formula is given as follows. ###### Proposition 2.7 (Integral representation formula). Suppose that $\varphi\in C({\mathcal{S}})$ and $\psi\in C(\mathcal{E})$. Then, the function defined by (2.10) $v(x)=\int_{\mathcal{S}}K(x;y)\,\varphi(y)\,dS_{y}+\int_{\mathcal{E}}P_{D}(x^{\prime};y^{\prime})\,\psi(y^{\prime})\,dS_{y^{\prime}}\ \mbox{ for }\ x\in B$ coincides with the unique solution of class $C^{1}_{N}(\overline{B})\cap C^{2}(B)$ of the problem (1.4)-(1.5). ###### Proof. By the well-known representation formula for the Dirichlet problem for the Poisson equation, we know that $Z(x^{\prime})=\int_{D}G_{D}(x^{\prime};z^{\prime})\,\partial_{x_{N}}w(z^{\prime},0)\,dz^{\prime}+\int_{\mathcal{E}}P_{D}(x^{\prime};z^{\prime})\,\psi(z^{\prime})\,dS_{z^{\prime}}.$ With the definition of $w$ in mind, by the Fubini theorem we then infer that $\int_{D}G_{D}(x^{\prime};z^{\prime})\,\partial_{x_{N}}w(z^{\prime},0)\,dz^{\prime}=\int_{S}\left[\int_{D}G_{D}(x^{\prime};z^{\prime})\,\partial_{x_{N}}P_{B}(z^{\prime},0;y)\,dz^{\prime}\right]\varphi(y)\,dS_{y}.$ Being as $x^{\prime}\in D$, the Fubini theorem is applicable in this formula, because the function $F$ defined a.e. on $D\times{\mathcal{S}}$ by $F(z^{\prime},y)=\varphi(y)\,\partial_{x_{N}}P_{B}(z^{\prime},0;y)\,G_{D}(x^{\prime};z^{\prime})$ is in $L^{1}(D\times{\mathcal{S}})$. In fact, we have that $\int_{D\times{\mathcal{S}}}|F(z^{\prime},y)|(dz^{\prime}\times dS_{y})=\int_{D}G_{D}(x^{\prime};z^{\prime})\left[\int_{\mathcal{S}}|\partial_{x_{N}}P_{B}(z^{\prime},0;y)||\varphi(y)|\,dS_{y}\right]dz^{\prime}$ (see [11, Theorem 1.12]). The right-hand side is finite thanks to the properties of $G_{D}$ and Lemma 2.5 with $\kappa=0$ and $|\beta|=1$. Finally, that $W(x)=\int_{\mathcal{S}}\left[\int_{0}^{x_{N}}P_{B}(x^{\prime},t;y)\,dt\right]\varphi(y)\,dS_{y}$ follows from (2.8), again by a straightforward application of the Fubini theorem. We have thus proved that the right-hand side of (2.10) coincides with $W+Z$. ∎ ## 3\. A priori estimates for the linearized problem A crucial a priori bound we will use to prove Theorem 1.1 is contained in the following theorem. ###### Theorem 3.1. Let $\alpha\in(0,1)$ and suppose that $\varphi\in C^{1+\alpha}(\overline{B})$ and $\psi\in C^{3/2+\alpha}(\overline{D})$. Then a solution $v$ of (1.4)-(1.5) has the properties $v\in C^{1+\alpha}(\overline{B}),\quad\partial_{x_{N}}v\in C^{1+\alpha}(\overline{B}),\quad x_{N}D^{2}_{x^{\prime}}v\in C^{\alpha}(\overline{B}).$ Moreover, the following inequality holds for some positive constant $C$ independent of $\varphi$ and $\psi$: (3.1) $|v|_{1+\alpha,\overline{B}}+|\partial_{x_{N}}v|_{1+\alpha,\overline{B}}+|x_{N}D^{2}_{x^{\prime}}v|_{\alpha,\overline{B}}\leq C\left(|\varphi|_{1+\alpha,\overline{B}}+|\psi|_{3/2+\alpha,\overline{D}}\right).$ We will use the following simple estimate, which is a refinement of [1, Lemma 3.2]. ###### Lemma 3.2. Let $\kappa>0$. Then there exists a constant $C>0$ such that $|x_{N}|\int_{0}^{|x_{N}|}\frac{dt}{(1-|x^{\prime}|^{2}-t^{2})^{1+\kappa}}\leq\frac{C}{(1-|x|^{2})^{\kappa}}$ for all $x\in B$. ###### Proof. Set $|x_{N}|=\sigma\,\sqrt{1-|x^{\prime}|^{2}}$; it holds that $0\leq\sigma<1$ for $x=(x^{\prime},x_{N})\in B$. By the change of variable $t=s\,\sqrt{1-|x^{\prime}|^{2}}$, we have that $\int_{0}^{|x_{N}|}\frac{dt}{(1-|x^{\prime}|^{2}-t^{2})^{1+\kappa}}=\frac{1}{(1-|x^{\prime}|^{2})^{1/2+\kappa}}\int_{0}^{\sigma}\frac{ds}{(1-s^{2})^{1+\kappa}},$ and hence $|x_{N}|(1-|x|^{2})^{\kappa}\int_{0}^{|x_{N}|}\frac{dt}{(1-|x^{\prime}|^{2}-t^{2})^{1+\kappa}}=\sigma(1-\sigma^{2})^{\kappa}\int_{0}^{\sigma}\frac{ds}{(1-s^{2})^{1+\kappa}}.$ The right-hand side is bounded by some constant $C$, since L’Hôpital’s rule shows that its limit as $\sigma\to 1^{-}$ is equal to $1/(2\kappa)$. Thus the lemma follows. ∎ The following lemma is essentially shown in [1, Lemma 2.5]. For the sake of completeness, we give a proof. ###### Lemma 3.3. Let $v\in C^{1}(B)\cap C(\overline{B})$. Suppose that there exist a positive constant $M$ and an exponent $\alpha\in(0,1)$ such that $|\nabla v(x)|\leq M(1-|x|^{2})^{-1+\alpha}\ \mbox{ for all }\ x\in B.$ Then $v\in C^{\alpha}(\overline{B})$ and it holds that (3.2) $[v]_{\alpha,\overline{B}}\leq CM,$ for some positive constant $C$ only depending on $\alpha$. ###### Proof. The assumption on $v$ gives that (3.3) $|\nabla v(x)|\leq M\,(1-|x|)^{-1+\alpha}=M\,|x-\overline{x}|^{-1+\alpha}\ \mbox{ with }\ \overline{x}=x/|x|,$ since $\alpha<1$. (i) Let $\vartheta=|x-\overline{x}|$ and $\ell=(x-\overline{x})/\vartheta=-\overline{x}$. Then, we have that $|v(x)-v(\overline{x})|=\left|\int_{0}^{\vartheta}\nabla v(\overline{x}+t\,\ell)\cdot\ell\,dt\right|\leq M\,\int_{0}^{\vartheta}t^{-1+\alpha}dt=\frac{M}{\alpha}\vartheta^{\alpha}.$ (ii) Let $\overline{x}$ and $\overline{y}$ be arbitrary points on ${\mathcal{S}}$ with $\vartheta=|\overline{x}-\overline{y}|<1$. Take $x=(1-\vartheta)\,\overline{x}$ and $y=(1-\vartheta)\,\overline{y}$. Then, for an intermediate point $\xi$ between $x$ and $y$, we have that $|v(\overline{x})-v(\overline{y})|\leq|v(x)-v(\overline{x})|+|v(x)-v(y)|+|v(y)-v(\overline{y})|\leq\\\ \frac{2\,M}{\alpha}\vartheta^{\alpha}+|\nabla v(\xi)|\,|x-y|\leq\frac{2\,M}{\alpha}\vartheta^{\alpha}+M\,\vartheta^{-1+\alpha}|x-y|\leq\\\ \frac{2\,M}{\alpha}\vartheta^{\alpha}+M\,\vartheta^{\alpha}.$ Here, we have used (i), (3.3) and the fact that $|\xi-\overline{\xi}|\geq|x-\overline{x}|=\vartheta$. If $\vartheta=|\overline{x}-\overline{y}|\geq 1$, then one can choose a finite number of points on ${\mathcal{S}}$, say $\overline{x}=\overline{x}_{0},\overline{x}_{1},\overline{x}_{2},\overline{x}_{3},\overline{x}_{4}=\overline{y}$ with $|\overline{x}_{i}-\overline{x}_{i+1}|<1$ so that the previous estimate applies, and the combination of the estimates yields the desired inequality. Therefore, $|v(\overline{x})-v(\overline{y})|\leq CM\,|\overline{x}-\overline{y}|^{\alpha}$, for some constant $C$. (iii) Now, take $x,y\in B$, set $\vartheta=|x-y|$, and let $\overline{x}$ and $\overline{y}$ be the usual projections of $x$ and $y$ on ${\mathcal{S}}$. We can always assume that $|y-\overline{y}|\geq|x-\overline{x}|$. If $|x-\overline{x}|\geq\vartheta$, then, for an intermediate point $\xi$ between $x$ and $y$, $|v(x)-v(y)|\leq|\nabla v(\xi)|\,|x-y|\leq M\,\vartheta^{-1+\alpha}|x-y|=M\,\vartheta^{\alpha},$ thanks to the inequalities $|\xi-\overline{\xi}|\geq|x-\overline{x}|\geq\vartheta$. If $|x-\overline{x}|<\vartheta$ instead, we first infer that $|y-\overline{y}|\leq|y-\overline{x}|\leq|y-x|+|x-\overline{x}|<2\,\vartheta.$ Thus, (i) gives that $|v(x)-v(\overline{x})|\leq M\,\vartheta^{\alpha}/\alpha$ and $|v(y)-v(\overline{y})|\leq 2^{\alpha}\,M\,\vartheta^{\alpha}/\alpha$, while (ii) yields: $|v(\overline{x})-v(\overline{y})|\leq CM\,|\overline{x}-\overline{y}|^{\alpha}\leq 4^{\alpha}\,CM\,\vartheta^{\alpha}.$ We then conclude thanks to the triangle inequality. The bound (3.2) then follows at once. ∎ ###### Proof of Theorem 3.1. Throughout the proof $C$ will denote a generic positive constant only depending on $N$ and $\alpha$. From Proposition 2.6, the unique solution $v\in C^{1}_{N}(\overline{B})\cap C^{2}(B)$ of (1.4)-(1.5) is given by $v(x)=W(x)+Z(x^{\prime})=\int_{0}^{x_{N}}w(x^{\prime},t)\,dt+Z(x^{\prime}),\quad x=(x^{\prime},x_{N})\in B,$ $w$ and $Z$ being the solutions of (2.1) and (2.7), respectively. We note that $Z$ is expressed as $Z=Z_{1}+Z_{2}$, where $Z_{1}$ is the solution of (2.7) with $\partial_{x_{N}}w(x^{\prime},0)$ replaced by $0$ and $Z_{2}$ is the solution of (2.7) with $\psi=0$. Hence, it will be enough to prove the three estimates: (3.4) $\displaystyle|W|_{1+\alpha,\overline{B}}+|\partial_{x_{N}}W|_{1+\alpha,\overline{B}}+|x_{N}D^{2}_{x^{\prime}}W|_{\alpha,\overline{B}}\leq C\,|\varphi|_{1+\alpha,\overline{B}},$ (3.5) $\displaystyle|Z_{1}|_{1+\alpha,\overline{D}}+|x_{N}D^{2}_{x^{\prime}}Z_{1}|_{\alpha,\overline{B}}\leq C\,|\psi|_{3/2+\alpha,\overline{D}}.$ (3.6) $\displaystyle|Z_{2}|_{1+\alpha,\overline{D}}+|x_{N}D^{2}_{x^{\prime}}Z_{2}|_{\alpha,\overline{B}}\leq C\,|\varphi|_{1+\alpha,\overline{B}}.$ We first derive these inequalities under the additional assumptions $\varphi\in C^{2+\alpha}(\overline{B})$ and $\psi\in C^{2+\alpha}(\overline{D})$. We then have (3.7) $w\in C^{2+\alpha}(\overline{B}),\quad W\in C^{2+\alpha}(\overline{B}),\quad Z_{1}\in C^{2+\alpha}(\overline{D}).$ We note that the following inequality holds: (3.8) $|w|_{1+\alpha,\overline{B}}\leq C\,|\varphi|_{1+\alpha,\overline{B}}.$ Indeed, this is shown as follows. Since $\int_{{\mathcal{S}}}P_{B}(x;y)\,dS_{y}=1,\quad\int_{{\mathcal{S}}}P_{B}(x;y)\,y\,dS_{y}=x,$ we have $\nabla w(x)=\nabla\varphi(x)+\int_{{\mathcal{S}}}\nabla_{x}P_{B}(x;y)\,[\varphi(y)-\varphi(x)-\nabla\varphi(x)\cdot(y-x)]\,dS_{y}.$ This with Lemma 2.4 shows that $|\nabla w(x)|\leq|\nabla\varphi(x)|+C[\nabla\varphi]_{\alpha,\overline{B}}\int_{{\mathcal{S}}}|x-y|^{-N+1+\alpha}dS_{y}.$ We easily find that the integral on the right is finite and is bounded by some constant independent of $x$, and hence $|\nabla w|_{0,\overline{B}}\leq C|\nabla\varphi|_{\alpha,\overline{B}}$. Since Proposition 2.2 gives the inequality $|D^{2}w(x)|\leq C|\varphi|_{1+\alpha,\overline{B}}(1-|x|^{2})^{-1+\alpha}$, we have that $[\nabla w]_{\alpha,\overline{B}}\leq C|\varphi|_{1+\alpha,\overline{B}}$ by Lemma 3.3. Therefore (3.8) holds. First, we observe that (3.6) easily follows from the Schauder estimates for the Poisson equation and (3.8). In fact, we have that $|Z_{2}|_{2+\alpha,\overline{D}}\leq C\,|\partial_{x_{N}}w(\cdot,0)|_{\alpha,\overline{D}}\leq C\,|w|_{1+\alpha,\overline{B}}\leq C\,|\varphi|_{1+\alpha,\overline{B}}.$ Next, we prove (3.4). It is clear that $|W|_{1+\alpha,\overline{B}}\leq C|w|_{1+\alpha,\overline{B}}.$ This together with (3.8) and the fact that $\partial_{x_{N}}W=w$ then yields: $|W|_{1+\alpha,\overline{B}}+|\partial_{x_{N}}W|_{1+\alpha,\overline{\Omega}}\leq C\,|w|_{1+\alpha,\overline{B}}\leq C\,|\varphi|_{1+\alpha,\overline{B}}.$ Therefore, we only need to verify that (3.9) $|x_{N}D^{2}_{x^{\prime}}W|_{\alpha,\overline{B}}\leq C|\varphi|_{1+\alpha,\overline{B}}.$ To prove this inequality, we examine pointwise estimates of $D^{2}_{x^{\prime}}W$. Proposition 2.2 yields that $|D^{2}_{x^{\prime}}W(x)|=\left|\int_{0}^{x_{N}}D^{2}_{x^{\prime}}w(x^{\prime},t)dt\right|\leq C\,|\varphi|_{1+\alpha,\overline{B}}\int_{0}^{|x_{N}|}\frac{dt}{(1-|x^{\prime}|^{2}-t^{2})^{1-\alpha}}.$ We estimate the last integral in two ways. First, by monotonicity in $t$ and $|x_{N}|$, we see that the integral can be bounded by $(1-|x|^{2})^{-1+\alpha}$. From this, we infer that (3.10) $|D^{2}_{x^{\prime}}W(x)|\leq C\,|\varphi|_{1+\alpha,\overline{B}}\,(1-|x|^{2})^{-1+\alpha}.$ Secondly, by the inequality (3.11) $x_{N}^{2}\leq 1-|x^{\prime}|^{2}\ \mbox{ for }\ (x^{\prime},x_{N})\in B,$ we get that $\int_{0}^{|x_{N}|}\frac{dt}{(1-|x^{\prime}|^{2}-t^{2})^{1-\alpha}}\leq\int_{0}^{|x_{N}|}\frac{dt}{(x_{N}^{2}-t^{2})^{1-\alpha}}=|x_{N}|^{-1+2\alpha}\int_{0}^{1}\frac{ds}{(1-s^{2})^{1-\alpha}},$ after the change of variable $t=|x_{N}|\,s$. This gives the bound: (3.12) $|x_{N}D^{2}_{x^{\prime}}W|_{0,\overline{B}}\leq C|\varphi|_{1+\alpha,\overline{B}}.$ In order to estimate the Hölder seminorm $[x_{N}D^{2}_{x^{\prime}}W]_{\alpha,\overline{B}}$, we consider the derivatives of $D^{2}_{x^{\prime}}W$. We use Proposition 2.2 to obtain $|D^{3}_{x^{\prime}}W(x)|=\left|\int_{0}^{x_{N}}D^{3}_{x^{\prime}}w(x^{\prime},t)dt\right|\leq C\,|\varphi|_{1+\alpha,\overline{B}}\int_{0}^{|x_{N}|}\frac{dt}{(1-|x^{\prime}|^{2}-t^{2})^{2-\alpha}}.$ Applying Lemma 3.2 to the right-hand side, we infer that (3.13) $|x_{N}D^{3}_{x^{\prime}}W(x)|\leq C\,|\varphi|_{1+\alpha,\overline{B}}\,(1-|x|^{2})^{-1+\alpha}.$ Furthermore, we see from Proposition 2.2 that (3.14) $|\partial_{x_{N}}D^{2}_{x^{\prime}}W(x)|=|D^{2}_{x^{\prime}}w(x)|\leq C\,|\varphi|_{1+\alpha,\overline{B}}\,(1-|x|^{2})^{-1+\alpha}.$ Combining (3.10), (3.13) and (3.14), we deduce that $|x_{N}D^{3}_{x^{\prime}}W(x)|+|\partial_{x_{N}}(x_{N}D^{2}_{x^{\prime}}W(x))|\leq C\,|\varphi|_{1+\alpha,\overline{B}}\,(1-|x|^{2})^{-1+\alpha}.$ Thus, by (3.7) and Lemma 3.3, we obtain that $[x_{N}D^{2}_{x^{\prime}}W]_{\alpha,\overline{B}}\leq C\,|\varphi|_{1+\alpha,\overline{B}}$. This together with (3.12) shows that (3.9) holds. Finally, we verify (3.5). Note that the same estimates as in Proposition 2.2 and (3.8) hold if $w$, $\varphi$ and $B$ are replaced by $Z_{1}$, $\psi$ and $D$, respectively. By (3.8), we see that (3.15) $|Z_{1}|_{1+\alpha,\overline{D}}\leq C\,|\psi|_{1+\alpha,\overline{D}}\leq C\,|\psi|_{3/2+\alpha,\overline{D}}.$ Proposition 2.2 and (3.11) show that $\displaystyle|x_{N}D_{x^{\prime}}^{2}Z_{1}(x^{\prime})|\leq C\,|\psi|_{3/2,\overline{D}}\,|x_{N}|\,(1-|x^{\prime}|)^{-\frac{1}{2}}\leq C\,|\psi|_{3/2+\alpha,\overline{D}},$ $\displaystyle|x_{N}D_{x^{\prime}}^{3}Z_{1}(x^{\prime})|\leq C\,|\psi|_{3/2+\alpha,\overline{D}}\,|x_{N}|\,(1-|x^{\prime}|)^{-\frac{3}{2}+\alpha}\leq C\,|\psi|_{3/2+\alpha,\overline{D}}\,(1-|x|)^{-1+\alpha},$ and $|\partial_{x_{N}}(x_{N}D_{x^{\prime}}^{2}Z_{1}(x^{\prime}))|=|D_{x^{\prime}}^{2}Z_{1}(x^{\prime})|\leq\\\ C\,|\psi|_{1+\alpha,\overline{D}}\,(1-|x^{\prime}|)^{-1+\alpha}\leq C\,|\psi|_{3/2+\alpha,\overline{D}}\,(1-|x|)^{-1+\alpha}.$ Hence it follows from (3.7) and Lemma 3.3 that $|x_{N}D_{x^{\prime}}^{2}Z_{1}|_{\alpha,\overline{B}}\leq C\,|\psi|_{3/2+\alpha,\overline{D}}.$ This and (3.15) gives (3.5). Now, if $\varphi\in C^{1+\alpha}(\overline{B})$ and $\psi\in C^{3/2+\alpha}(\overline{D})$, we take sequences $\\{\varphi_{j}\\}\subset C^{2+\alpha}(\overline{B})$ and $\\{\psi_{j}\\}\subset C^{2+\alpha}(\overline{D})$ such that (3.16) $\displaystyle\varphi_{j}\to\varphi\ \mbox{ in }\ C(\overline{B}),\quad\psi_{j}\to\psi\ \mbox{ in }\ C(\overline{D})\ \mbox{ as }\ j\to\infty,$ (3.17) $\displaystyle|\varphi_{j}|_{1+\alpha,\overline{B}}\leq C\,|\varphi|_{1+\alpha,\overline{B}},\quad|\psi_{j}|_{3/2+\alpha,\overline{D}}\leq C\,|\psi|_{3/2+\alpha,\overline{D}},\quad j=1,2,\ldots.$ Let $w_{j}$ be a solution of (2.1) with $\varphi=\varphi_{j}$. Then, (3.16) and the Schauder interior estimates for Poisson’s equation give that $w_{j}\to w$ in $C^{2}_{\rm loc}(B)$ as $j\to\infty$. Hence $W_{j}(x)=\int_{0}^{x_{N}}w_{j}(x^{\prime},t)\,dt$ converges to $W$ in $C^{2}_{\rm loc}(B)$. Since the inequality $|x_{N}D^{2}_{x^{\prime}}W_{j}|_{\alpha,\overline{B}}\leq C|\varphi_{j}|_{1+\alpha,\overline{B}}$ is valid, we obtain (3.9) by using (3.17) and letting $j\to\infty$. In a similar way, approximating $\psi$ by $\psi_{j}$ gives that (3.5) is still valid for $\psi\in C^{3/2+\alpha}(\overline{D})$. Thus the proof is complete. ∎ ## 4\. Local existence for the interior Backus problem with symmetric data This section is devoted to the proof of Theorem 1.1. ### 4.1. The nonlinear operator ${\mathcal{T}}$ We define an operator ${\mathcal{T}}$ by setting ${\mathcal{T}}[\varphi]=|\nabla v|^{2},$ where $v$ satisfies (1.4)-(1.5) with $\psi=0$. The next proposition shows that ${\mathcal{T}}$ is locally bounded and locally Lipschitz continuous on bounded subsets of $C_{\rm even}^{1+\alpha}(\overline{B})$. ###### Proposition 4.1. We have that ${\mathcal{T}}$ is a mapping from $C_{\rm even}^{1+\alpha}(\overline{B})$ into itself. Furthermore, there are positive constants $C_{1}$ and $C_{2}$ such that (4.1) $\displaystyle\bigl{|}{\mathcal{T}}[\varphi]\bigr{|}_{1+\alpha,\overline{B}}\leq C_{1}\,|\varphi|_{1+\alpha,\overline{B}}^{2},$ (4.2) $\displaystyle\bigl{|}{\mathcal{T}}[\varphi_{1}]-{\mathcal{T}}[\varphi_{2}]\bigr{|}_{1+\alpha,\overline{B}}\leq C_{2}\,\left(|\varphi_{1}|_{1+\alpha,\overline{B}}+|\varphi_{2}|_{1+\alpha,\overline{B}}\right)|\varphi_{1}-\varphi_{2}|_{1+\alpha,\overline{B}}$ for all $\varphi,\varphi_{1},\varphi_{2}\in C_{\rm even}^{1+\alpha}(\overline{B})$. To prove this proposition, we need the following simple lemma. ###### Lemma 4.2. Let $k$ be a non-negative integer and take $\alpha\in(0,1)$. Suppose that a function $v$ defined on $\overline{B}$ is such that $\partial_{x_{N}}v$ in $C^{k+\alpha}(\overline{B})$ and is zero on $\overline{D}\times\\{0\\}$. Then the function defined by (4.3) $\omega(x)=\begin{cases}\displaystyle\frac{v(x)}{x_{N}}&\mbox{ for }\ x_{N}\neq 0,\\\ \partial_{x_{N}}v(x^{\prime},0)&\mbox{ for }\ x_{N}=0,\end{cases}$ belongs to $C^{k+\alpha}(\overline{B})$ and satisfies the inequality: (4.4) $|\omega|_{k+\alpha,\overline{B}}\leq|\partial_{x_{N}}v|_{k+\alpha,\overline{B}}.$ ###### Proof. By the fundamental theorem of calculus, we have that $v(x^{\prime},x_{N})=x_{N}\int_{0}^{1}\partial_{x_{N}}v(x^{\prime},x_{N}t)\,dt,$ and hence $\omega$ can be written as $\omega(x)=\int_{0}^{1}\partial_{x_{N}}v(x^{\prime},x_{N}t)\,dt.$ Thus, we see that the partial derivative $D_{x}^{\beta}\omega$ exists for any multi-index $\beta=(\beta_{1},\ldots,\beta_{N})$ with $|\beta|\leq k$ and is given by $D^{\beta}_{x}\omega(x)=\int_{0}^{1}t^{\beta_{N}}D^{\beta}_{x}\partial_{x_{N}}v(x^{\prime},x_{N}t)dt.$ The assertion then easily follows from this formula. ∎ ###### Proof of Proposition 4.1. Throughout the proof, $i$ is any index in $\\{1,2,\ldots,N-1\\}$ and $C$ is a generic positive constant only depending on $N$ and $\alpha$. (i) Let $\varphi\in C_{\rm even}^{1+\alpha}(\overline{B})$ and let $v$ denote the solution of the problem (1.4)–(1.5) with $\psi=0$. Since $\varphi$ is even in $x_{N}$, we see that the function $-v(x^{\prime},-x_{N})$ is also a solution of (1.4)–(1.5) with $\psi=0$. By uniqueness, we infer that $v$ is odd in $x_{N}$. In particular, $v(x^{\prime},0)=0$, and hence (4.5) $\partial_{x_{j}}(\partial_{x_{i}}v)^{2}=2\partial_{x_{i}}v\,\partial^{2}_{x_{i}x_{j}}v=2\left(\partial_{x_{i}}\omega\right)\left(x_{N}\partial^{2}_{x_{i}x_{j}}v\right)\ \mbox{ for }\ j=1,\ldots,N,$ where $\omega$ is a function given by (4.3). From Theorem 3.1 and Lemma 4.2, we see that the right-hand side of this equality is in $C^{\alpha}(\overline{B})$. Therefore, we have that $(\partial_{x_{i}}v)^{2}\in C^{1+\alpha}(\overline{B})$. This together with the fact that $(\partial_{x_{N}}v)^{2}\in C^{1+\alpha}(\overline{B})$, which follows from Theorem 3.1, gives: ${\mathcal{T}}[\varphi]=|\nabla v|^{2}\in C^{1+\alpha}(\overline{B}).$ Since the fact that $v$ is odd in $x_{N}$ yields that $|\nabla v(x^{\prime},x_{N})|=|\nabla v(x^{\prime},-x_{N})|$, we conclude that ${\mathcal{T}}$ is a mapping from $C_{\rm even}^{1+\alpha}(\overline{B})$ to itself. (ii) Let us derive (4.1). By (4.5), we have: $\left|(\partial_{x_{i}}v)^{2}\right|_{1+\alpha,\overline{B}}=\left|(\partial_{x_{i}}v)^{2}\right|_{0,\overline{B}}+\sum_{j=1}^{N}\left|\partial_{x_{j}}(\partial_{x_{i}}v)^{2}\right|_{\alpha,\overline{B}}=\\\ \left|(\partial_{x_{i}}v)^{2}\right|_{0,\overline{B}}+2\sum_{j=1}^{N-1}\left|\left(\partial_{x_{i}}\omega\right)\left(x_{N}\partial^{2}_{x_{i}x_{j}}v\right)\right|_{\alpha,\overline{B}}+2\left|\partial_{x_{i}}v\,\partial^{2}_{x_{i}x_{N}}v)\right|_{\alpha,\overline{B}}.$ From (3.1), the first and third terms of the rightest-hand side are handled as $\displaystyle\left|(\partial_{x_{i}}v)^{2}\right|_{0,\overline{B}}\leq\left|\partial_{x_{i}}v\right|_{0,\overline{B}}^{2}\leq|v|_{1+\alpha,\overline{B}}^{2}\leq C\,|\varphi|_{1+\alpha,\overline{B}}^{2},$ $\displaystyle\left|\partial_{x_{i}}v\,\partial^{2}_{x_{i}x_{N}}v\right|_{\alpha,\overline{B}}\leq\left|\partial_{x_{i}}v\right|_{\alpha,\overline{B}}\left|\partial^{2}_{x_{i}x_{N}}v\right|_{\alpha,\overline{B}}\leq|v|_{1+\alpha,\overline{B}}\left|\partial_{x_{N}}v\right|_{1+\alpha,\overline{B}}\leq C|\varphi|_{1+\alpha,\overline{B}}^{2}.$ Furthermore, (3.1) and (4.4) show that $\left|\left(\partial_{x_{i}}\omega\right)\left(x_{N}\partial^{2}_{x_{i}x_{j}}v\right)\right|_{\alpha,\overline{B}}\leq\left|\partial_{x_{i}}\omega\right|_{\alpha,\overline{B}}\left|x_{N}\partial^{2}_{x_{i}x_{j}}v\right|_{\alpha,\overline{B}}\leq\\\ \left|\omega\right|_{1+\alpha,\overline{B}}\left|x_{N}\partial^{2}_{x_{i}x_{j}}v\right|_{\alpha,\overline{B}}\leq\left|\partial_{x_{N}}v\right|_{1+\alpha,\overline{B}}\left|x_{N}\partial^{2}_{x_{i}x_{j}}v\right|_{\alpha,\overline{B}}\leq C\,|\varphi|_{1+\alpha,\overline{B}}^{2},$ for $j=1,\ldots,N-1$. From these estimates it follows that (4.6) $\left|(\partial_{x_{i}}v)^{2}\right|_{1+\alpha,\overline{B}}\leq C\,|\varphi|_{1+\alpha,\overline{B}}^{2}.$ In order to estimate $(\partial_{x_{N}}v)^{2}$, we use (3.1) to find that (4.7) $\left|(\partial_{x_{N}}v)^{2}\right|_{1+\alpha,\overline{B}}\leq C\,\left|\partial_{x_{N}}v\right|_{1+\alpha,\overline{B}}^{2}\leq C\,|\varphi|_{1+\alpha,\overline{B}}^{2}.$ The combination of (4.6) and (4.7) then gives (4.1). (iii) It remains to prove (4.2). For $m=1,2$, let $v_{m}$ be the solution of the problem (1.4)-(1.5) with $\varphi=\varphi_{m}\in C_{\rm even}^{1+\alpha}(\overline{B})$ and $\psi=0$. Also, let $\omega_{m}$ be defined by (4.3) with $v=v_{m}$. It is clear that $\left|(\partial_{x_{i}}v_{1})^{2}-(\partial_{x_{i}}v_{2})^{2}\right|_{0,\overline{B}}\leq\\\ \left(|v_{1}|_{0,\overline{B}}+|v_{2}|_{0,\overline{B}}\right)|v_{1}-v_{2}|_{0,\overline{B}}\leq\left(|v_{1}|_{1+\alpha,\overline{B}}+|v_{2}|_{1+\alpha,\overline{B}}\right)|v_{1}-v_{2}|_{1+\alpha,\overline{B}}.$ Hence, (3.1) easily gives that (4.8) $\left|(\partial_{x_{i}}v_{1})^{2}-(\partial_{x_{i}}v_{2})^{2}\right|_{0,\overline{B}}\leq C\,\left(|\varphi_{1}|_{1+\alpha,\overline{B}}+|\varphi_{2}|_{1+\alpha,\overline{B}}\right)|\varphi_{1}-\varphi_{2}|_{1+\alpha,\overline{B}}.$ Next, we write the differential identity: $\sum_{j=1}^{N}\partial_{x_{j}}\left[(\partial_{x_{i}}v_{1})^{2}-(\partial_{x_{i}}v_{2})^{2}\right]=2\,\sum_{j=1}^{N}\bigl{[}\partial_{x_{i}}v_{1}\,\partial^{2}_{x_{i}x_{j}}v_{1}-\partial_{x_{i}}v_{2}\,\partial^{2}_{x_{i}x_{j}}v_{2}\bigr{]}=\\\ 2\,\sum_{j=1}^{N-1}\bigl{[}(\partial_{x_{i}}\omega_{1})(x_{N}\partial^{2}_{x_{i}x_{j}}v_{1}-x_{N}\partial^{2}_{x_{i}x_{j}}v_{2})+x_{N}\,(\partial^{2}_{x_{i}x_{j}}v_{2})(\partial_{x_{i}}\omega_{1}-\partial_{x_{i}}\omega_{2})\bigr{]}+\\\ 2\,\bigl{[}(\partial_{x_{i}}v_{1})(\partial^{2}_{x_{i}x_{N}}v_{1}-\partial^{2}_{x_{i}x_{N}}v_{2})+(\partial^{2}_{x_{i}x_{N}}v_{2})(\partial_{x_{i}}v_{1}-\partial_{x_{i}}v_{2})\bigr{]}.$ We then take care of the last summand: $\left|(\partial_{x_{i}}v_{1})(\partial^{2}_{x_{i}x_{N}}v_{1}-\partial^{2}_{x_{i}x_{N}}v_{2})+(\partial^{2}_{x_{i}x_{N}}v_{2})(\partial_{x_{i}}v_{1}-\partial_{x_{i}}v_{2})\right|_{\alpha,\overline{B}}\leq\\\ \left|\partial_{x_{i}}v_{1}\right|_{\alpha,\overline{B}}\left|\partial^{2}_{x_{i}x_{N}}v_{1}-\partial^{2}_{x_{i}x_{N}}v_{2}\right|_{\alpha,\overline{B}}+\left|\partial^{2}_{x_{i}x_{N}}v_{2}\right|_{\alpha,\overline{B}}\left|\partial_{x_{i}}v_{1}-\partial_{x_{i}}v_{2}\right|_{\alpha,\overline{B}}\leq\\\ \left|v_{1}\right|_{1+\alpha,\overline{B}}\left|\partial_{x_{N}}v_{1}-\partial_{x_{N}}v_{2}\right|_{1+\alpha,\overline{B}}+\left|\partial_{x_{N}}v_{2}\right|_{1+\alpha,\overline{B}}\left|v_{1}-v_{2}\right|_{1+\alpha,\overline{B}}.$ Thus, we have that $\left|(\partial_{x_{i}}v_{1})(\partial^{2}_{x_{i}x_{N}}v_{1}-\partial^{2}_{x_{i}x_{N}}v_{2})+(\partial^{2}_{x_{i}x_{N}}v_{2})(\partial_{x_{i}}v_{1}-\partial_{x_{i}}v_{2})\right|_{\alpha,\overline{B}}\leq\\\ C\,\left(|\varphi_{1}|_{1+\alpha,\overline{B}}+|\varphi_{2}|_{1+\alpha,\overline{B}}\right)|\varphi_{1}-\varphi_{2}|_{1+\alpha,\overline{B}}.$ Moreover, for $j=1,\ldots,N-1$, we have: $\left|(\partial_{x_{i}}\omega_{1})(x_{N}\partial^{2}_{x_{i}x_{j}}v_{1}-x_{N}\partial^{2}_{x_{i}x_{j}}v_{2})+x_{N}\,(\partial^{2}_{x_{i}x_{j}}v_{2})(\partial_{x_{i}}\omega_{1}-\partial_{x_{i}}\omega_{2})\right|_{\alpha,\overline{B}}\leq\\\ \left|\omega_{1}\right|_{1+\alpha,\overline{B}}\left|x_{N}\partial^{2}_{x_{i}x_{j}}v_{1}-x_{N}\partial^{2}_{x_{i}x_{j}}v_{2}\right|_{\alpha,\overline{B}}+\left|x_{N}\partial^{2}_{x_{i}x_{j}}v_{2}\right|_{\alpha,\overline{B}}\left|\omega_{1}-\omega_{2}\right|_{1+\alpha,\overline{B}}\leq\\\ \left|\partial_{x_{N}}v_{1}\right|_{1+\alpha,\overline{B}}\left|x_{N}\partial^{2}_{x_{i}x_{j}}v_{1}-x_{N}\partial^{2}_{x_{i}x_{j}}v_{2}\right|_{\alpha,\overline{B}}+\left|x_{N}\partial^{2}_{x_{i}x_{j}}v_{2}\right|_{\alpha,\overline{B}}\left|\partial_{x_{N}}v_{1}-\partial_{x_{N}}v_{2}\right|_{1+\alpha,\overline{B}}.$ where we have used (4.4). Hence, (3.1) gives: (4.9) $\left|(\partial_{x_{i}}\omega_{1})(x_{N}\partial^{2}_{x_{i}x_{j}}v_{1}-x_{N}\partial^{2}_{x_{i}x_{j}}v_{2})+x_{N}\,(\partial^{2}_{x_{i}x_{j}}v_{2})(\partial_{x_{i}}\omega_{1}-\partial_{x_{i}}\omega_{2})\right|_{\alpha,\overline{B}}\leq\\\ C\,\left(|\varphi_{1}|_{1+\alpha,\overline{B}}+|\varphi_{2}|_{1+\alpha,\overline{B}}\right)|\varphi_{1}-\varphi_{2}|_{1+\alpha,\overline{B}}.$ All in all, by (4.8)–(4.9) we deduce that $\left|(\partial_{x_{i}}v_{1})^{2}-(\partial_{x_{i}}v_{2})^{2}\right|_{1+\alpha,\overline{B}}\leq C\,\left(|\varphi_{1}|_{1+\alpha,\overline{B}}+|\varphi_{2}|_{1+\alpha,\overline{B}}\right)|\varphi_{1}-\varphi_{2}|_{1+\alpha,\overline{B}},$ for $i=1,\dots,N-1$. On the other hand, we can also use (3.1) to infer that $\left|(\partial_{x_{N}}v_{1})^{2}-(\partial_{x_{N}}v_{2})^{2}\right|_{1+\alpha,\overline{B}}\leq\\\ C\,\left(|\partial_{x_{N}}v_{1}|_{1+\alpha,\overline{B}}+|\partial_{x_{N}}v_{2}|_{1+\alpha,\overline{B}}\right)|\partial_{x_{N}}v_{1}-\partial_{x_{N}}v_{2}|_{1+\alpha,\overline{B}}\leq\\\ C\,\left(|\varphi_{1}|_{1+\alpha,\overline{B}}+|\varphi_{2}|_{1+\alpha,\overline{B}}\right)|\varphi_{1}-\varphi_{2}|_{1+\alpha,\overline{B}}.$ In conclusion, we obtain (4.2), and the lemma follows. ∎ ### 4.2. The nonlinear operator $\tilde{\mathcal{T}}_{\psi}$ We fix a cut-off function $\eta\in C^{\infty}(\mathbb{R})$ satisfying $\eta(t)=1\ \mbox{ if }\ |t|\leq\frac{1}{3},\qquad\eta(t)=0\ \mbox{ if }\ |t|\geq\frac{2}{3}.$ For a function $\phi$ defined on $\overline{B}$, we set ${{\mathcal{J}}}[\phi](x)=\eta(x_{N})\phi\left(\sqrt{1-x_{N}^{2}}e_{1}^{\prime},x_{N}\right)+(1-\eta(x_{N}))\phi(x),\quad x=(x^{\prime},x_{N})\in\overline{B}.$ Here $e_{1}^{\prime}=(1,0,\ldots,0)\in\mathbb{R}^{N-1}$. Then, for fixed $\psi\in C^{3/2+\alpha}(\overline{D})$, we define an operator $\tilde{\mathcal{T}}_{\psi}$ by $\tilde{\mathcal{T}}_{\psi}[\varphi]={\mathcal{J}}\left[|\nabla v|^{2}\right],$ where $v$ is the solution of (1.4)-(1.5). Our goal here is to obtain the properties of $\tilde{\mathcal{T}}_{\psi}$, which enables us to solve problem (1.1) in $C^{1+\alpha}_{\rm ax}(\overline{B})$. ###### Proposition 4.3. Let $\psi\in C^{3/2+\alpha}(\overline{D})$. Then the following hold. * (i) The operator $\tilde{\mathcal{T}}_{\psi}$ is a mapping from $C^{1+\alpha}(\overline{B})$ into itself. Moreover, there exist positive constants $C_{3}$ and $C_{4}$ such that (4.10) $\bigl{|}\tilde{\mathcal{T}}_{\psi}[\varphi]\bigr{|}_{1+\alpha,\overline{B}}\leq C_{3}\,\left(|\varphi|_{1+\alpha,\overline{B}}^{2}+|\psi|_{3/2+\alpha,\overline{D}}^{2}\right),$ (4.11) $\bigl{|}\tilde{\mathcal{T}}_{\psi}[\varphi_{1}]-\tilde{\mathcal{T}}_{\psi}[\varphi_{2}]\bigr{|}_{1+\alpha,\overline{B}}\leq\\\ C_{4}\,\left(|\varphi_{1}|_{1+\alpha,\overline{B}}+|\varphi_{2}|_{1+\alpha,\overline{B}}+|\psi|_{3/2+\alpha,\overline{D}}\right)|\varphi_{1}-\varphi_{2}|_{1+\alpha,\overline{B}}$ for all $\varphi,\varphi_{1},\varphi_{2}\in C^{1+\alpha}(\overline{B})$. * (ii) If $\varphi\in C^{1+\alpha}_{\rm ax}(\overline{B})$ and $\psi$ is constant, then (4.12) $\tilde{\mathcal{T}}_{\psi}[\varphi]\in C^{1+\alpha}_{\rm ax}(\overline{B})\ \mbox{ and }\ \tilde{\mathcal{T}}_{\psi}[\varphi]=|\nabla v|^{2}\ \mbox{ on }\ {\mathcal{S}}.$ Here $v$ is the solution of (1.4)-(1.5). Before proving the proposition, we examine properties of ${\mathcal{J}}$. ###### Lemma 4.4. If a function $\phi$ defined on $\overline{B}$ satisfies $\phi(x^{\prime},x_{N})=\phi(|x^{\prime}|e_{1}^{\prime},x_{N})$, then ${{\mathcal{J}}}[\phi]=\phi\ \mbox{ on }\ {\mathcal{S}}.$ ###### Proof. Let $x=(x^{\prime},x_{N})\in{\mathcal{S}}$. Then, since $|x^{\prime}|^{2}+x_{N}^{2}=1$, we have that $\phi\left(\sqrt{1-x_{N}^{2}}e_{1}^{\prime},x_{N}\right)=\phi(|x^{\prime}|e_{1}^{\prime},x_{N})=\phi(x).$ This gives that ${{\mathcal{J}}}[\phi](x)=\eta(x_{N})\phi(x)+(1-\eta(x_{N}))\phi(x)=\phi(x)$, as desired. ∎ ###### Lemma 4.5. Suppose that $\phi\in C^{\alpha}(\overline{B})\cap C^{1}(B)$ satisfies $\partial_{x_{N}}\phi\in C^{\alpha}(\overline{B})$ and $x_{N}\nabla_{x^{\prime}}\phi\in C^{\alpha}(\overline{B})$. Then ${{\mathcal{J}}}[\phi]\in C^{1+\alpha}(\overline{B})$ and $|{{\mathcal{J}}}[\phi]|_{1+\alpha,\overline{B}}\leq C\left(|\phi|_{\alpha,\overline{B}}+|\partial_{x_{N}}\phi|_{\alpha,\overline{B}}+|x_{N}\nabla_{x^{\prime}}\phi|_{\alpha,\overline{B}}\right),$ where $C$ is a positive constant independent of $\phi$. ###### Proof. Let $B_{1}=\left\\{(x^{\prime},x_{N})\in B:|x_{N}|<\frac{2}{3}\right\\},\qquad B_{2}=\left\\{(x^{\prime},x_{N})\in B:|x_{N}|>\frac{1}{3}\right\\},$ and define the mapping $\xi:\overline{B}\ni(x^{\prime},x_{N})\mapsto\xi(x)=\left(\sqrt{1-x_{N}^{2}}e_{1}^{\prime},x_{N}\right)\in\overline{B}.$ We first show that (4.13) $\displaystyle\phi\circ\xi\in C^{1+\alpha}(\overline{B_{1}}),\qquad\phi\in C^{1+\alpha}(\overline{B_{2}}),$ (4.14) $\displaystyle|\phi\circ\xi|_{1+\alpha,\overline{B_{1}}}+|\phi|_{1+\alpha,\overline{B_{2}}}\leq C\left(|\phi|_{\alpha,\overline{B}}+|\partial_{x_{N}}\phi|_{\alpha,\overline{B}}+|x_{N}\nabla_{x^{\prime}}\phi|_{\alpha,\overline{B}}\right).$ Here and subsequently, $C$ denotes a positive constant independent of $\phi$. To prove $\phi\circ\xi\in C^{1+\alpha}(\overline{B_{1}})$, we observe that $\phi\circ\xi$ depends only on $x_{N}$ and $\partial_{x_{N}}(\phi\circ\xi)=(\partial_{x_{N}}\phi)\circ\xi-\frac{1}{\sqrt{1-x_{N}^{2}}}(x_{N}\partial_{x_{1}}\phi)\circ\xi.$ By assumption and the fact that $\xi$ and $1/\sqrt{1-x_{N}^{2}}$ are smooth on $\overline{B_{1}}$, we see that the right-hand side of this equality is in $C^{\alpha}(\overline{B_{1}})$. Moreover, $|\partial_{x_{N}}(\phi\circ\xi)|_{\alpha,\overline{B_{1}}}\leq\left|(\partial_{x_{N}}\phi)\circ\xi\right|_{\alpha,\overline{B_{1}}}+\left|\frac{1}{\sqrt{1-x_{N}^{2}}}(x_{N}\partial_{x_{1}}\phi)\circ\xi\right|_{\alpha,\overline{B_{1}}}\leq\\\ C|\partial_{x_{N}}\phi|_{\alpha,\overline{B}}+C|(x_{N}\partial_{x_{1}}\phi)\circ\xi|_{\alpha,\overline{B_{1}}}\leq C|\partial_{x_{N}}\phi|_{\alpha,\overline{B}}+C|x_{N}\partial_{x_{1}}\phi|_{\alpha,\overline{B}}.$ Therefore, we have that $\phi\circ\xi\in C^{1+\alpha}(\overline{B_{1}})$ and (4.15) $|\phi\circ\xi|_{1+\alpha,\overline{B_{1}}}=|\phi\circ\xi|_{0,\overline{B_{1}}}+|\partial_{x_{N}}(\phi\circ\xi)|_{\alpha,\overline{B}}\leq\\\ |\phi|_{0,\overline{B}}+C|\partial_{x_{N}}\phi|_{\alpha,\overline{B}}+C|x_{N}\partial_{x_{1}}\phi|_{\alpha,\overline{B}}.$ Since $x_{N}^{-1}$ is smooth on $\overline{B_{2}}$, we deduce that $\phi=x_{N}^{-1}\cdot(x_{N}\phi)\in C^{1+\alpha}(\overline{B_{2}})$ and (4.16) $|\phi|_{1+\alpha,\overline{B_{2}}}=|\phi|_{0,\overline{B_{2}}}+|\partial_{x_{N}}\phi|_{\alpha,\overline{B_{2}}}+|x_{N}^{-1}\cdot x_{N}\nabla_{x^{\prime}}\phi|_{\alpha,\overline{B_{2}}}\leq\\\ |\phi|_{0,\overline{B}}+|\partial_{x_{N}}\phi|_{\alpha,\overline{B}}+C|x_{N}\nabla_{x^{\prime}}\phi|_{\alpha,\overline{B}}.$ By (4.15) and (4.16), we obtain (4.14). We note that $\eta$ and $1-\eta$ vanish on $\overline{B}\setminus B_{1}$ and $\overline{B}\setminus B_{2}$, respectively. This with (4.13) shows that ${\mathcal{J}}[\phi]=\eta(\phi\circ\xi)+(1-\eta)\phi\in C^{1+\alpha}(\overline{B})$. Furthermore, $|{{\mathcal{J}}}[\phi]|_{1+\alpha,\overline{B}}\leq C|\eta(\phi\circ\xi)|_{1+\alpha,\overline{B_{1}}}+C|(1-\eta)\phi|_{1+\alpha,\overline{B_{2}}}\leq C|\phi\circ\xi|_{1+\alpha,\overline{B_{1}}}+C|\phi|_{1+\alpha,\overline{B_{2}}}.$ Combining this and (4.14) proves the lemma. ∎ ###### Proof of Proposition 4.3. Let $\varphi,\varphi_{1},\varphi_{2}\in C^{1+\alpha}(\overline{B})$ and $\psi\in C^{3/2+\alpha}(\overline{D})$. In the proof, $C$ stands for a generic positive constant only independent of these functions. Let $v$ stand for the solution of (1.4)-(1.5). Then Theorem 3.1 shows that the function $\phi=|\nabla v|^{2}$ satisfies $\phi\in C^{\alpha}(\overline{B})\cap C^{1}(B)$, $\partial_{x_{N}}\phi=2\nabla v\cdot\nabla\partial_{x_{N}}v\in C^{\alpha}(\overline{B})$ and $x_{N}\partial_{x_{j}}\phi=2x_{N}\nabla v\cdot\nabla\partial_{x_{j}}v\in C^{\alpha}(\overline{B})$ for $j=1,\ldots,N-1$. Hence, using Lemma 4.5, we see that $\tilde{\mathcal{T}}_{\psi}[\varphi]={{\mathcal{J}}}[\phi]\in C^{1+\alpha}(\overline{B})$ and $\bigl{|}\tilde{\mathcal{T}}_{\psi}[\varphi]\bigr{|}_{1+\alpha,\overline{B}}\leq C\left(\bigl{|}|\nabla v|^{2}\bigl{|}_{\alpha,\overline{B}}+\bigl{|}\nabla v\cdot\nabla\partial_{x_{N}}v\bigl{|}_{\alpha,\overline{B}}+\sum_{j=1}^{N-1}\bigl{|}x_{N}\nabla v\cdot\nabla\partial_{x_{j}}v\bigl{|}_{\alpha,\overline{B}}\right).$ Each term of the right-hand side is estimated as $\bigl{|}|\nabla v|^{2}\bigl{|}_{\alpha,\overline{B}}\leq|v|_{1+\alpha,\overline{B}}^{2},\qquad\bigl{|}\nabla v\cdot\nabla\partial_{x_{N}}v\bigl{|}_{\alpha,\overline{B}}\leq C|v|_{1+\alpha,\overline{B}}|\partial_{x_{N}}v|_{1+\alpha,\overline{B}},$ and $\sum_{j=1}^{N-1}\bigl{|}x_{N}\nabla v\cdot\nabla\partial_{x_{j}}v\bigl{|}_{\alpha,\overline{B}}\leq C\sum_{j=1}^{N-1}|v|_{1+\alpha,\overline{B}}|x_{N}\nabla\partial_{x_{j}}v|_{\alpha,\overline{B}}\leq\\\ C|v|_{1+\alpha,\overline{B}}\left(|x_{N}D_{x^{\prime}}^{2}v|_{\alpha,\overline{B}}+|\partial_{x_{N}}v|_{1+\alpha,\overline{B}}\right).$ Therefore, by (3.1), we conclude that $\tilde{\mathcal{T}}_{\psi}$ is a mapping from $C^{1+\alpha}(\overline{B})$ into itself and that (4.10) holds. Inequality (4.11) is shown as follows. For $i=1,2$, we denote by $v_{i}$ the solution of (1.4)-(1.5) with $\varphi=\varphi_{i}$ and $\psi=\psi_{i}$. We see from Lemma 4.5 that (4.17) $\bigl{|}\tilde{\mathcal{T}}_{\psi}[\varphi_{1}]-\tilde{\mathcal{T}}_{\psi}[\varphi_{2}]\bigr{|}_{1+\alpha,\overline{B}}=\bigl{|}{\mathcal{J}}\bigl{[}|\nabla v_{1}|^{2}-|\nabla v_{2}|^{2}\bigr{]}\bigr{|}_{1+\alpha,\overline{B}}\leq\\\ C\left(\bigl{|}|\nabla v_{1}|^{2}-|\nabla v_{2}|^{2}\bigr{|}_{0,\overline{B}}+\bigl{|}\partial_{x_{N}}\left(|\nabla v_{1}|^{2}-|\nabla v_{2}|^{2}\right)\bigr{|}_{\alpha,\overline{B}}+\right.\\\ \left.\sum_{j=1}^{N-1}\bigl{|}x_{N}\partial_{x_{j}}\left(|\nabla v_{1}|^{2}-|\nabla v_{2}|^{2}\right)\bigr{|}_{\alpha,\overline{B}}\right).$ For abbreviation, we write $w_{1}=v_{1}+v_{2}$ and $w_{2}=v_{1}-v_{2}$. Then the first and second terms on the right of the above inequality can be handled as (4.18) $\bigl{|}|\nabla v_{1}|^{2}-|\nabla v_{2}|^{2}\bigr{|}_{0,\overline{B}}=\bigl{|}\nabla w_{1}\cdot\nabla w_{2}\bigr{|}_{0,\overline{B}}\leq C|w_{1}|_{1+\alpha,\overline{B}}|w_{2}|_{1+\alpha,\overline{B}},$ and $\bigl{|}\partial_{x_{N}}\left(|\nabla v_{1}|^{2}-|\nabla v_{2}|^{2}\right)\bigr{|}_{\alpha,\overline{B}}=\bigl{|}\nabla\partial_{x_{N}}w_{1}\cdot\nabla w_{2}+\nabla w_{1}\cdot\nabla\partial_{x_{N}}w_{2}\bigr{|}_{\alpha,\overline{B}}\leq\\\ C|\partial_{x_{N}}w_{1}|_{1+\alpha,\overline{B}}|w_{2}|_{1+\alpha,\overline{B}}+C|w_{1}|_{1+\alpha,\overline{B}}|\partial_{x_{N}}w_{2}|_{1+\alpha,\overline{B}}.$ The other terms are estimated as $\bigl{|}x_{N}\partial_{x_{j}}\left(|\nabla v_{1}|^{2}-|\nabla v_{2}|^{2}\right)\bigr{|}_{\alpha,\overline{B}}=\bigl{|}x_{N}\nabla\partial_{x_{j}}w_{1}\cdot\nabla w_{2}+x_{N}\nabla w_{1}\cdot\nabla\partial_{x_{j}}w_{2}\bigr{|}_{\alpha,\overline{B}}\leq\\\ |x_{N}\nabla\partial_{x_{j}}w_{1}|_{\alpha,\overline{B}}|w_{2}|_{1+\alpha,\overline{B}}+|w_{1}|_{1+\alpha,\overline{B}}|x_{N}\nabla\partial_{x_{j}}w_{2}|_{\alpha,\overline{B}},$ and hence $\sum_{j=1}^{N-1}\bigl{|}x_{N}\partial_{x_{j}}\left(|\nabla v_{1}|^{2}-|\nabla v_{2}|^{2}\right)\bigr{|}_{\alpha,\overline{B}}\leq\\\ \left(|x_{N}D_{x^{\prime}}^{2}w_{1}|_{\alpha,\overline{B}}+|\partial_{x_{N}}w_{1}|_{1+\alpha,\overline{B}}\right)|w_{2}|_{1+\alpha,\overline{B}}+\\\ |w_{1}|_{1+\alpha,\overline{B}}\left(|x_{N}D_{x^{\prime}}^{2}w_{2}|_{\alpha,\overline{B}}+|\partial_{x_{N}}w_{2}|_{1+\alpha,\overline{B}}\right).$ We note that $w_{1}$ satisfies (1.4)-(1.5) with $\varphi$ and $\psi$ replaced by $\varphi_{1}+\varphi_{2}$ and $2\psi$, respectively. Hence it follows from Theorem 3.1 that $|w_{1}|_{1+\alpha,\overline{B}}+|\partial_{x_{N}}w_{1}|_{1+\alpha,\overline{B}}+|x_{N}D_{x^{\prime}}^{2}w_{1}|_{\alpha,\overline{B}}\leq\\\ C\left(|\varphi_{1}|_{1+\alpha,\overline{B}}+|\varphi_{2}|_{1+\alpha,\overline{B}}+|\psi|_{3/2+\alpha,\overline{D}}\right).$ Moreover, since $w_{2}$ solves (1.4)-(1.5) with $\varphi=\varphi_{1}-\varphi_{2}$ and $\psi=0$, Theorem 3.1 shows that (4.19) $|w_{2}|_{1+\alpha,\overline{B}}+|\partial_{x_{N}}w_{2}|_{1+\alpha,\overline{B}}+|x_{N}D_{x^{\prime}}^{2}w_{2}|_{\alpha,\overline{B}}\leq C|\varphi_{1}-\varphi_{2}|_{1+\alpha,\overline{B}}.$ We thus obtain (4.11) by plugging (4.18)-(4.19) into (4.17). It remains to prove (ii). To this end, we assume that $\varphi\in C^{1+\alpha}_{\rm ax}(\overline{B})$ and that $\psi$ is constant. Then, we can directly check that for any $(N-1)\times(N-1)$ orthogonal matrix ${{\mathcal{R}}}$, the function $v({{\mathcal{R}}}x^{\prime},x_{N})$ also satisfies (1.4)-(1.5). Hence Proposition 2.1 gives that $v(x^{\prime},x_{N})=v({{\mathcal{R}}}x^{\prime},x_{N})$. In particular, we have that $|\nabla v(x^{\prime},x_{N})|^{2}=|\nabla v(|x^{\prime}|e_{1}^{\prime},x_{N})|^{2}$. Therefore (i) of Proposition 4.3 and Lemma 4.4 show that (4.12) holds, and the proof is complete. ∎ ### 4.3. Contraction mappings and the proof of Theorem 1.1 For $g\in C^{1+\alpha}(\overline{B})$ and $h\in\mathbb{R}$, we define operators $\Psi_{g}$ and $\tilde{\Psi}_{g,h}$ by $\Psi_{g}[\varphi]=\frac{1}{2}\left(g^{2}-1-{\mathcal{T}}[\varphi]\right),\qquad\tilde{\Psi}_{g,h}[\varphi]=\frac{1}{2}\left(g^{2}-1-\tilde{\mathcal{T}}_{h}[\varphi]\right).$ Note that, by Propositions 4.1 and 4.3, $\Psi_{g}$ is a mapping from $C_{\rm even}^{1+\alpha}(\overline{B})$ into itself and $\tilde{\Psi}_{g,h}$ is a mapping from $C_{\rm ax}^{1+\alpha}(\overline{B})$ into itself. We also define closed sets of $C^{1+\alpha}(\overline{B})$ by $\displaystyle X_{g}=\\{\varphi\in C_{\rm even}^{1+\alpha}(\overline{B}):|\varphi|_{1+\alpha,\overline{B}}\leq|g^{2}-1|_{1+\alpha,\overline{B}}\\},$ $\displaystyle\tilde{X}_{g,h}=\\{\varphi\in C_{\rm ax}^{1+\alpha}(\overline{B}):|\varphi|_{1+\alpha,\overline{B}}\leq|g^{2}-1|_{1+\alpha,\overline{B}}+|h|\\},$ and positive constants $\delta_{1}$ and $\delta_{2}$ by $\delta_{1}=\min\left\\{\frac{1}{C_{1}},\frac{\lambda}{C_{2}}\right\\},\qquad\delta_{2}=\min\left\\{\frac{1}{2C_{3}},\frac{2\lambda}{3C_{4}}\right\\}.$ Here, $\lambda\in(0,1)$ is an arbitrary fixed constant, and $C_{1}$, $C_{2}$, $C_{3}$, $C_{4}$ are the constants given in Propositions 4.1 and 4.3. ###### Lemma 4.6. The following hold. * (i) If $g\in C_{\rm even}^{1+\alpha}(\overline{B})$ satisfies $|g^{2}-1|_{1+\alpha,\overline{B}}\leq\delta_{1}$, then $\Psi_{g}$ has a unique fixed point in $X_{g}$. * (ii) If $g\in C_{\rm ax}^{1+\alpha}(\overline{B})$ and $h\in\mathbb{R}$ satisfy $|g^{2}-1|_{1+\alpha,\overline{B}}+|h|\leq\delta_{2}$, then $\tilde{\Psi}_{g.h}$ has a unique fixed point in $\tilde{X}_{g.h}$. ###### Proof. We show (i). Set $\delta=|g^{2}-1|_{1+\alpha,\overline{B}}$, so that $\delta\leq\delta_{1}$. For $\varphi\in X_{g}$, we see from Proposition 4.1 that $\Psi_{g}[\varphi]\in C_{\rm even}^{1+\alpha}(\overline{B})$ and $\bigl{|}\Psi_{g}[\varphi]\bigr{|}_{1+\alpha,\overline{B}}\leq\frac{1}{2}\left(\delta+\bigl{|}{\mathcal{T}}[\varphi]\bigr{|}_{1+\alpha,\overline{B}}\right)\leq\frac{1}{2}\left(\delta+C_{1}|\varphi|_{1+\alpha,\overline{B}}^{2}\right)\leq\frac{1}{2}\left(1+C_{1}\delta_{1}\right)\delta\leq\delta.$ Hence we have that $\Psi_{g}(X_{g})\subseteq X_{g}$. Furthermore, inequality (4.2) shows that for $\varphi_{1},\varphi_{2}\in X_{g}$, $\bigl{|}\Psi_{g}[\varphi_{1}]-\Psi_{g}[\varphi_{2}]\bigr{|}_{1+\alpha,\overline{B}}=\frac{1}{2}\,\bigl{|}{\mathcal{T}}[\varphi_{1}]-{\mathcal{T}}[\varphi_{2}]\bigr{|}_{1+\alpha,\overline{B}}\leq\\\ \frac{1}{2}\,C_{2}\,\left(|\varphi_{1}|_{1+\alpha,\overline{B}}+|\varphi_{2}|_{1+\alpha,\overline{B}}\right)|\varphi_{1}-\varphi_{2}|_{1+\alpha,\overline{B}}\leq C_{2}\,\delta_{1}|\varphi_{1}-\varphi_{2}|_{1+\alpha,\overline{B}}.$ Thus, we infer that $\bigl{|}\Psi_{g}[\varphi_{1}]-\Psi_{g}[\varphi_{2}]\bigr{|}_{1+\alpha,\overline{B}}\leq\lambda\,|\varphi_{1}-\varphi_{2}|_{1+\alpha,\overline{B}},$ i.e. we have proved that $\Psi_{g}$ is a contraction mapping in $X_{g}$. The Banach fixed point theorem then gives the desired conclusion. The assertion (ii) can be proved in the same way, by applying Proposition 4.3 instead of Proposition 4.1. ∎ We are now in a position to prove Theorem 1.1. ###### Proof of Theorem 1.1. We first prove (i). Note that (4.20) $|g^{2}-1|_{1+\alpha,\overline{\Omega}}\leq C_{0}|g+1|_{1+\alpha,\overline{\Omega}}|g-1|_{1+\alpha,\overline{\Omega}}\leq\\\ C_{0}\left(|g-1|_{1+\alpha,\overline{\Omega}}+2\right)|g-1|_{1+\alpha,\overline{\Omega}}$ for some constant $C_{0}>0$. Hence, if $|g-1|_{1+\alpha,\overline{\Omega}}\leq\sqrt{1+\frac{\delta_{1}}{C_{0}}}-1,$ we have that $|g^{2}-1|_{1+\alpha,\overline{\Omega}}\leq\delta_{1}$. Suppose that the above condition is satisfied. Then, $\Psi_{g}$ has a unique fixed point $\varphi\in X_{g}$, by Lemma 4.6. Let $v$ be the solution of (1.4)-(1.5) with $\psi=0$. Then $u=f+v$ is harmonic in $B$ and (4.21) $|\nabla u|^{2}=|\nabla f|^{2}+2\nabla f\cdot\nabla v+|\nabla v|^{2}=1+2\partial_{x_{N}}v+{\mathcal{T}}[\varphi]=\\\ 1+2\varphi+(g^{2}-1-2\Psi_{g}[\varphi])=g^{2}\ \mbox{ on }\ {\mathcal{S}}.$ This shows that $u$ is a solution of (1.1). Moreover, Theorem 3.1 and the fact that $\varphi\in X_{g}$ give that $u$ is in $C_{\rm odd}^{1+\alpha}(\overline{B})$ and $|u-f|_{1+\alpha,\overline{B}}=|v|_{1+\alpha,\overline{B}}\leq C\,|\varphi|_{1+\alpha,\overline{B}}\leq C\,|g^{2}-1|_{1+\alpha,\overline{B}}\leq C\,|g-1|_{1+\alpha,\overline{B}},$ where $C>0$ is a constant. The assertion (i) thus holds if we take $\delta_{0}$ such that $\delta_{0}\leq\sqrt{1+\delta_{1}/C_{0}}-1$. For (ii), we may assume $h=0$, by considering $\tilde{u}=u-h$ instead of $u$. Then, (ii) can be shown in a similar way. We see from (4.20) that $|g^{2}-1|_{1+\alpha,\overline{\Omega}}\leq\delta_{2}$ if $|g-1|_{1+\alpha,\overline{\Omega}}\leq\sqrt{1+\frac{\delta_{2}}{C_{0}}}-1.$ Under this condition, Lemma 4.6 shows that $\tilde{\Psi}_{g,0}$ has a fixed point $\tilde{\varphi}\in\tilde{X}_{g,0}$. We put $\tilde{u}=f+\tilde{v}$ for the solution $\tilde{v}$ of (1.4)-(1.5) with $\varphi=\tilde{\varphi}$ and $\psi=0$. Then $\tilde{u}$ solves (1.1), since the same computation as in (4.21) is valid thanks to (4.12). The fact that $\tilde{u}\in C_{\rm ax}^{1+\alpha}(\overline{B})$ and the inequality $|\tilde{u}-f|_{1+\alpha,\overline{B}}\leq C\,|g-1|_{1+\alpha,\overline{B}}$ follow from Theorem 3.1. We have thus shown (ii), and the proof is complete. ∎ ## Acknowledgements The authors would like to thank the anonymous referee for useful suggestions for improving the readability of the presentation. The first author was partially supported by the Grant-in-Aid for Early-Career Scientists 19K14574, Japan Society for the Promotion of Science. The second author was partially supported by the Gruppo Nazionale per l’Analisi Matematica, la Probabilità e le loro Applicazioni (GNAMPA) dell’Istituto Nazionale di Alta Matematica (INdAM). The third author was supported in part by the Grant-in-Aid for Scientific Research (C) 20K03673, Japan Society for the Promotion of Science. This research started while the second author was visiting the Department of Mathematics of Tokyo Institute of Technology. He wants to thank their kind hospitality. ## References * [1] Sh. A. Alimov, _On a problem with an oblique derivative_ , (Russian) Differentsial’nye Uravneniya 17 (1981), no. 10, 1738–1751, 1915. English translation: Differential Equations 17 (1981), no. 10, 1073–1083 (1982). * [2] G. E. Backus, _Application of a non-linear boundary value problem for the Laplace’s equation to gravity and geomagnetic surveys_ , Quart. J. Mech. Appl. Math. XXI (1968), 195–221. * [3] D. Colton, R. Kress, Integral equation methods in scattering theory. John Wiley & Sons, Inc., New York, 1983. * [4] G. Díaz, J. J. Díaz, J. Otero, _Construction of the maximal solution of Backus’ problem in geodesy and geomagnetism_ , Stud. Geophys. Geod. 55 (2011), 415-440. * [5] D. Gilbarg, N. S. Trudinger, Elliptic partial differential equations of second order. Reprint of the 1998 edition, Springer-Verlag, Berlin, 2001. * [6] M. Hämäläinen, R. Hari, R. J. Ilmoniemi, J. Knuutila, O. V. Lounasmaa, _Magnetoencephalography – theory, instrumentation, and applications to noninvasinve studies of the working human brain_ , Review of Modern Physics, 65 (1993), 413–497. * [7] R. S. Hamilton, _The inverse function theorem of Nash and Moser_ , Bull. Amer. Math. Soc. 7 (1982), 65–222. * [8] M. C. Jorge, _Local existence of the solution to a nonlinear inverse problem in gravitation_ , Quart. Appl. Math. 45 (1987), 287–292. * [9] M. C. Jorge, R. Magnanini, _Explicit calculation of the solution to Backus problem with condition for uniqueness_ , J. Math. Anal. Appl. 173 (1993), 515–522. * [10] T. Kan, R. Magnanini, M. Onodera, _Backus problem in geophysics: a resolution near the dipole in fractional Sobolev spaces_ , NoDEA Nonlinear Differential Equations Appl. 29 (2022), no. 3, Paper No. 21, 29 pp. * [11] E. H. Lieb, M. Loss, Analysis. American Mathematical Society, Providence, RI, 1997. * [12] R. Magnanini, _A fully nonlinear boundary value problem for the Laplace equation in dimension two_ , Appl. Anal. 39 (2-3), 185–192. * [13] J. Moser, _A new technique for the construction of solutions of nonlinear differential equations_ , Proc. N. A. S. 47 (1961), 1824–1831. * [14] J. Moser, _A rapidly convergent iteration method and non-linear partial differential equations – I_ , Ann. Sc. Norm. Sup. Pisa (3) 20 (1966), 499–435. * [15] J. Nash, _$C^{1}$ isometric imbeddings_, Ann. of Math. 60 (1954), 383–396. * [16] W. Rudin, Real and Complex Analysis. Second edition. McGraw-Hill Book Co., New York-Düsseldorf-Johannesburg, 1974.
# Conservative algebras of $2$-dimensional algebras, IV Amir Fernández Ouaridi A. Fernández Ouaridi: CMUC, Universidade de Coimbra, Coimbra, Portugal; Universidad de Cádiz, Cádiz, Spain. <EMAIL_ADDRESS>, Ivan Kaygorodov I. Kaygorodov: CMA-UBI, Universidade da Beira Interior, Covilhã, Portugal; Moscow Center for Fundamental and Applied Mathematics, Moscow, Russia; Saint Petersburg State University, Russia<EMAIL_ADDRESS>and Cándido Martín González C. Martín: Departamento de Álgebra Geometría y Topología, Facultad de Ciencias, Universidad de Málaga, Campus de Teatinos s/n. 29071 Málaga. Spain. <EMAIL_ADDRESS> ###### Abstract. The notion of conservative algebras appeared in a paper of Kantor in 1972. Later, he defined the conservative algebra $W(n)$ of all algebras (i.e. bilinear maps) on the $n$-dimensional vector space. If $n>1$, then the algebra $W(n)$ does not belong to any well-known class of algebras (such as associative, Lie, Jordan, or Leibniz algebras). It looks like that $W(n)$ in the theory of conservative algebras plays a similar role with the role of $\mathfrak{gl}_{n}$ in the theory of Lie algebras. Namely, an arbitrary conservative algebra can be obtained from a universal algebra $W(n)$ for some $n\in\mathbb{N}.$ The present paper is a part of a series of papers, which dedicated to the study of the algebra $W(2)$ and its principal subalgebras. Keywords: bilinear maps, conservative algebra, contraction, identities. MSC2020: 17A30111Corresponding author: Ivan Kaygorodov <EMAIL_ADDRESS> ## Introduction A multiplication on a vector space $W$ is a bilinear mapping $W\times W\to W$. We denote by $(W,P)$ the algebra with underlining space $W$ and multiplication $P$. Given a vector space $W$, a linear mapping ${\bf A}:W\rightarrow W$, and a bilinear mapping ${\bf B}:W\times W\to W$, we can define a multiplication $[{\bf A},{\bf B}]:W\times W\to W$ by the formula $[{\bf A},{\bf B}](x,y)={\bf A}({\bf B}(x,y))-{\bf B}({\bf A}(x),y)-{\bf B}(x,{\bf A}(y))$ for $x,y\in W$. For an algebra ${\bf A}$ with a multiplication $P$ and $x\in{\bf A}$ we denote by $L_{x}^{P}$ the operator of left multiplication by $x$. If the multiplication $P$ is fixed, we write $L_{x}$ instead of $L_{x}^{P}$. In 1990 Kantor [14] defined the multiplication $\cdot$ on the set of all algebras (i.e. all multiplications) on the $n$-dimensional vector space $V_{n}$ as follows: ${\bf A}\cdot{\bf B}=[L_{e}^{\bf A},{\bf B}],$ where ${\bf A}$ and ${\bf B}$ are multiplications and $e\in V_{n}$ is some fixed vector. Let $W(n)$ denote the algebra of all algebra structures on $V_{n}$ with multiplication defined above. If $n>1$, then the algebra $W(n)$ does not belong to any well-known class of algebras (such as associative, Lie, Jordan, or Leibniz algebras). The algebra $W(n)$ turns out to be a conservative algebra (see below). In 1972 Kantor [11] introduced conservative algebras as a generalization of Jordan algebras (also, see a good written survey about the study of conservative algebras and superalgebras [25]). Namely, an algebra ${\bf A}=(W,P)$ is called a conservative algebra if there is a new multiplication $F:W\times W\rightarrow W$ such that (1) $\displaystyle[L_{b}^{P},[L_{a}^{P},P]]=-[L_{F(a,b)}^{P},P]$ for all $a,b\in W$. In other words, the following identity holds for all $a,b,x,y\in W$: (2) $b(a(xy)-(ax)y-x(ay))-a((bx)y)+(a(bx))y+(bx)(ay)\\\ -a(x(by))+(ax)(by)+x(a(by))=-F(a,b)(xy)+(F(a,b)x)y+x(F(a,b)y).$ The algebra $(W,F)$ is called an algebra associated to ${\bf A}$. The main subclass of conservative algebras is the variety of terminal algebras, which defined by the identity (2) with $F(a,b)=\frac{1}{3}(2ab+ba).$ It includes the varieties of Leibniz and Jordan algebras as subvarieties. Let us recall some well-known results about conservative algebras. In [11] Kantor classified all simple conservative algebras and triple systems of second-order and defined the class of terminal algebras as algebras satisfying some certain identity. He proved that every terminal algebra is a conservative algebra and classified all simple finite-dimensional terminal algebras with left quasi-unit over an algebraically closed field of characteristic zero [12]. Terminal trilinear operations were studied in [13]. After that, Cantarini and Kac classified simple finite-dimensional (and linearly compact) super-commutative and super-anticommutative conservative superalgebras and some generalization of these algebras (also known as “rigid” or quasi- conservative superalgebras) over an algebraically closed field of characteristic zero [3]. The classification of all $2$-dimensional conservative and rigid (in sense of Kac-Cantarini) algebras is given in [2]; and also, the algebraic and geometric classification of nilpotent low dimensional terminal algebras is given in [16, 17]. The algebra $W(n)$ plays a similar role in the theory of conservative algebras as the Lie algebra of all $n\times n$ matrices $\mathfrak{gl}_{n}$ plays in the theory of Lie algebras. Namely, in [14] Kantor considered the category $\mathscr{S}_{n}$ whose objects are conservative algebras of non-Jacobi dimension $n$. It was proven that the algebra $W(n)$ is the universal attracting object in this category, i.e., for every $M\in\mathscr{S}_{n}$ there exists a canonical homomorphism from $M$ into the algebra $W(n)$. In particular, all Jordan algebras of dimension $n$ with unity are contained in the algebra $W(n)$. The same statement also holds for all noncommutative Jordan algebras of dimension $n$ with unity. Some properties of the product in the algebra $W(n)$ were studied in [5, 15]. The universal conservative superalgebra was constructed in [20]. The study of low dimensional conservative algebras was started in [18]. The study of properties of $2$-dimensional algebras is also one of popular topic in non-associative algebras (see, for example, [6, 22, 23, 4, 24]) and as we can see the study of properties of the algebra $W(2)$ could give some applications on the theory of $2$-dimensional algebras. So, from the description of idempotents of the algebra $W(2)$ it was received an algebraic classification of all $2$-dimensional algebras with left quasi-unit [21]. Derivations and subalgebras of codimension 1 of the algebra $W(2)$ and of its principal subalgebras $W_{2}$ and $S_{2}$ were described [18]. Later, the automorphisms, one-sided ideals, idempotents, local (and $2$-local) derivations and automorphisms of $W(2)$ and its principal subalgebras were described in [1, 21]. Note that $W_{2}$ and $S_{2}$ are simple terminal algebras with left quasi-unit from the classification of Kantor [12]. The present paper is devoted to continuing the study of properties of $W(2)$ and its principal subalgebras. Throughout this paper, unless stated otherwise, $\mathbb{F}$ denotes a field of characteristic zero. All algebras are defined over $\mathbb{F}$. The multiplication table of $W(2)$ is given by the following table. | $e_{1}$ | $e_{2}$ | $e_{3}$ | $e_{4}$ | $e_{5}$ | $e_{6}$ | $e_{7}$ | $e_{8}$ ---|---|---|---|---|---|---|---|--- $e_{1}$ | $-e_{1}$ | $-3e_{2}$ | $e_{3}$ | $3e_{4}$ | $-e_{5}$ | $e_{6}$ | $e_{7}$ | $-e_{8}$ $e_{2}$ | $3e_{2}$ | $0$ | $2e_{1}$ | $e_{3}$ | $0$ | $-e_{5}$ | $e_{8}$ | $0$ $e_{3}$ | $-2e_{3}$ | $-e_{1}$ | $-3e_{4}$ | $0$ | $e_{6}$ | $0$ | $0$ | $-e_{7}$ $e_{4}$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ $e_{5}$ | $-2e_{1}$ | $-3e_{2}$ | $-e_{3}$ | $0$ | $-2e_{5}$ | $-e_{6}$ | $-e_{7}$ | $-2e_{8}$ $e_{6}$ | $2e_{3}$ | $e_{1}$ | $3e_{4}$ | $0$ | $-e_{6}$ | $0$ | $0$ | $e_{7}$ $e_{7}$ | $2e_{3}$ | $e_{1}$ | $3e_{4}$ | $0$ | $-e_{6}$ | $0$ | $0$ | $e_{7}$ $e_{8}$ | $0$ | $e_{2}$ | $-e_{3}$ | $-2e_{4}$ | $0$ | $-e_{6}$ | $-e_{7}$ | $0$ ## 1\. Inönü-Wigner contractions of $W(2)$ and its subalgebras The class of conservative algebras includes the variety of terminal algebras, which includes all Leibniz and Jordan algebras. On the other hand, the variety of terminal algebras is ”dual” to the variety of commutative algebras (in the sense of generalized TTK-functor). The algebra $W(2)$ is not terminal, but its principal subalgebras $W_{2}$ and $S_{2}$ are terminal. Our main aim is to try to understand how the algebra $W(2)$ ”far” from terminal algebras. For a particular answer for our question, we will consider contractions to some certain subalgebras of $W(2)$ and study its relations with the variety of terminal algebras. The standard Inönü-Wigner contraction was introduced in [8]. We will call it IW contraction for short. ###### Definition 1. Let $\mu,\chi$ represent algebras ${\bf A}$ and ${\bf B}$ respectively defined on a vector space $V$. Suppose that there are some elements $E_{i}^{t}\in V$ $(1\leq i\leq n$, $t\in{\mathbb{F}}^{*})$ such that $E^{t}=(E_{1}^{t},\dots,E_{n}^{t})$ is a basis of $V$ for any $t\in{\mathbb{F}}^{*}$ and the structure constants of $\mu$ in this basis are $\mu_{i,j}^{k}(t)$ for some polynomials $\mu_{i,j}^{k}(t)\in{\mathbb{F}}[t]$. If $\mu_{i,j}^{k}(0)=\chi_{i,j}^{k}$ for all $1\leq i,j,k\leq n$, then ${\bf A}\to{\bf B}$. To emphasize that the parametrized basis $E^{t}=(E_{1}^{t},\dots,E_{n}^{t})$ $(t\in{\mathbb{F}}^{*})$ gives a degeneration between the algebras represented by the structures $\mu$ and $\chi$, we will write $\mu\xrightarrow{E^{t}}\chi$. Suppose that ${\bf A}_{0}$ is an $(n-m)$-dimensional subalgebra of the $n$-dimensional algebra ${\bf A}$ and $\mu$ is a structure representing ${\bf A}$ such that ${\bf A}_{0}$ corresponds to the subspace $\langle e_{m+1},\dots,e_{n}\rangle$ of $V$. Then $\mu\xrightarrow{(te_{1},\dots,te_{m},e_{m+1},\dots,e_{n})}\chi$ for some $\chi$ and the algebra ${\bf B}$ represented by $\chi$ is called the IW contraction of ${\bf A}$ with respect to ${\bf A}_{0}$. ### 1.1. IW contraction of $W(2)$ #### 1.1.1. The algebra $\overline{W(2)}$ The description of all subalgebras of codimension $1$ for the algebra $W(2)$ is given in [18]. Namely, $W(2)$ has only one $7$-dimensional subalgebra. It is generated by elements $e_{1},e_{3},e_{4},e_{5},e_{6},e_{7},e_{8},$ and it is terminal. Let us consider the IW contraction $W(2)\xrightarrow{(e_{1},te_{2},e_{3},e_{4},e_{5},e_{6},e_{7},e_{8})}\overline{W(2)}.$ It is easy to see, that the multiplication table of $\overline{W(2)}$ is given by the following table. | $e_{1}$ | $e_{2}$ | $e_{3}$ | $e_{4}$ | $e_{5}$ | $e_{6}$ | $e_{7}$ | $e_{8}$ ---|---|---|---|---|---|---|---|--- $e_{1}$ | $-e_{1}$ | $-3e_{2}$ | $e_{3}$ | $3e_{4}$ | $-e_{5}$ | $e_{6}$ | $e_{7}$ | $-e_{8}$ $e_{2}$ | $3e_{2}$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ $e_{3}$ | $-2e_{3}$ | $0$ | $-3e_{4}$ | $0$ | $e_{6}$ | $0$ | $0$ | $-e_{7}$ $e_{4}$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ $e_{5}$ | $-2e_{1}$ | $-3e_{2}$ | $-e_{3}$ | $0$ | $-2e_{5}$ | $-e_{6}$ | $-e_{7}$ | $-2e_{8}$ $e_{6}$ | $2e_{3}$ | $0$ | $3e_{4}$ | $0$ | $-e_{6}$ | $0$ | $0$ | $e_{7}$ $e_{7}$ | $2e_{3}$ | $0$ | $3e_{4}$ | $0$ | $-e_{6}$ | $0$ | $0$ | $e_{7}$ $e_{8}$ | $0$ | $e_{2}$ | $-e_{3}$ | $-2e_{4}$ | $0$ | $-e_{6}$ | $-e_{7}$ | $0$ After a carefully checking of the dimension of the algebra of derivation of $\overline{W(2)},$ we have $\dim\mathfrak{Der}(\overline{W(2)})=3.$ Since $\dim\mathfrak{Der}(W(2))=2,$ it follows that the degeneration $W(2)\to\overline{W(2)}$ is primary, that is, there is no algebra ${\bf A}$ such that ${W(2)}\to{\bf A}$ and ${\bf A}\to\overline{W(2)}$, where ${\bf A}$ is neither isomorphic to $W(2)$ or $\overline{W(2)}$ (see [9]). ###### Lemma 2. The algebra $\overline{W(2)}$ is a non-terminal conservative non-simple algebra. ###### Proof. The subspace $\langle e_{2},e_{3},e_{4},e_{6},e_{7},e_{8}\rangle$ gives a $6$-dimensional ideal, it gives that $\overline{W(2)}$ is non-simple. The non- terminal property is following from the direct verification of the terminal identity (for example, using a modification of the Wolfram code presented in [10]). The conservative property is following from the direct verification of the conservative identity with the additional multiplication $*:$ $e_{1}*e_{1}=-e_{1}$ | $e_{1}*e_{2}=-e_{2}$ | $e_{1}*e_{5}=-2e_{1}$ | $e_{2}*e_{1}=e_{2}$ | $e_{2}*e_{5}=-e_{2}$ ---|---|---|---|--- $e_{2}*e_{8}=-e_{2}$ | $e_{5}*e_{1}=-2e_{1}$ | $e_{5}*e_{2}=-2e_{2}$ | $e_{5}*e_{5}=-4e_{1}.$ | ∎ ###### Lemma 3. Let $\mathscr{S}$ be a subalgebra of $\overline{W(2)}$ of codimension $1,$ the $\mathscr{S}$ is one of the following conservative subalgebras ${\mathscr{S}}_{1}=\langle e_{2},e_{3},e_{4},e_{5},e_{6},e_{7},e_{8}\rangle$ | ${\mathscr{S}}_{2}=\langle e_{1},e_{3},e_{4},e_{5},e_{6},e_{7},e_{8}\rangle$ ---|--- ${\mathscr{S}}_{5}=\langle e_{1},e_{2},e_{3},e_{4},e_{6},e_{7},e_{8}\rangle$ | ${\mathscr{S}}_{\alpha,\beta}=\langle e_{1}+\alpha e_{8},e_{2},e_{3},e_{4},e_{5}+\beta e_{8},e_{6},e_{7}\rangle_{\alpha,\beta\in{\mathbb{F}}},$ where only ${\mathscr{S}}_{1},$ ${\mathscr{S}}_{2}$, ${\mathscr{S}}_{0,0}$ and ${\mathscr{S}}_{-1,1}$ are terminal. ###### Proof. Let $\mathscr{S}$ be generated by the following set $\\{e_{1},\ldots,\widehat{e_{i}},\ldots,e_{8}\\}.$ By some easy verification of $8$ possibilities, we have that there are only $4$ subalgebras of this type: for $i=1,2,5,8.$ Let us consider the situation when $\mathscr{S}$ is generated by seven vectors of the following type: $\\{\sum\alpha_{i1}e_{i},\ldots,\sum\alpha_{i7}e_{i}\\}.$ By some linear combinations, we can reduce this basis to a basis considered above, or a basis of the following type: $\\{e_{1}+\alpha_{1}e_{8},\ldots,e_{7}+\alpha_{1}e_{8}\\}.$ It is easy to see, that $(e_{2}+\alpha_{2}e_{8})^{2}=\alpha_{2}e_{2}\in{\mathscr{S}}$ | $(e_{3}+\alpha_{3}e_{8})^{2}=-3e_{4}-\alpha_{3}(e_{3}+e_{7})\in{\mathscr{S}}$ ---|--- $(e_{4}+\alpha_{4}e_{8})^{2}=-2\alpha_{4}e_{4}\in{\mathscr{S}}$ | $(e_{6}+\alpha_{6}e_{8})^{2}=-\alpha_{6}(e_{6}-e_{7})\in{\mathscr{S}},$ which gives that $e_{2},e_{4}\in{\mathscr{S}}$ and there are four cases: I. $e_{3},e_{6}\in{\mathscr{S}}$ | II. $e_{3},e_{6}-e_{7}\in{\mathscr{S}}$ ---|--- III. $e_{3}+e_{7},e_{6}\in{\mathscr{S}}$ | IV. $e_{3}+e_{7},e_{6}-e_{7}\in{\mathscr{S}}$ Analysing all these cases, we have that ${\mathscr{S}}$ is a subalgebra considered above, or it has the following basis $\langle e_{1}+\alpha e_{8},e_{2},e_{3},e_{4},e_{5}+\beta e_{8},e_{6},e_{7}\rangle_{\alpha,\beta\in{\mathbb{F}}}.$ The conservative property of the subalgebra ${\mathscr{S}}_{5}$ is following from the direct verification of the conservative identity with the additional multiplication $*:$ $e_{1}*e_{1}=-e_{1}$ | $e_{1}*e_{2}=-e_{2}$ | $e_{2}*e_{1}=e_{2}$ | $e_{2}*e_{8}=-e_{2}$ ---|---|---|--- Let us give the multiplication table of $\overline{W(2)}$ in more useful way (here the subalgebra $\langle e_{1},\ldots,e_{7}\rangle$ gives ${\mathscr{S}}_{\alpha,\beta}$): $e_{1}e_{1}=-e_{1}$ | $e_{1}e_{2}=(-3+\alpha)e_{2}$ | $e_{1}e_{3}=(1-\alpha)e_{3}$ | $e_{1}e_{4}=(3-2\alpha)e_{4}$ ---|---|---|--- $e_{1}e_{5}=-e_{5}$ | $e_{1}e_{6}=(1-\alpha)e_{6}$ | $e_{1}e_{7}=(1-\alpha)e_{7}$ | $e_{1}e_{8}=-e_{8}$ $e_{2}e_{1}=3e_{2}$ | $e_{3}e_{1}=-2e_{3}-\alpha e_{7}$ | $e_{3}e_{3}=-3e_{4}$ | $e_{3}e_{5}=e_{6}-\beta e_{7}$ $e_{3}e_{8}=-e_{7}$ | $e_{5}e_{1}=-2e_{1}$ | $e_{5}e_{2}=(-3+\beta)e_{2}$ | $e_{5}e_{3}=(-1-\beta)e_{3}$ $e_{5}e_{4}=-2\beta e_{4}$ | $e_{5}e_{5}=-2e_{5}$ | $e_{5}e_{6}=(-1-\beta)e_{6}$ | $e_{5}e_{7}=(-1-\beta)e_{7}$ $e_{5}e_{8}=-2e_{8}$ | $e_{6}e_{1}=2e_{3}+\alpha e_{7}$ | $e_{6}e_{3}=3e_{4}$ | $e_{6}e_{5}=-e_{6}+\beta e_{7}$ $e_{6}e_{8}=e_{7}$ | $e_{7}e_{1}=2e_{3}+\alpha e_{7}$ | $e_{7}e_{3}=3e_{4}$ | $e_{7}e_{5}=-e_{6}+\beta e_{7}$ $e_{7}e_{8}=e_{7}$ | $e_{8}e_{2}=e_{2}$ | $e_{8}e_{3}=-e_{3}$ | $e_{8}e_{4}=-2e_{4}$ | $e_{8}e_{6}=-e_{6}$ | $e_{8}e_{7}=-e_{7}$ | The conservative property of the subalgebra ${\mathscr{S}}_{\alpha,\beta}$ is following from the direct verification of the conservative identity with the additional multiplication $*:$ $e_{1}*e_{1}=-e_{1}$ | $e_{1}*e_{2}=-e_{2}$ | $e_{1}*e_{5}=-2e_{1}$ | $e_{2}*e_{1}=(1-\alpha)e_{2}$ ---|---|---|--- $\par e_{2}*e_{5}=(-1-\beta)e_{2}$ | $e_{5}*e_{1}=-2e_{1}$ | $e_{5}*e_{2}=-2e_{2}$ | $e_{5}*e_{5}=-4e_{1}$ ∎ #### 1.1.2. Algebras $\overline{{\mathscr{S}}}$ In the present subsection, we have to talk about contractions of the algebra $\overline{W(2)}$ to its subalgebra of codimension $1.$ $\bullet$ $\overline{W(2)}\xrightarrow{(te_{1},{\mathscr{S}}_{1})}\overline{{\mathscr{S}}_{1}}.$ It is easy to see that the multiplication of $\overline{{\mathscr{S}}_{1}}$ is given by the following table. $e_{3}e_{3}=-3e_{4}$ | $e_{3}e_{5}=e_{6}$ | $e_{3}e_{8}=-e_{7}$ | $e_{5}e_{1}=-2e_{1}$ | $e_{5}e_{2}=-3e_{2}$ | $e_{5}e_{3}=-e_{3}$ | $e_{5}e_{5}=-2e_{5}$ ---|---|---|---|---|---|--- $e_{5}e_{6}=-e_{6}$ | $e_{5}e_{7}=-e_{7}$ | $e_{5}e_{8}=-2e_{8}$ | $e_{6}e_{3}=3e_{4}$ | $e_{6}e_{5}=-e_{6}$ | $e_{6}e_{8}=e_{7}$ | $e_{7}e_{3}=3e_{4}$ $e_{7}e_{5}=-e_{6}$ | $e_{7}e_{8}=e_{7}$ | $e_{8}e_{2}=e_{2}$ | $e_{8}e_{3}=-e_{3}$ | $e_{8}e_{4}=-2e_{4}$ | $e_{8}e_{6}=-e_{6}$ | $e_{8}e_{7}=-e_{7}$ ###### Lemma 4. The algebra $\overline{{\mathscr{S}}_{1}}$ is terminal. $\bullet$ $\overline{W(2)}\xrightarrow{(te_{5},{\mathscr{S}}_{5})}\overline{{\mathscr{S}}_{5}}.$ It is easy to see that the multiplication of $\overline{{\mathscr{S}}_{5}}$ is given by the following table. $e_{1}e_{1}=-e_{1}$ | $e_{1}e_{2}=-3e_{2}$ | $e_{1}e_{3}=e_{3}$ | $e_{1}e_{4}=3e_{4}$ | $e_{1}e_{5}=-e_{5}$ | $e_{1}e_{6}=e_{6}$ ---|---|---|---|---|--- $e_{1}e_{7}=e_{7}$ | $e_{1}e_{8}=-e_{8}$ | $e_{2}e_{1}=3e_{2}$ | $e_{3}e_{1}=-2e_{3}$ | $e_{3}e_{3}=-3e_{4}$ | $e_{3}e_{8}=-e_{7}$ $\par e_{6}e_{1}=2e_{3}$ | $e_{6}e_{3}=3e_{4}$ | $e_{6}e_{8}=e_{7}$ | $e_{7}e_{1}=2e_{3}$ | $e_{7}e_{3}=3e_{4}$ | $e_{7}e_{8}=e_{7}$ $\par e_{8}e_{2}=e_{2}$ | $e_{8}e_{3}=-e_{3}$ | $e_{8}e_{4}=-2e_{4}$ | $e_{8}e_{6}=-e_{6}$ | $e_{8}e_{7}=-e_{7}$ | ###### Lemma 5. The algebra $\overline{{\mathscr{S}}_{5}}$ is a non-terminal conservative algebra. ###### Proof. The conservative property of the algebra $\overline{{\mathscr{S}}_{5}}$ is following from the direct verification of the conservative identity with the additional multiplication $*:$ $e_{1}*e_{1}=-e_{1}$ | $e_{1}*e_{2}=-e_{2}$ | $e_{2}*e_{1}=e_{2}$ | $e_{2}*e_{8}=-e_{2}$ ---|---|---|--- ∎ $\bullet$ $\overline{W(2)}\xrightarrow{(te_{8},{\mathscr{S}}_{\alpha,\beta})}\overline{{\mathscr{S}}_{\alpha,\beta}}.$ It is easy to see that the multiplication of $\overline{{\mathscr{S}}_{\alpha,\beta}}$ is given by the following table. $e_{1}e_{1}=-e_{1}$ | $e_{1}e_{2}=(-3+\alpha)e_{2}$ | $e_{1}e_{3}=(1-\alpha)e_{3}$ | $e_{1}e_{4}=(3-2\alpha)e_{4}$ ---|---|---|--- $\par e_{1}e_{5}=-e_{5}$ | $e_{1}e_{6}=(1-\alpha)e_{6}$ | $e_{1}e_{7}=(1-\alpha)e_{7}$ | $e_{1}e_{8}=-e_{8}$ $\par e_{2}e_{1}=3e_{2}$ | $e_{3}e_{1}=-2e_{3}-\alpha e_{7}$ | $e_{3}e_{3}=-3e_{4}$ | $e_{3}e_{5}=e_{6}-\beta e_{7}$ $\par e_{5}e_{1}=-2e_{1}$ | $e_{5}e_{2}=(-3+\beta)e_{2}$ | $e_{5}e_{3}=(-1-\beta)e_{3}$ | $e_{5}e_{4}=-2\beta e_{4}$ $\par e_{5}e_{5}=-2e_{5}$ | $e_{5}e_{6}=(-1-\beta)e_{6}$ | $e_{5}e_{7}=(-1-\beta)e_{7}$ | $e_{5}e_{8}=-2e_{8}$ $\par e_{6}e_{1}=2e_{3}+\alpha e_{7}$ | $e_{6}e_{3}=3e_{4}$ | $e_{6}e_{5}=-e_{6}+\beta e_{7}$ | $e_{7}e_{1}=2e_{3}+\alpha e_{7}$ | $\par e_{7}e_{3}=3e_{4}$ | $e_{7}e_{5}=-e_{6}+\beta e_{7}$ | ###### Lemma 6. The algebra $\overline{{\mathscr{S}}_{\alpha,\beta}}$ is a conservative algebra; and it is a terminal algebra if and only if $(\alpha,\beta)=(-1,1)$ or $(\alpha,\beta)=(0,0)$. ###### Proof. The conservative property of the algebra $\overline{{\mathscr{S}}_{\alpha,\beta}}$ is following from the direct verification of the conservative identity with the additional multiplication $*:$ $e_{1}*e_{1}=-e_{1}$ | $e_{1}*e_{2}=-e_{2}$ | $e_{1}*e_{5}=-2e_{1}$ | $e_{2}*e_{1}=(1-\alpha)e_{2}$ ---|---|---|--- $\par e_{2}*e_{5}=(-1-\beta)e_{2}$ | $e_{5}*e_{1}=-2e_{1}$ | $e_{5}*e_{2}=-2e_{2}$ | $e_{5}*e_{5}=-4e_{1}$ ∎ #### 1.1.3. The algebra $\widehat{W(2)}$ The second interesting ”big” subalgebra of $W(2)$ is $W_{2},$ which is generated by $e_{1},\ldots,e_{6}.$ Let us consider the IW contraction $W(2)\xrightarrow{(e_{1},e_{2},e_{3},e_{4},e_{5},e_{6},te_{7},te_{8})}{\widehat{W(2)}}.$ It is easy to see, that the multiplication table (for nonzero products) of ${\widehat{W(2)}}$ is given by the following table. | $e_{1}$ | $e_{2}$ | $e_{3}$ | $e_{4}$ | $e_{5}$ | $e_{6}$ | $e_{7}$ | $e_{8}$ ---|---|---|---|---|---|---|---|--- $e_{1}$ | $-e_{1}$ | $-3e_{2}$ | $e_{3}$ | $3e_{4}$ | $-e_{5}$ | $e_{6}$ | $e_{7}$ | $-e_{8}$ $e_{2}$ | $3e_{2}$ | $0$ | $2e_{1}$ | $e_{3}$ | $0$ | $-e_{5}$ | $e_{8}$ | $0$ $e_{3}$ | $-2e_{3}$ | $-e_{1}$ | $-3e_{4}$ | $0$ | $e_{6}$ | $0$ | $0$ | $-e_{7}$ $e_{4}$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ $e_{5}$ | $-2e_{1}$ | $-3e_{2}$ | $-e_{3}$ | $0$ | $-2e_{5}$ | $-e_{6}$ | $-e_{7}$ | $-2e_{8}$ $e_{6}$ | $2e_{3}$ | $e_{1}$ | $3e_{4}$ | $0$ | $-e_{6}$ | $0$ | $0$ | $e_{7}$ ###### Lemma 7. The algebra ${\widehat{W(2)}}$ is terminal. #### 1.1.4. The algebra $\widehat{\widehat{W(2)}}$ The next interesting subalgebra of $W(2)$ is $S_{2},$ which is generated by $e_{1},\ldots,e_{4}.$ Let us consider the IW contraction $W(2)\xrightarrow{(e_{1},e_{2},e_{3},e_{4},te_{5},te_{6},te_{7},te_{8})}{\widehat{\widehat{W(2)}}}.$ It is easy to see, that the multiplication table (for nonzero products) of ${\widehat{\widehat{W(2)}}}$ is given by the following table. | $e_{1}$ | $e_{2}$ | $e_{3}$ | $e_{4}$ | $e_{5}$ | $e_{6}$ | $e_{7}$ | $e_{8}$ ---|---|---|---|---|---|---|---|--- $e_{1}$ | $-e_{1}$ | $-3e_{2}$ | $e_{3}$ | $3e_{4}$ | $-e_{5}$ | $e_{6}$ | $e_{7}$ | $-e_{8}$ $e_{2}$ | $3e_{2}$ | $0$ | $2e_{1}$ | $e_{3}$ | $0$ | $-e_{5}$ | $e_{8}$ | $0$ $e_{3}$ | $-2e_{3}$ | $-e_{1}$ | $-3e_{4}$ | $0$ | $e_{6}$ | $0$ | $0$ | $-e_{7}$ ###### Corollary 8. The algebra ${\widehat{\widehat{W(2)}}}$ is terminal. #### 1.1.5. The algebra $\widetilde{W(2)}$ The next interesting subalgebra of $W(2)$ is generated by $e_{1}$ and $e_{2}.$ Let us consider the IW contraction $W(2)\xrightarrow{(e_{1},e_{2},te_{3},te_{4},te_{5},te_{6},te_{7},te_{8})}{\widetilde{W(2)}}.$ It is easy to see, that the multiplication table (for nonzero products) of ${\widetilde{W(2)}}$ is given by the following table. | $e_{1}$ | $e_{2}$ | $e_{3}$ | $e_{4}$ | $e_{5}$ | $e_{6}$ | $e_{7}$ | $e_{8}$ ---|---|---|---|---|---|---|---|--- $e_{1}$ | $-e_{1}$ | $-3e_{2}$ | $e_{3}$ | $3e_{4}$ | $-e_{5}$ | $e_{6}$ | $e_{7}$ | $-e_{8}$ $e_{2}$ | $3e_{2}$ | $0$ | $0$ | $e_{3}$ | $0$ | $-e_{5}$ | $e_{8}$ | $0$ $e_{3}$ | $-2e_{3}$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ $e_{6}$ | $2e_{3}$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ $e_{7}$ | $2e_{3}$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ ###### Lemma 9. The algebra ${\widetilde{W(2)}}$ is a non-Leibniz, non-Jordan terminal algebra. #### 1.1.6. The algebra $\widetilde{\widetilde{W(2)}}$ The last interesting subalgebra of $W(2)$ is generated by $e_{1}.$ Let us consider the IW contraction $W(2)\xrightarrow{(e_{1},te_{2},te_{3},te_{4},te_{5},te_{6},te_{7},te_{8})}\widetilde{\widetilde{W(2)}}.$ It is easy to see, that the multiplication table (for nonzero products) of $\widetilde{\widetilde{W(2)}}$ is given by the following table. | $e_{1}$ | $e_{2}$ | $e_{3}$ | $e_{4}$ | $e_{5}$ | $e_{6}$ | $e_{7}$ | $e_{8}$ ---|---|---|---|---|---|---|---|--- $e_{1}$ | $-e_{1}$ | $-3e_{2}$ | $e_{3}$ | $3e_{4}$ | $-e_{5}$ | $e_{6}$ | $e_{7}$ | $-e_{8}$ $e_{2}$ | $3e_{2}$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ $e_{3}$ | $-2e_{3}$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ $e_{6}$ | $2e_{3}$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ $e_{7}$ | $2e_{3}$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ | $0$ ###### Lemma 10. The algebra $\widetilde{\widetilde{W(2)}}$ is a non-Leibniz, non-Jordan terminal algebra. ### 1.2. IW contraction of $S_{2}$ and $W_{2}$ Thanks to [18], algebras $S_{2}$ and $W_{2}$ have also only one subalgebra of codimension $1,$ which are $\langle e_{1},e_{3},e_{4}\rangle$ and $\langle e_{1},e_{3},e_{4},e_{5},e_{6}\rangle.$ Other important subalgebras of $S_{2}$ and $W_{2}$ are $\langle e_{1}\rangle,$ $\langle e_{1},e_{2}\rangle$ and $\langle e_{1},e_{2},e_{3},e_{4}\rangle.$ All contractions of $S_{2}$ (and $W_{2}$) with respected to all cited subalgebras can be obtained as $4$-dimensional subalgebras $\langle e_{1},e_{2},e_{3},e_{4}\rangle$ ($6$-dimensional subalgebras $\langle e_{1},\ldots,e_{6}\rangle$) of the following algebras $\overline{W(2)},$ ${\widehat{\widehat{W(2)}}},$ ${\widetilde{W(2)}}$ and $\widetilde{\widetilde{W(2)}}.$ All these subalgebras are non-Leibniz, non-Jordan terminal algebras. ## 2\. Varieties related to $W(2)$ and its subalgebras ### 2.1. Identities Let ${\mathbb{F}}$ be a field of characteristic zero and ${\mathbb{F}}\langle x_{1},x_{2},\ldots,x_{n}\rangle$ the free nonassociative ${\mathbb{F}}$-algebra in $n$ indeterminates. Let ${\bf A}$ be any algebra and ${\mathfrak{S}}^{n}_{\bf A}$ the subspace of ${\mathbb{F}}\langle x_{1},x_{2},\ldots,x_{n}\rangle$ of all $n$-linear $w(x_{1},x_{2},\ldots,x_{n})$ which vanish on ${\bf A}$ (so $w(x_{1},x_{2},\ldots,x_{n})$ is of degree one in each variable). We have studied by direct verification the subspace ${\mathfrak{S}}^{n}_{\bf A}$ for $n=3,4$ of $W(2)$ and of its IW contractions mentioned in this paper. ###### Proposition 11. In the following table, we summarize the dimension of the subspaces ${\mathfrak{S}}^{n}_{\bf A}$ for ${\bf A}\in\\{W(2),\overline{W(2)},\overline{\mathscr{S}_{1}},\overline{\mathscr{S}_{5}},\overline{\mathscr{S}_{-1,1}},\overline{\mathscr{S}_{0,0}},\widehat{W(2)},\widehat{\widehat{W(2)}},{\widetilde{W(2)}}\\}$ and $n=3,4$. Algebra (${\bf A}$) $\textrm{dim}({\mathfrak{S}}^{3}_{\bf A})$ $\textrm{dim}({\mathfrak{S}}^{4}_{\bf A})$ Comments $W(2)$ 0 0 non-terminal conservative $\overline{W(2)}$ 0 20 non-terminal, conservative $\overline{\mathscr{S}_{1}}$ 0 64 terminal $\overline{\mathscr{S}_{5}}$ 0 40 non-terminal, conservative $\overline{\mathscr{S}_{-1,1}}$ 0 64 terminal $\overline{\mathscr{S}_{0,0}}$ 0 44 terminal $\widehat{W(2)}$ 0 24 terminal $\widehat{\widehat{W(2)}}$ 0 47 terminal ${\widetilde{W(2)}}$ 0 82 terminal ${\widetilde{\widetilde{W(2)}}}$ 2 101 terminal Moreover, if ${\bf A}=\overline{\mathscr{S}_{\alpha,\beta}}$ then $\dim({\mathfrak{S}}^{3}_{\bf A})=0$ for $(\alpha,\beta)\neq(2,1)$ and $(\alpha,\beta)\neq(0,-3)$. ###### Proof. We have determined the spaces ${\mathfrak{S}}^{n}_{\bf A}$ for $n=3,4$ by constructing an arbitrary $n$-linear map $w(x_{1},\ldots,x_{n})$ and solving $w(x_{1},\ldots,x_{n})=0$ in ${\bf A}$ using Wolfram. ∎ ###### Proposition 12. If ${\bf A}=\overline{\mathscr{S}_{2,1}}$ then $\dim({\mathfrak{S}}^{3}_{\bf A})=3$ and a basis of the ${\mathbb{F}}$-vector space ${\mathfrak{S}}^{3}_{\bf A}$ is the set of identities: 1. (1) $x_{1}(x_{2}x_{3})-x_{2}(x_{1}x_{3}),$ 2. (2) $x_{2}(x_{3}x_{1})-x_{3}(x_{2}x_{1}),$ 3. (3) $x_{3}(x_{1}x_{2})-x_{1}(x_{3}x_{2}).$ Now, consider the following identities: $\mathfrak{st}^{n}_{1}=\sum\limits_{\sigma\in\mathbb{S}_{n}}(-1)^{\sigma}(\ldots(x_{\sigma(1)}x_{\sigma(2)})x_{\sigma(3)}\ldots)x_{\sigma(n)}\mbox{ and }\mathfrak{st}^{n}_{2}=\sum\limits_{\sigma\in\mathbb{S}_{n}}(-1)^{\sigma}x_{\sigma(n)}(\ldots x_{\sigma(3)}(x_{\sigma(2)}x_{\sigma(1)})\ldots).$ It is clear that if an algebra satisfies $\mathfrak{st}^{n}_{1}$ (resp. $\mathfrak{st}^{n}_{2}$) then it also satisfies $\mathfrak{st}^{n+1}_{1}$ (resp. $\mathfrak{st}^{n+1}_{2}$). ###### Proposition 13. If ${\bf A}=\overline{\mathscr{S}_{0,-3}}$ then $\dim({\mathfrak{S}}^{3}_{\bf A})=1$. This ${\mathbb{F}}$-vector space is generated by $2\mathfrak{st}^{3}_{1}-3\mathfrak{st}^{3}_{2}.$ ###### Proposition 14. If ${\bf A}=\widetilde{\widetilde{W(2)}}$ then $\dim({\mathfrak{S}}^{3}_{\bf A})=2$ and a basis of the ${\mathbb{F}}$-vector space ${\mathfrak{S}}^{3}_{\bf A}$ is the set of identities $\mathfrak{st}^{3}_{1},\mathfrak{st}^{3}_{2}$. We have studied the space ${\mathfrak{S}}^{n}_{\bf A}$ for the subalgebras of $W(2)$ mentioned in this paper. We have also studied the identities $\mathfrak{st}^{n}_{1}$ and $\mathfrak{st}^{n}_{2}$ for these subalgebras for $n=3,4,5$. ###### Proposition 15. In the following table, we summarize the dimension of the subspaces ${\mathfrak{S}}^{n}_{\bf A}$ for ${\bf A}$ a subalgebra of $W(2)$ and $n=3,4$. Subalgebra $({\bf A})$ $\textrm{dim}({\mathfrak{S}}^{3}_{\bf A})$ $\textrm{dim}({\mathfrak{S}}^{4}_{\bf A})$ Comments $B_{2}:=\langle e_{1},e_{3},e_{4},e_{5},e_{6},e_{7},e_{8}\rangle$ 0 64 no $\mathfrak{st}^{4}_{1}$, no $\mathfrak{st}^{4}_{2}$, $\mathfrak{st}^{5}_{1},\mathfrak{st}^{5}_{2}$ $W_{2}=\langle e_{1},e_{2},e_{3},e_{4},e_{5},e_{6}\rangle$ 0 24 no $\mathfrak{st}^{5}_{1}$, no $\mathfrak{st}^{4}_{2},$ $\mathfrak{st}^{5}_{2}$ $C_{2}:=\langle e_{1},e_{3},e_{4},e_{5},e_{6}\rangle$ 0 64 no $\mathfrak{st}^{4}_{1}$, no $\mathfrak{st}^{4}_{2}$, $\mathfrak{st}^{5}_{1},$ $\mathfrak{st}^{5}_{2}$ $S_{2}=\langle e_{1},e_{2},e_{3},e_{4}\rangle$ 3 86 no $\mathfrak{st}^{4}_{1}$, no $\mathfrak{st}^{4}_{2}$, $\mathfrak{st}^{5}_{1},$ $\mathfrak{st}^{5}_{2}$ $D_{2}:=\langle e_{1},e_{3},e_{4}\rangle$ 6 110 $\mathfrak{st}^{3}_{1},\mathfrak{st}^{3}_{2}$ $E_{2}:=\langle e_{1},e_{2}\rangle$ 8 115 $\mathfrak{st}^{3}_{1},\mathfrak{st}^{3}_{2}$ Moreover, ${\mathfrak{S}}^{4}_{B_{2}}={\mathfrak{S}}^{4}_{C_{2}};$ all present algebras are terminal, non-Leibniz and non-Jordan. ###### Proposition 16. If ${\bf A}=S_{2}$, the subalgebra of $W(2)$ generated by $e_{1},\ldots,e_{4}$, then a basis of the ${\mathbb{F}}$-vector space ${\mathfrak{S}}^{3}_{\bf A}$ is the set of identities: 1. (1) $30x_{1}(x_{2}x_{3})-42x_{1}(x_{3}x_{2})-25x_{2}(x_{1}x_{3})+7x_{2}(x_{3}x_{1})+39x_{3}(x_{1}x_{2})-9x_{3}(x_{2}x_{1})+10(x_{1}x_{3})x_{2}+15(x_{2}x_{1})x_{3}-19(x_{3}x_{1})x_{2}-6(x_{3}x_{2})x_{1}$, 2. (2) $-5x_{1}(x_{2}x_{3})+11x_{1}(x_{3}x_{2})+5x_{2}(x_{1}x_{3})-11x_{2}(x_{3}x_{1})-12x_{3}(x_{1}x_{2})+12x_{3}(x_{2}x_{1})-5(x_{1}x_{3})x_{2}+5(x_{2}x_{3})x_{1}+2(x_{3}x_{1})x_{2}-2(x_{3}x_{2})x_{1}$, 3. (3) $-3x_{1}(x_{2}x_{3})-3x_{1}(x_{3}x_{2})+4x_{2}(x_{1}x_{3})-4x_{2}(x_{3}x_{1})+3x_{3}(x_{1}x_{2})+3x_{3}(x_{2}x_{1})+3(x_{1}x_{2})x_{3}+2(x_{1}x_{3})x_{2}-2(x_{3}x_{1})x_{2}-3(x_{3}x_{2})x_{1}$. Finally, we have studied the family of identities $\mathfrak{st}^{n}_{1}$ and $\mathfrak{st}^{n}_{2}$ for $W(2)$ and its contractions. ###### Proposition 17. In the following table, we summarize which identities from the families $\mathfrak{st}^{n}_{1}$ and $\mathfrak{st}^{n}_{1}$ are satisfies for every contraction of $W(2)$, for $n=3,4,5$. Algebra $\mathfrak{st}^{3}_{1}$ $\mathfrak{st}^{4}_{1}$ $\mathfrak{st}^{5}_{1}$ $\mathfrak{st}^{3}_{2}$ $\mathfrak{st}^{4}_{2}$ $\mathfrak{st}^{5}_{2}$ $W(2)$ ✗ ✗ ✗ ✗ ✗ ✓ $\overline{W(2)}$ ✗ ✗ ✓ ✗ ✗ ✓ $\overline{\mathscr{S}_{1}}$ ✗ ✗ ✓ ✗ ✗ ✓ $\overline{\mathscr{S}_{5}}$ ✗ ✗ ✓ ✗ ✗ ✓ $\overline{\mathscr{S}_{\alpha,\beta}}$ ✗ $\text{\char 51}_{\alpha=\frac{3+\beta}{2}}$ ✓ $\text{\char 51}_{(\alpha,\beta)=(2,1)}$ $\text{\char 51}_{\alpha=\frac{3+\beta}{2}}$ ✓ $\widehat{W(2)}$ ✗ ✗ ✗ ✗ ✗ ✓ $\widehat{\widehat{W(2)}}$ ✗ ✗ ✓ ✗ ✗ ✓ ${\widetilde{W(2)}}$ ✗ ✓ ✓ ✗ ✓ ✓ ###### Corollary 18. ${\mathfrak{S}}^{4}_{W(n)}=0$ and ${\mathfrak{S}}^{5}_{W(2)}\neq 0$. The present corollary gives the following question. ###### Open question. Find minimal $k,$ such that ${\mathfrak{S}}^{k}_{W(n)}\neq 0$. In this case, is $W(n)$ satisfying $\mathfrak{st}^{k}_{2}$? ### 2.2. Other degree five identities for $W(2)$ In this subsection we are interested in finding other degree five identities for $W(2)$. Consider the set of free monomials $w(x_{1},x_{2},x_{3},x_{4},x_{5})$ of degree five up to permutations of the variables. There are exactly fourteen monomials: $w_{1}(x_{1},x_{2},x_{3},x_{4},x_{5})=(((x_{1}x_{2})x_{3})x_{4})x_{5}$ | $w_{2}(x_{1},x_{2},x_{3},x_{4},x_{5})=((x_{1}x_{2})x_{3})(x_{4}x_{5})$ ---|--- $w_{3}(x_{1},x_{2},x_{3},x_{4},x_{5})=((x_{1}x_{2})(x_{3}x_{4}))x_{5}$ | $w_{4}(x_{1},x_{2},x_{3},x_{4},x_{5})=(x_{1}x_{2})((x_{3}x_{4})x_{5})$ $w_{5}(x_{1},x_{2},x_{3},x_{4},x_{5})=(x_{1}x_{2})(x_{3}(x_{4}x_{5}))$ | $w_{6}(x_{1},x_{2},x_{3},x_{4},x_{5})=((x_{1}(x_{2}x_{3}))x_{4})x_{5}$ $w_{7}(x_{1},x_{2},x_{3},x_{4},x_{5})=(x_{1}(x_{2}x_{3}))(x_{4}x_{5})$ | $w_{8}(x_{1},x_{2},x_{3},x_{4},x_{5})=(x_{1}((x_{2}x_{3})x_{4}))x_{5}$ $w_{9}(x_{1},x_{2},x_{3},x_{4},x_{5})=x_{1}(((x_{2}x_{3})x_{4})x_{5})$ | $w_{10}(x_{1},x_{2},x_{3},x_{4},x_{5})=x_{1}((x_{2}x_{3})(x_{4}x_{5}))$ $w_{11}(x_{1},x_{2},x_{3},x_{4},x_{5})=(x_{1}(x_{2}(x_{3}x_{4})))x_{5}$ | $w_{12}(x_{1},x_{2},x_{3},x_{4},x_{5})=x_{1}((x_{2}(x_{3}x_{4}))x_{5})$ $w_{13}(x_{1},x_{2},x_{3},x_{4},x_{5})=x_{1}(x_{2}((x_{3}x_{4})x_{5}))$ | $w_{14}(x_{1},x_{2},x_{3},x_{4},x_{5})=x_{1}(x_{2}(x_{3}(x_{4}x_{5})))$ Now, consider the $\mathbb{F}$-vector spaces ${\mathfrak{Z}}^{i}$ generated by the set: $\left\\{w_{i}(x_{\sigma(1)},x_{\sigma(2)},x_{\sigma(3)},x_{\sigma(4)},x_{\sigma(5)}):\sigma\in\mathbb{S}_{5}\right\\}.$ Denote by ${\mathfrak{Z}}^{i}_{\bf A}$ the subspace of ${\mathfrak{Z}}^{i}$ of all $5$-linear polynomials vanishing on $\bf A$. Then we have the following result regarding the dimension of these subspaces: ###### Proposition 19. If ${\bf A}=W(2)$, then $\textrm{dim}({\mathfrak{Z}}^{i}_{\bf A})=0$ for $1\leq i\leq 13$ and $\textrm{dim}({\mathfrak{Z}}^{14}_{\bf A})=5$. A basis of the ${\mathbb{F}}$-vector space ${\mathfrak{Z}}^{14}_{\bf A}$ is the set of identities: 1. (1) $\sum\limits_{\sigma\in\mathbb{S}_{4}}(-1)^{\sigma}x_{\sigma(2)}(x_{\sigma(3)}(x_{\sigma(4)}(x_{\sigma(5)}x_{1}))).$ 2. (2) $\sum\limits_{\sigma\in\mathbb{S}_{4}}(-1)^{\sigma}x_{\sigma(1)}(x_{\sigma(3)}(x_{\sigma(4)}(x_{\sigma(5)}x_{2}))).$ 3. (3) $\sum\limits_{\sigma\in\mathbb{S}_{4}}(-1)^{\sigma}x_{\sigma(1)}(x_{\sigma(2)}(x_{\sigma(4)}(x_{\sigma(5)}x_{3}))).$ 4. (4) $\sum\limits_{\sigma\in\mathbb{S}_{4}}(-1)^{\sigma}x_{\sigma(1)}(x_{\sigma(2)}(x_{\sigma(3)}(x_{\sigma(5)}x_{4}))).$ 5. (5) $\sum\limits_{\sigma\in\mathbb{S}_{4}}(-1)^{\sigma}x_{\sigma(1)}(x_{\sigma(2)}(x_{\sigma(3)}(x_{\sigma(4)}x_{5}))).$ Moreover, the linear combination with parameters $(1,-1,1,-1,1)$ is $\mathfrak{st}^{5}_{2}$. ### 2.3. Central extensions The notion of central extensions appeared in the study of Lie algebras, but it can be considered in an arbitrary variety of algebras (see, for example, [19]). The calculation of central extensions of an algebra ${\bf A}$ of dimension $n$ from a certain variety of algebras gives the classification of all algebras with $(k-n)$-dimensional annihilator, such that its factor algebra by the annihilator is isomorphic to ${\bf A}$ (see, for example, [7]). These calculations are carried out by studying the cohomology, with respect to a polynomial identity, of the algebra ${\bf A}$. In this section, we are interested in the central extensions of the contractions and subalgebras of $W(2)$ considered in this paper. Some of these contractions and subalgebras have turned out to be terminal, i.e., they satisfy the terminal identity (degree four). The following result is about these particular algebras. ###### Proposition 20. There are no terminal central extensions of the terminal contractions of $W(2)$: $\overline{\mathscr{S}_{1}}$, $\overline{\mathscr{S}_{-1,1}}$, $\overline{\mathscr{S}_{0,0}}$, $\widehat{W(2)}$, $\widehat{\widehat{W(2)}}$, $\widetilde{W(2)}$ and $\widetilde{\widetilde{W(2)}}.$ ###### Proof. Recall that if $Z^{2}_{P}\left({\bf A},{\mathbb{F}}\right)$ denotes the space of cocycles with respect to the polynomial identity $P$ of the algebra ${\bf A}$, $B^{2}\left({\bf A},{\mathbb{F}}\right)$ denotes the space of coborders of the algebra ${\bf A}$ and $H^{2}_{P}\left({\bf A},{\mathbb{F}}\right):=Z^{2}_{P}\left({\bf A},{\mathbb{F}}\right)/B^{2}\left({\bf A},{\mathbb{F}}\right)$ denotes the cohomology space with respect to the polynomial identity $P$ of the algebra ${\bf A}$, then if $H^{2}_{P}\left({\bf A},{\mathbb{F}}\right)$ is trivial, we have that ${\bf A}$ has no central extensions for the identity $P$. Now, fix $P=T$ the terminal identity. Thus, the result is proven by direct calculation of the cohomology space, obtaining that $H^{2}_{T}\left({\bf A},{\mathbb{F}}\right)$ is trivial for any of the terminal contractions ${\bf A}$ considered. ∎ ###### Proposition 21. There are no terminal central extensions of the terminal subalgebras of $W(2)$: $B_{2},W_{2},C_{2},S_{2},D_{2},E_{2}$. ###### Proof. The result follows by the direct calculation of the cohomology space, obtaining that $H^{2}_{T}\left({\bf A},{\mathbb{F}}\right)$ is trivial for any of the terminal subalgebras ${\bf A}$ considered. ∎ Similarly, we can determine if there are central extensions for the rest of identities mentioned in the previous section. ###### Proposition 22. By calculating the correspoding cohomology space, we conclude the following. 1. (1) There are no central extensions of $\overline{\mathscr{S}_{2,1}}$ in the variety defined by one identity from the proposition 12. 2. (2) $\textrm{dim}\,Z^{2}_{P}\left(\overline{\mathscr{S}_{0,-3}},{\mathbb{F}}\right)=31$ and $\textrm{dim}\,H^{2}_{P}\left(\overline{\mathscr{S}_{0,-3}},{\mathbb{F}}\right)=23$, where $P$ is the identity $2\mathfrak{st}^{3}_{1}-3\mathfrak{st}^{3}_{2}$ from proposition 13. 3. (3) There are no central extensions of $S_{2}$ in the variety defined by one identity from the proposition 16. Regarding the central extensions with respect to the identities $\mathfrak{st}^{n}_{1}$ and $\mathfrak{st}^{n}_{2}$ for $n=3,4,5$ (see Proposition 15 and Proposition 17), we have the following result. ###### Proposition 23. The dimensions of the spaces of cocycles and coborders of the subalgebras of $W(2)$ are given. Algebra $\textrm{dim}\,B^{2}$ $\textrm{dim}\,Z^{2}_{\mathfrak{st}^{3}_{1}}$ $\textrm{dim}\,Z^{2}_{\mathfrak{st}^{4}_{1}}$ $\textrm{dim}\,Z^{2}_{\mathfrak{st}^{5}_{1}}$ $\textrm{dim}\,Z^{2}_{\mathfrak{st}^{3}_{2}}$ $\textrm{dim}\,Z^{2}_{\mathfrak{st}^{4}_{2}}$ $\textrm{dim}\,Z^{2}_{\mathfrak{st}^{5}_{2}}$ $B_{2}$ 7 - - 44 - - 44 $W_{2}$ 6 - - - - - 30 $C_{2}$ 5 - - 24 - - 24 $S_{2}$ 4 - - 16 - - 16 $D_{2}$ 3 8 9 9 8 9 9 $E_{2}$ 2 4 4 4 4 4 4 ###### Proposition 24. The dimensions of the spaces of cocycles and coborders of the subalgebras of $W(2)$ are given. Algebra $\textrm{dim}\,B^{2}$ $\textrm{dim}\,Z^{2}_{\mathfrak{st}^{3}_{1}}$ $\textrm{dim}\,Z^{2}_{\mathfrak{st}^{4}_{1}}$ $\textrm{dim}\,Z^{2}_{\mathfrak{st}^{5}_{1}}$ $\textrm{dim}\,Z^{2}_{\mathfrak{st}^{3}_{2}}$ $\textrm{dim}\,Z^{2}_{\mathfrak{st}^{4}_{2}}$ $\textrm{dim}\,Z^{2}_{\mathfrak{st}^{5}_{2}}$ $W(2)$ 8 - - - - - 38 $\overline{W(2)}$ 8 - - 54 - - 54 $\overline{\mathscr{S}_{1}}$ 8 - - 60 - - 60 $\overline{\mathscr{S}_{5}}$ 8 - - 60 - - 60 $\overline{\mathscr{S}_{-1,1}}$ 8 - - 60 - - 60 $\overline{\mathscr{S}_{0,0}}$ 8 - - 60 - - 60 $\widehat{W(2)}$ 8 - - - - - 40 $\widehat{\widehat{W(2)}}$ 8 - - 50 - - 54 ${\widetilde{W(2)}}$ 8 - 52 64 - 52 64 $\widetilde{\widetilde{W(2)}}$ 8 43 64 64 43 64 64 By Proposition 23 and Proposition 24, we can conclude that for any of the subalgebras and contractions considered there are central extensions with respect to the identities $\mathfrak{st}^{n}_{1}$ and $\mathfrak{st}^{n}_{2}$ for $n=3,4,5$. ## Acknowledgements Funding The first part of this work is supported by the Spanish Government through the Ministry of Universities grant ‘Margarita Salas’, funded by the European Union - NextGenerationEU; by the Junta de Andalucía through projects UMA18-FEDERJA-119 and FQM-336 and by the Spanish Ministerio de Ciencia e Innovación through project PID 2019-104236GB-I00, all of them with FEDER funds; FCT UIDB/00212/2020 and UIDP/00212/2020. The second part of this work is supported by the Russian Science Foundation under grant 22-11-00081. ## References * [1] F. Arzikulov, N. Umrzaqov, Conservative algebras of 2-dimensional algebras. III, Communications in Mathematics, 29 (2021), 2, 255–267. * [2] A. Calderón, A. Fernández Ouaridi, I. Kaygorodov, The classification of $2$-dimensional rigid algebras, Linear and Multilinear Algebra, 68 (2020), 4, 828–844. * [3] N. Cantarini, V. Kac, Classification of linearly compact simple rigid superalgebras, International Mathematics Research Notices, 17 (2010), 3341–3393. * [4] M. Cardoso Gonçalves, D. Gonçalves, D. Martín Barquero, C. Martín González, M. Siles Molina, Squares and associative representations of two-dimensional evolution algebras, Journal of Algebra and its Applications 20 (2021), 6, 2150090. * [5] R. Fehlberg Júnior, I. Kaygorodov, On the Kantor product, II, Carpatian Mathematical Publications, 2022, to appear, arXiv:2201.00174 * [6] A. Giambruno, S. Mishchenko, M. Zaicev, Codimension growth of two-dimensional non-associative algebras, Proceedings of the American Mathematical Society, 135 (2007), 11, 3405–3415. * [7] A. Hegazi, H. Abdelwahab, A. J. Calderon Martin, The classification of n-dimensional non-Lie Malcev algebras with $(n-4)$-dimensional annihilator, Linear Algebra and its Applications, 505 (2016), 32–56. * [8] E. Inönü, E.P. Wigner, On the contraction of groups and their representations, Proceedings of the National Academy of Sciences USA, 39 (1953), 510–524. * [9] A. Fernández Ouaridi, I. Kaygorodov, M. Khrypchenko, Yu. Volkov, Degenerations of nilpotent algebras, Journal of Pure and Applied Algebra, 226 (2022), 3, 106850, 21 pp. * [10] Sh. Kadyrov, F. Mashurov, Unified computational approach to nilpotent algebra classification problems, Communications in Mathematics, 29 (2021), 2, 215–226. * [11] I. Kantor, Certain generalizations of Jordan algebras (Russian), Trudy Sem. Vektor. Tenzor. Anal., 16 (1972), 407–499. * [12] I. Kantor, On an extension of a class of Jordan algebras, Algebra and Logic, 28 (1989), 2, 117–121. * [13] I. Kantor, Terminal trilinear operations, Algebra and Logic, 28 (1989), 1, 25–40. * [14] I. Kantor, An universal conservative algebra, Siberian Mathematical Journal, 31 (1990), 3, 388–395. * [15] I. Kaygorodov, On the Kantor Product, Journal of Algebra and Its Applications, 16 (2017), 9, 1750167 (17 pages). * [16] I. Kaygorodov, M. Khrypchenko, Yu. Popov, The algebraic and geometric classification of nilpotent terminal algebras, Journal of Pure and Applied Algebra, 225 (2021), 6, 106625. * [17] I. Kaygorodov, A. Khudoyberdiyev, A. Sattarov, One-generated nilpotent terminal algebras, Communications in Algebra, 48 (2020), 10, 4355–4390. * [18] I. Kaygorodov, A. Lopatin, Yu. Popov, Conservative algebras of 2-dimensional algebras, Linear Algebra and its Applications, 486 (2015), 255–274. * [19] I. Kaygorodov, S. Lopes, P. Páez-Guillán, Non-associative central extensions of null-filiform associative algebras, Journal of Algebra, 560 (2020), 1190–1210. * [20] I. Kaygorodov, Yu. Popov, A. Pozhidaev, The universal conservative superalgebra, Communications in Algebra, 47 (2019), 10, 4064–4074. * [21] I. Kaygorodov, Yu. Volkov, Conservative algebras of 2-dimensional algebras. II, Communications in Algebra, 45 (2017), 8, 3413-3421. * [22] I. Kaygorodov, Yu. Volkov, The variety of $2$-dimensional algebras over algebraically closed fields, Canadian Journal of Mathematics, 71 (2019), 4, 819–842. * [23] H. Petersson, The classification of two-dimensional nonassociative algebras, Results in Mathematics, 37 (2000), 1-2, 120–154. * [24] H. Petersson, M. Scherer, The number of nonisomorphic two-dimensional algebras over a finite field, Results in Mathematics, 45 (2004), no. 1-2, 137–152. * [25] Yu. Popov, Conservative algebras and superalgebras: a survey, Communications in Mathematics, 28 (2020), 2, 231–251.
# Microscopic Optical Potentials: recent achievements and future perspectives Paolo Finelli Dipartimento di Fisica ed Astronomia, Università degli Studi di Bologna and INFN, Sezione di Bologna, Via Irnerio 46, I-40126 Bologna, Italy <EMAIL_ADDRESS>Matteo Vorabbi Department of Physics, University of Surrey, Guildford, Surrey, GU2 7XH, UK<EMAIL_ADDRESS>Carlotta Giusti Dipartimento di Fisica, Università degli Studi di Pavia and INFN, Sezione di Pavia, I-27100 Pavia, Italy<EMAIL_ADDRESS> ###### Abstract Few years ago we started the investigation of microscopic Optical Potentials (OP) in the framework of chiral effective field theories [1, 2] and published our results in a series of manuscripts. Starting from the very first work [3], where a microscopic OP was introduced following the multiple scattering procedure of Watson [4], and then followed by Refs. [5, 6], where the agreement with experimental data and phenomenological approaches was successfully tested, we finally arrived at a description of elastic scattering processes off non-zero spin nuclei [7]. Among our achievements, it is worth mentioning the partial inclusion of three-nucleon forces [8], and the extension of our OP to antiproton-nucleus elastic scattering [9]. Despite the overall good agreement with empirical data obtained so far, we do believe that several improvements and upgrades of the present approach are still to be achieved. In this short essay we would like to address some of the most relevant achievements and discuss an interesting development that, in our opinion, is needed to further improve microscopic OPs in order to reach in a near future the same level of accuracy of the phenomenological ones. ## 1 Introduction It is well known that a suitable and successful framework to describe nucleon- nucleus processes is provided by the concept of the nuclear optical potential. Within this approach it is possible to compute the scattering observables across wide regions of the nuclear landscape and to extend calculations to inelastic channels or to a wide variety of nuclear reactions, e.g., nucleon transfer, knockout, capture, or breakup. Even if it is true that a phenomenological approach is generally preferred to achieve a more accurate description of the available experimental data, nowadays, with the upcoming facilities for exotic nuclei (FRIB at MSU just to mention one of the most important projects [10]), we strongly believe that a microscopic approach will be the preferred way to make robust predictions and to assess the unavoidable theoretical uncertainties already in a near future [11]. ## 2 Rooting optical potentials within ab initio approaches The calculation of a microscopic optical potential requires, in principle, the solution of the full many-body nuclear problem for the incident nucleon and the A nucleons of the target. In practice, with suitable approximations, microscopic optical potentials are usually calculated from two basic quantities: the nucleon-nucleon ($NN$) $t$ matrix and the matter distribution of the nucleus in the coordinate $\rho({\bm{r}},{\bm{r}^{\prime}})$, or in the momentum $\rho({\bm{k}},{\bm{k}^{\prime}})$ representation space. Formally we have to start from the the full $(A+1)$-body Lippmann-Schwinger equation that allows to determine the two-body transition matrix $T$ defined as $V|\Psi\rangle=T|\phi\rangle\;,$ (1) where $|\Psi\rangle$ is the scattered state and $|\Phi\rangle$ is a free wave (asymptotic) solution, in terms of an external two-body interaction $V$ $T=V+VG_{0}(E)T\,.$ (2) The previous relation can be manipulated in order to obtain a set of two equations in terms of an auxiliary quantity, i.e. the optical potential $U$. The first approximation we need is given by the spectator expansion [12], in which the scattering relation is expanded in a finite series of terms where the target nucleons interact directly with the incident proton. In particular, the first term of this series only involves the interaction of the projectile with a single target nucleon, the second term involves the interaction of the projectile with two target nucleons, and so on to the subsequent orders. As a consequence we can split Eq. (2) into a set of two equations. The first one is an integral equation for $T$ $T=U+UG_{0}(E)PT\,,$ (3) where $U$ is the optical potential operator, and the second one is an integral equation for $U$ $U=V+VG_{0}(E)QU\,.$ (4) $G_{0}(E)$ is the free propagator for the $(A+1)$-nucleon system and the operators $P$ and $Q$ in Eqs. (3) and (4) are projection operators, $P+Q=\hbox{$1\hskip-1.2pt\vrule depth=0.0pt,height=6.88889pt,width=0.7pt\vrule depth=0.0pt,height=0.3pt,width=1.19995pt$}\,.$ (5) For elastic scattering processes $P$ selects only the ground state $|\Phi_{0}\rangle$, i.e. $P=|\Phi_{0}\rangle\langle\Phi_{0}|\;.$ (6) Eq. (4) can be numerically solved if we assume the validity of another approximation, namely the impulse approximation where, since the projectile energy scales involved are large compared to nuclear binding forces of the interacting target nucleon, we can safely use the free $NN$ $t$ matrix in the Eq. (3). Most of the theoretical derivations can be found in Refs. [3, 12]. Figure 1: In this infographic the main building blocks of the optical potential are schematically shown along the main features of each components. For practical reasons the auxiliary optical potential $U$ is expressed as a function of the center-of-mass momentum ${\bm{K}}\equiv 1/2({\bm{k}}+{\bm{k}}^{\prime})$, and the momentum transfer ${\bm{q}}\equiv({\bm{k}}-{\bm{k}}^{\prime})$. It can be factorized into three pieces [13, 14, 15, 16]: the Möller correction $\eta({\bm{q}},{\bm{K}},{\bm{P}})$ that ensures the conservation of the transmitted flux after the transformation in the center-of-mass reference frame, the projectile-target potential $t_{NN}({\bm{q}},{\bm{K}};\omega)$ that depends also on the energy of the projectile $\omega$, and the structure component $\rho_{N}({\bm{q}},{\bm{P}})$, i.e. the Fourier transform of the non-local matter density in the momentum space. For both inputs we show into the coloured insertions the main features. For the potential part, at the moment, we are able to include the full $NN$ interaction and averaged $NNN$ contributions in terms of a medium density-dependent $NN$ interaction [17]. This solution requires to fix a density $\bar{\rho}$ at which the calculations are performed. So far [7, 8] we decided to explore the impact of $NNN$ terms performing calculations for several values of $\bar{\rho}$ (usually in the range $0.08$ fm${}^{-3}\leq\bar{\rho}\leq 0.13$ fm-3, see Figs 1-6 of Ref. [7]). Investigations in this direction are currently underway [18]. For the structure part we used the NCSM [21] for light nuclei that allows us to fully use both $NN$ and $NNN$ terms (see Sect II and III. of Ref. [7]). For medium nuclei other approaches should be explored, like Self-Consistent Green Functions (SCGF) [22, 23]. Then, from a practical point of view, the OP can be computed in momentum space as follows (after some mathematical manipulations that can be found in Refs. [13, 14, 15, 16]) $\begin{split}U({\bm{q}},{\bm{K}};E)=\sum_{N=p,n}&\int d{\bm{P}}\;\eta({\bm{q}},{\bm{K}},{\bm{P}})\,t_{NN}\left[{\bm{q}},\frac{1}{2}\left(\frac{A+1}{A}{\bm{K}}+\sqrt{\frac{A-1}{A}}{\bm{P}}\right);E\right]\\\ &\times\rho_{N}\left({\bm{P}}+\sqrt{\frac{A-1}{A}}\frac{{\bm{q}}}{2},{\bm{P}}-\sqrt{\frac{A-1}{A}}\frac{{\bm{q}}}{2}\right)\,,\end{split}$ (7) where ${\bm{q}}$ and ${\bm{K}}$ represent the momentum transfer and the average momentum, respectively. Here ${\bm{P}}$ is an integration variable, $t_{NN}$ is the $NN$ $t$ matrix and $\rho_{N}$ is the one-body nuclear density matrix. The parameter $\eta$ is the Möller factor, that imposes the Lorentz invariance of the flux when we pass from the $NA$ to the $NN$ frame in which the $t$ matrices are evaluated. Finally, $E$ is the energy at which the $t$ matrices are evaluated and it is fixed at one half the kinetic energy of the incident nucleon in the laboratory frame. As shown in Fig. 1, the calculation of Eq. (7) requires two external inputs: the $NN$ $t$ matrix and the one-body density $\rho_{N}$. The calculation of the density matrix could be performed using different ab initio approaches. The method followed in Refs. [19, 20], where one-body translationally invariant densities were computed within the ab initio No-Core Shell Model (NCSM) [21] approach proved to be very successful in our own works [7, 9]. Alternatively other methods could be used, i.e. Self-Consistent Green Functions (SCGF) [22, 23], that allow to overcome the main limitation of the NCSM method, namely the severe limits on the atomic mass of the target nucleus due to the computational overload. Work in this direction is currently underway and a thorough investigation of nickel and calcium isotopes will soon appear [24]. For heavier nuclei, the only available method is Density Functional Theory (DFT). Even if most of the parametrizations have a phenomenological derivation, in recent years some authors [25] started to employ such approach from an ab initio point of view. The main advantages of using ab initio methods in our framework is twofold. On one hand we can preserve the consistency of our theoretical framework if the same realistic interactions are used both for the structure part and the projectile-target interaction (with some caveats explained in Refs. [3, 5, 6, 8, 7]) and, on the other hand, we are able to provide more theoretically founded predictions for exotic nuclei since we do not have to rely on any fitting procedure on any selections of stable nuclei. To calculate the projectile-target potential we rely on modern $NN$ potentials derived in the framework of Chiral Perturbation Theory (ChPT). Among the many available choices of realistic potentials, i.e. those that reproduce phase- shifts and binding energies of light nuclei with $\chi^{2}|_{\rm data}\simeq 1$, the choice of chiral potentials over phenomenological ones responds to some important requirements that will be listed shortly below. Generally speaking, ChPT is a perturbative technique for the description of hadron scattering amplitudes based on expansions in powers of a parameter that can be generally defined as $(p,m_{\pi})/\Lambda_{b}$, where $p$ is the magnitude of three-momenta of the external particles, $m_{\pi}$ is the pion mass, and $\Lambda_{b}$ is the chiral symmetry breaking scale [26]. ChPT respects the low-energy symmetries of Quantum ChromoDynamics (QCD) and, up to a certain extent, is model independent and systematically improvable by an order-by-order expansion, with controlled uncertainties from neglected higher- order terms [27]. Both the fact that ChPT is directly connected to QCD and that the $NN$ components (and also $NNN$) are organized in terms of a systematic theoretical expansion is a huge improvement respect to conventional $NN$ potentials, like CD-Bonn [28] or Nijmegen [29]. The hierarchy of terms could also be exploited in order to provide an estimate for the theoretical uncertainties. From a practical point of view, for our own calculations [3, 5, 6, 7, 8, 9] we mainly employed the $NN$ chiral interactions developed by Entem et al. [30, 31] up to the fifth order (N4LO) or, alternatively, the ones developed by Epelbaum et al. [32, 33] following a different strategy. The main difference between the two theoretical approaches of the chiral potentials concerns the renormalization procedures. The strategy followed in Refs. [32, 33] consists in a coordinate space regularization for the long-range contributions, and a conventional momentum space regularization for the contact (short-range) terms. On the other hand, for Refs. [30, 31], a slightly more conventional approach was pursued. A spectral function regularization was employed to regularize the loop contributions and a conventional regulator function to deal with divergences in the Lippman-Schwinger equation. The three-nucleon forces deserve a separate treatment since so far we can employ genuine $3N$ forces only to compute the one-body densities of the target nuclei (up to the third order N2LO with a local and nonlocal regularization [34, 35]). At N2LO order we include the $2\pi$ exchange diagram between three nucleons, a one-$\pi$-exchange plus a $NN$ contact term and a $3N$ contact term. For more details and an explicit derivation of the relevant formulae, we refer the reader to Refs. [34]. For the projectile-target part we follow the prescriptions by Holt et al. [17] to derive a suitable density- dependent NN interaction. ### 2.1 A selection of results To show the robustness of our proposed approach we will show a couple of non trivial calculations perfomed in the last years. The first case is the elastic proton scattering on 7Li, in which the target nucleus has spin and parity quantum numbers $J^{\pi}=3/2^{-}$. Theoretical predictions along with the experimental data are presented in Fig. 2. Figure 2: Differential cross section (upper panel) and analyzing power (lower panel), as functions of the center-of-mass scattering angle, for 200 MeV protons elastically scattered from 7Li (Jπ = 3/2-). The results were obtained using the $NN$ $t$ matrix computed with the $NN$ chiral interaction at N3LO order supplemented by a density dependent NN interaction (where the baryon density is varied in the range between 0.08 fm-3 and 0.13 fm-3) and the one- body nonlocal density matrices computed with the NCSM method using realistic nucleon-nucleon and three-nucleon local-nonlocal chiral interactions. The differential cross section and analyzing power were measured at the Indiana University Cyclotron Facility using a polarized proton beam at a laboratory bombarding energy of 200.4 MeV [36]. The agreement between the theoretical prediction and the empirical data is good for the differential cross section, over all the angular distribution shown in the figure, and satisfactory for the analyzing power for values of the scattering angle up to $\sim 45^{\rm o}$. The second test case is an extension much needed for future applications: inverse kinematics. In fact, in recent years, experimental efforts have multiplied to develop the technologies necessary to study the elastic scattering of protons (and ions) by using inverse kinematics. This configuration is necessary to study exotic nuclei that have very short average lifetimes. A very interesting experiment performed at the GSI was the study of the scattering of 56,58Ni nuclei on hydrogen targets at energies near 400 MeV/nucleon in inverse kinematics for the determination of the distribution of nuclear matter [37]. Since the only theoretical approach used so far to analyze the experimental data is the Glauber model which contains phenomenological inputs, it is mandatory to establish a framework in which microscopic calculations could help understading experimental results. In Fig. 3 we show the results only for $t\leq 0.1$ (GeV/c)2 in the laboratory system because, due to the challenges of inverse kinematics, the regime of the elastic scattering needs to be focused at small center-of-mass angles. The agreement with experimental data is impressive despite the different kinematics and the high-energy regime at which the measurements are performed. Figure 3: Differential cross section as a function of the Mandelstam variable $t$ in the laboratory frame for 56Ni(p, p)56Ni elastic scattering at $E=400$ MeV/u. Calculations are performed with different microscopic OPs derived from the $NN$ chiral potentials of Epelbaum et al. (EKM) [32, 33] and Entem et al. (EMN) [30, 31] at N4LO in comparison with the well known CD-Bonn [28]. With empty black squares and circles are shown the empirical data along with the corresponding error bars [37]. ## 3 Necessary extensions Among the different extensions under development, we would like to mention probably the most relevant one: the inclusion in our framework of inelastic reactions. In the previous sections we mentioned a class of experiments that usually need the subtraction of contributions from the inelastic channel to perform a correct data analysis. In this perspective, if we wish to establish a consistent microscopic approach for inelastic NA scattering, which is our long-term goal, it is mandatory to develop a two-step theory program: first a reliable description of OPs for states with spin-parity quantum numbers J${}^{\pi}\neq 0^{+}$ (already under study with good results [7]) and then an extension of our formalism that, so far, is restricted to the elastic channel. Let’s suppose to extend the definition of the projection operator $P$ $P\equiv P_{0}+P_{1}=\ket{\Psi_{0}}\bra{\Psi_{0}}+\ket{\Psi_{1}}\bra{\Psi_{1}}\,,$ (8) with still $P+Q=\hbox{$1\hskip-1.2pt\vrule depth=0.0pt,height=6.88889pt,width=0.7pt\vrule depth=0.0pt,height=0.3pt,width=1.19995pt$}$. Inserting this relation into the many-body Lippmann-Schwinger equation we could follow the same prescriptions adopted for elastic case to define the following quantities $T_{00}\equiv P_{0}TP_{0}\,,\quad T_{01}\equiv P_{0}TP_{1}\,,\dots\quad G_{00}\equiv P_{0}G_{0}(E)P_{0}\,,\ldots$ (9) that represent the new scattering amplitudes that can be accomodated in a matrix form as follows (see Fig. 4) ${\bm{T}}={\bm{U}}+{\bm{U}}{\bm{G}}{\bm{T}}\,,$ (10) with the previous terms defined as ${\bm{T}}\equiv\begin{pmatrix}T_{00}&T_{01}\\\ T_{10}&T_{11}\end{pmatrix}\,,\qquad{\bm{U}}\equiv\begin{pmatrix}U_{00}&U_{01}\\\ U_{10}&U_{11}\end{pmatrix}\,,\qquad{\bm{G}}\equiv\begin{pmatrix}G_{00}&0\\\ 0&G_{11}\end{pmatrix}\,.$ (11) When projected on the momentum basis, the equation for the $00$ component of the scattering matrix becomes $T_{00}({\bm{k}}^{\prime},{\bm{k}})=U_{00}({\bm{k}}^{\prime},{\bm{k}})+\int d{\bm{p}}\frac{U_{00}({\bm{k}}^{\prime},{\bm{p}})T_{00}({\bm{p}},{\bm{k}})}{E-E(p)+i\epsilon}+\int d{\bm{p}}\frac{U_{01}({\bm{k}}^{\prime},{\bm{p}})T_{10}({\bm{p}},{\bm{k}})}{E-E(p)-e_{1}+i\epsilon}\,,$ (12) and similarly for the other equations. Here $e_{1}$ is the excitation energy of the first excited state. Of course the previous equation has to be expanded in partial waves along with the other 3 equations. Work along this direction is in progress [38]. Figure 4: This infographic shows how the structure of Eq. 2 must change in order to include the description of inelastic processes. In order to accommodate the inclusion of the relevant excited state, it is necessary to calculate four optical potentials that describe all the available scattering channels. According to this picture we would need both stationary and transition densities. ## References ## References * [1] Epelbaum E, Hammer H W and Meissner U G 2009 Rev. Mod. Phys. 81 1773 * [2] Machleidt R and Entem D R 2011 Phys. Rep. 503 1 * [3] Vorabbi M, Finelli P and Giusti C 2016 Phys. Rev. C 93 034619 * [4] Riesenfeld W B and Watson K M 1956 Phys. Rev. 102 1157 * [5] Vorabbi M, Finelli P and Giusti C 2017 Phys. Rev. C 96 044001 * [6] Vorabbi M, Finelli P and Giusti C 2018 Phys. Rev. C 98 064602 * [7] Vorabbi M, Gennari M, Finelli P, Giusti C, Navrátil P and Machleidt R 2022 Phys. Rev. C 105 014621 * [8] Vorabbi M, Gennari M, Finelli P, Giusti C, Navrátil P and Machleidt R 2021 Phys. Rev. C 103 024604 * [9] Vorabbi M, Gennari M, Finelli P, Giusti C and Navrátil P 2020 Phys. Rev. Lett. 124 162501 * [10] Castelvecchi D 2022 Nature 605 201 * [11] Hebborn C et al., arXiv:2210.07293 [nucl-th] * [12] Chinn C R, Elster Ch, Thaler R M and Weppner S P 1992 Phys. Rev. C 52 1992 * [13] Elster Ch and Tandy P C 1989 Phys. Rev. C 40 81 * [14] Elster Ch, Cheon T, Redish E F and Tandy P C 1990 Phys. Rev. C 41 814 * [15] Elster Ch, Weppner S P and Chinn C R 1997 Phys. Rev. C 56 2080 * [16] Weppner S P, Elster Ch and Huber D 1998 Phys. Rev. C 57 1378 * [17] Holt J W, Kaiser N and Weise W 2010 Phys. Rev. C 81 024002 * [18] Vorabbi M, Finelli P and Giusti C, in preparation * [19] Burrows M, Elster Ch, Popa G, Launey K, Nogga A and Maris P, 2017 Phys. Rev. C 97 024325 * [20] Gennari M, Vorabbi M, Calci A and Navrátil P 2017 Phys. Rev. C 97 034619 * [21] Barrett B R, Navrátil P and Vary J P 2013 Prog. Part. Nucl. Phys. 69 131 * [22] Dickhoff W H and Barbieri C 2004 Prog. in Part. and Nuc. Phys. 52 377 * [23] Carbone A, Cipollone A, Barbieri C, Rios A and Polls A 2013 Phys. Rev. C 88 054326 * [24] Vorabbi M, Barbieri C, Somà V, Finelli P and Giusti C, in preparation * [25] Marino F, Barbieri C, Carbone A, Colò G, Lovato A, Pederiva F, Roca-Maza X and Vigezzi E, 2021 Phys. Rev. C 104 024315 * [26] Scherer S 2003 Adv. Nucl. Phys. 27 277 * [27] Georgi G 1993 Ann. Rev. Nucl. Part. Sci. 43 209 * [28] Machleidt R 2001 Phys. Rev. C 63 024001 * [29] de Swart J J, Klomp R A M M, Rentmeester M C M and Rijken T A, 1995 Few Body Syst. Suppl. 8 438 * [30] Entem D R, Kaiser N, Machleidt R and Nosyk Y 2015 Phys. Rev. C 91 014002 * [31] Entem D R, Machleidt R and Nosyk Y 2017 Phys. Rev. C 96 024004 * [32] Epelbaum E, Krebs H and Meißner U - G 2015 Phys. Rev. Lett. 115 122301 * [33] Epelbaum E, Krebs H and Meißner U - G 2015 Eur. Phys. J. A 51 53 * [34] Navràtil P 2007 Few-Body Syst. 41 117 * [35] Gysbers P, Hagen G, Holt J D, Jansen G R , Morris T D, Navrátil P, Papenbrock T, Quaglioni S, Schwenk A, Stroberg S R and Wendt K A 2019 Nat. Phys. 15 428 * [36] Glover C W, Foster C C, Schwandt P, Comfort J R, Rapaport J, Taddeucci T N, Wang D, Wagner G J, Seubert J, Carpenter A W, Carr J A, Petrovich F, Philpott R J and Threapleton M J 1991 Phys. Rev. C 43 1664 * [37] Von Schmid M 2015 Nuclear matter distribution of 56Ni measured with EXL, https://d-nb.info/1112044388/34, Ph. D. Thesis * [38] Vorabbi M, Gennari M, Finelli P, Giusti C and Navrátil P, in preparation
# What Physical Layer Security Can Do for 6G Security Miroslav Mitev1, Arsenia Chorti1,2, H. Vincent Poor3, Gerhard Fettweis1,4 1 Barkhausen Institut, Dresden, Germany; 2 ETIS, UMR 8051 CY Cergy Paris Université, ENSEA, CNRS, Cergy, France; 3 School of Engineering and Applied Science, Princeton University, Princeton, NJ, USA; 4 Vodafone Chair for Mobile Communications Systems, Technische Universität Dresden, Dresden, Germany; Corresponding author<EMAIL_ADDRESS> ###### Abstract While existing security protocols were designed with a focus on the core network, the enhancement of the security of the B5G access network becomes of critical importance. Despite the strengthening of 5G security protocols with respect to LTE, there are still open issues that have not been fully addressed. This work is articulated around the premise that rethinking the security design bottom up, starting at the physical layer, is not only viable in 6G but importantly, arises as an efficient way to overcome security hurdles in novel use cases, notably massive machine type communications (mMTC), ultra reliable low latency communications (URLLC) and autonomous cyberphysical systems. Unlike existing review papers that treat physical layer security orthogonally to cryptography, we will try to provide a few insights of underlying connections. Discussing many practical issues, we will present a comprehensive review of the state-of the-art in i) secret key generation from shared randomness, ii) the wiretap channels and fundamental limits, iii) authentication of devices using physical unclonable functions (PUFs), localization and multi-factor authentication, and, iv) jamming attacks at the physical layer. We finally conclude with the proposers’ aspirations for the 6G security landscape, in the hyper-connectivity and semantic communications era. ## I Introduction The rollout of fifth-generation (5G) mobile networks and the forthcoming sixth-generation (6G) will bring about fundamental changes in the way we communicate, access services and entertainment. In the context of security, inarguably, 5G security enhancements present a big improvement with respect to LTE. However, as the complexity of the application scenarios increases with the introduction of novel use cases, notably ultra-reliable low latency (URLLC), massive machine type communications (mMTC) and autonomous cyberphysical systems (drones, autonomous cars, robots, etc.), novel security challenges arise that might be difficult to address using the standard paradigm of complexity based classical cryptographic solutions. Specific use cases with open security issues are described in detail in a number of 3GPP technical reports, e.g., on the false base station attack scenario [1] and on the security issues in URLLC [2]. Indeed, for beyond 5G (B5G) systems, there exist security aspects that can be further enhanced by exploiting different approaches, as classical mechanisms either fall short in guaranteeing all the security and privacy relevant aspects, or, can be strengthened with mechanisms that could provide a second layer of protection. In the past years, physical layer security (PLS) [3] has been studied and indicated as a possible way to emancipate networks from classical, complexity based, security approaches. Multiple white papers on the vision for 6G incorporate physical layer security, e.g., [4, 5, 6], as well as in the IEEE International Network Generations Roadmap (INGR) 1st and 2nd Editions [7]. Motivated by the above, a key point of this paper is to showcase how PLS and in general security controls at the PHY level can be exploited towards securing future networks. One of the most promising and mature PLS technologies concern the distillation of symmetric keys from shared randomness, typically in the form of wireless fading coefficients. Within the channel’s coherence time, small scale fading is reciprocal, time-varying and random in nature and therefore, offers a valid, inherently secure source for key agreement (KA) protocols between two communicating parties. This is pertinent to many forthcoming B5G applications that will require strong, but nevertheless, lightweight KA mechanisms, notably in the realm of Internet of things (IoT). With respect to authentication, there are multiple PLS possibilities, including physical unclonable functions (PUFs), wireless fingerprinting and high precision localization. Combined with more classical approaches, these techniques could enhance authentication in demanding scenarios, including (but not limited to) device to device (D2D) and Industry 4.0. Note that according to the 6G vision, as a network of (sub)networks, authentication might be required independently for access to the local (sub)network and to the core network, making the adoption of RF and device fingerprints a viable alternative for fast authentication of local wireless connections. In parallel, mmWave and subTHz bands require the use of a huge number of antennas and pencil sharp beamforming. Consequently, a viable scenario for the wiretap channel can be substantiated, without any assumptions regarding the hardware (number of antennas, noise figure, etc.) or the position of a potential eavesdropper. Similarly, visible light communications (VLC) systems offer respective use cases. It is therefore pertinent to discuss advancements in wiretap secrecy encoders. The interplay between secrecy and privacy in finite blocklengths is another aspect that emerged from recent fundamental results in finite blocklength secrecy coding and should be highlighted. Furthermore, new types of attacks have to be accounted for. In particular, there is mounting concern for potential jamming attacks and pilot contamination attacks during beam allocation and entry phases of nodes into the network [8]. Clearly, such attacks cannot be addressed with standard cryptographic tools and the required solutions can only emerge at the PHY, potentially in the form of jamming-resilient waveform and code design. Finally, a less considered aspect relates to anomaly / intrusion detection by monitoring hardware metrics. This can be either used for distributed anomaly detection in low-end IoT networks, i.e., by monitoring memory usage, Tx and Rx time, debug interface of devices, or, for more generalized anomaly detection of devices of untrusted manufacturers, etc. Such approaches could help lessen the monitoring overhead of centralized approaches and could provide new approaches towards the identification of the source of the anomaly [9]. Looking at the bigger picture, future security controls will be adaptive and context-aware [10]. In this framework, rethinking the security design bottom up can provide low-cost alternatives. In particular, 1. 1 PLS can provide information-theoretic security guarantees with lightweight mechanisms (e.g., using LDPC, Polar codes, etc.); 2. 2 Hybrid crypto-PLS protocols can provide fast, low-footprint and low-complexity solutions for issues such as in [1] and [2]; 3. 3 PLS can act as an extra security layer, complementing other approaches, enhancing the trustworthiness of the radio access network (RAN); 4. 4 PLS is inherently adaptive and can leverage the context and the semantics of the data exchanged. In the following we will provide a comprehensive review of fundamental, cutting edge results in PLS and showcase how PLS can be employed to achieve many of the standard security goals, notably confidentiality, authentication, integrity. To this end, and, in order to provide a platform for a fair comparison to standard crypto schemes and a discussion on the potential advantages of hybrid PLS-crypto systems, we will first review fundamental cryptographic concepts and goals in Section II. Next, Section III gives a brief motivation on why PLS should be considered for the 6G. In Section IV the wiretap channel theory will be presented (focusing on information theoretic characterizations for the finite blocklength) along with some recent results for privacy in sensing systems. Subsequently Section V discusses the topic of secret key generation (SKG) from shared randomness and highlights two subtle points concerning the pre-processing of the observation channel coefficients and coding methods in the short blocklength, furthermore, jamming attacks and countermeasures are discussed [11, 12]. In Section VI hardware based and statistical methods used in authentication will be visited, focusing on localization based authentication [13, 14] and physical unclonable functions. Finally, future directions and the authors’ aspirations for security controls at all layers in 6G will be presented in Section VII. ## II Background Concepts in Cryptography and Network Security Starting with some fundamental concepts in cryptography, we will address questions that arise in the systematic study of any system. In particular, we will provide answers to the following questions: ”what do we want to achieve?”; ”what is the system model?”; ”what are the underlying assumptions, and what are the desirable properties?” With respect to what we aim to achieve, typically any security system aims at reaching one or multiple of four fundamental goals. The first goal is to be able to provide data confidentiality, i.e., security against eavesdropping (passive attackers). The corresponding threat model involves two legitimate parties communicating in the presence of an eavesdropper. Typically, with the aid of encryption, confidentiality is ensured against passive attackers. The second major goal is that of data integrity, i.e., providing guarantees that as the data traverses through the network, any modification or alteration of a message will be perceptible at the destination. The corresponding threat model involves an active attacker that in addition to intercepting messages also performs modifications. The third major security goal is authentication (user or device), while access control is a closely related topic. The threat model involves again an active attacker that potentially attempts to gain unauthorized access. Finally, the fourth goal is that of availability, i.e., users should not be denied services. The network should be resilient to active attacks that fall in the general category of “denial of service”. With respect to the system model, as noted above, the basic system setting includes three nodes. Two legitimate parties, that are referred here Alice and Bob and an adversarial node that is typically referred to as Eve (passive eavesdropper) or Mallory (active attacker, i.e., man-in-the-middle). To securely transmit a message (plaintext) to Bob, Alice uses a secret key to first encrypt it to a ciphertext. The ciphertext is then propagated through the transmission medium and received at Bob. Bob can decrypt the ciphertext by using the same or a different type of key, depending on the underlying algorithm. ### II-A Confidentiality To perform the operations above, i.e., encryption / decryption, Alice and Bob rely on the use of ciphers. A key feature of modern block ciphers is to exploit highly non-linear operations to induce confusion, i.e., to render statistical inference attacks impossible. A textbook example of a linear cipher that is badly broken is the substitution cipher in which each letter of the alphabet is moved $k$ positions to the right (or to the left), with $k$ changing per letter. Considering the English alphabet, this results in $25!$ possible key combinations, making a brute force attack impractical. However, due to the linearity of the operations (permutations), a frequency analysis of a (long enough) ciphertext suffices to guess the plaintext. A revolutionizing result in security was presented by Shannon in $1949$ [15], when he demonstrated that perfect secrecy can be achieved if and only if (iff) the entropy of the secret key is greater or equal to the entropy of the plaintext. The corresponding scheme, known as one-time-pad, is implemented by xor-ing the plaintext with the key. Unfortunately, to perform the above, the key size must be at least equal to that of the data which raises the problem of key distribution. While one-time pad is impractical, it provided insight into how secrecy can be achieved. In particular, it inspired the family of stream ciphers that rely on the idea of inflating short key sequences to psedorandom sequences of the same size as the plaintext and xor-ing them. This is achieved through the use of pseudorandom number generators (PRNGs). Although they cannot provide perfect secrecy (entropy cannot increase by data processing as a consequence of the data processing inequality), their usage led to the introduction of a more practical concept, i.e., semantic security. The definition of semantic security for PRGNs relies on the indistinguishability between their output and the output of a truly random source. More generally, semantic security ensures that a non-negligible statistical advantage cannot be accumulated by an adversary in polynomial time. For all practical purposes, if a statistical advantage happens with probability higher that $2^{-30}$, e.g., one bit is leaked in one gigabyte of data, the system is considered broken (not semantically secure). A canonical example of modern block ciphers is the advanced encryption standard (AES). AES is a semantically secure symmetric block cipher which takes a $n$-bit plaintext ($n=128$) and a $k$-bit key ($k$ chosen from $128$, $192$, or $256$ bits, with AES-$256$ considered to be quantum resistant) as input and outputs a $n$-bit ciphertext. AES relies on a set of substitution and permutation operations including the use of substitution (S) boxes. A well structured S-box removes the relation and dependency between bits, making a (linear or differential) cryptanalysis attack impossible. To allow the re-use of a single key for multiple blocks, nonces can be used. Nonces are deterministic (e.g., a counter) or random (initialization vectors), chosen such that a pair (key, nonce) never repeats. The important message here is that, today’s cryptographic mechanisms allow the use of a short key sequence (e.g., 96 Bytes of key material in TLS v1.3) for the encryption of very long data sequences (in the order of GBs), allowing to overcome the key issue with one-time pad. ### II-B Data integrity Data integrity is achieved with message authentication codes (MACs). The principle of MACs is to append a small label (tag) to each message, which validates its integrity. A MAC consists of two algorithms: signing and verification. Similarly to confidentiality schemes, there are historical examples of broken integrity algorithms in which linear functions (e.g., cyclic redundancy checks) have been used to generate MACs. Modern signing algorithms (tag generation) leverage the use of secret keys and symmetric block ciphers to generate a $t$-bit tag for a $n$-bit message, with $t<<n$. Upon reception, the verification algorithm uses the key, the received message and the tag and outputs a binary decision, i.e., the integrity check is either successful or not. Building on the above, a naturally arising concept is the one of authenticated encryption (AE) which combines both confidentiality and integrity. Various options exist on how to perform the two operations. One approach, that is always correct and provably secure, is the so called encrypt-then-sign, i.e., after a plaintext is encrypted a tag is generated over the ciphertext. The receiver would first check the integrity and iff successful would continue with decryption. ### II-C Authentication The process of authentication relies on digital signatures, which in turn, are used to produce digital certificates. Digital certificate is data signed by a trusted third party (certificate authority (CA)) that ensures the authenticity of the its owner. A certificate contains information about the CA, the owner of the certificate, the validity of the certificate, etc. As an example, when a user accesses a public server, the server proves its authenticity by presenting a certificate signed from a CA. To achieve mutual authentication the user must enter a password information, provide biometric data, etc. ## III Motivation for Considering Physical Layer Security Given the fact that all schemes discussed in the previous section are widely deployed and trusted, one question remains: What is the motivation in considering PLS? PLS technologies can offer multiple security techniques: i) secrecy encoders for wiretap channels, ii) privacy preserving transmission, iii) secret key generation from shared randomness iv) physical unclonable functions for device authentication, and v) localization or RF fingerprinting based authentication. While crypto solutions can provide these functionalities for current standards, they face number of challenges when considering new and emerging technologies. First, latency requirements are getting more stringent than ever, bringing the need for faster authentication and integrity checks. Second, large scale IoT deployment requires flexible and easily scalable security solutions that could simultaneously satisfy different security levels. A third element comes from the rise of quantum computing which opens the need for quantum secure algorithms. Finally, a fourth motivation comes from the new PHY infrastructures where the number of operations performed at the edge are expected to rise dramatically. Therefore, it is of utmost importance to separate the security of the core network from the one at the edge and introduce new faster and lightweight security algorithms. The statements above are complemented with the following list: 1. 1. Regarding latency, 3GPP has recently noted that delays should be minimized in two directions, delays incurred by the communication and delays incurred due to computational overhead. A particular case where computational overhead of current standards do not comply with the requirements is security. As an example, it has been shown that the verification of a digital signature, in a vehicular networking scenario using a $400$ MHz processor, exceeds the tolerated delays and requires approximately $20$ ms [16]. Such results hint that a revolutionizing actions are needed in that direction. 2. 2. Next, deploying billions of IoT devices is not inconceivable anymore. In 2016, it has been demonstrated that a Mirai sized attack (e.g., $6\times 10^{5}$ bots) is plausible. The attack has been demonstrated over simple machines, e.g. water heater, however, controlling $6\times 10^{5}$ can instantly change the demand in the smart grid by $3$ GW, which is comparable to having an access to a nuclear plant. Examples like this raise a lot of questions on the security of the IoT. 3. 3. In 2017, the NIST started the investigation on the topic of quantum resistance and post-quantum cryptography. However, as it stands now, the state of the art is based on using longer keys and increased complexity. This makes the mechanisms heavier which contradicts with the need for low latency and low footprint. Hence, post-quantum innovations at the moment are not well aligned to the expectations towards 6G networks. 4. 4. Finally, new PHY and networking structures are being developed for the next generation of communication technologies. The central idea is to enhance the role of AI edge intelligence. This is a key component, that can enable the use of PLS in 6G. More details regarding this point will be discussed in Sec. VII. In the following sections it will be discussed how PLS technologies can be employed and some fundamental results in the area will be showed. ## IV Confidentiality and Privacy Using PLS ### IV-A Confidential transmission In this section two aspects of physical layer security will be discussed, i.e., data confidentiality and data privacy. In detail, the information theoretic formulations of these problems will be investigated. As noted in Section I secure data transmission tends to be a higher layer issue, e.g., enabled by encryption. However, confidential data transmission becomes difficult when considering massive numbers of low cost and low complexity devices. This is where physical layer security can play an important role. The idea is, instead of having reliability encoding, i.e., error control coding separated from the encryption, we can use joint encoding schemes that provide both reliability and security. This approach, known as wiretap coding, was proposed approximately half a century ago by A. Wyner [17]. Wyner looked at a three terminal wireless channel, i.e., two legitimate users Alice and Bob, and an eavesdropper, Eve. He recognized that the channels between the terminals are not perfect, i.e., their transmission will be impacted by noise. Therefore, when Alice transmits, Bob and Eve will not see exactly what has been transmitted. Moreover, Bob and Eve will have different received signals as they have different noisy channels. Wyner was interested in whether Alice could send a message reliably to Bob, while keeping it secret from Eve. To answer, he looked at the reliable rate to Bob, versus the equivocation at Eve (conditional entropy of the message at Eve’s receiver). Note that, perfect secrecy can be achieved if the reliable rate at which data is being transmitted to Bob equals to the equivocation of Eve. To measure these quantities Wyner introduced a new metric, named secrecy capacity, which is the maximum reliable rate that equals the equivocation. He further showed that, achieving positive secrecy capacity is possible, hence, confidential transmission can be performed without the use of secret keys. However, achieving positive secrecy capacity is possible iff, the measurements at Eve are degraded with respect to those at Bob. A plausible example is when the signal to noise ratio (SNR) at Bob is higher than the SNR at Eve. Now, thinking about the physical layer, it is clear that the properties of radio propagation, i.e., diffusion and superposition, provide opportunities to achieve positive secrecy capacity. For example, by using the natural degradeness over time (e.g., fading), by introducing an artificial degradeness to the eavesdropper (e.g., interference and jamming), or, by leveraging spatial diversity (e.g., multiple antenna systems and relays can create secrecy degrees of freedom). Based on the above, over the last fifteen years the idea of wiretap coding has been further examined considering several fundamental channel models: broadcast channel (one transmitter, multiple receivers), multiple access channel (multiple transmitters, one receiver), interference channels (multiple transmitters, multiple receivers); see e.g. [18]. To illustrate the main results in the area, this work focuses on the broadcast channel [19]. First, consider a Gaussian broadcast channel with Alice being a transmitter and Bob and Eve receivers. Assume two messages are transmitted: $M_{1}$ intended for both receivers and $M_{2}$ a secret message that is intended only for Bob. To define the capacity region we consider a degraded channel at Eve. In particular, it is assumed that the SNR level at Bob equals $10$ dB, and the SNR at Eve is $5$ dB. This is illustrated in Figure 1 where the horizontal axis gives the range of possible rates for the common message $M_{1}$, and the vertical axis gives the range of possible rates for the secret message $M_{2}$. The capacity region without secrecy constraints is shown with red solid curve and the secrecy capacity is indicated by the dashed blue curve. It can be observed that, if secrecy is required, part of the available capacity must be sacrificed in order to confuse the eavesdropper for that message. It is important to note that the amount to be sacrificed depends upon choosing a codeword that randomizes the message w.r.t. Eve, but allows Bob to successfully verify it. Figure 1: Achievable rates for the Gaussian broadcast channel. Next, Figure 2 shows the impact when the SNR at Eve varies. Similarly, the horizontal axis gives the common rate and the vertical axis gives the secrecy rate. The arrow shows that, if the SNR at Eve decreases, the range for the common rate shrinks and the range of secrecy rates increases. On the other hand, if the SNR at Eve reaches $10$ dB, the same level as Bob’s SNR, the secrecy region collapses. That is, if the second receiver is not degraded, secrecy rate becomes zero. Figure 2: Achievable rates for the Gaussian broadcast channel considering variable SNR at Eve. Interestingly, things change when looking at a fading Gaussian broadcast channel. To illustrate this scenario we consider the same model, i.e., one transmitter, two receivers, one common message, and one secret message, but we assume that both the receivers have the same level of Gaussian noise, i.e., Bob and Eve have $5$ dB SNR. This is given in Figure 3. The difference between Bob and Eve is the fading parameter, i.e., Bob’s experiences Rayleigh fading with a unit parameter, and Eve has Rayleigh fading with parameter $\sigma_{2}$. Note, a smaller $\sigma_{2}$, results in more intense fading. As before, when Eve’s channel gets worse, i.e., $\sigma_{2}$ decreases, it can be seen that the range of common rates on the horizontal axis shrinks and the range of secret rates on the vertical axis increases. However, a distinction here is that if the two receivers observe the statistically identical channels (this is the case when $\sigma_{2}=1$), the secrecy capacity does not collapse as in the case of the Gaussian channel. This result holds under the assumption of perfect channel knowledge and follows from the fact that fading provides additional degrees of freedom leading to advantage during the time when other receivers experience deeper fade. Figure 3: Achievable rates for the Rayleigh fading broadcast channel considering variable $\sigma^{2}$ at Eve. (From [19].). A major issue concerning the results above comes from an information theoretic perspective. In particular, they are based on the assumption of infinite coding blocklength. Hence, it concerns the following scenario. Assume that a message $W$, that is encoded into a length-$n$ codeword, is transmitted into the channel. After passing through the wireless medium noisy instances of the codeword are obtained by Bob and Eve. These codewords are then fed into Bob’s and Eve’s decoders. The desired property for this scenario is that for Bob to be able to reconstruct the codeword perfectly while at the same time, the leakage of the codeword to Eve is bounded by the quantity $\delta$. In the original formulation by Wyner, the considered blocklength is infinity, i.e, $n$, the number of channel uses, is infinity. When $n\rightarrow\infty$, the probability of error at Bob, i.e., probability that he decodes to a $\hat{W}$ which is different compared to $W$ goes to zero. Additionally, the information leakage $\delta$ also goes to zero. The secrecy capacity for this case has been formulated as the difference between the mutual information between Alice, $X_{A}$, and Bob, $X_{B}$, and the mutual information between Alice and Eve, $X_{E}$, when considering the maximum from the channel input distribution $P_{X}$, i.e.,: $C_{S}=\underset{P_{X}}{\max}\\{I(X_{A};X_{B})-I(X_{A};X_{E})\\}.$ (1) This is an intuitive result, i.e., achieving positive secrecy capacity relies on the degradation of Eve’s channel. The limitation of this theory is that it gives only asymptotic results that are not suitable for low latency applications, such as in an IoT scenario. This opens the question: What is achievable in the non-asymptotic case?, and the answer depends on the finite blocklength information theory. Assume we have a source $W$, which can take $1,2,\dots,M$ possible values, i.e., it has $\log_{2}M$ bits. The source is mapped using an encoder to a sequence, $X^{n}$, which is then passed through a channel. Due to noise, the receiver will observe a corrupted version of the transmission, i.e., $Y^{n}$, which is then decoded to $\hat{W}$. If the errors between $\hat{W}$ and $W$ are less than a particular value, $\epsilon$, the decoder could reconstruct the original source. In systems like this, the design of $nM\epsilon$ codes is of particular interest: $M$ the number of source symbols, $n$ the number of channel uses, and $\epsilon$ the upper bound on the reconstruction fidelity of the source at the output of the decoder. The fundamental limit for such a system is defined by the maximum $M$, i.e., the largest possible number of source symbols that can be transmitted through the channel in $n$ channel uses and be reconstructed at the decoder with error probability $\leq\epsilon$. Note that, $\underset{n\rightarrow\infty}{\lim}\frac{1}{n}\log_{2}(M)$ gives the Shannon’s capacity where $\epsilon\rightarrow 0$. However, in an actual system $n$ and $\epsilon$ are finite values. Considering this, an approximation for $M^{*}$ was derived in [20], and it is given as $\log M^{*}(n,\epsilon)=nC-\sqrt{n}VQ^{-1}(\epsilon)+\mathcal{O}(\log n),$ (2) where $C$ gives the Shannon’s capacity, $Q^{-1}(\epsilon)$ defines the tail of a standard Gaussian distribution evaluated at $\epsilon$, and $V$ is the channel dispersion, which is the variance of the information density (note that Shannon’s capacity is the mean of the information density). The result from Equation (2) is illustrated in Figure 4, where an AWGN channel is assumed with SNR equal to $0$ dB, $\epsilon=10^{-3}$ and $C=1/2$. The figure shows the upper bound and lower bound for the capacity for finite block lengths, denoted here by “Converse”, and “Best achievability”, respectively. Hence, the actual capacity, which remains to be found, lies between those two curves. While the gap between the curves is small for high values of $n$, it can be observed that for small values of $n$ the gap remains large, hence, further work in the area is required to obtain a more precise solution. Figure 4: Upper and lower bounds on the capacity regions for short block length communication. SNR is equal to $0$ dB and $\epsilon=10^{-3}$. (From [20]) Following the result for channel capacity, it has been just recently shown that the secrecy capacity in the finite blocklength scenario can also be approximated [21]. Fixing the error probability at Bob, $\epsilon$, the leakage at Eve, $\delta$, and the block length $n$, an approximation for the secrecy capacity is given as $R^{*}(n,\epsilon,\delta)=C_{S}-\sqrt{\frac{V}{n}}Q^{-1}\left(\frac{\delta}{1-\epsilon}\right)+\mathcal{O}\left(\frac{\log n}{n}\right),$ (3) where $V$ is defined similarly to the channel dispersion of (2). The result from Equation (3) is illustrated in Figure 5. The figure considers a binary symmetric wiretap channel with crossover probability $p=0.11$, $\delta=\epsilon=10^{-3}$ and $C_{S}=1/2$. A similar trend is observed as in the previous figure, the gap between upper bound (Converse) and lower bound (Best achievability) shrinks and widens as $n$ gets larger or smaller, respectively. Figure 5: Upper and lower bounds on the secrecy capacity for short block length communication in a binary symmetric channel with crossover probability $p=0.11$ and $\delta=\epsilon=10^{-3}$. (From [21].) This has also been evaluated for a Gaussian wiretap channel and the result is illustrated in Figure 6. The SNR at Bob here equals $3$ dB, and the SNR at Eve equals $-3$ dB. It can be observed that the gap between achievability and converse is even larger for this scenario. However, what is important to mention here is that the upper bound, when considering finite block lengths, is far from the asymptotic secrecy capacity, $C_{S}$. This shows that research on emerging IoT technologies should not rely on asymptotic results and should focus on the investigation of short block length communications. Figure 6: Upper and lower bounds on the secrecy capacity for short block length communication in a Gaussian wiretap channel. SNR at Bob is equal to $3$ dB and SNR at Eve equals to $-3$ dB. (From [21].) ### IV-B Privacy in sensing systems Differently from secrecy, where the concern is about restricting a malicious party from getting access to the transmission, in the case of privacy, the goal is to keep part of the information secret from other parties, including the legitimate receiver (Bob). A simple way to ensure there is no privacy leakage is to deny access to Bob, however, without having a recipient the data source becomes useless. Therefore, it is important to study, which part of the data can be shared, such that the message is successfully and securely transmitted, while the privacy leakage is minimized. This section focuses on the problem of privacy leakage with particular focus on sensing systems. Such systems include smart meters, cameras, motion sensors, i.e., devices that generate useful data for companies who provide users with particular service (alarm, power supply, etc.). While companies can use the data to improve their services, the full access to it endangers the privacy of users. The above hints towards that, there is a fundamental trade-off between privacy and usefulness of data (distortion). This is illustrated in Figure 7. If the data is completely private, i.e., its equivocation at Bob is high, the data becomes useless and it is fully distorted. Contrarily, if the data is fully accessible, i.e., it has low distortion at Bob, then its equivocation goes to zero and the data is not private. Figure 7: Trade-off between privacy and usefulness of data. Now, when considering a specific application, i.e., smart meters, the trade- off can be specified as follows: a smart meter measures the electricity usage in almost real time, hence, having the utility of providing users with information on their usage, but in the same time it leaks this information to the power supply company who can use it to trace in-home activities [22]. One way to model this problem is through a hidden Gauss-Markov model. This is given in Figure 8 where the hidden state is the intermittent state, e.g., turning your toaster on, your kettle on, etc. The figure captures a smart meter trace, and shows that the privacy-utility trade-off for this model can be characterized by a reverse water-filling [23]. The trade-off here is defined by the water level $\phi$, such that all signals with power lower than $\phi$ are being suppressed by the meter, while all signals above are being be transmitted (and leaked) by the meter. Therefore, the value of $\phi$ defines the amount of privacy that the user is willing to sacrifice to increase his utility. Figure 8: Privacy-utility trade-off characterized by a reverse water-filling. Another way to approach the same problem is through using control, i.e., actively controlling what the meter sees based on storage and energy harvesting [24]. This is illustrated in Figure 9, where the utility-privacy trade-off for this model is captured by measuring wasted energy versus information leakage. Presenting this control approach as a Markov model allows to numerically determine the efficient frontier. This is given in Figure 10, where the red curve gives the optimal trade-off of wasted power versus information leakage. Figure 9: Privacy-utility trade-off characterized by a measuring wasted energy versus information leakage. Figure 10: Wasted power versus information leakage when considering a control approach. Another example is when considering the case of competitive privacy. In competitive privacy, there are multiple agents (Bobs) each having own privacy utility trade-offs. On one hand, there are multiple interacting agents who are competing with one another, but, on the other hand, the agents have coupled measurements. In detail, each agent wants to estimate its own parameters and can help other agents by sharing data but does not want to compromise his own privacy. This competitive scenario can be represented as a linear measurement model [25]. Utility can be measured in terms of mean squared error on the state estimation and privacy can be measured in terms of information leakage. In fact, it has been shown that this reduces to a classical problem, known as the Wyner-Ziv problem or the distributed source coding problem. Particularly, it has not been discussed what is the optimal amount of information that must be exchanged, but it has been shown that the optimal way to exchange information is by using Wyner-Ziv coding. Next, depending on the scenario a simple way to find the optimal amount of information is through the use of game theory. Finally, an important conclusion for this section is that information theory can help us understand the fundamental limits of security and privacy. While mainly theoretical constructs have been discussed, it is clear that there is a need to connect the theoretical analyses to real networks. Building on the above, some emerging research directions include finite blocklength analysis (short packet low latency communication), scaling laws for large networks (channel models that consider massive networks) and practical coding schemes. ## V Secret Key Generation Using PLS This section focuses on several aspects concerning SKG. First, it provides an overview on how to extract symmetric keys from shared randomness, then it shows how SKG can be incorporated in actual crypto systems, and finally, it discusses how the SKG process can be made resilient to active attacks. ### V-A Secret key generation Generally, the SKG protocol consists of three steps: advantage distillation, information reconciliation, and, privacy amplification. Assuming two legitimate parties, e.g., Alice and Bob, the steps can be summarized as follows. In the first step, Alice and Bob exchange pilot signals during the coherence time of the channel, and obtain correlated observations $Z_{A}$ and $Z_{B}$, respectively. In the second step, their observations are first quantized and then passed through a distributed source code type of decoder. During this step Alice (or Bob) shares side information, which is used by Bob to correct errors at the output of his decoder. Hence, at the end of this step both parties obtain a common binary sequence. Finally, to produce a maximum entropy key and suppress the leaked information, privacy amplification is performed. In this last step, Alice and Bob apply an irreversible compression function (e.g., hash function) over the reconciled bit sequence. This produces a uniform key that is unobservable by adversaries. There are few important points that need to be taken into account for the success of the SKG process. First, channel measurements represent a mixture of large scale and small scale fading components. In multiple studies, it has been demonstrated that the large scale component is strongly dependent on the location and the distance between users, which makes it predictable for eavesdroppers. Therefore, to distill a secret key, Alice and Bob should either remove this part from their measurements and generate the key using the unpredictable small scale components or should compress more at the privacy amplification. This point is further discussed in Section VII. Second, the SKG protocol should follow all the steps described above, and no steps should be skipped. As an example, skipping the privacy amplification would give Alice and Bob longer key sequence, however, the key sequence is vulnerable to different attacks [26]. Third, it is important that, Alice and Bob do not transmit information related to their observations, as this could be exposed to eavesdroppers in the vicinity. Forth, Alice and Bob should respect the coherence time and coherence bandwidth of the channel, such that their subsequent measurements are decorrelated in time and frequency. This allows them to generate random and unpredictable bit sequences. Finally, as mentioned in the previous section, further testing of short blocklength encoders is necessary in order to identify the optimal solution for SKG. Regarding the last point, Figures 11 and 12 show a comparison between an upper bound, evaluated in [27], versus information reconciliation rates achieved using of LDPC, polar codes and BCH codes [28]. Both figures $n=128$ and $n=512$ show that polar codes with CRC and BCH codes with list decoding outperform the other approaches, making them good candidates for reconciliation decoding. Note that such type of encoders are already used in 5G for different purposes. Figure 11: FER performance of reconciliation codes compared to the lower bound from [27] for $n=128$. (From [28].) Figure 12: FER performance of reconciliation codes compared to the lower bound from [27] for $n=512$. (From [28].) ### V-B Secret key generation in hybrid crypto systems Building on the above, we continue with a particular example on how SKG can be incorporated in hybrid security cryptographic schemes. In detail, it will be discussed how to build a SKG-based authenticated encryption. Three ingredients are needed to formulate this problem: * • A SKG scheme $\verb"G":\mathbb{C}\rightarrow\mathcal{K}\times\mathcal{S}$, that takes channel measurements as input and generates a key $\mathbf{k}$ and side information $\mathbf{s}$. * • A symmetric encryption algorithm, i.e., a pair of functions $\verb"Es":\mathcal{K}\times\mathcal{M}\rightarrow\mathcal{C_{T}}$ and $\verb"Ds":\mathcal{K}\times\mathcal{C_{T}}\rightarrow\mathcal{M}$, for encryption and decryption, respectively, where $\mathcal{C_{T}}$ defines the ciphertext space and $\mathcal{M}$ the message space. * • A message authentication code (MAC) algorithm, given as $\verb"Sign":\mathcal{K}\times\mathcal{M}\rightarrow\mathcal{T}$, for signing and $\verb"Ver":\mathcal{K}\times\mathcal{M}\times\mathcal{T}\rightarrow(yes,no)$, for verification, where $\mathcal{T}$ defines the tag space. Now, the components can be combined as follows: 1. 1. SKG is performed between Alice and Bob as: $\verb"G"(\mathbf{h})=\left(\mathbf{k},\mathbf{s_{A}}\right),$ (4) where $\mathbf{h}$ represents the channel measurements, $\mathbf{k}$ the generated key after privacy amplification and $\mathbf{s_{A}}$ is Alice’s side information that has to be transmitted to Bob to finalize the process. 2. 2. Before transmitting $\mathbf{s_{A}}$ to Bob, Alice breaks her key into two parts $\mathbf{k}=\\{\mathbf{k}_{e},\mathbf{k}_{i}\\}$, generates a ciphertext as $\mathbf{c}=\verb"Es"(\mathbf{k}_{e},\mathbf{m})$ and signs it as $\mathbf{t}=\verb"Sign"(\mathbf{k}_{i},\mathbf{c})$. Afterwards she transmits to Bob the concatenation of $[\mathbf{s}_{A}||\mathbf{c}||\mathbf{t}]$, i.e., in a single message she can transmit the side information and her message. 3. 3. Upon receiving the above, Bob uses the side information $\mathbf{s}_{A}$, to finish the SKG process, i.e., to obtains the key $\mathbf{k}$. Then, he checks the integrity of the received ciphertext as $\verb"Ver"(\mathbf{k}_{i},\mathbf{c},\mathbf{t})$ and if successful he decrypts and obtain the message $\mathbf{m}$. Differently from the standard SKG scheme, where SKG is performed in parallel at both nodes and data exchange happens only after the key generation is finalized, in the scheme above Alice completes the SKG locally and then transmits in a single go the ciphertext, the tag, and, the side information (e.g., syndrome). Then Bob uses the syndrome to complete the SKG and performs the authenticated decryption. This small change in the standard procedure shows how PLS can be easily combined with standard crypto schemes. Such approaches bring new opportunities. For example, the scheme above opens the problem of transmission optimization. Consider a scenario with multiple subcarriers used for transmission. The subcarriers can then be split into two subsets, a subset $\mathcal{D}$ used for transmitting encrypted data and a subset $\bar{\mathcal{D}}$ used for transmitting side information (syndromes). This transmission scheme can be optimized considering several constraints. The first constraint comes from the world of cryptography, i.e., based on the choice of cryptographic cipher we can define the amount of data to be encrypted with a single key. This can be captured by the following constraint: $C_{SKG}\geq\beta C_{D},\quad 0<\beta\leq 1,$ (5) where $C_{SKG}$ defines the key generation rate, $C_{D}$ defines the data rate and $\beta$ is a quantity that relates the key size to the data size that will be encrypted, e.g., $\beta=1$ corresponds to a one-time pad cipher. The second constraint comes from the world of information theory. It relates the necessary (side information) syndrome rate $C_{R}$ and the SKG rate as follows: $C_{R}\geq\kappa C_{SKG},$ (6) where $\kappa$ defines minimum number of reconciliation bits with respect to the key bits. It is a parameter defined by the type of the encoder/decoder used for SKG, e.g., for a $\frac{k}{n}$ block encoder $\kappa=\frac{n-k}{k}$. Further constraints that can be incorporated are power constraint: $\sum_{j=1}^{N}p_{j}\leq NP,\;\;p_{j}\geq 0,\;\forall j\in\\{1,\dots,N\\},$ (7) and a channel capacity constraint, i.e., $C_{D}+C_{R}\leq C,$ (8) where $N$ gives the number of subcarriers, $P$ is the power limit per subcarrier and $C$ is the total capacity of the channel. The objective of the problem can then be defined as: $\max_{p_{j},j\in\mathcal{D}}C_{D}\quad\text{s.t. \eqref{eq:beta_constraint}, \eqref{eq:kappa}, \eqref{eq:power}, and \eqref{eq:C}}$ (9) The problem can be turned into a combinatorial optimization problem which can be solved optimally using dynamic programming techniques or sub-optimally using heuristic approaches. Overall, this problem shows how physical layer aspects can be related to cryptographic schemes, in the form of a hybrid security scheme, and provide new opportunities for cross layer optimization. The problem was solved in [29] and the main result is depicted in Figure 13. The figure shows the long term efficiency (expected sum data rate normalized to the capacity of the channel) of the proposed parallel approach, i.e., the transmission of side information and encrypted data are done simultaneously on $\mathcal{\bar{D}}$ and $\mathcal{D}$, respectively, versus a standard sequential transmission approach. It can be seen that, for most values of $\beta$, the parallel approach outperforms the sequential one. Another observations is that as $\beta$ increases, the efficiency decreases. This is expected result as higher $\beta$ will required more frequent key generation, hence, less data transmission. Finally, an important result that can be observed on the graph is that the authors proposed a simple heuristic approach for the parallel scheme that gives an equivalent efficiency to the optimal solution solved using dynamic programming approach (i.e., as a Knapsack problem). Further interesting aspects that can be included in this analysis are factors such as handover or other aspects that may cause frequent key generation. Figure 13: Efficiency comparison for $N=64$, SNR$=10$ dB and $\kappa=2$. (From [29].) This problem has been further investigated in [30], where a general quality of service (QoS) delay constraint was introduced. The work is based on leveraging the theory of the effective capacity and identifies the maximum supported transmission rate when considering a delay constraints, i.e., instead of maximizing the data rate $C_{D}$ the problem focuses on maximizing the effective data rate $E_{C}(\alpha)$, given as $E_{C}(\alpha)=-\frac{1}{\alpha}\log_{2}\left(\mathbb{E}\left[e^{-\alpha C_{D}}\right]\right),$ (10) where $\alpha=\frac{\theta T_{f}B}{\ln(2)}$ with $\theta$ being a MAC sub- layer parameter that captures the packet arrival rate and introduces a delay requirement into the problem, $T_{f}$ is the frame duration and $B$ denotes the bandwidth. Considering that, [30] identified the optimal power allocation policy that maximizes $E_{C}(\alpha)$ as $p_{i}^{*}=\frac{1}{g_{0}^{\frac{N}{\alpha+N}}\hat{g}_{i}^{\frac{\alpha}{\alpha+N}}}-\frac{1}{\hat{g}_{i}},$ (11) where $g_{0}$ is a cut-off value that can be found from the power constraint and $\hat{g}_{i}$ $i=1,\dots,N$ denote the imperfectly estimated channel gains. If the system can tolerate looser delay requirements, i.e., $\theta\rightarrow 0$ the result above converges to the well-known water- filling algorithm and if stringent delay constraints are implied, i.e., $\theta\rightarrow\infty$ the optimal power allocation converges to total channel inversion. Similarly to the previous case, it has been demonstrated that the parallel approach outperforms the sequential approach, in terms of efficiency, regardless of the values of $\theta$ and $\beta$ [30]. ### V-C Secret key generation under active attacks The previous section discussed how SKG can be used to build authenticated encryption protocols. However, the above scheme could only be secure under the assumption that the advantage distillation phase is robust against active attacks. Therefore, this section focuses on active attacks during SKG, in particular the injection attack is investigated. The idea of this attack is illustrated in Figure 14. AliceMalloryBob$XH$$W=\mathbf{H_{A}}^{T}\mathbf{P}\ X_{J}$$W=\mathbf{H_{B}}^{T}\mathbf{P}\ X_{J}$ Figure 14: Alice and Bob have single transmit and receive antennas and exchange pilot signals $X$ over a Rayleigh fading channel $H$. A MiM, Mallory, with multiple transmit antennas injects a pre-coded signal $\mathbf{P}X_{J}$, such that the received signals at Alice and Bob are equal $W=\mathbf{H_{A}}^{T}\mathbf{P}=\mathbf{H_{B}}^{T}\mathbf{P}$. Differently from previous sections, instead of an eavesdropper, an active man- in-the-middle (MiM) attacker is considered, referred to as Mallory. The system model assumes two legitimate users, Alice and Bob, each having a single antenna and Mallory, who has two antennas. The goal of the attacker is to inject an equivalent signal $W$ at both, Alice and Bob, such that their channel observations $Z_{A}$ and $Z_{B}$, respectively, will also include the injected signal: $\displaystyle Z_{A}=XH+W+N_{A}$ (12) $\displaystyle Z_{B}=XH+W+N_{B},$ (13) where the channel realization between Alice-Bob is denoted by $H\sim\mathcal{CN}(0,\sigma^{2})$, the exchanged signal over this channel is given as $X$, $\mathbb{E}[|X|^{2}]\leq P$, the noise observations at Alice and Bob are given as $N_{A},N_{B}\sim\mathcal{CN}\left(0,1\right)$ and the injected signals over the link Eve-Alice (given as $\mathbf{H}_{A}$) and Eve- Bob (given as $\mathbf{H}_{B}$) are given as $W=\mathbf{H_{A}}^{T}\mathbf{P}X_{J}=\mathbf{H_{B}}^{T}\mathbf{P}X_{J}$. The received signals are equal, thanks to the precoding matrix $\mathbf{P}$. A simple mathematical operation can reveal that, as long as Mallory has one extra antenna, as compared to Alice and Bob, the design of the pre-coding matrix is straight forward, i.e., $\displaystyle\mathbf{H_{A}}^{T}\mathbf{P}X_{J}$ $\displaystyle=\mathbf{H_{B}}^{T}\mathbf{P}X_{J}\Rightarrow$ $\displaystyle P_{1}$ $\displaystyle=\frac{H_{B2}-H_{A2}}{H_{A1}-H_{B1}}P_{2}.$ (14) Overall, this is a simple attack to mount its consequences are crucial. As it can be seen in Equations (12) and (13), by injecting the signals, Mallory adds additional term to the shared randomness between Alice and Bob, turning it into $XH+W$. Hence, this allows Mallory to obtain partial information with respect to the generated key. Fortunately, a simplistic countermeasure has been proposed in [11]. The idea is instead of using deterministic pilot signals $X$, as described above, Alice and Bob can transmit independent and randomized probe signals $X$ and $Y$, respectively. This turns their observations into $\displaystyle Z_{A}=YH+W+N_{A},$ (15) $\displaystyle Z_{B}=XH+W+N_{B},$ (16) which allows them to simply post-multiply by their own transmission resulting into the following: $\displaystyle\tilde{Z}_{A}$ $\displaystyle=$ $\displaystyle XZ_{A}=XYH+XW+XN_{A},$ (17) $\displaystyle\tilde{Z}_{B}$ $\displaystyle=$ $\displaystyle YZ_{B}=XYH+YW+YN_{B},$ (18) where, as it can be seen, $W$ is not anymore part of the shared randomness. Therefore, as long as $X$ and $Y$ are uncorrelated this simple approach can successfully reduce an injection attack to a less harmful uncorrelated jamming attack. In detail, the jamming attack has impact on the achievable key rate but does not reveal anything about the key to Mallory. Now, when Mallory’s attack is reduced to jamming, a smart thing she can do, is to act as a reactive jammer. A reactive jammer would first sense the spectrum and jam only subcarriers where she detects a transmission. Considering a multicarrier system, Mallory can choose a sensing threshold and jam only subcarriers where she detects signals with power greater than the chosen threshold. A thorough analysis considering this scenario has be performed in [11], where this problem has been investigated using game theory. In fact, the scenario can be formulated as a non-cooperative zero-sum game with two players, i.e., player $L$, (legitimate users act as a single player), and player $J$, (the jammer). Based on the fact that player $J$ jams only after observing the action from player $L$, this is formed as a hierarchical game with $L$ being the leader of the game and $J$ being the follower. Note that in hierarchical games, the optimal action is the Stackelberg equilibrium (SE). What was shown in this study is that the SE is based on two things: i) the sensitivity of the receiver at player $J$, and more specifically how well the sensing threshold is chosen, and ii) the available power at the legitimate users. The SE is defined as: * • If the jammer has badly chosen threshold, depending on the available power at the legitimate users they would optimally: 1. 1. equally distribute their power below the sensing threshold and do not comprise their communication. 2. 2. transmit with full power on all subcarriers, hence being sensed and jammed. * • If the jammer has chosen a low threshold that allows to detect all ongoing transmissions, Alice and Bob have no choice but to transmit at full power. Overall, SKG is a promising PLS technology and could help solving the key distribution issue for emerging 6G applications, e.g., addressing scalability for massive IoT [31]. ## VI Authentication Using PLS One of the main motivations to look at PLS authentication schemes is the increasing complexity of standard crypto schemes. In fact, it has been shown in multiple studies that there exists a trade-off between delay and key sizes used in the cryptographic schemes. A particular example that focuses on addressing such issues is the zero-round- trip-time (0-RTT) protocol introduced in the TLS version $1.3$ for session resumption. The idea is based on using resumption keys to quickly resume a session, in a 0-RTT, as opposed to re-authenticating users every subsequent session. Unfortunately, it has been shown that this scheme is vulnerable a set of attacks (e.g., replay attack), however, the community answer was “But too big a win not to do” [32]. This section gives a hint on what PLS can do in terms of authentication for 6G systems. In particular, it first gives a brief background on physical unclonable functions (PUFs), then discusses how localization can be used as an authentication factor, and finally, it introduces a secure 0-RTT authentication protocol that leverages multiple PLS technologies. ### VI-A Physical unclonable functions PUFs can be referred to as device fingerprints. The idea is that, the manufacturing of a circuit is a process with unique characteristics (e.g., due to change in the temperature, vibrations), which makes each device unique on its own. While devices operate in a similar manner, they always have small variations in terms of delays, power-on-state, jitter, etc. This gives an opportunity to leverage these uniqueness, and use it for authentication. Given that, a standard PUF based authentication protocol follows two phases. An enrolment phase which takes place offline, and an authentication phase which is performed online. During the enrolment phase, a set of challenges are run on a device’s PUF. A set of challenge could refer to measuring propagation delays over different propagation paths. Due to the presence of noise, these measurements are passed through a suitable encoder to generate helper data. Following that, a verifier (e.g., a server) creates a database where challenge-response pairs (CRPs) are stored along with the corresponding helper data. Next, during the online authentication phase, the verifier sends a random challenge to the device, and the device replies with a new PUF measurement. The authentication is successful if the verifier can regenerate the response saved during enrolment by using the new response and the helper data in its database. Note that, to avoid replay type of attacks a CRP should not be re-used. A major advantage of the scheme above is that the device does not need to store any key information and relies only on PUF measurements. Hence, if the device is compromised (e.g., “captured by an enemy”), no useful information can be extracted. ### VI-B Location-based authentication Localization precision is continuously increasing and the goal of 6G technologies is to achieve centimeter level accuracy. Popular approaches for fingerprinting rely on measuring received signal strength (RSS), carrier frequency offsets, I-Q imbalances, CSI measurements and more. This section presents a lightweight example for location based authentication, through a low-complexity proximity estimation. Consider a mobile low-end device with a single antenna and low computational power. Assume that the device has a map of a premise and knows the location of the access points within this premise. A simple strategy to perform reverse authentication (i.e., the device authenticates an access point) is to move in an unpredictable manner and measure the RSS from multiple positions. As the RSS is strongly related to the distance between devices, this simple approach allows to confirm the location of the access point. Typically, localization would require either the deployment of multiple nodes that measure the RSS simultaneously or advanced hardware/computational capabilities when considering a single device. The approach above does not have such requirements and can still be used as an authentication factor. In fact, the proximity detection described above can provide resilience to impersonation type of attacks, e.g., in the presence of a malicious access point. Now, we summarize some open research issues in the direction of using fingerprint based authentication. A concern that naturally arises is about the resilience of such schemes to jamming and man-in-the-middle type of attacks. In particular, how to cope with interference transmissions, or pilot contamination type of attacks, both of which can alter the precision of the localization information. Another issue concerns the trustworthiness of the localization information, i.e., depending whether we operate at short or long distance, the variability of measurements can change, hence, bringing uncertainty into the system. Finally, another aspect concerns the type of application where such approach could be useful. A good example comes from the idea presented above, e.g., reverse authentication. Reverse authentication can help in mitigating attacks that fall into the general category of false base station attacks (which are open issues in 5G). However, we note that before deploying location-based authentication technologies all concerns must be addressed. ### VI-C Multi-factor PLS authentication A recent publication [14], has shown how three PLS credentials (PUFs, SKG and location fingerprints) can be combined into a multi-factor PLS based authentication protocol. The proposed scheme uses PUFs as a mutual authentication factor between a mobile node (Alice) and a static server (Bob). The protocol is realized following a typical PUF approach, i.e., following two steps, enrolment and authentication. The use of PUFs provides several security guaranties, including protection against physical and cloning attacks. Next, Alice uses proximity estimation as a second authentication factor. This simple technique re-assures her for the legitimacy of Bob and provides resistance to impersonation attacks (e.g., false base station attacks). To provide anonymity for Alice, the scheme introduces one-time alias IDs. After a successful authentication, both parties exchange resumption secrets, following a standard TLS 1.3 procedure. The resumption secrets are used for a fast 0-RTT re- authentication between Alice and Bob, i.e., session resumption (as opposed to performing a full authentication procedure). While the standard approach for session resumption is not forward secure and is vulnerable to replay attacks, the scheme in [14] uses SKG keys to randomize the resumption secrets. It is shown that adding SKG ensures both perfect forward security and resistance against replay attacks. In general, using the physical layer for authentication is a well investigated topic. Schemes like the one above, show that there are already multiple PHY schemes which can contribute for the system security. Some of the research problems in the area include design of high-entropy PUFs and accurate and privacy-preserving location-based authentication. ## VII Conclusions and Future Directions This paper highlights the role that PLS could play in 6G, in view of the evolution in terms of security, with the concepts of trust, context awareness, and quality of security. 6G is expected to introduce new features to communication standards including sensing, subTHz communication, massive MIMO, extreme beamforming, learning and actuating, ultra reliable low latency computing and more. While it is still not clear how the transition from 5G to 6G will look like, there is growing interest on the use of semantics, semantic communications, semantic compression, and context awareness in 6G. Another perspective was introduced with quality of security (QoSec), i.e., different slices of the network have different security and privacy requirements. This brings the need of adaptive security levels. A series of questions arise based on the above: How to define other security levels? How to perform adaptive identity management? How to make an intelligent risk assessment? PLS emerges as a contestant for the next generation of security systems in 6G. One key advantage of PLS is that it is inherently adaptive. This is due to the fact that in physical technologies, the secrecy outage probability can be directly tuned through adjusting the transmission rate. In particular, wireless channels can be treated as a source of two things, a source of uniqueness, and a source of entropy. For example, in a slow flat fading scenario (e.g. LoS) then the channel could be treated as a good source of uniqueness. As discussed in Section VI, uniqueness can be easily used for authentication purposes. On the other hand, if the channel changes very fast, due to small scale fading, it could be treated as a good source of entropy. The variability of the channel can then be directly used to either distill keys, or perform keyless transmission. An important observation is that if one is not available, e.g., uniqueness, then the other will be, e.g., entropy. Following the above, an open research question is, how to characterize the channel properties and particularly, which part of the channel should be considered as predictable and which as unpredictable. It is not an easy question to answer as it would require the characterization of the channel correlations in time, frequency and space domains; but it is an important one as it would allow the alignment of PLS metrics to semantic security metrics. Finally, we believe it is now time to start defining the security levels based on the usage of multiple elements. Here, we list several elements: 1. 1. Criticality of information - how important the information is from user or the network perspective; 2. 2. Value of information for the attacker - this captures, who is the attacker and how much effort is expected to put into compromising the system; 3. 3. System resilience - this includes the stability and repair time after an attack; 4. 4. Threat level - the usage of context to recognize “abnormal” events (could include location, behavior and communication information); 5. 5. QoS constraint - systems are expected to comply with particular QoS index. Today, the deployment of PLS in systems is still lacking traction. However, there is a growing interest by industry and academia. This paper shows the potential of PLS for upcoming wireless system designs. It gives concrete examples of use cases for PLS, reaching far beyond addressing encryption. By doing so, greatly improving the security of 6G networks. For PLS it is instrumental to characterize and exploit the wireless channel from a security point of view. A key advantage is seen for developing light-weight security solutions for low-latency and massive IoT use cases. ## Acknowledgement This work is financed by the Saxon State government out of the State budget approved by the Saxon State Parliament. ## References * [1] 3GPP TR33.809, “Study on 5G security enhancements against false base stations (Rel 16),” 2018. * [2] “3GPP TR 33.825 V0.3.0, Study on the Security for 5G URLLC (Release 16),” 3rd Generation Partnership Project; Technical Specification Group Services and System Aspects. * [3] M. Bloch and J. Barros, _Physical-Layer Security: From Information Theory to Security Engineering_. Cambridge University Press, 2011. * [4] M. Latva-aho and K. Leppänen, “Key drivers and research challenges for 6G ubiquitous wireless intelligence,” October 2019, published online by the University of Oulu. * [5] V. Ziegler, P. Schneider, H. Viswanathan, M. Montag, S. Kanugovi, and A. Rezaki, “Security and trust in the 6G era,” _IEEE Access_ , vol. 9, pp. 142 314–142 327, 2021. * [6] “Deliverable D1.3 Targets and requirements for 6G - initial E2E architecture, 2022,” A flagship for B5G/6G vision and intelligent fabric of technology enablers connecting human, physical, and digital worlds, Hexa-X. * [7] “International Network Generations Roadmap (INGR),” IEEE , available online https://futurenetworks.ieee.org/roadmap. * [8] L. Sanguinetti, E. Björnson, and J. Hoydis, “Toward massive MIMO 2.0: Understanding spatial correlation, interference suppression, and pilot contamination,” _IEEE Transactions on Communications_ , vol. 68, no. 1, pp. 232–257, 2020. * [9] G. A. N. Segura, A. Chorti, and C. B. Margi, “Centralized and distributed intrusion detection for resource-constrained wireless SDN networks,” _IEEE Internet of Things Journal_ , vol. 9, no. 10, pp. 7746–7758, 2022. * [10] A. Chorti, A. N. Barreto, S. Köpsell, M. Zoli, M. Chafii, P. Sehier, G. Fettweis, and H. V. Poor, “Context-aware security for 6G wireless: The role of physical layer security,” _IEEE Communications Standards Magazine_ , vol. 6, no. 1, pp. 102–108, 2022. * [11] M. Mitev, A. Chorti, E. V. Belmega, and M. Reed, “Man-in-the-middle and denial of service attacks in wireless secret key generation,” in _IEEE Global Communication Conference (GLOBECOM)_ , Dec 2019, pp. 1–6. * [12] E. Belmega and A. Chorti, “Protecting secret key generation systems against jamming: Energy harvesting and channel hopping approaches,” _IEEE Transactions on Information Forensics and Security_ , vol. 12, no. 11, pp. 2611–2626, 2017. * [13] L. Xiao, Q. Yan, W. Lou, G. Chen, and Y. T. Hou, “Proximity-based security techniques for mobile users in wireless networks,” _IEEE Transactions on Information Forensics and Security_ , vol. 8, no. 12, pp. 2089–2100, 2013. * [14] M. Mitev, M. Shakiba-Herfeh, A. Chorti, M. Reed, and S. Baghaee, “A physical layer, zero-round-trip-time, multifactor authentication protocol,” _IEEE Access_ , vol. 10, pp. 74 555–74 571, 2022. * [15] C. E. Shannon, “Communication theory of secrecy systems,” _The Bell System Technical Journal_ , vol. 28, no. 4, pp. 656–715, 1949. * [16] A. Teniou and B. A. Bensaber, “Efficient and dynamic elliptic curve qu‐vanstone implicit certificates distribution scheme for vehicular cloud networks,” _Security and Privacy_ , Jan. 2018. * [17] A. D. Wyner, “The wire-tap channel,” _The Bell System Technical Journal_ , vol. 54, no. 8, pp. 1355–1387, 1975. * [18] H. V. Poor and R. F. Schaefer, “Wireless physical layer security,” _Proceedings of the National Academy of Sciences_ , vol. 114, no. 1, pp. 19–26, 2017. * [19] Y. Liang, H. V. Poor, and S. Shamai, “Secure communication over fading channels,” _IEEE Transactions on Information Theory_ , vol. 54, no. 6, pp. 2470–2492, 2008. * [20] Y. Polyanskiy, H. V. Poor, and S. Verdu, “Channel coding rate in the finite blocklength regime,” _IEEE Transactions on Information Theory_ , vol. 56, no. 5, pp. 2307–2359, 2010. * [21] W. Yang, R. F. Schaefer, and H. V. Poor, “Wiretap channels: Nonasymptotic fundamental limits,” _IEEE Transactions on Information Theory_ , vol. 65, no. 7, pp. 4069–4093, 2019. * [22] G. Giaconi, D. Gunduz, and H. V. Poor, “Privacy-aware smart metering: Progress and challenges,” _IEEE Signal Processing Magazine_ , vol. 35, no. 6, pp. 59–78, 2018. * [23] L. Sankar, S. R. Rajagopalan, S. Mohajer, and H. Poor, “Smart meter privacy: A theoretical framework,” _IEEE Transactions on Smart Grid_ , vol. 4, no. 2, pp. 837–846, 2013. * [24] G. Giaconi, D. Gündüz, and H. V. Poor, “Privacy-cost trade-offs in smart electricity metering systems,” _IET Smart Grid_ , vol. 3, no. 5, pp. 596–604, 2020. * [25] E. V. Belmega, L. Sankar, and H. V. Poor, “Enabling data exchange in two-agent interactive systems under privacy constraints,” _IEEE Journal of Selected Topics in Signal Processing_ , vol. 9, no. 7, pp. 1285–1297, 2015. * [26] C. H. Bennett, G. Brassard, C. Crepeau, and U. M. Maurer, “Generalized privacy amplification,” _IEEE Transactions on Information Theory_ , vol. 41, no. 6, 1995. * [27] H. Tyagi and S. Watanabe, “Converses for secret key agreement and secure computing,” _IEEE Transactions on Information Theory_ , vol. 61, no. 9, 2015\. * [28] M. Shakiba-Herfeh and A. Chorti, “Comparison of short blocklength Slepian-Wolf coding for key reconciliation,” in _2021 IEEE Statistical Signal Processing Workshop (SSP)_ , 2021, pp. 111–115. * [29] M. Mitev, A. Chorti, and M. Reed, “Subcarrier scheduling for joint data transfer and key generation schemes in multicarrier systems,” in _IEEE Global Communication Conference (GLOBECOM)_ , Dec 2019, pp. 1–6. * [30] M. Mitev, A. Chorti, M. Reed, L. Musavian, “Authenticated secret key generation in delay-constrained wireless systems,” _EURASIP Journal on Wireless Communications and Networking_ , no. 122, June 2020. * [31] M. Mitev, T. M. Pham, A. Chorti, A. N. Barreto, and G. Fettweis, “Physical Layer Security - from Theory to Practice,” _TechRxiv preprint_ , 11 2022\. * [32] E. Rescorla, “TLS 1.3,” Stanford University, November 2015.
# Locally Frobenius algebras and Hopf algebras Andrew Baker School of Mathematics & Statistics, University of Glasgow, Glasgow G12 8QQ, Scotland<EMAIL_ADDRESS>http://www.maths.gla.ac.uk/$\sim$ajb (Date: this version 26/12/2022 – version 3 arXiv:2212.00437 ) ###### Abstract. We develop a theory of _locally Frobenius algebras_ which are colimits of certain directed systems of Frobenius algebras. A major goal is to obtain analogues of the work of Moore & Peterson and Margolis on _nearly Frobenius algebras_ and _$P$ -algebras_ which was applied to graded Hopf algebras such as the Steenrod algebra for a prime. Such locally Frobenius algebras are coherent and in studying their modules we are naturally led to focus on coherent and finite dimensional modules. Indeed, the category of coherent modules over locally Frobenius algebra $A$ is abelian with enough projectives and injectives since $A$ is injective relative to the coherent modules; however it only has finite limits and colimits. The finite dimensional modules also form an abelian category but finite dimensional modules are never coherent. The minimal ideals of a locally Frobenius algebra are precisely the ones which are isomorphic to coherent simple modules; in particular it does not contain a copy of any finite dimensional simple module so it is not a Kasch algebra. We discuss possible versions of stable module categories for such algebras. We also discuss possible monoidal structures on module categories of a locally Frobenius Hopf algebra: for example tensor products of coherent modules turn out to be pseudo-coherent. Examples of locally Frobenius Hopf algebras include group algebras of locally finite groups, already intensively studied in the literature. ###### Key words and phrases: Frobenius algebra, Hopf algebra, stable module category ###### 2020 Mathematics Subject Classification: Primary 16T05; Secondary 16S99, 57T05 I would like to thank the following: The Max-Planck-Institut für Mathematik in Bonn for supporting my visit during April and May 2022; Scott Balchin, Tobias Barthel, Ken Brown, Bob Bruner, John Rognes and Chuck Weibel for sharing their mathematical knowledge and insights. This project really took off during the first Covid-19 lock-down in the Spring of 2020, the social isolation in that strange period at least proved conducive to mathematical research ###### Contents 1. 1 Recollections on rings and modules 2. 2 Frobenius algebras and Frobenius extensions 3. 3 Locally Frobenius algebras 4. 4 Modules over a locally Frobenius algebra 5. 5 Stable module categories for locally Frobenius algebras 6. 6 Locally Frobenius Hopf algebras 7. 7 Some examples of locally Frobenius Hopf algebras 1. 7.1 Group algebras of locally finite groups 2. 7.2 Dual profinite group algebras ## Introduction The aim of the paper is to develop a non-graded version of the theories of nearly _Frobenius algebras_ and _$P$ -algebras_ introduced half a century ago by Moore & Peterson, and Margolis [JCM&FPP:NearlyFrobAlg, HRM:Book, AB:Palgebras]; these were motivated by topological applications involving the Steenrod algebra for a prime number. In these graded connected versions, the properties obtained for the algebras and their modules are reminiscent of properties of Poincaré duality algebras (the graded equivalent of Frobenius algebras) and we are able to prove similar results. However there are some difficulties which are overcome in the graded context by concentrating on bounded below modules, and it is not clear how to obtain analogous results in our setting. Our most complete results involve coherent modules over _locally Frobenius algebras_ which are themselves coherent rings, although some results on finite dimensional modules are also obtained. One motivation for setting up this theory is to introduce stable module categories for such algebras and we discuss options for doing this, exploiting the fact that a locally Frobenius algebra is self injective at least relative to coherent modules and in some cases also relative to all finite dimensional modules. Instead of indexing a family of subalgebras on natural numbers as in the graded theory, we use a family of augmented Frobenius subalgebras indexed on a directed set; this allows us to include examples such as group algebras of locally finite groups in our theory. When the indexing is countable this has implications for vanishing of derived functors of limits taken over the indexing set but otherwise we do not make use of its cardinality. We give some general examples of locally Frobenius algebras, but we leave detailed exploration of examples to future work. The special case of group algebras of locally finite groups and its literature was drawn to my attention by my Glasgow colleague Ken Brown. Notation & conventions: Throughout, $\Bbbk$ will denote a field of characteristic $p\geqslant 0$. All rings, algebras, modules and their homomorphisms will be unital. A _directed set_ $(\Lambda,\preccurlyeq)$ will mean a filtered partially ordered set, i.e., every pair (or finite set) of elements of $\Lambda$ has an upper bound. A system of objects and morphisms in a category indexed on such a $(\Lambda,\preccurlyeq)$ will be referred to as a _$(\Lambda,\preccurlyeq)$ -filtered system_ or _$\Lambda$ -filtered system_. ## 1\. Recollections on rings and modules We will require results on projective, injective and flat modules over Noetherian and coherent rings which are thoroughly covered in Lam [TYL:LectModules&Rings]. The following definitions are standard except for projectively finitely related. ###### Definition 1.1. Let $R$ be a ring and $M$ a left/right $R$-module. * • $M$ is _finitely generated_ (f.g.) if there is an exact sequence $F\to M\to 0$ with $F$ finitely generated and free (or equivalently f.g. projective). * • $M$ is _finitely related_ (f.r.) if there is a short exact sequence $0\to K\to F\to M\to 0$ with $K$ finitely generated and $F$ free. * • $M$ is _projectively finitely related_ (p.f.r.) if there is a short exact sequence $0\to K\to P\to M\to 0$ with $K$ finitely generated and $P$ projective. * • $M$ is _finitely presented_ (f.p.) if there is a short exact sequence $0\to K\to F\to M\to 0$ with $K$ f.g. and $F$ f.g. and free (or equivalently f.g. projective). The notion of p.f.r. is of course redundant for local rings since every projective module is free by a celebrated result of Kaplansky [IK:ProjMod]. The next result is a basic tool used in arguments about such conditions, see for example Lam [TYL:LectModules&Rings]*lemma (5.1). ###### Lemma 1.2 (Schanuel’s Lemma). Let $M$ be an $R$-module. Suppose that there are short exact sequences $0\to K\to P\to M\to 0,\quad 0\to L\to Q\to M\to 0$ where $P$ is projective. Then there is a short exact sequence of the form $\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{K\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{L\oplus P\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{Q\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{0}$ and moreover, if $Q$ is also projective, $P\oplus L\cong K\oplus Q.$ ###### Corollary 1.3. Suppose that $M$ is f.p. and that there is a short exact sequence $0\to L\to Q\to M\to 0$ where $Q$ is f.g., then so is $L$. Using Schanuel’s Lemma, it is easy to see that $\text{(f.g. \& f.r.)}\Longleftrightarrow\text{f.p.}$, so we use these descriptions interchangeably as well as using f.g. projective modules in place of f.g. free modules. The first two parts of the following result are taken from [TYL:LectModules&Rings]*chapter 2§4. ###### Theorem 1.4. Let $R$ be a ring and $M$ a left/right $R$-module. _(a)_ $M$ is f.r. if and only if $M\cong M_{0}\oplus F$ where $M_{0}$ is f.p. and $F$ is free. _(b)_ If $M$ is f.r. then it is flat if and only if it is projective. _(c)_ $M$ is p.f.r. if and only if $M\oplus F^{\prime}\cong M_{0}\oplus F^{\prime\prime}$ where $M_{0}$ is f.p. and $F^{\prime},F^{\prime\prime}$ are free. ###### Proof. (a) See [TYL:LectModules&Rings]*theorem 2.4.26(c). (b) See [TYL:LectModules&Rings]*theorem 2.4.30. (c) If $M$ is p.f.r. there is a short exact sequence $0\to K\to P\to M\to 0$ with $K$ finitely generated and $P$ projective. Choose a projective module so that $F=Q\oplus P$ is free and write $\displaystyle F^{\infty}=\bigoplus_{i\in\mathbb{N}}F$. By the Eilenberg swindle, $P\oplus F^{\infty}\cong F^{\infty}$ so there is an exact sequence $0\to K\to F^{\infty}\to M\oplus F^{\infty}\to 0.$ Therefore $M\oplus F^{\infty}$ is f.r. and the result follows from (a). ∎ Part (c) can be rephrased as saying that saying that being p.f.r. is equivalent to being _stably f.p._ , or _stably coherent_ when the ring itself is coherent. We also recall the characterisation of flat modules provided by the theorem of Lazard and Govorov [TYL:LectModules&Rings]*theorem 4.34. ###### Theorem 1.5 (Lazard-Govorov theorem). Let $R$ be a ring and $M$ an $R$-module. Then $M$ is flat if and only if it is a filtered colimit of f.g. free modules. ### Coherence for rings and modules We recall the notion of (pseudo-)coherence since we will make heavy use of it. For basic properties of (pseudo-)coherent modules see Bourbaki [Bourbaki:HomAlg]*X.§3, ex. 10. ###### Definition 1.6. Let $R$ be a ring. * • An $R$-module $M$ is _pseudo-coherent_ if every f.g. submodule is f.p.. * • A f.g. pseudo-coherent module is called _coherent_. * • $R$ is _left/right coherent_ if it is coherent as a left/right $R$-module. * • $R$ is _coherent_ if it is coherent as a left and right $R$-module. ###### Remark 1.7. Over a coherent ring, a module is f.p. if and only if it is coherent, and every p.f.r. module is pseudo-coherent. Moreover, its coherent modules form an abelian category (see Theorem 4.1). The notion of coherence has an obvious meaning for algebras over a field, and we will use it without further comment. Of course every Noetherian ring is coherent, so Frobenius algebras are coherent. Here is a well known result that can be used to produce many more examples of coherent rings, this version appears in Bourbaki [Bourbaki:HomAlg]*X.§3, ex. 11e, and it can be used to show that an infinitely generated polynomial ring over a coherent commutative ring is coherent. ###### Proposition 1.8. Let $(\Lambda,\preccurlyeq)$ be a directed set and let $R(\lambda)$ $(\lambda\in\Lambda)$ be a $\Lambda$-directed system of rings and homomorphisms $\varphi_{\alpha}^{\beta}\colon R(\alpha)\to R(\beta)$ for $\alpha\preccurlyeq\beta$. Assume that * • whenever $\alpha\preccurlyeq\beta$, $R(\beta)$ is flat as a right/left $R(\alpha)$-module; * • each $R(\lambda)$ is coherent. Then the ring $\displaystyle R=\operatorname*{colim}_{(\Lambda,\preccurlyeq)}R(\lambda)$ is left/right coherent and for every $\lambda\in\Lambda$, $R$ is a flat right/left $R(\lambda)$-module. Noetherianness and coherence of rings are characterised by homological properties. ###### Theorem 1.9. Let $R$ be a ring. _(a)_ $R$ is left/right Noetherian if and only if all coproducts and directed colimits of left/right injectives are injective. _(b)_ $R$ is left/right coherent if and only if all products and directed limits of flat left/right modules are flat. ###### Proof. (a) See [TYL:LectModules&Rings]*theorem 1.3.46. (b) This is a result of Chase, see [TYL:LectModules&Rings]*theorem 1.4.47. ∎ We will frequently make use of faithful flatness, so for the convenience of the reader we state some of the main properties. ###### Proposition 1.10. Let $R$ be a ring and $P$ a flat right $R$-module. _(a)_ $P$ is faithfully flat if it satisfies any and hence all of the following equivalent conditions. * • A sequence of left $R$-modules $0\to L\to M\to N\to 0$ is short exact if and only if the induced sequence $0\to P\otimes_{R}L\to P\otimes_{R}M\to P\otimes_{R}N\to 0$ is short exact. * • For a left $R$-module $M$, $P\otimes_{R}M=0$ if and only if $M=0$. * • A homomorphism of left $R$-modules $\varphi\colon M\to N$ is zero if and only if the induced homomorphism $\operatorname{Id}_{P}\otimes\varphi\colon P\otimes_{R}M\to P\otimes_{R}N$ is zero. _(b)_ If $P$ is faithfully flat then $P\otimes_{R}(-)$ reflects monomorphisms, epimorphisms and isomorphisms. _(c)_ Let $R\to S$ be a ring homomorphism so that $S$ is a faithfully flat right $R$-module, and let $M$ be a left $R$-module. If $S\otimes_{R}M$ is a simple $S$-module, then $M$ is simple. ###### Proof. (a) See [TYL:LectModules&Rings]*theorem 4.70. (b) This is immediate from (a). (c) By flatness, a short exact sequence of $R$-modules $0\to L\to M\to N\to 0$ on tensoring with $S$ yields a short exact sequence of $S$-modules $0\to S\otimes_{R}L\to S\otimes_{R}M\to S\otimes_{R}N\to 0.$ If $S\otimes_{R}M$ is simple then one of $S\otimes_{R}L$ or $S\otimes_{R}N$ must be trivial, so by faithful flatness of $S$, one of $L$ or $N$ must be trivial. Hence $M$ is simple. ∎ ## 2\. Frobenius algebras and Frobenius extensions Recall that a Frobenius $\Bbbk$-algebra $R$ is self-injective, i.e., injective as a left/right $R$-module. The following is a fundamental consequence. Recall that there are induction and coinduction functors $\operatorname{ind}_{\Bbbk}^{R}\colon\mathbf{Mod}_{\Bbbk}\to\mathbf{Mod}_{R},\quad\operatorname{coind}_{\Bbbk}^{R}:\mathbf{Mod}_{\Bbbk}\to\mathbf{Mod}_{R}$ where $\operatorname{ind}_{\Bbbk}^{R}(-)=R\otimes_{\Bbbk}(-),\quad\operatorname{coind}_{\Bbbk}^{R}(-)=\operatorname{Hom}_{\Bbbk}(R,-).$ Notice that for a $\Bbbk$-vector space $W$, the $R$-module $\operatorname{ind}_{\Bbbk}^{R}W$ is projective (and in fact free) while $\operatorname{coind}_{\Bbbk}^{R}W$ is injective. For future use we recall that since we are working over a field $\Bbbk$, every $R$-module $M$ admits an embedding into an injective, (2.1) $M\to\operatorname{coind}_{\Bbbk}^{R}M;\quad m\longmapsto(r\mapsto rm).$ The next result is a standard reformulation of what it means to be a Frobenius algebra. ###### Theorem 2.1. If $R$ is a Frobenius $\Bbbk$-algebra, then the functors $\operatorname{ind}_{\Bbbk}^{R}$ and $\operatorname{coind}_{\Bbbk}^{R}$ are naturally isomorphic. Hence for a $\Bbbk$-vector space $W$, $\operatorname{ind}_{\Bbbk}^{R}W\cong\operatorname{coind}_{\Bbbk}^{R}W$ is both projective and injective. ###### Corollary 2.2. An $R$-module is injective if and only if it is projective. ###### Proof. Let $J$ be an injective $R$-module. As a special case of (2.1) there is a monomorphism of $R$-modules $J\to\operatorname{coind}_{\Bbbk}^{R}J$ and by injectivity there is a commutative diagram with exact row $\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{J\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\operatorname{Id}_{J}}$$\textstyle{\operatorname{coind}_{\Bbbk}^{R}J\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{J}$ so $J$ is a retract of a projective module, hence projective. A similar argument shows that for a projective $P$, there is a commutative diagram $\textstyle{\operatorname{ind}_{\Bbbk}^{R}P\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{P\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{0}$$\textstyle{P\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\operatorname{Id}_{P}}$ which shows that $P$ is injective. ∎ A useful consequence of this result is that every monomorphism of $R$-modules $R\to M$ splits. The next result summarises the fundamental properties of modules over a Frobenius algebra which motivate much of our work on modules over locally Frobenius algebras. ###### Proposition 2.3. Let $R$ be a Frobenius $\Bbbk$-algebra and $M$ an $R$-module. Then the following are equivalent: * • $M$ is injective; * • $M$ is projective; * • $M$ is flat. ###### Proof. The first two are equivalent by Corollary 2.2, while projectivity implies flatness. Since $R$ is self-injective and Noetherian, by using the Lazard-Govorov Theorem 1.5, Corollary 2.2 and Theorem 1.9(a) we find that flatness implies injectivity. ∎ Now recall that if $S$ is a simple $R$-module then Schur’s Lemma implies that then $\operatorname{End}_{R}(S)$ is a division algebra central over $\Bbbk$, so $S$ can be viewed as a vector space over $\operatorname{End}_{R}(S)$ and $\dim_{\Bbbk}S=\dim_{\Bbbk}\operatorname{End}_{R}(S)\,\dim_{\operatorname{End}_{R}(S)}S.$ ###### Proposition 2.4. Let $R$ be a Frobenius $\Bbbk$-algebra. _(a)_ Every $R$-module embeds into a free module. In particular, every f.g. $R$-module embeds into a f.g. free module. _(b)_ Let $S$ be a simple left $R$-module. Then $S$ embeds as a submodule of $R$ with multiplicity equal to $\dim_{\operatorname{End}_{R}(S)}S$. ###### Proof. We focus on left modules but the case of right modules is similar. (a) For an $R$-module $M$, there is an injective homomorphism as in (2.1), $M\to\operatorname{coind}_{\Bbbk}^{R}M\cong\operatorname{ind}_{\Bbbk}^{R}M$ and the induced module is a free $R$-module. When $M$ is f.g. $R$-module it is a f.d. $\Bbbk$-vector space so the induced module is a f.g. free $R$-module. (b) Every non-trivial homomorphism $S\to R$ is injective. By Schur’s Lemma its endomorphism algebra is a division algebra which is central over $\Bbbk$. The multiplicity of $S$ in $R$ is equal to $\dim_{\operatorname{End}_{R}(S)^{\mathrm{op}}}\operatorname{Hom}_{R}(S,R)$, where $\operatorname{End}_{R}(S)$ acts on the right of $\operatorname{Hom}_{R}(S,R)$ by precomposition. The coinduced module $\operatorname{coind}_{\Bbbk}^{R}\Bbbk=\operatorname{Hom}_{\Bbbk}(R,\Bbbk)$ is a left $R$-module through the right action on the domain, and then there are isomorphisms of right $\operatorname{End}_{R}(S)$-vector spaces $\displaystyle\operatorname{Hom}_{R}(S,R)$ $\displaystyle\cong\operatorname{Hom}_{R}(S,\operatorname{coind}_{\Bbbk}^{R}\Bbbk)$ $\displaystyle\cong\operatorname{Hom}_{\Bbbk}(R\otimes_{R}S,\Bbbk)$ $\displaystyle\cong\operatorname{Hom}_{\Bbbk}(S,\Bbbk),$ hence $\operatorname{Hom}_{R}(S,R)$ is non-trivial and the multiplicity of $S$ is $\dim_{\operatorname{End}_{R}(S)^{\mathrm{op}}}\operatorname{Hom}_{R}(S,R)=\dim_{\operatorname{End}_{R}(S)}S.\qed$ An important property of finite dimensional Hopf algebras is that they have non-zero _integrals_. Although in general Frobenius algebras do not, Frobenius algebras augmented over the ground field do. If $\varepsilon\colon R\to\Bbbk$ is the augmentation, its subspaces of left and right _integrals with respect to $\varepsilon$_ are defined by ${\smallint}^{\mathrm{l}}_{R}=\\{h\in R:\forall y\in R,\;yh=\varepsilon(y)h\\},\quad{\smallint}^{\mathrm{r}}_{R}=\\{h\in R:\forall y\in R,\;hy=\varepsilon(y)h\\}.$ Then there is an isomorphism ${\smallint}^{\mathrm{l}}_{R}\cong\operatorname{Hom}_{R}(\Bbbk,R)$ and a similar identification of ${\smallint}^{\mathrm{r}}_{R}$ with right module homomorphisms. ###### Proposition 2.5. Let $R$ be a Frobenius $\Bbbk$-algebra which is augmented over $\Bbbk$. Then viewing $R$ and $\Bbbk$ as left or right $R$-modules we have $\dim_{\Bbbk}\operatorname{Hom}_{R}(\Bbbk,R)=1$ and therefore $\dim_{\Bbbk}{\smallint}^{\mathrm{l}}_{R}=1=\dim_{\Bbbk}{\smallint}^{\mathrm{r}}_{R}.$ ###### Proof. This follows from Proposition 2.4(b). ∎ If ${\smallint}^{\mathrm{r}}_{R}={\smallint}^{\mathrm{l}}_{R}$ then the augmented Frobenius algebra is _unimodular_ ; see Farnsteiner [RF:FrobExtnHopfAlgs] for more on this. We will be interested in Frobenius algebras which are _symmetric_ , i.e., they have a Frobenius form $\lambda\colon A\to\Bbbk$ which induces a symmetric bilinear form. It follows that the associated Nakayama automorphism is the identity and so by Farnsteiner [RF:FrobExtnHopfAlgs]*lemma 1.1 it is unimodular. For a Hopf algebra the counit is the natural choice of augmentation and we will always choose it. If the Hopf algebra is commutative or cocommutative then its antipode is a self-inverse automorphism making it _involutive_ , and since the identity function is inner, by [RF:FrobExtnHopfAlgs]*proposition 2.3 it is a symmetric Frobenius algebra if and only if it is unimodular. For finite dimensional Hopf algebras there is an analogue of Maschke’s Theorem. ###### Theorem 2.6. Let $H$ be a finite dimensional Hopf algebra and $\varepsilon$ its counit. Then $H$ is semisimple if and only if $\varepsilon{\smallint}^{\mathrm{l}}_{H}\neq\\{0\\}\neq\varepsilon{\smallint}^{\mathrm{r}}_{H}.$ ###### Proof. See [ML:TourRepThy]*page 553 for example. The proof of semisimplicity uses a non-zero idempotent $e\in{\smallint}^{\mathrm{l}}_{H}$ or $e\in{\smallint}^{\mathrm{r}}_{H}$ and the coproduct applied to it. ∎ ###### Remark 2.7. This result does not apply to all augmented Frobenius algebras. As an example, consider $R=\Bbbk\times R_{0}$ with the augmentation being projection onto the first factor and $R_{0}$ a local Frobenius algebra that is not semisimple. Let $\lambda_{0}$ be a Frobenius form on $R_{0}$; then the form on $R$ given by $\lambda(x,y)=x+\lambda_{0}(y)$ is Frobenius and $(1,0)\in{\smallint}^{\mathrm{l}}_{R}$ but $R$ is not semisimple. Now we turn to _Frobenius extensions_ , introduced by Kasch [FK:ProjFrobExtns]. For a concise account which highlights aspects relevant to our work, see Lorenz [ML:TourRepThy], especially the exercises for sections 2.2 and 12.4. Other useful sources are Farnsteiner [RF:FrobExtnHopfAlgs] and Fischman et al [DF&SM&H-JS:FrobExtns]. We adopt the notation $A:B$ rather than $A/B$ to indicate an extension of algebras. Suppose given a homomorphism $B\to A$ of $\Bbbk$-algebras so we can view $A$ as a $B$-bimodule and consider the extension of $\Bbbk$-algebras $A:B$. ###### Definition 2.8. The extension $A:B$ is a _Frobenius extension_ if there is a $B$-bimodule homomorphism $\Lambda\colon A\to B$ and elements $x_{i},y_{i}\in A$ $(1\leqslant i\leqslant n)$ such that for all $a\in A$, $\sum_{1\leqslant i\leqslant n}y_{i}\Lambda(x_{i}a)=a=\sum_{1\leqslant i\leqslant n}\Lambda(ay_{i})x_{i}.$ ###### Proposition 2.9. Suppose that $A:B$ is a Frobenius extension with associated $B$-bimodule homomorphism $\Lambda\colon A\to B$ and elements $x_{i},y_{i}\in A$ $(1\leqslant i\leqslant n)$. Then $A$ is projective as a left and right $B$-module. Furthermore, there is an isomorphism of functors $\operatorname{coind}^{A}_{B}\xrightarrow{\;\cong\;}\operatorname{ind}^{A}_{B};$ which is defined on each $B$-module $M$ by $\operatorname{coind}^{A}_{B}M=\operatorname{Hom}_{B}(A,M)\mapsto\operatorname{coind}^{A}_{B}M;\quad f\mapsto\sum_{1\leqslant i\leqslant n}y_{i}f(x_{i}).$ Conversely, if $A$ is projective as a left and right $B$-module and $\operatorname{coind}^{A}_{B}\cong\operatorname{ind}^{A}_{B}$, then $A:B$ is a Frobenius extension. In practice we will be considering the situation in following result. ###### Proposition 2.10. Let $\varepsilon\colon A\to\Bbbk$ be an augmented symmetric Frobenius algebra and let $B\subseteq A$ be a subalgebra augmented by the restriction of $\varepsilon$ and which is also a symmetric Frobenius algebra. Then $A:B$ is a Frobenius extension. ###### Proof. Since the Nakayama automorphisms of $A$ and $B$ are the identity functions, the result is an immediate consequence of [RF:FrobExtnHopfAlgs]*theorem 1.3. ∎ ## 3\. Locally Frobenius algebras Let $(\Lambda,\preccurlyeq)$ be an infinite partially ordered set which has the following properties, in particular it is a _directed set_ : * • it is filtered: for any $\lambda^{\prime},\lambda^{\prime\prime}\in\Lambda$ there is a common upper bound $\lambda\in\Lambda$, so $\lambda^{\prime}\preccurlyeq\lambda$ and $\lambda^{\prime\prime}\preccurlyeq\lambda$; * • there is a unique initial element $\lambda_{0}$; * • for every $\lambda\in\Lambda$ there is a $\lambda^{\prime\prime\prime}\in\Lambda$ with $\lambda\precneqq\lambda^{\prime\prime\prime}$. We will denote such a directed set by $(\Lambda,\preccurlyeq,\lambda_{0})$ but often just refer to it as $\Lambda$. If $\Lambda^{\prime}\subseteq\Lambda$ is an infinite filtered subset containing $\lambda_{0}$, then will refer to $(\Lambda^{\prime},\preccurlyeq,\lambda_{0})$ as a _subdirected set_ and write $(\Lambda^{\prime},\preccurlyeq,\lambda_{0})\subseteq(\Lambda,\preccurlyeq,\lambda_{0})$. ###### Definition 3.1. A (necessarily infinite dimensional) $\Bbbk$-algebra $A$ is _locally Frobenius of shape $\Lambda$_ if it satisfies the following conditions. (a) There is a $\Lambda$-directed system of symmetric Frobenius algebras $A(\lambda)$ ($\lambda\in\Lambda$) augmented over $\Bbbk$ and proper inclusion homomorphisms $\iota_{\lambda^{\prime}}^{\lambda^{\prime\prime}}\colon A(\lambda^{\prime})\hookrightarrow A(\lambda^{\prime\prime})$ with $A(\lambda_{0})=\Bbbk$ and $\displaystyle A=\bigcup_{\lambda\in\Lambda}A(\lambda)$. (b) Each inclusion $\iota_{\lambda^{\prime}}^{\lambda^{\prime\prime}}$ makes $A(\lambda^{\prime\prime}):A(\lambda^{\prime})$ a free Frobenius extension (i.e., $A(\lambda^{\prime\prime})$ is free as a left and right $A(\lambda^{\prime})$-module). Of course the Frobenius condition in (b) is a consequence of Proposition 2.10, while the freeness is an additional assumption. ###### Definition 3.2. Let $A$ be a locally Frobenius algebra of shape $\Lambda$ and let $(\Lambda^{\prime},\preccurlyeq,\lambda_{0})\subseteq(\Lambda,\preccurlyeq,\lambda_{0})$ be a subdirected set. A subalgebra $B\subseteq A$ is a _locally Frobenius subalgebra of shape $\Lambda^{\prime}$_ if $\displaystyle B=\bigcup_{\lambda\in\Lambda^{\prime}}B(\lambda)$ is locally Frobenius algebra of shape $\Lambda^{\prime}$ and for every $\lambda\in\Lambda^{\prime}$, $B(\lambda)\subseteq A(\lambda)$. ###### Notation. We will denote the kernel of the augmentation $A(\lambda)\to\Bbbk$ by $A(\lambda)^{+}$; clearly $A$ is also augmented so we similarly denote its augmentation ideal by $A^{+}$. All of these are completely prime maximal ideals (recall that an ideal $P$ in a ring is _completely prime_ if $xy\in P$ implies $x\in P$ or $y\in P$); in general this is stronger the notion of prime (an ideal $Q$ in a ring $R$ is _prime_ if $xRy\subseteq P$ implies $x\in P$ or $y\in P$). All nilpotent elements of $A$ are contained in $A^{+}$, and if $e\in A$ is an idempotent then exactly one of $e$ or $1-e$ is in $A^{+}$. The augmentation condition in (a) may seem unnecessary but we do require it for some technical results and it is automatically satisfied in the Hopf algebra case because the counit is an augmentation. The Frobenius extension condition in (b) might be weakened, and even in the Hopf algebra case it is not otherwise guaranteed except in special circumstances such as when the $A(\lambda)$ are all local. We make a trivial observation. ###### Lemma 3.3. Let $A$ be a locally Frobenius algebra. Then for every finite subset $Z\subseteq A$ there is a $\lambda\in\Lambda$ such that $Z\subseteq A(\lambda)$. Hence every finite dimensional subspace $V\subseteq A$ is contained in some $A(\lambda)$. ###### Proof. Each element $z\in Z$ is contained in some $A(\lambda_{z})$ and by the filtering condition there is an upper bound $\lambda$ of the $\lambda_{z}$, so that $Z\subseteq A(\lambda)$. ∎ Of course a f.d. subalgebra of $A$ is f.d. subspace so it is a subalgebra of some $A(\lambda)$. The next result provides an important class of examples of locally Frobenius algebras which we will refer to as _locally Frobenius Hopf algebras_. ###### Proposition 3.4. Let $H$ be a Hopf algebra over $\Bbbk$. Then $H$ is a locally Frobenius algebra of shape $\Lambda$ if the following conditions are satisfied. _(a)_ There is a $\Lambda$-directed system of f.d. subHopf algebras $H(\lambda)\subseteq H$ $(\lambda\in\Lambda)$ and inclusion homomorphisms $\iota_{\lambda^{\prime}}^{\lambda^{\prime\prime}}\colon H(\lambda^{\prime})\to H(\lambda^{\prime\prime})$ with $H(\lambda_{0})=\Bbbk$ and $\displaystyle H=\bigcup_{\lambda\in\Lambda}H(\lambda)$. _(b)_ Each $H(\lambda)$ is involutive and unimodular. ###### Proof. The Larson-Sweedler theorem [LarsonSweedlerThm] implies that each $H(\lambda)$ is Frobenius, while [RF:FrobExtnHopfAlgs]*proposition 2.3 implies that it is symmetric since the square of the antipode is the identity which is also a Nakayama automorphism. Finally, the Nichols-Zoeller theorem [WDN&MBZ:HAfreeness]. implies that each pair $H(\lambda^{\prime\prime}):H(\lambda^{\prime})$ is free. ∎ Of course if each Hopf algebra $H(\lambda)$ is commutative or cocommutative then it is involutive. It is not necessarily true that an infinite dimensional Hopf algebra $H$ is free as a left/right module over a finite dimensional subHopf algebra $K$, although some results on when this holds are known; for the case where $K$ is semisimple see [WDN&MBR:HAfreenessinfdim], while for the case where $K$ is a normal subalgebra see [H-JS:RemQGps]. However, for our purposes this is not required. Later we will discuss some examples of locally Frobenius Hopf algebras such as group algebras of locally finite groups. The next definition builds on Definition 3.2. ###### Definition 3.5. Let $H$ be a locally Frobenius Hopf algebra of shape $\Lambda$ and let $(\Lambda^{\prime},\preccurlyeq,\lambda_{0})\subseteq(\Lambda,\preccurlyeq,\lambda_{0})$ be a subdirected set. A subHopf algebra $K\subseteq H$ is a _locally Frobenius subHopf algebra of shape $\Lambda^{\prime}$_ if $\displaystyle K=\bigcup_{\lambda\in\Lambda^{\prime}}K(\lambda)$ is locally Frobenius Hopf algebra of shape $\Lambda^{\prime}$ and for every $\lambda\in\Lambda^{\prime}$, $K(\lambda)\subseteq H(\lambda)$. We end this section with a result suggested by Richardson [JSR:GpRngsnonzerosocle]*lemma 2.2. This shows that the augmentation ideal of a locally Frobenius subalgebra generates a ‘large’ submodule. ###### Proposition 3.6. Let $A$ be a locally Frobenius algebra of shape $\Lambda$ and $B$ a locally Frobenius subalgebra of shape $\Lambda^{\prime}\subseteq\Lambda$. Then the submodule $AB^{+}\subseteq A$ is essential. ###### Proof. This proof is a straightfoward adaption of that in Richardson [JSR:GpRngsnonzerosocle]. Suppose that $Az\cap AB^{+}=0$ for some non-zero $z\in A$. Choose $\alpha\in\Lambda$ such that $z\in A(\alpha)$. As $\Lambda^{\prime}$ is infinite we can choose a $\beta\in\Lambda^{\prime}$ for which $\dim_{\Bbbk}A(\beta)>\frac{\dim_{\Bbbk}A(\beta)}{\dim_{\Bbbk}A(\beta)z}.$ Now choose $\gamma\in\Lambda^{\prime}$ so that $A(\alpha)\subseteq A(\gamma)\supseteq A(\beta).$ and then $A(\gamma)z\cap A(\gamma)A(\beta)^{+}\subseteq Az\cap A(\gamma)B^{+}=0.$ Now recall that $A(\gamma)$ is a free module over each of $A(\alpha)$ and $A(\beta)$, so $A(\gamma)z\cong A(\gamma)\otimes_{A(\alpha)}A(\alpha)z,\quad A(\gamma)A(\beta)^{+}\cong A(\gamma)\otimes_{A(\beta)}A(\beta)^{+},$ giving $\displaystyle\dim_{\Bbbk}A(\gamma)z$ $\displaystyle=\frac{\dim_{\Bbbk}A(\gamma)}{\dim_{\Bbbk}A(\alpha)}\dim_{\Bbbk}A(\alpha)z,$ $\displaystyle\dim_{\Bbbk}A(\gamma)A(\beta)^{+}$ $\displaystyle=\frac{\dim_{\Bbbk}A(\gamma)}{\dim_{\Bbbk}A(\beta)}\bigl{(}\dim_{\Bbbk}A(\beta)-1\bigr{)}.$ Using these we obtain $\displaystyle\dim_{\Bbbk}A(\gamma)$ $\displaystyle\geqslant\dim_{\Bbbk}\bigl{(}A(\gamma)z\oplus A(\gamma)A(\beta)^{+}\bigr{)}$ $\displaystyle=\dim_{\Bbbk}A(\gamma)z+\dim_{\Bbbk}A(\gamma)A(\beta)^{+}$ $\displaystyle=\frac{\dim_{\Bbbk}A(\gamma)\dim_{\Bbbk}A(\alpha)z}{\dim_{\Bbbk}A(\alpha)}+\frac{\dim_{\Bbbk}A(\gamma)\bigl{(}\dim_{\Bbbk}A(\beta)-1\bigr{)}}{\dim_{\Bbbk}A(\beta)}$ $\displaystyle>\frac{\dim_{\Bbbk}A(\gamma)}{\dim_{\Bbbk}A(\beta)}+\frac{\dim_{\Bbbk}A(\gamma)\bigl{(}\dim_{\Bbbk}A(\beta)-1\bigr{)}}{\dim_{\Bbbk}A(\beta)}$ $\displaystyle=\dim_{\Bbbk}A(\gamma),$ which is impossible. Therefore $AB^{+}\subseteq A$ must be an essential submodule. ∎ ## 4\. Modules over a locally Frobenius algebra Now we will describe some basic properties of modules over locally Frobenius algebras. Throughout this section we will suppose that $A$ is a locally Frobenius of shape $\Lambda$. First we state a result for coherent $A$-modules; since $A$ is coherent these are exactly the f.p. $A$-modules. ###### Theorem 4.1. The category $\mathbf{Mod}^{\mathrm{coh}}_{A}$ of coherent $A$-modules and their homomorphisms is an abelian category with all finite limits and colimits as well as enough projectives and injectives. ###### Proof. Clearly this category has finite products and coproducts, so it is sufficient to check that it has kernels, images and cokernels; this is an exercise in Bourbaki [Bourbaki:HomAlg]*ex. §3.10(b), see also Cohen [JMC:Coherent]*section 1. For projectives and injectives see Lemma 4.17 below. ∎ ###### Proposition 4.2. _(a)_ For each $\lambda\in\Lambda$, $A$ is injective, projective and flat as a left or right $A(\lambda)$-module. _(b)_ The $\Bbbk$-algebra $A$ is coherent. _(c)_ Suppose that $M$ is a coherent $A$-module. Then for some $\lambda$ there is an $A(\lambda)$-module $M^{\prime}$ with a finite presentation $0\longleftarrow M^{\prime}\longleftarrow A(\lambda)^{k}\longleftarrow A(\lambda)^{\ell}$ inducing a finite presentation $0\longleftarrow A\otimes_{A(\lambda)}M^{\prime}\longleftarrow A^{k}\longleftarrow A^{\ell}$ where $A\otimes_{A(\lambda)}M^{\prime}\cong M$. _(d)_ Let $\varphi\colon M\to N$ be a homomorphisms of coherent $A$-modules. Then there is a $\lambda\in\Lambda$, and a homomorphism of finitely generated $A(\lambda)$-modules $\varphi^{\prime\prime}\colon M^{\prime\prime}\to N^{\prime\prime}$ fitting into a commutative diagram of $A$-module homomorphisms. $\textstyle{A\otimes_{A(\lambda)}M^{\prime\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\operatorname{Id}\otimes\varphi^{\prime\prime}}$$\scriptstyle{\cong}$$\textstyle{A\otimes_{A(\lambda)}N^{\prime\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\cong}$$\textstyle{M\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\varphi}$$\textstyle{N}$ ###### Proof. We will give brief indications of the proofs. (a) By Proposition 1.8 $A$ is a flat $A(\lambda)$-module, hence by Proposition 2.3 it is also injective and projective. (b) Since each $A(\lambda)$ is Noetherian over $\Bbbk$, Proposition 1.8 implies that $A$ is left and right coherent. (c) If $M$ is an f.p. $A$-module, there is an exact sequence of $A$-modules $0\longleftarrow M\xleftarrow{\rho_{0}}A^{m}\xleftarrow{\rho_{1}}A^{n}$ for some $m,n$. The image of $\rho_{1}$ must be contained in $A(\lambda)^{m}\subseteq A^{m}$ for some $\lambda\in\Lambda$, so we obtain an exact sequence of $A(\lambda)$-modules $0\longleftarrow M^{\prime}\xleftarrow{\rho^{\prime}_{0}}A(\lambda)^{m}\xleftarrow{\rho^{\prime}_{1}}A(\lambda)^{n}$ and on tensoring with $A$, by (a) this yields an exact sequence of $A$-modules $0\longleftarrow A\otimes_{A(\lambda)}M^{\prime}\xleftarrow{\operatorname{Id}\otimes\rho^{\prime}_{0}}A\otimes_{A(\lambda)}A(\lambda)^{m}\xleftarrow{\operatorname{Id}\otimes\rho^{\prime}_{1}}A\otimes_{A(\lambda)}A(\lambda)^{n}$ which is equivalent to the original one. (d) Using (c) we can represent $M$ and $N$ as induced up from f.g. $A(\lambda_{1})$-modules $M^{\prime},N^{\prime}$ for some $\lambda_{1}$. The image of the restriction of $\varphi$ to $M^{\prime}$ in $A\otimes_{A(\lambda_{1})}N^{\prime}$ lies in some $A(\lambda_{2})\otimes_{A(\lambda_{1})}N^{\prime}$ where $\lambda_{1}\preccurlyeq\lambda_{2}$. Base changing gives a homomorphism $\varphi^{\prime\prime}\colon M^{\prime\prime}=A(\lambda_{2})\otimes_{A(\lambda_{1})}M^{\prime}\to A(\lambda_{2})\otimes_{A(\lambda_{1})}N^{\prime}=N^{\prime\prime}$ with the required properties. ∎ The faithful flatness condition of Proposition 4.9 gives another useful property. ###### Corollary 4.3. Every short exact sequence of coherent $A$-modules (4.1) $0\to L\xrightarrow{\varphi}M\xrightarrow{\theta}N\to 0$ is induced up from a short exact sequence of f.g. $A(\alpha)$-modules $0\to L^{\prime}\xrightarrow{\varphi^{\prime}}M^{\prime}\xrightarrow{\theta^{\prime}}N^{\prime}\to 0$ for some $\alpha\in\Lambda$, i.e., there is a commutative diagram of $A$-modules of the following form. $\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{L\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\varphi}$$\scriptstyle{\cong}$$\textstyle{M\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\theta}$$\scriptstyle{\cong}$$\textstyle{N\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\cong}$$\textstyle{0}$$\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{A\otimes_{A(\alpha)}L^{\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\operatorname{Id}\otimes\varphi^{\prime}}$$\textstyle{A\otimes_{A(\alpha)}M^{\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\operatorname{Id}\otimes\theta^{\prime}}$$\textstyle{A\otimes_{A(\alpha)}N^{\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{0}$ In particular, an isomorphism $\varphi\colon L\to M$ is induced up from an isomorphism of f.g. $A(\alpha)$-modules for some $\alpha\in\Lambda$. ###### Proof. Choose $\alpha\in\Lambda$ so that there are homomorphisms of f.g. $A(\alpha)$-modules $L^{\prime}\xrightarrow{\varphi^{\prime}}M^{\prime}\xrightarrow{\theta^{\prime}}N^{\prime}$ such that $0\to A\otimes_{A(\alpha)}L^{\prime}\xrightarrow{\operatorname{Id}\otimes\varphi^{\prime}}A\otimes_{A(\alpha)}M^{\prime}\xrightarrow{\operatorname{Id}\otimes\theta^{\prime}}A\otimes_{A(\alpha)}N^{\prime}\to 0$ corresponds to the short exact sequence (4.1). Then by Proposition 1.10(a), $0\to L^{\prime}\xrightarrow{\varphi^{\prime}}M^{\prime}\xrightarrow{\theta^{\prime}}N^{\prime}\to 0$ is a short exact sequence of $A(\alpha)$-modules. The statement about isomorphisms also follows using faithful flatness of the $A(\alpha)$-module $A$ and Proposition 1.10(b). ∎ Now we give some results on the dimension of coherent $A$-modules. ###### Lemma 4.4. Let $\lambda\in\Lambda$ and let $L\subseteq A(\lambda)$ be a left ideal. Then as left $A$-modules, $A\otimes_{A(\lambda)}A(\lambda)/L\cong A/AL.$ and the $\Bbbk$-vector space $A/AL$ is infinite dimensional. ###### Proof. By a standard criterion for flatness of [TYL:LectModules&Rings]*(4.12), multiplication induces an isomorphism $A\otimes_{A(\lambda)}L\xrightarrow{\cong}AL$ where $AL\subseteq A$ is the left ideal generated by $L$. Therefore there is a commutative diagram of left $A$-modules $\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{A\otimes_{A(\lambda)}L\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\cong}$$\textstyle{A\otimes_{A(\lambda)}A(\lambda)\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\cong}$$\textstyle{A\otimes_{A(\lambda)}A(\lambda)/L\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{0}$$\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{AL\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{A\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{A/AL\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{0}$ with exact rows. It follows that the right hand vertical arrow is an isomorphism. Whenever $\alpha\preccurlyeq\beta$, $A(\beta)$ is a finite rank free $A(\alpha)$-module, therefore $\dim_{\Bbbk}A(\beta)=\dim_{\Bbbk}A(\alpha)\operatorname{rank}_{A(\alpha)}A(\beta)$ and $\operatorname{rank}_{A(\alpha)}A(\beta)\leqslant\operatorname{rank}_{A(\alpha)}A$ where is strictly increasing as a function of $\beta$. ∎ ###### Proposition 4.5. Let $M$ be a coherent left/right $A$-module. Then $M$ is an infinite dimensional $\Bbbk$-vector space. ###### Proof. We will assume that $M$ is a left module, the proof when it is a right module is similar. When $M$ is cyclic, $M\cong A/L$ for some f.g. left ideal $L$. A finite set of generators of $L$ must lie in some $A(\lambda)$ so the ideal $L(\lambda)=A(\lambda)\cap L$ satisfies $AL(\lambda)=L$ and by Lemma 4.4, $M\cong AL(\lambda)$ is infinite dimensional. Now for a general coherent module, suppose that $M$ has $m$ generators $x_{1},\dots,x_{m}$ where $m$ is minimal. Then for the proper submodule $M^{\prime}\subseteq M$ generated by $x_{1},\dots,x_{m-1}$, $M/M^{\prime}$ is a non-trivial cyclic coherent module which is infinite dimensional. Since there is an epimorphism $M\to M/M^{\prime}$, $M$ must be infinite dimensional. ∎ This gives an important fact about f.d. modules. ###### Proposition 4.6. A non-trivial f.d. $A$-module is not coherent. For simple modules we have the following. ###### Corollary 4.7. Let $S=A/L$ be a f.d. simple left/right $A$-module, where $L$ is a maximal left/right ideal of $A$. Then $S$ is not coherent and $L$ is not finitely generated. In particular, the trivial $A$-module $\Bbbk$ is not coherent and the augmentation ideal $A^{+}$ is not f.g. as a left/right module. Of course this shows that a locally Frobenius algebra is never Noetherian. As we will see later, it also implies that the trivial module is not isomorphic to a submodule of $A$. ###### Corollary 4.8. Let $z\in A$ be nilpotent. Then $A/Az$ and $A/zA$ are both infinite dimensional. ###### Proof. Every nilpotent element of $A$ is contained in the completely prime ideal $A^{+}$. ∎ We end this section with another important observation. ###### Proposition 4.9. For $\alpha,\beta\in\Lambda$, * • if $\alpha\preccurlyeq\beta$ then $A(\beta)$ is a faithfully flat $A(\alpha)$-module; * • $A$ is a faithfully flat $A(\alpha)$-module. ###### Proof. We know that $A(\beta)$ and $A$ are flat as $A(\alpha)$-modules; we also know that the inclusions $A(\alpha)\hookrightarrow A(\beta)$ and $A(\alpha)\hookrightarrow A$ split as $A(\alpha)$-module homomorphisms since $A(\alpha)$ is self-injective. For an $A(\alpha)$-module $M$, the unit induces a split homomorphism of $A(\alpha)$-modules $M\to A(\beta)\otimes_{A(\alpha)}M$, showing that $A(\beta)$ is a faithful $A(\alpha)$-module. $\textstyle{M\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\cong}$$\scriptstyle{\cong}$$\textstyle{A(\alpha)\otimes_{A(\alpha)}M\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{A(\beta)\otimes_{A(\alpha)}M\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{A(\alpha)\otimes_{A(\alpha)}M}$ A similar argument with $A$ in place of $A(\beta)$ shows that $A$ is also faithful. ∎ Combining Lemma 1.10(c) with Proposition 4.9 we obtain ###### Corollary 4.10. Let $\alpha,\beta\in\Lambda$ and let $M$ be a left $A(\alpha)$-module. Then * • for $\alpha\preccurlyeq\beta$, if $A(\beta)\otimes_{A(\alpha)}M$ is a simple $A(\beta)$-module then $M$ is simple; * • if $A\otimes_{A(\alpha)}M$ is a simple $A$-module then $M$ is simple. Here are some useful consequences of faithful flatness of $A$. ###### Corollary 4.11. Let $M$ be a simple coherent $A$-module. Then there is a $\lambda\in\Lambda$ for which $M$ has the form $M\cong A\otimes_{A(\lambda)}M^{\prime}$ for a simple $A(\lambda)$-module $M^{\prime}$. ###### Proof. By Proposition 4.2 there is a $\lambda\in\Lambda$ such that $M$ is induced up from an $A(\lambda)$-module $M^{\prime}$. By Corollary 4.10, $M^{\prime}$ is simple. ∎ ###### Corollary 4.12. Let $\lambda\in\Lambda$, and let $\varphi\colon M_{\lambda}\to N_{\lambda}$ be a homomorphism of f.g. left $A(\lambda)$-modules. If $\operatorname{Id}\otimes\varphi\colon A\otimes_{A(\lambda)}M_{\lambda}\to A\otimes_{A(\lambda)}N_{\lambda}$ is a monomorphism/an epimorphism/an isomorphism, then so is $\varphi$. ###### Proof. See Proposition 1.10(b). ∎ The Frobenius extension condition for each inclusion $A(\lambda)\subseteq A(\lambda^{\prime})$ ensures that the induction and coinduction functors $\operatorname{ind}_{A(\lambda)}^{A(\lambda^{\prime})}\colon\mathbf{Mod}_{A(\lambda)}\to\mathbf{Mod}_{A(\lambda^{\prime})}$ and $\operatorname{coind}_{A(\lambda)}^{A(\lambda^{\prime})}:\mathbf{Mod}_{A(\lambda)}\to\mathbf{Mod}_{A(\lambda^{\prime})}$ are naturally isomorphic, where for a left $A(\lambda)$-module $M$, $\operatorname{ind}_{A(\lambda)}^{A(\lambda^{\prime})}M=A(\lambda^{\prime})\otimes_{A(\lambda)}M,\quad\operatorname{coind}_{A(\lambda)}^{A(\lambda^{\prime})}M=\operatorname{Hom}_{A(\lambda)}(A(\lambda^{\prime}),M).$ Here we use the right multiplication of $A(\lambda^{\prime})$ on itself to define the left $A(\lambda^{\prime})$-module structure on $\operatorname{Hom}_{A(\lambda)}(A(\lambda^{\prime}),M)$. Of course we can specialise to the case $\lambda=\lambda_{0}$ and $A(\lambda_{0})=\Bbbk$. For an $A(\lambda)$-module $M$, there is injective composition $M\xrightarrow{\cong}\operatorname{Hom}_{\Bbbk}(\Bbbk,M)\to\operatorname{Hom}_{\Bbbk}(A(\lambda),M);\quad x\mapsto(a\mapsto ax)$ which is an $A(\lambda)$-module homomorphism with $\operatorname{Hom}_{\Bbbk}(A(\lambda),M)\cong\operatorname{coind}_{\Bbbk}^{A(\lambda)}M\cong\operatorname{ind}_{\Bbbk}^{A(\lambda)}M$ being both an injective $A(\lambda)$-module and a free module $A(\lambda)$-module. Of course if $M$ is f.g. then so is $\operatorname{ind}_{\Bbbk}^{A(\lambda)}M$. ###### Proposition 4.13. Let $M$ be a coherent $A$-module. Then there is an embedding of $M$ into a f.g. free $A$-module. ###### Proof. We know that $M\cong\operatorname{ind}_{A(\lambda)}^{A}M^{\prime}$ for some f.g. $A(\lambda)$-module with $\lambda\in\Lambda$. There is also an embedding of $M^{\prime}$ into a f.g. free $A(\lambda)$-module $F^{\prime}$ say. Inducing up and using flatness of $A$ over $A(\lambda)$, we obtain an injection $M\xrightarrow{\cong}\operatorname{ind}_{A(\lambda)}^{A}M^{\prime}\to\operatorname{ind}_{A(\lambda)}^{A}F^{\prime}=F$ where $F$ is a f.g. free $A$-module. ∎ ###### Lemma 4.14. If $M$ is a coherent $A$-module, then for $s>0$, $\operatorname{Ext}_{A}^{s}(M,A)=0$. Hence $A$ is injective in the category of coherent $A$-modules. More generally, this holds if $M$ is a coproduct of coherent $A$-modules. ###### Proof. By Proposition 4.2(c), for some $\lambda\in\Lambda$ there is an $A(\lambda)$-module $M^{\prime}$ such that $A\otimes_{A(\lambda)}M^{\prime}\cong M$. Then $\operatorname{Ext}^{*}_{A}(M,A)\cong\operatorname{Ext}^{*}_{A}(A\otimes_{A(\lambda)}M^{\prime},A)\cong\operatorname{Ext}^{*}_{A(\lambda)}(M^{\prime},A).$ Now recall that by Proposition 4.2(a), for $\lambda\in\Lambda$, $A$ is an injective $A(\lambda)$-module, hence for $s>0$, $\operatorname{Ext}^{s}_{A(\lambda)}(M^{\prime},A)=0$. An alternative argument uses Proposition 4.9. A sequence of coherent $A$-modules $0\to U\to V$ is induced up from a sequence of $A(\lambda)$-modules for some $\Lambda\in\Lambda$, $0\to A\otimes_{A(\lambda)}U^{\prime}\to A\otimes_{A(\lambda)}V^{\prime}$ which is exact if and only if the sequence $0\to U^{\prime}\to V^{\prime}$ is exact. Now given the solid diagram of $A$-modules $\textstyle{A\otimes_{A(\lambda)}U^{\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\cong}$$\textstyle{A\otimes_{A(\lambda)}V^{\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\cong}$$\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{U\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{V\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{A}$ with exact row, by using the adjunction there is a solid diagram of $A(\lambda)$-modules $\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{U^{\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{V^{\prime}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{A}$ with exact row. Since $A$ is a flat $A(\lambda)$-module, it is injective and it follows that we can complete this diagram with a dashed arrow. Again using the adjunction, we can complete the original diagram, showing that $A$ is relatively injective. For a coproduct of coherent modules, it is standard that $\operatorname{Ext}_{A}^{s}(-,N)$ sends coproducts to products. ∎ ###### Corollary 4.15. If $M$ is a f.r. $A$-module, then for $s>0$, $\operatorname{Ext}_{A}^{s}(M,A)=0$. ###### Proof. By Theorem 1.4(a), such a module has the form $M\cong M_{0}\oplus F$, with $M_{0}$ f.p. and $F$ free. Then for $s>0$, $\operatorname{Ext}_{A}^{s}(M,A)\cong\operatorname{Ext}_{A}^{s}(M_{0},A)=0.\qed$ ###### Proposition 4.16. Let $A$ be a locally Frobenius $\Bbbk$-algebra and $M$ an $A$-module. _(a)_ If $M$ is f.p. then there is an embedding of $M$ into a finite rank free module. _(b)_ If $M$ is f.r. then there is an embedding of $M$ into a free module. ###### Proof. (a) As in the proof of Lemma 4.14, a finite presentation $A^{m}\to A^{n}\to M\to 0$ is induced up from an exact sequence of $A(\lambda)$-modules $A(\lambda)^{m}\to A(\lambda)^{n}\to M^{\prime}\to 0$ for some $\lambda\in\Lambda$, where $A\otimes_{A(\lambda)}M^{\prime}\cong M$. By Proposition 2.4(a), there is a monomorphism $M^{\prime}\to F^{\prime}$ where $F^{\prime}$ is a f.g. free $A(\lambda)$-module. By flatness of $A$ over $A(\lambda)$, this gives a monomorphism $M\xrightarrow{\cong}A\otimes_{A(\lambda)}M^{\prime}\to A\otimes_{A(\lambda)}F^{\prime}=F,$ where $F$ is a f.g. free $A$-module. (b) By Theorem 1.4(a), $M\cong F\oplus M_{0}$ with $F$ free and $M_{0}$ f.p., so the result follows using (a). ∎ We now have a result which completes the proof of Theorem 4.1. ###### Lemma 4.17. The abelian category $\mathbf{Mod}^{\mathrm{coh}}_{A}$ has enough projectives and injectives which are the summands of f.g. free modules. ###### Proof. The existence of projectives is obvious. Lemma 4.14 implies the existence of injectives in $\mathbf{Mod}^{\mathrm{coh}}_{A}$, and by Proposition 4.16(a), every coherent module embeds in a f.g. free module which is also injective. ∎ ###### Remark 4.18. Of course we should not expect $A$ to be an injective in $\mathbf{Mod}_{A}$ so it is only relatively injective with respect to the full subcategory $\mathbf{Mod}^{\mathrm{coh}}_{A}$; the basic notions of relative homological algebra can be found in Eilenberg & Moore [SE&JCM:RelHomAlg]. We can use finitely generated free modules to build projective resolutions in $\mathbf{Mod}^{\mathrm{coh}}_{A}$ for computing $\operatorname{Ext}_{A}$ since for a coherent module $M$, $\operatorname{Ext}_{A}^{*}(-,M)$ are isomorphic to derived functors on $\mathbf{Mod}^{\mathrm{coh}}_{A}$. On the category of all $A$-modules the left exact functor $\operatorname{Hom}_{A}(M,-)$ has the right derived functors $\operatorname{Ext}_{A}^{*}(M,-)$. By Lemma 4.14, for each finitely generated free module $F$, $\operatorname{Ext}_{A}^{s}(M,F)=0$ if $s>0$. This means that $F$ is $\operatorname{Hom}_{A}(M,-)$-acyclic, and it is well-known that these right derived functors can be computed using resolutions by such modules which always exist here; see [CAW:HomAlg] on $F$-acyclic objects and dimension shifting. ### Injective, projective and flat modules Lemma 4.14 shows that $A$ is relatively injective with respect to the category of coherent $A$-modules; it follows that f.g. free and projective modules are also relatively injective. We can relate flatness to relatively injectivity. ###### Lemma 4.19. Let $P$ be a flat $A$-module. Then $P$ is relatively injective with respect to the category of coherent $A$-modules. ###### Proof. By [TYL:LectModules&Rings]*theorem 4.32, every homomorphism $M\to P$ from a finitely presented $A$-module factors as $\textstyle{M\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{F\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{P}$ where $F$ is a finitely generated free module. Then given a diagram of solid arrows with exact row where $U,V$ are coherent, $\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{U\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{V\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{F\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{P}$ we can extend it with the dashed arrows using the factorisation above, and since $F$ is relatively injective there is a dotted arrow making the whole diagram commute. This can also be proved using the Lazard-Govorov Theorem 1.5. ∎ Since products of (relative) injectives are (relative) injectives, this is consistent with Chase’s Theorem 1.9(b) which says that products of flat modules over a coherent ring are flat. ###### Lemma 4.20. Let $J$ be a relatively injective $A$-module with respect to the category of coherent $A$-modules. Then $J$ is flat. In particular, every injective $A$-module is flat. ###### Proof. Let $M$ be a coherent $A$-module. By Proposition 4.13 there is a monomorphism $i\colon M\to F$ where $F$ is a f.g. free module. Now for any homomorphism $f\colon M\to J$, the diagram of solid arrows with exact row $\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{M\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{i}$$\scriptstyle{f}$$\textstyle{F\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{J}$ can be extended by a dashed arrow by injectivity. By Lam [TYL:LectModules&Rings]*theorem 4.32, $J$ is flat. ∎ Here is a summary of what we have established. ###### Proposition 4.21. Let $M$ be an $A$-module. Consider the following conditions: * (A) $M$ is relatively injective for the category of coherent $A$-modules; * (B) $M$ is flat; * (C) $M$ is projective; * (D) $M$ is a product of flat modules. Then we have the implications shown. (A)(B)(C)(D) Furthermore, if $M$ is coherent or more generally p.f.r., then $\text{\rm(B)}\Longrightarrow\text{\rm(C)}$. ###### Proof. The equivalence of (A) and (B) follows from Lemmas 4.19 and 4.20. The implication $\text{(D)}\Longrightarrow\text{(B)}$ follows from Chase’s theorem [TYL:LectModules&Rings]*theorem 4.47; in particular it applies to the case of a product of free modules. Actually we don’t really need to use Chase’s Theorem since a product of (relative) injectives is a (relative) injective. The last statement follows by Theorem 1.4(a) and Lam [TYL:LectModules&Rings]*theorem (4.30). ∎ We also mention another result on flat modules whose proof requires Corollary 4.27 which will be given later. ###### Proposition 4.22. Let $M$ be a f.d. $A$-module and $P$ a flat $A$-module. Then $\operatorname{Hom}_{A}(M,P)=0.$ ###### Proof. By the Lazard-Govorov Theorem 1.5, $P$ is a filtered colimit of f.g. free modules $F_{\alpha}$, $P=\operatorname*{colim}_{\alpha}F_{\alpha}.$ Since $M$ is f.g. and by Corollary 4.27, $\operatorname{Hom}_{A}(M,P)=\operatorname*{colim}_{\alpha}\operatorname{Hom}_{A}(M,F_{\alpha})=0.\qed$ ### Jacobson Radicals Recall for $\alpha\preccurlyeq\beta$, $A(\alpha)$ is left/right self-injective and so each inclusion $A(\alpha)\hookrightarrow A(\beta)$ is split as a left/right $A(\alpha)$-module homomorphism. So by Lam [TYL:NonCommRings]*proposition 5.6, (4.2) $A(\alpha)\cap\operatorname{rad}A(\beta)\subseteq\operatorname{rad}A(\alpha).$ Similarly, the inclusion $A(\alpha)\hookrightarrow A$ is split since $A$ is an injective $A(\alpha)$-module and (4.3) $A(\alpha)\cap\operatorname{rad}A\subseteq\operatorname{rad}A(\alpha).$ Since $A(\alpha)$ is Artinian, $\operatorname{rad}A(\alpha)$ is actually nilpotent and so nil, therefore $\operatorname{rad}A$ is also a nil ideal (in fact the largest one); see also Lam [TYL:NonCommRings]*proposition 4.19. If all of the $A(\alpha)$ are semisimple, then for $z\in\operatorname{rad}A$ there must be some $\gamma$ such that $z\in A(\gamma)$, hence $z\in A(\gamma)\cap\operatorname{rad}A\subseteq\operatorname{rad}A(\gamma)=\\{0\\}.$ Therefore $\operatorname{rad}A=\\{0\\}$ and $A$ is semiprimitive (or Jacobson semisimple in the terminology of Lam [TYL:NonCommRings]). In fact since each $A(\alpha)$ is von Neumann regular, it easily follows that $A$ is too; see Lam [TYL:NonCommRings]*corollary 4.24. Since $A$ is not Noetherian, it cannot be semisimple by Lam [TYL:NonCommRings]*corollary 4.25. For primitive ideals we have the following. If $\operatorname{ann}_{A}(S)$ is the annihilator of a simple $A$-module $S$ which restricts to a simple $A(\alpha)$-module for some $\alpha$, then for any $\beta\succcurlyeq\alpha$, $A(\beta)\cap\operatorname{ann}_{A}(S)=\operatorname{ann}_{A(\beta)}(S).$ ###### Proposition 4.23. Suppose that each $A(\alpha)$ is local with maximal ideal $A(\alpha)^{+}$, so that each homomorphism $A(\alpha)\to A(\beta)$ is local. Then $A$ is also local. ###### Proof. The kernel of the augmentation $\varepsilon\colon A\to\Bbbk$ is the maximal ideal $A^{+}\lhd A$ for which $A(\alpha)\cap A^{+}\subseteq A(\alpha)^{+}=\operatorname{rad}A(\alpha).$ Now given $z\in A^{+}$ and $a\in A$ we can assume that $z,a\in A(\gamma)$ for some $\gamma$ and so $1+az$ has a left inverse in $A(\gamma)\subseteq A$; a similar argument shows that $1+za$ has a right unit, therefore $A^{+}=\operatorname{rad}A$. ∎ ### Annihilators We will discuss left annihilators and modules, but similar considerations apply to right annihilators and modules. Let $a\in A$. Then the left $A$-module homomorphism $A\to Aa;\quad x\mapsto xa$ fits into a short exact sequence $0\to\operatorname{ann}^{\mathrm{l}}_{A}(a)\to A\to Aa\to 0$ where $Aa\subseteq A$ is a f.g. submodule and so f.p., therefore $\operatorname{ann}^{\mathrm{l}}_{A}(a)$ is also a f.p. module. More generally we have ###### Proposition 4.24. Let $M$ be a coherent $A$-module and $m\in M$. Then the left ideal $\operatorname{ann}_{A}(m)\subseteq A$ is a f.p. submodule, hence it is a coherent $A$-module. More generally, if $W\subseteq M$ is a f.d. $\Bbbk$-subspace then its annihilator $\operatorname{ann}_{A}(W)$ is f.g. as a left ideal, hence $\operatorname{ann}_{R}(W)$ and $M/\operatorname{ann}_{R}(W)$ are coherent modules. ###### Proof. For the first part, see [TYL:LectModules&Rings]*theorem 2.4.58. If $n_{1},\ldots,n_{k}$ span the $\Bbbk$-vector space $W$, then $\operatorname{ann}_{R}(W)=\operatorname{ann}_{R}(n_{1})\cap\cdots\cap\operatorname{ann}_{R}(n_{k}).$ By a well-known argument, the intersection of two f.g. submodules of a coherent module is f.g. and so coherent. Hence, $\operatorname{ann}_{R}(W)$ and $M/\operatorname{ann}_{R}(W)$ are coherent modules. ∎ ### Finite and simple modules Although $A$ will have simple modules, in general they will not all be coherent. Frobenius algebras are Kasch algebras (i.e., every left/right simple module is isomorphic to a left/right ideal). The situation for a locally Frobenius algebra $A$ is less straightforward. For example, the trivial module $\Bbbk$ can never be isomorphic to a submodule because of the next result characterising the simple modules which occur as minimal left/right ideals of $A$. ###### Proposition 4.25. Let $S$ be a simple $A$-module. Then $S$ is isomorphic to submodule of $A$ if and only if it is coherent. ###### Proof. Assume that $S\subseteq A$ and $0\neq s\in S$. Since $As=S$, Proposition 4.24 implies that there is a short exact sequence $0\to\operatorname{ann}^{\mathrm{l}}_{A}(s)\to A\to S\to 0$ with $\operatorname{ann}^{\mathrm{l}}_{A}(s)$ f.p., so $S$ is coherent. For the converse, suppose that $S$ is a coherent simple $A$-module. By Proposition 4.13, there is an embedding $j\colon S\hookrightarrow F$ into a f.g. free module. At least one of the compositions of $j$ with the projection onto a copy of $A$ must be non-zero and so a monomorphism by simplicity, hence $S$ is isomorphic to a submodule of $A$. ∎ We already know that f.d. simple modules are not coherent by Corollary 4.7, so combining this with Proposition 4.25 we obtain an important consequence. ###### Corollary 4.26. A f.d. simple module $S$ is not isomorphic to a submodule of $A$, or equivalently $\operatorname{Hom}_{A}(S,A)=0$. More generally, for a coherent $A$-module $N$, $\operatorname{Hom}_{A}(S,N)=0$. In particular this applies to the trivial module $\Bbbk$. ###### Proof. For the second statement, recall that there is an embedding $N\hookrightarrow F$ into a finitely generated free module. The short exact sequence $0\to N\to F\to F/N\to 0$ induces a long exact sequence beginning with $\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{\operatorname{Hom}_{A}(S,N)\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{\operatorname{Hom}_{A}(S,F)\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{\operatorname{Hom}_{A}(S,F/N)\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{\cdots}$$\textstyle{0}$ so $\operatorname{Hom}_{A}(S,M)=0$. ∎ We can extend this result to arbitrary f.d. modules. ###### Corollary 4.27. Suppose that $M$ is a f.d. $A$-module. Then for any coherent module $N$, $\operatorname{Hom}_{A}(M,N)=0.$ ###### Proof. This can be proved by induction on the length of a composition series of $M$. The inductive step uses the fact that for a f.d. simple module $S$, we have $\operatorname{Hom}_{A}(S,N)=0$ by Propositions 4.25 and 4.13. ∎ We also have the following. ###### Proposition 4.28. Suppose that $L\subseteq A$ is a minimal left ideal and $0\neq s\in L$. Then at least one of $L\subseteq A^{+}$ or $\operatorname{ann}^{\mathrm{l}}_{A}(s)\subseteq A^{+}$ must be true. Furthermore, if $\operatorname{ann}^{\mathrm{l}}_{A}(s)\subseteq A^{+}$ then $\dim_{\Bbbk}L=\infty$. Similar results are true for minimal right ideals. ###### Proof. By Proposition 4.25, $\operatorname{ann}^{\mathrm{l}}_{A}(s)$ is f.g., and since $\operatorname{ann}^{\mathrm{l}}_{A}(s)s=0$ we have $s\in A^{+}$ or $\operatorname{ann}_{A}(s)\subseteq A^{+}$ because $A^{+}$ is completely prime. By Proposition 4.5, if $\operatorname{ann}_{A}(s)\subseteq A^{+}$ then $L\cong A/\operatorname{ann}_{A}(s)$ is infinite dimensional. ∎ ###### Proposition 4.29. Suppose that $S$ is a finite dimensional simple $A$-module. Then there is a $\lambda\in\Lambda$ for which the restriction of $S$ to $A(\lambda)$ is simple. Hence for $\lambda^{\prime}\in\Lambda$ with $\lambda\preccurlyeq\lambda^{\prime}$, the restriction of $S$ to $A(\lambda^{\prime})$ is also simple. ###### Proof. I would like to thank Ken Brown for the following proof. The quotient algebra $A/\operatorname{ann}_{A}(S)$ is a finite dimensional simple $\Bbbk$-algebra with dimension $\dim_{\Bbbk}A/\operatorname{ann}_{A}(S)=m$ say. Lift a $\Bbbk$-basis of $A/\operatorname{ann}_{A}(S)$ to a linearly independent set $E=\\{e_{1},\ldots,e_{m}\\}\subseteq A$ which is contained in some $A(\lambda)$. We also have the ideal $J=A(\lambda)\cap\operatorname{ann}_{A}(S)\lhd A(\lambda)$ and the composition $A(\lambda)\hookrightarrow A\to A/\operatorname{ann}_{A}(S)$ factors through an injective homomorphism $A(\lambda)/J\to A/\operatorname{ann}_{A}(S)$. Since the images of the elements of $E$ in $A(\lambda)/J$ are still linearly independent, we have $\dim_{\Bbbk}(A(\lambda)/J)\geqslant m$; but as $\dim_{\Bbbk}A(\lambda)/J\leqslant m$ we obtain $\dim_{\Bbbk}A(\lambda)/J=\dim_{\Bbbk}A/\operatorname{ann}_{A}(S)$. Therefore we have $A(\lambda)/J\cong A/\operatorname{ann}_{A}(S)$, making a $A(\lambda)/J$ simple Artinian ring. But we know that $S$ is the (unique) simple $A/\operatorname{ann}_{A}(S)$-module so it is also simple as an $A(\lambda)/J$-module and as an $A(\lambda)$-module. The other statement is clear. ∎ ###### Corollary 4.30. Suppose that $S$ is a f.d. non-coherent simple $A$-module. Then there is an $\alpha\in\Lambda$ such that * • $S$ is a simple $A(\alpha)$-module; * • there is an embedding of $A(\alpha)$-modules $S\hookrightarrow A(\alpha)$; * • there is a $\beta\succcurlyeq\alpha$ such that the composition $S\hookrightarrow A(\alpha)\hookrightarrow A(\beta)$ is not a homomorphism of $A(\beta)$-modules. Therefore there is an $a\in A(\beta)$ such that $aS\nsubseteq S$ and in particular $aS\setminus S\neq\varnothing$. ###### Proof. The first statement follows from Proposition 4.29 and the second is a consequence of a Frobenius algebras being a Kasch algebra. If for every $\beta\succcurlyeq\alpha$ this composition were a $A(\beta)$-module homomorphism then we would obtain an $A$-module embedding $S\hookrightarrow A$, contradicting Proposition 4.25. ∎ Of course there may be infinite dimensional minimal left/right ideals which satisfy one of the conditions $L^{2}=0$ or $L^{2}=L$. If $L=Az$ and $L^{2}=0$ then $z\in A^{+}$ since this ideal is completely prime. The next results were suggested by analogous results of Richardson [JSR:GpRngsnonzerosocle]. ###### Lemma 4.31. Let $L$ be a minimal left/right ideal in $A$. Then there is a $\lambda\in\Lambda$ for which $L(\lambda)=A(\lambda)\cap L$ is a non-trivial minimal left/right ideal in $A(\lambda)$ and $L=AL(\lambda)\cong A\otimes_{A(\lambda)}L(\lambda).$ ###### Proof. We outline the proof for left ideals. Write $L=Az$ and choose $\lambda\in\Lambda$ so that $z\in A(\lambda)$. Then $L(\lambda)=A(\lambda)\cap L$ is a non-trivial left ideal in $A(\lambda)$ and $AL(\lambda)=L$. By the well-known criterion for flatness of [TYL:LectModules&Rings]*(4.12), multiplication gives an isomorphism $A\otimes_{A(\lambda)}L(\lambda)\xrightarrow{\cong}AL(\lambda)=L$ and Corollary 4.11 implies that $L(\lambda)$ is simple. ∎ ###### Corollary 4.32. Let $L$ be a minimal left ideal in $A$. Then either $L^{2}=0$ or there is an idempotent $e\in A$ such that $L=Ae$. Similarly for a minimal right ideal in $A$, either $L^{2}=0$ or $L$ is generated by an idempotent. ###### Proof. If $L^{2}\neq 0$ then suppose $L(\lambda)=A(\lambda)\cap L\neq 0$. In the Artinan ring $A(\lambda)$, the minimal ideal $L(\lambda)=L(\lambda)^{2}$ is generated by an idempotent $e$ and $L=Ae$. ∎ Our next result and its proof are direct generalisations of Richardson [JSR:GpRngsnonzerosocle]*proposition 2.8. ###### Proposition 4.33. Let $0\neq w\in A$. Then $Aw$ is a minimal left ideal of $A$ if and only if $wA$ is a minimal right ideal. Hence the left and right socles of $A$ agree. ###### Proof. If $Aw$ is a minimal left ideal then by Lemma 4.31 there is a $\lambda\in\Lambda$ such that $w\in A(\lambda)$ and $A(\lambda)\cap Aw$ is a minimal left ideal in $A(\lambda)$. Since $wA$ is the union of the $wA(\lambda)$ over all such $\lambda$, it suffices to show that $wA(\lambda)$ is a minimal right ideal. There is an isomorphism of left $A(\lambda)$-modules $A(\lambda)w\xrightarrow{\cong}A(\lambda)/\operatorname{ann}^{\mathrm{l}}_{A(\lambda)}(w)$ so $\operatorname{ann}^{\mathrm{l}}_{A(\lambda)}(w)=\operatorname{ann}_{A(\lambda)}(wA(\lambda))$ which is a maximal left ideal. As $A(\lambda)$ is Frobenius its left and right submodule lattices correspond under an inclusion reversing bijection, therefore $wA(\lambda)$ is the annihilator ideal of $A(\lambda)w$ which is a minimal right ideal. ∎ Another result of Richardson [JSR:GpRngsnonzerosocle]*lemma 2.7 (see also Hartley & Richardson [BH&JSR:SocleGpRngs]) can be extended to a locally Frobenius algebra. We know that a minimal left ideal of $A$ is a coherent $A$-module and so infinite dimensional; nevertheless this result gives some sort of finiteness albeit over a division algebra which is infinite dimensional over $\Bbbk$. ###### Proposition 4.34. Let $L=Aw$ be a minimal left ideal of $A$. _(a)_ The ideal $L$ is finite dimensional over the division algebra $\operatorname{End}_{A}(L)$. _(b)_ Every finite subset of $\operatorname{End}_{A}(L)$ lies in a finite dimensional separable $\Bbbk$-subalgebra. _(c)_ If $\operatorname{char}\Bbbk>0$, then $\operatorname{End}_{A}(L)$ is a field. ###### Proof. This involves a routine reworking of that for locally finite group algebras and is left to the reader, the main point to notice is that group algebras of finite subgroups need to be replaced by Frobenius algebras $A(\lambda)$. ∎ For locally Frobenius Hopf algebras our results on f.d. modules should be compared with the following. ###### Proposition 4.35 (See [ML:TourRepThy]*proposition 10.6). Suppose that $H$ is a Hopf algebra over $\Bbbk$. If $H$ is infinite dimensional then it has no non-trivial f.d. left/right ideals. Combining this with Proposition 4.25 we obtain a result which also follows from Proposition 4.5. ###### Corollary 4.36. If $H$ is a locally Frobenius Hopf algebra then no f.d. simple module is coherent. ### Coherent projective covers In general f.g. modules over a coherent ring may not have projective covers in the usual sense. But we can define something similar for coherent modules over a locally Frobenius algebra $A$. Consider the coherent $A$-module $A\otimes_{A(\alpha)}M_{\alpha}$ where $M_{\alpha}$ is a f.g. (hence f.d.) $A(\alpha)$-module. If $\alpha\preccurlyeq\beta$ then $M_{\beta}=A(\beta)\otimes_{A(\alpha)}M_{\alpha}$ has dimension $\dim_{\Bbbk}M_{\beta}=\frac{\dim_{\Bbbk}A(\beta)\dim_{\Bbbk}M_{\alpha}}{\dim_{\Bbbk}A(\alpha)}.$ so $\frac{\dim_{\Bbbk}M_{\beta}}{\dim_{\Bbbk}A(\beta)}=\frac{\dim_{\Bbbk}M_{\alpha}}{\dim_{\Bbbk}A(\alpha)}.$ By Corollary 4.3, if there is an isomorphism of $A$-modules $A\otimes_{A(\alpha)}M_{\alpha}\cong A\otimes_{A(\beta)}N_{\beta}$ then for some large enough $\gamma$ there is an isomorphism of $A(\gamma)$-modules $M_{\gamma}=A(\gamma)\otimes_{A(\alpha)}M_{\alpha}\cong A(\gamma)\otimes_{A(\beta)}N_{\beta}=N_{\gamma},$ therefore $\frac{\dim_{\Bbbk}M_{\gamma}}{\dim_{\Bbbk}A(\gamma)}=\frac{\dim_{\Bbbk}N_{\gamma}}{\dim_{\Bbbk}A(\gamma)}.$ This shows that to any coherent $A$-module we can assign an isomorphism invariant rational number by taking its _coherent dimension_ to be $\operatorname{coh- dim}(A\otimes_{A(\alpha)}M_{\alpha})=\frac{\dim_{\Bbbk}M_{\alpha}}{\dim_{\Bbbk}A(\alpha)}.$ Now let $M$ be a coherent $A$-module. Then there exist epimorphisms of $A$-modules $P\to M$ where $P$ is a coherent projective module. By Corollary 4.12, such an epimorphism can be taken to be induced up from an epimorphism of $A(\alpha)$-modules $P_{\alpha}\to M_{\alpha}$ for some $\alpha\in\Lambda$, where $P_{\alpha}$ is projective; we will refer to such epimorphisms as _coherent epimorphisms_. Notice also that $\dim_{\Bbbk}P_{\alpha}\geqslant\dim_{\Bbbk}M_{\alpha}$ so $\operatorname{coh- dim}(A\otimes_{A(\alpha)}P_{\alpha})\geqslant\operatorname{coh- dim}(A\otimes_{A(\alpha)}M_{\alpha}).$ In fact we have the following useful observation. Let $\varphi\colon P\to N$ be a homomorphism of coherent $A$-modules with $P$ coherent projective, and let $\theta\colon M\to N$ be a coherent epimorphism. Then we can factor $\varphi$ through $M$. $\textstyle{M\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\theta}$$\textstyle{P\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\varphi}$$\scriptstyle{\widetilde{\varphi}}$$\textstyle{N}$ By choosing a large enough $\alpha\in\Lambda$ we can find underlying homomorphisms of f.g. $A(\alpha)$-modules $\textstyle{M_{\alpha}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\theta_{\alpha}}$$\textstyle{P_{\alpha}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\varphi_{\alpha}}$$\scriptstyle{\widetilde{\varphi}_{\alpha}}$$\textstyle{N_{\alpha}}$ which on applying $A\otimes_{A(\alpha)}(-)$ induce the diagram of $A$-modules. We need to check that this second diagram commutes. Since $\theta_{\alpha}\circ\widetilde{\varphi}_{\alpha}$ and $\varphi_{\alpha}$ induce the same homomorphism $P\to M$, so by Proposition 1.10(a) $\theta_{\alpha}\circ\widetilde{\varphi}_{\alpha}-\varphi_{\alpha}$ must be trivial and this diagram also commutes. Now suppose that we have two coherent epimorphisms $P\xrightarrow{\varphi}M\xleftarrow{\theta}Q$ where $P$ and $Q$ are coherent projectives. The preceding discussion shows that we can assume that for some $\lambda\in\Lambda$ there are there are projective $A(\lambda)$-modules $P_{\lambda}$ and $Q_{\lambda}$ for which $P\cong A\otimes_{A(\lambda)}P_{\lambda}$ and $Q\cong A\otimes_{A(\lambda)}Q_{\lambda}$ together with a f.g. $A(\lambda)$-module $M_{\lambda}$ for which $M\cong A\otimes_{A(\lambda)}M_{\lambda}$; furthermore there is a commutative diagram of $A(\lambda)$-modules $\textstyle{P_{\lambda}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\varphi_{\lambda}}$$\scriptstyle{\rho}$$\textstyle{Q_{\lambda}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\theta_{\lambda}}$$\scriptstyle{\sigma}$$\textstyle{P_{\lambda}\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\varphi_{\lambda}}$$\textstyle{M_{\lambda}}$ where $\varphi_{\lambda}$ and $\theta_{\lambda}$ induce $\varphi$ and $\theta$. On tensoring with $A$ and using faithful flatness we see that $\sigma\circ\rho=\operatorname{Id}-{P_{\lambda}}$, so $P_{\lambda}$ is a summand of $Q_{\lambda}$ and $\dim_{\Bbbk}P_{\lambda}\geqslant\dim_{\Bbbk}Q_{\lambda}$. We can now define the _coherent rank of $M$_ to be $\operatorname{coh-rank}M=\min\\{\operatorname{coh-dim}P:\text{there is a coherent epi. $P\to M$}\\}\geqslant\operatorname{coh-dim}(M).$ If we realise this minimum with a coherent epimorphism $P\to M$, then for any other such coherent epimorphism $Q\to M$ the above discussion shows that $\operatorname{coh-dim}P=\operatorname{coh-dim}Q$ and so $P\cong Q$. Thus we can regard $P\to M$ as a _coherent projective cover_. ### Pseudo-coherent modules In Bourbaki [Bourbaki:HomAlg]*ex. §3.10, as well as coherent modules, pseudo- coherent modules are considered, where a module is _pseudo-coherent_ if every finitely generated submodule is finitely presented (so over a coherent ring it is coherent). The reader is warned that pseudo-coherent is sometimes used in a different sense; for example, over a coherent ring the definition of Weibel [CAW:Ktheory]*example II.7.1.4 corresponds to our coherent. Examples of pseudo-coherent modules over a coherent ring $A$ include coproducts of coherent modules such as the f.r. modules of Definition 1.1. However pseudo-coherent modules do not form a full abelian subcategory of $\mathbf{Mod}_{A}$ since for example cokernels of homomorphisms between pseudo-coherent modules need not be pseudo-coherent. Nevertheless, pseudo- coherent modules do occur quite commonly when working with coherent modules over locally Frobenius algebras. ###### Lemma 4.37. Let $A$ be a ring and $M$ a pseudo-coherent $A$-module. Then $M$ is the union of its coherent submodules and therefore their colimit. ###### Proof. Every element $m\in M$ generates a cyclic submodule which is coherent. ∎ ###### Lemma 4.38. Let $A$ be a coherent ring and $B$ an Artinian subring where $A$ is flat as a right $B$-module. Then every extended $A$-module $A\otimes_{B}N$ is pseudo- coherent. ###### Proof. Let $U\subseteq A\otimes_{B}N$ be a finitely generated submodule. Taking a finite generating set and expressing each element as a sum of basic tensors we find that $U\subseteq A\otimes_{B}U^{\prime}$ where $U^{\prime}\subseteq N$ is a finitely generated submodule. As $B$ is Artinian and so Noetherian, $U^{\prime}$ is finitely presented, so by flatness of $A$, $A\otimes_{B}U^{\prime}$ is a finitely presented $A$-module. Since $A$ is coherent, $U$ is also finitely presented. ∎ An important special case of this occurs when $A$ is a coherent algebra over a field $\Bbbk$ and $B$ is a finite dimensional subalgebra. If $N$ is a $B$-module then it is locally finite, i.e., every element is contained in a finite dimensional submodule. We will discuss pseudo-coherence for modules over a locally Frobenius Hopf algebra in Section 6. The following stronger notions of pseudo-coherence are perhaps more likely to be important for example for computational purposes. ###### Definition 4.39. * • A module over a coherent ring is _strongly pseudo-coherent_ if it is a coproduct of coherent modules. * • A module $M$ over a locally Frobenius algebra $A$ indexed on $\Lambda$ is _$\lambda$ -strongly pseudo-coherent_ for $\lambda\in\Lambda$ if $M\cong A\otimes_{A(\lambda)}M^{\prime}$ where the $A(\lambda)$-module $M^{\prime}$ is a coproduct of finitely generated $A(\lambda)$-modules. Over a coherent ring every free module is strongly pseudo-coherent as is every finitely related module since it is the sum of a coherent module and a free module. Clearly a $\lambda$-strongly pseudo-coherent module over a locally Frobenius algebra $A$ is strongly pseudo-coherent. ### Cohomology for finite dimensional $A$-modules To end this section we discuss $\operatorname{Ext}$ groups for finite dimensional modules. We will work with left modules, but a similar discussion applies to right modules. If $W$ is an $A(\lambda)$-module for some $\lambda\in\Lambda$ and $N$ is an $A$-module (hence an $A(\lambda)$-module via restriction), then $A\otimes_{A(\lambda)}W$ is an $A$-module and $\operatorname{Hom}_{A(\lambda)}(W,N)\cong\operatorname{Hom}_{A(\lambda)}(A\otimes_{A(\lambda)}W,N).$ More generally, since $A$ is $A(\lambda)$-flat, $\operatorname{Ext}^{s}_{A(\lambda)}(W,N)\cong\operatorname{Ext}^{s}_{A}(A\otimes_{A(\lambda)}W,N).$ If $M,N$ are $A$-modules viewed as $A(\lambda)$-modules through restriction, then $\operatorname*{colim}_{(\Lambda,\preccurlyeq)}A\otimes_{A(\lambda)}M\cong M$ as $A$-modules and (4.4) $\operatorname{Hom}_{A}(M,N)\cong\operatorname{Hom}_{A}(\operatorname*{colim}_{(\Lambda,\preccurlyeq)}A\otimes_{A(\lambda)}M,N)\cong\lim_{(\Lambda,\preccurlyeq)}\operatorname{Hom}_{A(\lambda)}(M,N).$ There is a spectral sequence of Jensen [LNM254]*théorème 4.2 of the form $\mathrm{E}_{2}^{s,t}={\lim_{(\Lambda,\preccurlyeq)}}^{\\!s}\operatorname{Ext}^{t}_{A}(A\otimes_{A(\lambda)}M,N)\Longrightarrow\operatorname{Ext}^{s+t}_{A}(M,N),$ where $\displaystyle{\lim_{(\Lambda,\preccurlyeq)}}^{\\!s}$ is the $s$-th right derived functor of $\displaystyle\lim_{(\Lambda,\preccurlyeq)}$. The $\mathrm{E}_{2}$-term can be rewritten to give (4.5) $\mathrm{E}_{2}^{s,t}={\lim_{(\Lambda,\preccurlyeq)}}^{\\!s}\operatorname{Ext}^{t}_{A(\lambda)}(M,N)\Longrightarrow\operatorname{Ext}^{s+t}_{A}(M,N).$ When $\Lambda$ is countable, by a result of Jensen [LNM254], for $s>1$, $\displaystyle{\lim_{\lambda\in\Lambda}}^{\\!s}$ is trivial, so for each $n\geqslant 1$ there is an exact sequence (4.6) $0\to{\lim_{(\Lambda,\preccurlyeq)}}^{\\!1}\operatorname{Ext}^{n-1}_{A(\lambda)}(M,N)\to\operatorname{Ext}^{n}_{A}(M,N)\to\lim_{(\Lambda,\preccurlyeq)}\operatorname{Ext}^{n}_{A(\lambda)}(M,N).\to 0$ Suppose that $J$ is an injective $A$-module and $\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{U\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{V}$$\textstyle{J}$ is a diagram of $A(\lambda)$-modules with an exact row. By flatness of $A$ as an $A(\lambda)$-module, $\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{A\otimes_{A(\lambda)}U\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{A\otimes_{A(\lambda)}V}$ is an exact sequence of $A$-modules and on applying $\operatorname{Hom}_{A(\lambda)}(-,J)$ we obtain a commutative diagram $\textstyle{\operatorname{Hom}_{A(\lambda)}(U,J)\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\cong}$$\textstyle{\ignorespaces\ignorespaces\ignorespaces\ignorespaces\operatorname{Hom}_{A(\lambda)}(V,J)\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{\cong}$$\textstyle{0}$$\textstyle{\ignorespaces\ignorespaces\ignorespaces\ignorespaces\operatorname{Hom}_{A}(A\otimes_{A(\lambda)}U,J)}$$\textstyle{\ignorespaces\ignorespaces\ignorespaces\ignorespaces\operatorname{Hom}_{A}(A\otimes_{A(\lambda)}V,J)}$ where the lower row is exact by injectivity of the $A$-module $J$. It follows that $J$ is an injective $A(\lambda)$-module. Now taking $N=J$ and using (4.6) with $n=1$ we obtain ${\lim_{(\Lambda,\preccurlyeq)}}^{\\!1}\operatorname{Hom}_{A(\lambda)}(M,J)=0,$ and by using (4.4), $\operatorname{Hom}_{A}(M,J)\cong\lim_{(\Lambda,\preccurlyeq)}\operatorname{Hom}_{A(\lambda)}(M,J).$ We know that $A$ is relatively injective with respect to coherent $A$-modules. For a finite dimensional non-coherent $A$-module $M$, the spectral sequence (4.5) gives $\mathrm{E}_{2}^{s,t}={\lim_{(\Lambda,\preccurlyeq)}}^{\\!s}\operatorname{Ext}^{t}_{A(\lambda)}(M,A)\Longrightarrow\operatorname{Ext}^{s+t}_{A}(M,A)$ and as $A$ is $A(\lambda)$-injective for every $\lambda$, while if $t=0$, $\mathrm{E}_{2}^{s,t}=\begin{dcases*}{\lim_{(\Lambda,\preccurlyeq)}}^{\\!s}\operatorname{Hom}_{A(\lambda)}(M,A)&if $t=0$,\\\ \quad\quad\quad\quad\quad 0&if $t>0$.\end{dcases*}$ So we have $\operatorname{Ext}^{n}_{A}(M,A)\cong{\lim_{(\Lambda,\preccurlyeq)}}^{\\!n}\operatorname{Hom}_{A(\lambda)}(M,A).$ In particular, if $\Lambda$ is countable, $\operatorname{Ext}^{n}_{A}(M,A)=0$ when $n>1$. ###### Definition 4.40. If $S$ is a finite dimensional simple $A$-module then $A$ is _$S$ -Margolisian_ if $\operatorname{Ext}^{n}_{A}(S,A)=0$ for all $n>0$. If $A$ is $S$-Margolisian for all finite dimensional simple $A$-modules then $A$ is _Margolisian_. So $A$ is Margolisian if for every finite dimensional non-coherent simple $A$-module $S$ and every $s>0$, ${\lim_{(\Lambda,\preccurlyeq)}}^{\\!s}\operatorname{Hom}_{A(\lambda)}(M,A)=0.$ We can summarise this in a result which is obtained by combining these ideas with what we already know for coherent finite dimensional $A$-modules. ###### Proposition 4.41. $A$ is Margolisian if and only if $A$ is relatively injective with respect to the abelian category $\mathbf{Mod}_{A}^{\mathrm{f.d.}}$ of all finite dimensional $A$-modules. Margolis [HRM:Book] shows that every $P$-algebra is $\Bbbk$-Margolisian in our sense. In this case $\Bbbk$ is the only simple module. His proof makes essential use of the fact that its modules are graded and we have neither been able to find an argument that works in our situation nor a counter example, so it seems possible that every locally Frobenius algebra is Margolisian, or at least $\Bbbk$-Margolisian. ###### Remark 4.42. As in Remark 4.18, for a f.d. $A$-module $M$, the left exact functor $\operatorname{Hom}_{A}(M,-)$ on the category of all $A$-modules has the right derived functors $\operatorname{Ext}_{A}^{*}(M,-)$. If $A$ is Margolisian then for each finitely generated free module $F$, $\operatorname{Ext}_{A}^{*}(M,F)=0$ and $F$ is $\operatorname{Hom}_{A}(M,-)$-acyclic, so these right derived functors can be computed using resolutions by such modules which are the injectives in $\mathbf{Mod}_{A}^{\mathrm{coh}}$. So they are also right derived functors on $\mathbf{Mod}_{A}^{\mathrm{coh}}$. But then $\operatorname{Ext}_{A}^{*}(M,N)=0$ for every coherent module $N$. We now discuss the finite dual of a locally Frobenius $\Bbbk$-algebra $A$ and some related ideas. As we could not find a convenient reference we give details which are probably well known. Recall that the _finite dual_ of $A$ is $A^{\circ}=\\{f\in\operatorname{Hom}_{\Bbbk}(A,\Bbbk):\text{$\exists I\lhd A$ s.t. $I$ is cofinite and $I\subseteq\ker f$}\\}\subseteq\operatorname{Hom}_{\Bbbk}(A,\Bbbk).$ Here $I$ is _cofinite_ if $\dim_{\Bbbk}A/I<\infty$. It is clear that $A^{\circ}\subseteq\operatorname{Hom}_{\Bbbk}(A,\Bbbk)$ is a $\Bbbk$-subspace. There are two $A$-module structures on $\operatorname{Hom}_{\Bbbk}(A,\Bbbk)$ namely a left one induced by right multiplication on the domain and a right one induced by left multiplication on the domain. ###### Lemma 4.43. The left and right $A$-module structures on $A^{\circ}$ each restrict to make $A^{\circ}$ a submodule of $\operatorname{Hom}_{\Bbbk}(A,\Bbbk)$. Furthermore, if $M$ is a f.d. $A$-module the image of every $A$-module homomorphism $M\to\operatorname{Hom}_{\Bbbk}(A,\Bbbk)$ is contained in $A^{\circ}$. ###### Proof. We give the argument for the left module structure. Let $f\in A^{0}$ be trivial on a cofinite ideal $I\lhd A$. If $a\in A$ and $x\in I$ then $(af)(x)=f(xa)=0$ since $xa\in I$. Since $M$ is f.d., the associated algebra homomorphism $A\to\operatorname{End}_{\Bbbk}(M)$ has a cofinite kernel $J\lhd A$ say. Now for an $A$-module homomorphism $M\to\operatorname{Hom}_{\Bbbk}(A,\Bbbk)$, the image of each $m\in M$ must vanish on $J$. ∎ If $W$ is a f.d. vector space then we can similarly define $\operatorname{Hom}_{\Bbbk}^{\mathrm{fin}}(A,W)=\\{f\in\operatorname{Hom}_{\Bbbk}(A,W):\text{$\exists I\lhd A$ s.t. $I$ is cofinite and $I\subseteq\ker f$}\\}\subseteq\operatorname{Hom}_{\Bbbk}(A,W).$ Then the obvious left and right $A$-module structures on $\operatorname{Hom}_{\Bbbk}(A,W)$ also restrict to make $\operatorname{Hom}_{\Bbbk}^{\mathrm{fin}}(A,W)$ a left or right submodule. If we choose a basis for $W$ we obtain isomorphisms of left and right $A$-modules $\operatorname{Hom}_{\Bbbk}^{\mathrm{fin}}(A,W)\cong(A^{\circ})^{\dim_{\Bbbk}W}.$ An analogue of this is also true when $W$ is infinite dimensional. It is standard that injectives in the module category $\mathbf{Mod}_{A}$ are summands of modules of the form $\operatorname{Hom}_{\Bbbk}(A,W)$. In particular every $A$-module $M$ admits an embedding $M\hookrightarrow\operatorname{Hom}_{\Bbbk}(A,M)$ where the copy of $M$ in the codomain is viewed just as a vector space. ###### Lemma 4.44. Suppose that $W$ is a f.d. vector space. Then $\operatorname{Hom}_{\Bbbk}^{\mathrm{fin}}(A,W)$ is a relative injective with respect to the category of f.d. $A$-modules. Furthermore every f.d. $A$-module admits a resolution by such relative injectives. ###### Proof. It suffices to show this for $A^{\circ}$ itself. Suppose we have the solid diagram of $A$-modules $\textstyle{0\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{U\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{V\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{A^{\circ}\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\textstyle{\operatorname{Hom}_{\Bbbk}(A,\Bbbk)}$ where $U$ and $V$ are f.d. and the row is exact. Because $\operatorname{Hom}_{\Bbbk}(A,\Bbbk)$ is a genuine injective $A$-module there is a dotted arrow making the diagram commute. By the second part of Lemma 4.43 its image is contained in $A^{\circ}$ so we obtain a dashed arrow making the diagram commute. ∎ This means that to calculate $\operatorname{Ext}_{A}^{*}(M,N)$ where $M$ and $N$ are f.d., we can use a resolution of $N$ by $A$-modules of the form $\operatorname{Hom}_{\Bbbk}^{\mathrm{fin}}(A,W)$. Of course this is not a resolution in $\mathbf{Mod}_{A}^{\mathrm{f.d.}}$, but it is in the category of _locally finite_ $A$-modules $\mathbf{Mod}_{A}^{\mathrm{l.f.}}$ which is a full abelian subcategory of $\mathbf{Mod}_{A}$. ## 5\. Stable module categories for locally Frobenius algebras For an arbitrary ring there are two distinct ways to define a stable module category starting with the category of modules, namely by treating as trivial those homomorphisms which factor through either projectives or injectives; for a Frobenius algebra these coincide. A locally Frobenius algebra $A$ is not self-injective, but it is injective relative to certain types of $A$-modules, in particular coherent modules and f.r. modules. So starting with the abelian category of coherent $A$-modules we obtain a stable module category using either approach. However, there are grounds for thinking that f.r. or pseudo- coherent modules should also be included although neither form an abelian category in an obvious way. We adopt some standard notions for stable module categories for the stable module category of coherent $A$-modules which we will denote by $\mathbf{Stmod}^{\mathrm{coh}}_{A}$. We will also use the notation $\\{M,N\\}=\mathbf{Stmod}^{\mathrm{coh}}_{A}(M,N)=\mathbf{Mod}^{\mathrm{coh}}_{A}(M,N)/\approx$ for the morphisms, where for two homomorphisms $f,g\colon M\to N$, $f\approx g$ if and only if there is a factorisation $\textstyle{M\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$$\scriptstyle{f-g}$$\textstyle{N}$$\textstyle{F\ignorespaces\ignorespaces\ignorespaces\ignorespaces}$ where $F$ is a f.g. free module. For a coherent module $M$, the objects $\Omega M$ and $\mho M$ are defined by taking exact sequences $P\xrightarrow{p}M\to 0,\quad 0\to M\xrightarrow{j}J$ where $P$ and $J$ are f.g. free and hence injective modules. Using Schanuel’s Lemma, $\ker p$ and $\operatorname{coker}j$ are well-defined up to isomorphism in $\mathbf{Stmod}^{\mathrm{coh}}_{A}$ so we may use these as representatives of $\Omega M$ and $\mho M$. ###### Lemma 5.1. In $\mathbf{Stmod}^{\mathrm{coh}}_{A}$, every object $M$ is isomorphic to objects of the form $\Omega M^{\prime}$ and $\mho M^{\prime\prime}$ for some coherent modules $M^{\prime}$ and $M^{\prime\prime}$. Then $\Omega$ and $\mho$ define mutually inverse endofunctors of $\mathbf{Stmod}^{\mathrm{coh}}_{A}$. ###### Proof. By Proposition 4.16(a) we know that $M$ is a submodule of a f.g. free module $F$ so there is an exact sequence $0\to M\to F\to M^{\prime}\to 0$ showing that $M=\Omega M^{\prime}$. On the other hand there is also an epimorphism $F^{\prime}\to M$ for some f.g. free module $F^{\prime}$ and by Lemma 4.14 this is injective, hence there is an exact sequence $0\to M^{\prime\prime}\to F^{\prime}\to M\to 0$ and therefore $M=\mho M^{\prime\prime}$. This also shows that $M^{\prime\prime}\cong\Omega M$ and so $M\cong\mho\Omega M$. A similar argument shows that $M\cong\Omega\mho M$. ∎ ###### Proposition 5.2. The functors $\Omega,\mho\colon\mathbf{Stmod}_{A}^{\mathrm{coh}}\to\mathbf{Stmod}_{A}^{\mathrm{coh}}$ satisfy the following for all coherent $A$-modules $M$ and $N$ and these isomorphisms are natural: $\\{\Omega M,\Omega N\\}\cong\\{M,N\\}\cong\\{\mho M,\mho N\\}.$ We extend $\\{-,-\\}$ to a graded $\mathbb{Z}$-bifunctor by setting $\\{M,N\\}^{n}=\\{M,N\\}_{-n}=\begin{dcases*}\\{M,\mho^{n}N\\}&if $n\geqslant 0$,\\\ \\{\Omega^{-n}M,N\\}&if $n<0$.\end{dcases*}$ This leads to a triangulated structure on $\mathbf{Stmod}_{A}^{\mathrm{coh}}$. Of course the point of this definition is that when $n>0$, $\operatorname{Ext}_{A}^{n}(M,N)\cong\\{M,N\\}^{n}.$ A major defect of this stable module category in the case when $A$ is a locally Frobenius Hopf algebra is its lack of an obvious monoidal structure: the tensor product $M\otimes_{\Bbbk}N$ of two coherent modules with the diagonal action need not be coherent as an $A$-module. This problem disappears if one of the factors is instead a f.d. module. We will discuss this further in Section 6. ## 6\. Locally Frobenius Hopf algebras In this section we consider results on locally Frobenius Hopf algebras where the Hopf structure plays a rôle. We are particularly motivated by the goal of extending results on locally finite group algebras. A convenient source for background material on Hopf algebras is provided by Montgomery [SM:HopfAlgActions]. ###### Assumption 6.1. We will assume that $H$ is a Hopf algebra over the field $\Bbbk$ with coproduct $\psi\colon H\to H\otimes H$ and antipode $\chi\colon H\to H$, where will use the notation for these: $\psi(h)=\sum_{i}h^{\prime}_{i}\otimes h_{i}^{\prime\prime},\quad\overline{h}=\chi(h).$ We will also assume that $K\subseteq H$ is a subHopf algebra with $H$ being flat as a left and right $K$-module. Now given two left $H$-modules $L$ and $M$, their tensor product $L\otimes M$ is a left $H$-module with the diagonal action given by $h\cdot(\ell\otimes m)=\sum_{i}h^{\prime}_{i}\ell\otimes h^{\prime\prime}_{i}m,$ where the coproduct on $h$ is $\psi h=\sum_{i}h^{\prime}_{i}\otimes h^{\prime\prime}_{i}.$ In particular, given a left $H$-module $M$ and a left $K$-module $N$, the tensor product of $M$ and $H\otimes_{K}N$ is a left $H$-module $M\otimes(H\otimes_{K}N)$. There is an isomorphism of $H$-modules (6.1) $M\otimes(H\otimes_{K}N)\xrightarrow[\cong]{\Theta}H\otimes_{K}(M\otimes N);\quad m\otimes(h\otimes n)\mapsto\sum_{i}h^{\prime}_{i}\otimes(\overline{h^{\prime\prime}_{i}}m\otimes n)$ where $\overline{x}=\chi(x)$ and $M\otimes N$ is a left $K$-module with the diagonal action. A particular instance of this is (6.2) $(H/\\!/K)\otimes(H/\\!/K)\cong H\otimes_{K}(H/\\!/KB)$ where $H/\\!/K=H/HK^{+}\cong H\otimes_{K}\Bbbk.$ ###### Lemma 6.2. Suppose that $H$ is coherent and $K$ is finite dimensional. Let $M$ be a left $H$-module and $N$ a left $K$-module. Then the $H$-module $M\otimes(H\otimes_{K}N)$ is pseudo-coherent. ###### Proof. Using (6.1), $M\otimes(H\otimes_{K}N)\cong H\otimes_{K}(M\otimes N)$ which is pseudo-coherent by Lemma 4.38. ∎ In general, the tensor product of two coherent modules over a locally Frobenius Hopf algebra is not a coherent module. However, it turns out that it is pseudo-coherent. ###### Proposition 6.3. Suppose that $H$ is a locally Frobenius Hopf algebra and that $L$ and $M$ are two coherent left $H$-modules. Then the $H$-module $L\otimes M$ is pseudo- coherent. ###### Proof. Every coherent $H$-module is induced from a finitely generated $H(\lambda)$-module for some $\lambda$. By choosing a large enough $\lambda$ we can assume that $L\cong H\otimes_{H(\lambda)}L^{\prime}$ for some finitely generated $H(\lambda)$-module $L^{\prime}$. Then as $H$-modules, $L\otimes M\cong H\otimes_{H(\lambda)}(L^{\prime}\otimes M),$ which is a pseudo-coherent $H$-module. ∎ If $H$ is a locally Frobenius Hopf algebra, then for $\lambda$, the cyclic $H$-module $H/\\!/H(\lambda)=H/HH(\lambda)^{+}\cong H\otimes_{H(\lambda)}\Bbbk$ can be viewed as an $H(\lambda)$-module, and by (6.2) there is an isomorphism of $H$-modules $H/\\!/H(\lambda)\otimes H/\\!/H(\lambda)\cong H\otimes_{H(\lambda)}H/\\!/H(\lambda).$ So if $H/\\!/H(\lambda)$ is a coproduct of f.g./f.d. $H(\lambda)$-modules then $H/\\!/H(\lambda)\otimes H/\\!/H(\lambda)$ is $\lambda$-strongly pseudo- coherent $H$-module. Now we consider _normality_ for subHopf algebras of locally Frobenius Hopf algebras. ###### Definition 6.4. The left and right _adjoint actions_ of $h\in H$ on $x\in H$ are given by $h\cdot x=\sum_{i}h^{\prime}_{i}x\overline{h_{i}^{\prime\prime}},\quad x\cdot h=\sum_{i}\overline{h^{\prime}_{i}}xh_{i}^{\prime\prime}.$ When $H=\Bbbk G$ is a group algebra then if $g\in G$, the adjoint action agrees with conjugation by $g$. The adjoint action of $H$ on itself is compatible with the product $\varphi\colon H\otimes H\to H$ and coproduct $\psi\colon H\to H\otimes H$ provided $H\otimes H$ is given the diagonal coactions for which $h\cdot(x\otimes y)=\sum_{i}h^{\prime}_{i}\cdot x\otimes h^{\prime\prime}_{i}\cdot y,\quad(x\otimes y)\cdot h=\sum_{i}x\cdot h^{\prime}_{i}\otimes y\cdot h^{\prime\prime}_{i}.$ In the following definition, the notion of _strongly normalised_ is motivated by the case of the group algebra of a locally finite group $G$. If a finite subgroup $H\leqslant G$ normalises a subgroup $K$, then for any finite subgroup $K^{\prime}\leqslant K$, the finite set $\bigcup_{h\in H}hK^{\prime}h^{-1}\subseteq K$ is contained in some finite subgroup $K^{\prime}\leqslant K$. This has implications for the image of the adjoint action of $\Bbbk H$ on $\Bbbk K^{\prime}$ which is contained in $\Bbbk K^{\prime\prime}$. For a locally Frobenius Hopf algebra our definition specialises to this case. ###### Definition 6.5. A sub(Hopf) algebra $K\subseteq H$ is _normal_ if for all $h\in H$, $h\cdot K\subseteq K$ and $K\cdot h\subseteq K$. If $H^{\prime}\subseteq H$ is a subHopf algebra, then $K$ is _normalised by $H^{\prime}$_ if for every $h\in H^{\prime}$, $h\cdot K\subseteq K$ and $K\cdot h\subseteq K$. If $K\subseteq H$ is a locally Frobenius subHopf algebra of shape $\Lambda^{\prime}\subseteq\Lambda$ then $K$ is _strongly normalised_ by a subHopf algebra $H^{\prime}\subseteq H$ if for all $h\in H^{\prime}$ and $\lambda\in\Lambda^{\prime}$, there is a $\lambda_{h}\in\Lambda^{\prime}$ such that $h\cdot K(\lambda)\subseteq K(\lambda_{h}),\quad K(\lambda)\cdot h\subseteq K(\lambda_{h}).$ Notice that in the strongly normalised case, if $H^{\prime}$ is f.d., then $\sum_{h\in H^{\prime}}h\cdot K(\lambda)$ is a f.d. subspace so is contained in some $K(\widetilde{\lambda}_{h})$. Here is an omnibus proposition combining results found in Montgomery [SM:HopfAlgActions]*section 3.4 ###### Proposition 6.6. Let $K\subseteq H$ be a subHopf algebra. _(a)_ Let $K$ be normal. Then $HK^{+}=K^{+}H$ and this is a Hopf ideal in $H$; furthermore, the quotient homomorphism $H\to H/HK^{+}$ is a morphism of Hopf algebras. _(b)_ If $H$ is faithfully flat as a left/right $K$-module and $HK^{+}=K^{+}H$, then $K$ is normal. _(c)_ If $H$ is finite dimensional then $K$ is normal if and only if $HK^{+}=K^{+}H$. The next result is probably standard but we do not know a convenient reference. ###### Proposition 6.7. Suppose that $H$ is a Hopf algebra over a field and that $K,L$ are subHopf algebras where $K$ is normalised by $L$. Then $KL=LK$ and this is a subHopf algebra of $H$. ###### Proof. This is similar to the proof of Proposition 6.6(a). The key point is that for $k\in K$ and $\ell\in L$, using the definition of the counit and coassociativity we have, $\displaystyle\ell k$ $\displaystyle=\sum_{i}\ell^{\prime}_{i}k\varepsilon(\ell^{\prime\prime}_{i})$ $\displaystyle=\sum_{i}\ell^{\prime}_{i}\cdot k\ell^{\prime\prime}_{i}\in KL$ so $LK\subseteq KL$. Similarly, $KL\subseteq LK$. ∎ ###### Proposition 6.8. Let $H$ be a locally Frobenius Hopf algebra of shape $\Lambda$ and let $K\subseteq H$ be a locally Frobenius subHopf algebra if shape $\Lambda^{\prime}\subseteq\Lambda$. Suppose that $L\subseteq H$ is a finite dimensional subHopf algebra and $K$ is normalised by $L$. Then for any finite dimensional subHopf algebra $K^{\prime}\subseteq K$ there is a $\lambda\in\Lambda$ such that $H(\lambda)$ contains $K^{\prime}$ and $L$; furthermore $K\cap H(\lambda)$ is normalised by $L$ and $L(K\cap H(\lambda))=(K\cap H(\lambda))L\subseteq H(\lambda)$ is a finite dimensional subHopf algebra containing $K^{\prime}$ and $L$. ###### Proof. This is a straightforward consequence of Lemma 3.3 and Proposition 6.7. ∎ ### Locally finite $H$-modules and $H^{\circ}$-comodules For a locally Frobenius Hopf algebra $H$ the finite dual $H^{\circ}$ is also a Hopf algebra. It is standard that the categories of locally finite $H$-modules $\mathbf{Mod}_{H}^{\mathrm{l.f.}}$ and left $H^{\circ}$-comodules $\mathbf{Comod}_{H^{\circ}}$ are equivalent (in fact isomorphic). The latter category has as injectives summands of extended comodules $H^{\circ}\otimes W$, so when viewed as locally finite $H$-modules these are the injectives of $\mathbf{Mod}_{H}^{\mathrm{l.f.}}$. It is also well known that $\mathbf{Comod}_{H^{\circ}}$ and $\mathbf{Mod}_{H}^{\mathrm{l.f.}}$ lack projectives so only the right derived functors of $\operatorname{Hom}_{H}(M,-)\cong\operatorname{Cohom}_{H^{\circ}}(M,-)=\mathbf{Comod}_{H^{\circ}}(M,-)$ are defined and $\operatorname{Coext}_{H^{\circ}}(-,-)$ is not a balanced functor. We study the graded analogue of this for $P$-algebras in [AB:Palgebras]. Of course the existence of injectives makes it possible to define a stable module category using these. The tensor product of two locally finite $H$-modules is also locally finite so $\mathbf{Mod}_{H}^{\mathrm{l.f.}}$ has a monoidal structure and this passes to the stable module category since for any $H^{\circ}$-comodule $M$, $H^{\circ}\otimes M$ is isomorphic to an extended comodule. ## 7\. Some examples of locally Frobenius Hopf algebras In this section we decribe some examples which occur in the literature whose understanding might be aided by viewing them as locally Frobenius algebras but we leave detailed investigation for future work. ### 7.1. Group algebras of locally finite groups Recall that a countable group $G$ is _locally finite_ if every finite subset $S\subseteq G$ is contained in a finite subgroup. The group algebras of such groups have been studied in the literature, for example in the work of Hartley, Richardson and Musson [BH&JSR:SocleGpRngs, JSR:GpRngsnonzerosocle, IMM:InjModGpAlgLOcFinGps]. ###### Proposition 7.1. Let $\Bbbk$ be a field and $G$ a locally finite group. Then $\Bbbk G$ is a locally Frobenius Hopf algebra. ###### Proof. We take the indexing set $\lambda$ to be the set of finite subgroups ordered by inclusion and for each $H\in\lambda$ take the finite group algebras $\Bbbk H$ which is a subHopf algebra of $\Bbbk G$. A finite dimensional subspace $V\subseteq\Bbbk G$ has a basis whose elements are expressible as linear combinations of finitely many elements of $G$, therefore it must be contained in some finite subgroup $H\leqslant G$, so $V\subseteq\Bbbk H$. For two finite subgroups $H\leqslant K\leqslant G$, there are left and right integrals $\sum_{h\in H}h\in{\smallint}_{\Bbbk H},\quad\sum_{k\in K}h\in{\smallint}_{\Bbbk K}.$ Then taking a complete set of right coset representatives $k_{1},\ldots,k_{|K:H|}$ for $K/H$ we have $\biggl{(}\Bbbk_{1}+\cdots+k_{|K:H|}\biggr{)}\sum_{h\in H}h=\sum_{k\in K}k,$ so ${\smallint}_{\Bbbk K}\subseteq\Bbbk K{\smallint}_{\Bbbk H}$. The condition of Lorenz [ML:TourRepThy]* 12.4.1(a) shows that $\Bbbk K:\Bbbk H$ is a Frobenius extension. ∎ ### 7.2. Dual profinite group algebras Let $G$ be a profinite group, $G=\lim_{\begin{subarray}{c}N\lhd G\\\ |G:N|<\infty\end{subarray}}G/N.$ For a field $\Bbbk$ the pro-group algebra $\Bbbk G=\lim_{\begin{subarray}{c}N\lhd G\\\ |G:N|<\infty\end{subarray}}\Bbbk G/N$ is a complete topological Hopf algebra. However, there is also dual object $\Bbbk(G)=\operatorname*{colim}_{\begin{subarray}{c}N\lhd G\\\ |G:N|<\infty\end{subarray}}\Bbbk(G/N)$ where $\Bbbk(G/N)=\operatorname{Map}(G/N,\Bbbk)\cong\operatorname{Hom}_{\Bbbk}(\Bbbk G/N,\Bbbk)$ is the dual group ring. Alternatively, $\Bbbk(G)$ agrees with the algebra of locally constant (continuous) functions $G\to\Bbbk$ with respect to the profinite topology. Since each $\Bbbk(G/N)$ is a commutative Hopf algebra so is $\Bbbk(G)$. In fact, $\Bbbk(G)$ also agrees with the finite dual of $\Bbbk G$, commonly denoted by $(\Bbbk G)^{\mathrm{o}}$. ###### Proposition 7.2. For a profinite group $G$, $\Bbbk(G)$ is a commutative locally Frobenius Hopf algebra. ###### Proof. We need to check that for cofinite normal subgroups $M\lhd G$ and $N\lhd G$ with $M\lhd N$, $\Bbbk(G/M):\Bbbk(G/N)$ is a Frobenius extension. The functions $\delta_{1_{G/M}}\in\Bbbk(G/M)$ and $\delta_{1_{G/N}}\in\Bbbk(G/N)$ with $\displaystyle\delta_{1_{G/M}}(gM)$ $\displaystyle=\begin{dcases*}1&if $g\in M$,\\\ 0&otherwise,\end{dcases*}$ $\displaystyle\delta_{1_{G/N}}(gN)$ $\displaystyle=\begin{dcases*}1&if $g\in N$,\\\ 0&otherwise,\end{dcases*}$ are integrals for $\Bbbk(G/M)$ and $\Bbbk(G/N)$. The quotient homomorphism $\pi\colon G/M\to G/N$ satisfies $\pi^{*}\delta_{1_{G/N}}=\sum_{gM\in\ker\pi}\delta_{gM}$ and $\delta_{1_{G/M}}\pi^{*}\delta_{1_{G/N}}=\delta_{1_{G/M}},$ so the condition of Lorenz [ML:TourRepThy]* 12.4.1(a) shows that $\Bbbk(G/M):\Bbbk(G/N)$ is a Frobenius extension. ∎ ## References
# Metric approximation of set-valued functions of bounded variation by integral operators Elena E. Berdysheva 111 University of Cape Town, South Africa, Nira Dyn 222Tel-Aviv University, School of Mathematical Sciences, Elza Farkhi 22footnotemark: 2 Alona Mokhov 444Afeka, Tel-Aviv Academic College of Engineering Abstract. We introduce an adaptation of integral approximation operators to set-valued functions (SVFs, multifunctions), mapping a compact interval $[a,b]$ into the space of compact non-empty subsets of ${\mathbb{R}}^{d}$. All operators are adapted by replacing the Riemann integral for real-valued functions by the weighted metric integral for SVFs of bounded variation with compact graphs. For such a set-valued function $F$, we obtain pointwise error estimates for sequences of integral operators at points of continuity, leading to convergence at such points to $F$. At points of discontinuity of $F$, we derive estimates, which yield the convergence to a set, first described in our previous work on the metric Fourier operator. Our analysis uses recently defined one-sided local quasi-moduli at points of discontinuity and several notions of local Lipschitz property at points of continuity. We also provide a global approach for error bounds. A multifunction $F$ is represented by the set of all its metric selections, while its approximation (its image under the operator) is represented by the set of images of these metric selections under the operator. A bound on the Hausdorff distance between these two sets of single-valued functions in $L^{1}$ provides our global estimates. The theory is illustrated by presenting the examples of two concrete operators: the Bernstein-Durrmeyer operator and the Kantorovich operator. Key words: Set-valued functions, functions of bounded variation, metric integral, metric approximation, integral operators, positive linear operators, rate of convergence Mathematics Subject Classification 2020: 26E25, 28B20, 41A35, 41A36, 41A25, 26A45 ## 1 Introduction We study set-valued functions (SVFs, multifunctions) that map a compact interval $[a,b]\subset{\mathbb{R}}$ into the space of compact non-empty subsets of ${\mathbb{R}}^{d}$. These functions appear in different fields such as dynamical systems, control theory, optimization, game theory, differential inclusions, economy, geometric modeling. See the book [2] for foundations of set-valued analysis. Approximation methods for SVFs has been developing in the last decades. Older works deal mostly with convex-valued multifunctions and their approximation based on Minkowski linear combinations, e.g., [24, 13, 15, 23, 5, 9]. In [28], such an adapation of the classical Bernstein polynomial operator is proved to converge to SVFs with convex compact images (values). Yet, it is shown that this adaptation fails to approximate general SVFs (with general compact not necessarily convex images). In general, approximation methods developed for multifunctions with convex images usually are not suitable for general SVFs. A first successful attempt to approximate general SVFs from their samples is accomplished by Z. Artstein in [1], where piecewise linear approximants are constructed. This is done by replacing binary Minkowski average of two sets with the metric average, which is further extended in [16] to the metric linear combination of several sets. Based on the metric linear combination, N. Dyn, E. Farkhi and A. Mokhov developed in a series of works [14, 20, 16, 17, 18] adaptation of classical sample-based approximation operators to continuous general SVFs. For these adapted operators, termed metric operators, error estimates are obtained, which for most operators are similar to those obtained in the real-valued case. Special attention is given to Bernstein polynomial operators, Schoenberg spline operators and polynomial interpolation operators. Later in [6], the above metric approach is extended to SVFs of bounded variation. The metric approach is applied in [19] to introduce and study the metric integral for general SVFs of bounded variation. The metric integral is not necessarily convex in contrast to the Aumann integral, which is always convex, even if the integrand is not convex-valued [3]. In [7] the metric integral is extended to the weighted metric integral, which is used, with the Dirichlet kernels as weight functions, to define metric Fourier partial sums for SVFs of bounded variation. The convergence of these partial sums is analyzed at points of continuity of a multifunction as well as at points of discontinuity. An important tool in the analysis at points of discontinuity is the notion of one-sided local quasi-moduli of a function of bounded variation. In this paper we adapt integral approximation operators for real-valued functions to general SVFs of bounded variation. Previous adaptations of integral operators to SVFs are limited to convex-valued multifunctions and are based on the Aumann integral (see e.g. [4, 10]). Our adaptation is based on the weighted metric integral, and its analysis applies and extends the techniques developed in [7]. The outline of the paper is as follows. Section 2 gives a short overview of notions we use in the paper, and also discusses different regularity properties of functions with values in a metric space. In Section 3 we refine known results concerning approximation of real-valued functions by sequences of integral approximation operators, which are necessary for the adaptation of these operators to SVFs. The core part of the paper is Section 4 where we construct an adaptation of integral approximation operators to general SVFs. For set-valued functions of bounded variation with compact graphs, we study pointwise convergence, in the Hausdorff metric, of sequences of such operators at points of continuity of the function as well as at points of discontinuity, and derive estimates for the rate of convergence. In Section 5 we illustrate our theory by considering examples of two particular integral approximation operators, the Bernstein- Durrmeyer operator and the Kantorovich operator. In the final Section 6 we provide global error bounds. The multifunction $F$ is represented by the set of all its metric selections (see [18] for more information on representations of SVFs), while its approximation (its image under the operator) is represented by the set of images of these metric selections under the operator. A bound of the Hausdorff distance between these two sets of single-valued functions in $L^{1}$ is obtained using results from [8]. ## 2 Preliminaries In this section we introduce some notation and basic notions related to sets and set-valued functions. We discuss notions of regularity of functions in metric spaces. We review the notions of metric selections and the metric integral of set-valued functions. ### 2.1 On sets All sets considered from now on are sets in ${{\mathbb{R}}}^{d}$. We denote by $\mathrm{K}({{\mathbb{R}}}^{d})$ the collection of all compact non-empty subsets of ${{\mathbb{R}}}^{d}$. The metric in ${{\mathbb{R}}}^{d}$ is of the form $\rho(u,v)=|u-v|$, where $|\cdot|$ is any fixed norm on ${{\mathbb{R}}}^{d}$. Recall that ${{\mathbb{R}}}^{d}$ endowed with this metric is a complete metric space and that all norms on ${{\mathbb{R}}}^{d}$ are equivalent. To measure the distance between two non-empty sets ${A,B\in\mathrm{K}({{\mathbb{R}}}^{d})}$, we use the Hausdorff metric based on $\rho$ $\mathrm{haus}(A,B)_{\rho}=\max\left\\{\sup_{a\in A}\mathrm{dist}(a,B)_{\rho},\;\sup_{b\in B}\mathrm{dist}(b,A)_{\rho}\right\\},$ where the distance from a point $c$ to a set $D$ is $\mathrm{dist}(c,D)_{\rho}=\inf_{d\in D}\rho(c,d)$. It is well known that $\mathrm{K}({{\mathbb{R}}}^{d})$ endowed with the Hausdorff metric is a complete metric space [25, 27]. In the following, we keep the metric in ${{\mathbb{R}}}^{d}$ fixed, and omit the notation $\rho$ as a subscript. We denote by $|A|=\mathrm{haus}(A,\\{0\\})$ the “norm” of the set $A\in\mathrm{K}({{\mathbb{R}}}^{d})$. The set of projections of $a\in{{\mathbb{R}}}^{d}$ on a set $B\in\mathrm{K}({{\mathbb{R}}}^{d})$ is $\Pi_{B}{(a)}=\\{b\in B\ :\ |a-b|=\mathrm{dist}(a,B)\\},$ and the set of metric pairs of two sets $A,B\in\mathrm{K}({{\mathbb{R}}}^{d})$ is $\Pi\big{(}{A},{B}\big{)}=\\{(a,b)\in A\times B\ :\ a\in\Pi_{A}{(b)}\;\,\mbox{or}\;\,b\in\Pi_{B}{(a)}\\}.$ Using metric pairs, we can rewrite $\mathrm{haus}(A,B)=\max\\{|a-b|\ :\ (a,b)\in\Pi\big{(}{A},{B}\big{)}\\}.$ We recall the notions of a metric chain and of a metric linear combination [19]. ###### Definition 2.1. [19] Given a finite sequence of sets $A_{0},\ldots,A_{n}\in\mathrm{K}({{\mathbb{R}}}^{d})$, $n\geq 1$, a metric chain of $A_{0},\ldots,A_{n}$ is an $(n+1)$-tuple $(a_{0},\ldots,a_{n})$ such that $(a_{i},a_{i+1})\in\Pi\big{(}{A_{i}},{A_{i+1}}\big{)}$, $i=0,1,\ldots,n-1$. The collection of all metric chains of $A_{0},\ldots,A_{n}$ is denoted by ${\mathrm{CH}}(A_{0},\ldots,A_{n})=\left\\{(a_{0},\ldots,a_{n})\ :\ (a_{i},a_{i+1})\in\Pi\big{(}{A_{i}},{A_{i+1}}\big{)},\ i=0,1,\ldots,n-1\right\\}.$ The metric linear combination of the sets $A_{0},\ldots,A_{n}\in\mathrm{K}({{\mathbb{R}}}^{d})$, $n\geq 1$, is $\bigoplus_{i=0}^{n}\lambda_{i}A_{i}=\left\\{\sum_{i=0}^{n}\lambda_{i}a_{i}\ :\ (a_{0},\ldots,a_{n})\in{\mathrm{CH}}(A_{0},\ldots,A_{n})\right\\},\quad\lambda_{0},\ldots,\lambda_{n}\in{\mathbb{R}}.$ ###### Remark 2.2. For any $j\in{\mathbb{N}}$, $0\leq j\leq n$ and for any $a\in A_{j}$ there exists a metric chain $(a_{0},\ldots,a_{n})\in{\mathrm{CH}}(A_{0},\ldots,A_{n})$ such that $a_{j}=a$. For a possible construction see [16], Figure 3.2. Note that the metric linear combination depends on the order of the sets, in contrast to the Minkowski linear combination of sets which is defined by $\sum_{i=0}^{n}\lambda_{i}A_{i}=\left\\{\sum_{i=0}^{n}\lambda_{i}a_{i}\ :\ a_{i}\in A_{i}\right\\},\quad n\geq 1.$ The upper Kuratowski limit of a sequence of sets $\\{A_{n}\\}_{n=1}^{\infty}$ is the set of all limit points of converging subsequences $\\{a_{n_{k}}\\}_{k=1}^{\infty}$, where ${a_{n_{k}}\in A_{n_{k}}}$, $k\in{\mathbb{N}}$, namely $\limsup_{n\to\infty}A_{n}=\left\\{a\ :\ \exists\,\\{n_{k}\\}_{k=1}^{\infty},\,n_{k+1}>n_{k},\,k\in{\mathbb{N}},\ \exists\,a_{n_{k}}\in A_{n_{k}}\text{ such that }\lim_{k\to\infty}a_{n_{k}}=a\right\\}.$ (1) ### 2.2 Notions of regularity of functions with values in a metric space In this paper we consider functions defined on a fixed compact interval $[a,b]\subset{\mathbb{R}}$ with values in a complete metric space $(X,\rho)$, where $X$ is either ${{\mathbb{R}}}^{d}$ or $\mathrm{K}({{\mathbb{R}}}^{d})$. We recall the notion of the variation of ${f:[a,b]\rightarrow X}$. Let ${\,\chi=\\{x_{0},\ldots,x_{n}\\}}$, $a=x_{0}<x_{1}<\cdots<x_{n}=b$, be a partition of the interval $[a,b]$ with the norm ${\displaystyle|\chi|=\max_{0\leq i\leq n-1}(x_{i+1}-x_{i})}.$ The variation of $f$ on the partition $\chi$ is defined as $V(f,\chi)=\sum_{i=1}^{n}\rho(f(x_{i}),f(x_{i-1}))\,.$ The total variation of $f$ on $[a,b]$ is $V_{a}^{b}(f)=\sup_{\chi}V(f,\chi),$ where the supremum is taken over all partitions of $[a,b]$. A function $f$ is said to be of bounded variation on $[a,b]$ if ${V_{a}^{b}(f)<\infty}$. We call functions of bounded variation BV functions and write $f\in\mathrm{BV}[a,b]$. If $f$ is also continuous, we write $f\in\mathrm{CBV}[a,b]$. For $f\in\mathrm{BV}[a,b]$ the function $v_{f}:[a,b]\rightarrow{\mathbb{R}}$, $v_{f}(x)=V_{a}^{x}(f)$ is called the variation function of $f$. Note that $V_{z}^{x}(f)=v_{f}(x)-v_{f}(z)\quad\mbox{for}\quad a\leq z<x\leq b,$ and that $v_{f}$ is monotone non-decreasing. For a BV function ${f:{\mathbb{R}}\rightarrow X}$ the following property holds (see e.g. Lemma 2.4 in [6]), $\int_{a}^{b}V_{x-\delta}^{x+\delta}(f)dx\leq 2\delta V_{a}^{b}(f).$ (2) We recall the notion of the local modulus of continuity [26], which is central to the approximation of functions at continuity points. For $f:[a,b]\to X$ the local modulus of continuity at $x^{*}\in[a,b]$ is $\omega\big{(}{f},{x^{*}},{\delta}\big{)}=\sup\left\\{\,\rho(f(x_{1}),f(x_{2})):\;x_{1},x_{2}\in\left[x^{*}-\delta/2,x^{*}+\delta/2\right]\cap[a,b]\,\right\\},\quad\delta>0.$ It follows from the definition of the variation that ###### Result 2.3. For a function $f:[a,b]\to X$, $f\in\mathrm{BV}[a,b]$, $\omega(f,x^{*},\delta)\leq\omega(v_{f},x^{*},\delta),\quad x^{*}\in[a,b],\quad\delta>0.$ Moreover, $f$ is continuous at $x^{*}\in[a,b]$ if and only if $v_{f}$ is continuous at $x^{*}$. ###### Result 2.4. A function $f:[a,b]\to X$, $f\in\mathrm{BV}[a,b]$ is left continuous at $x^{*}\in(a,b]$ if and only if $v_{f}$ is left continuous at $x^{*}$. The function $f$ is right continuous at $x^{*}\in[a,b)$ if and only if $v_{f}$ is right continuous at $x^{*}$. A function $f:[a,b]\to X$ of bounded variation with values in a complete metric space $(X,\rho)$ is not necessarily continuous, but has right and left limits at any point $x$ [11]. We denote the one-sided limits by $f(x+)=\lim_{t\to x+0}f(t),\quad f(x-)=\lim_{t\to x-0}f(t).$ In [7], we introduced the notion of the left and right local quasi-moduli. For a function $f:[a,b]\to X$ of bounded variation, the left local quasi-modulus at point $x^{*}$ is $\varpi^{-}\big{(}{f},{x^{*}},{\delta}\big{)}=\sup{\big{\\{}\rho(f(x^{*}-),f(x))\ :\ x\in[x^{*}-\delta,x^{*})\cap[a,b]\big{\\}}},\quad\delta>0\ ,\;x^{*}\in(a,b].$ (3) Similarly, the right local quasi-modulus is $\varpi^{+}\big{(}{f},{x^{*}},{\delta}\big{)}=\sup{\\{\rho(f(x^{*}+),f(x))\ :\ x\in(x^{*},x^{*}+\delta]\cap[a,b]\\}},\quad\delta>0\ ,\;x^{*}\in[a,b).$ (4) Clearly, for $f\in\mathrm{BV}[a,b]$ the local quasi-moduli satisfy $\lim_{\delta\to 0^{+}}\varpi^{-}\big{(}{f},{x^{*}},{\delta}\big{)}=0,\quad x^{*}\in(a,b],\quad\text{and}\quad\lim_{\delta\to 0^{+}}\varpi^{+}\big{(}{f},{x^{*}},{\delta}\big{)}=0,\quad x^{*}\in[a,b).$ Defining $\varpi\big{(}{f},{x^{*}},{\delta}\big{)}=\max\\{\varpi^{-}\big{(}{f},{x^{*}},{\delta}\big{)},\varpi^{+}\big{(}{f},{x^{*}},{\delta}\big{)}\\},$ we obtain $\lim_{\delta\to 0^{+}}\varpi\big{(}{f},{x^{*}},{\delta}\big{)}=0.$ (5) ###### Lemma 2.5. Let $f:[a,b]\to X$, $f\in\mathrm{BV}[a,b]$, then for any $x^{*}\in(a,b]$ or $[a,b)$, respectively, and $\delta>0$ we have $\varpi^{-}\big{(}{f},{x^{*}},{\delta}\big{)}\leq\varpi^{-}\big{(}{v_{f}},{x^{*}},{\delta}\big{)},\quad\varpi^{+}\big{(}{f},{x^{*}},{\delta}\big{)}\leq\varpi^{+}\big{(}{v_{f}},{x^{*}},{\delta}\big{)}.$ ###### Proof. The first inequality follows from the fact that for $x^{*}\in(a,b]$ and $\max\\{x^{*}-\delta,a\\}\leq x<x^{*}$ we have $\rho(f(x^{*}-),f(x))\leq\lim_{t\to x^{*}-0}V_{x}^{t}(f)=\lim_{t\to x^{*}-0}v_{f}(t)-v_{f}(x)=\rho(v_{f}(x^{*}-),v_{f}(x)).$ Similarly one can show the second inequality. ∎ Below we discuss several notions of Lipschitz regularity. A functions $f:[a,b]\to(X,\rho)$ is Lipschitz continuous with a constant $\mathcal{L}>0$, if $\rho(f(x),f(y))\leq{\mathcal{L}}|x-y|,\quad\forall\,x,y\in[a,b].$ ###### Definition 2.6. Let $f:[a,b]\to(X,\rho)$. * (a) We say that $f$ is locally Lipschitz around a point $x$ with the Lipschitz constant $\mathcal{L}>0$ if there exists $\delta>0$ such that $\rho(f(x_{1}),f(x_{2}))\leq{\mathcal{L}}|x_{1}-x_{2}|,\quad\forall\,x_{1},x_{2}\in\left(x-\delta/2,x+\delta/2\right)\cap[a,b].$ (6) * (b) A function $f$ is locally Lipschitz at a point $x$ with the Lipschitz constant $\mathcal{L}>0$ if there exists $\delta>0$ such that $\rho(f(z),f(x))\leq{\mathcal{L}}|z-x|,\quad\forall\,z\in\left(x-\delta/2,x+\delta/2\right)\cap[a,b].$ (7) We denote by $\mathrm{Lip}\\{{x},{\mathcal{L}}\\}$ the collection of all functions $f$ satisfying (7). * (c) A function $f$ is globally Lipschitz at a point $x$ if there exists $\mathcal{L}>0$ such that $\rho(f(z),f(x))\leq{\mathcal{L}}|z-x|,\quad\forall\,z\in[a,b].$ (8) ###### Remark 2.7. Note that if $f$ is locally Lipschitz around a point $x$ with the Lipschitz constant $\mathcal{L}$, then $f\in\mathrm{Lip}\\{{x},{\mathcal{L}}\\}$, but the inverse implication does not hold. For example, the function $f(x)=x\sin(1/x)$ for $x\neq 0$ and $f(0)=0$ is not locally Lipschitz around $x=0$, while $f\in\mathrm{Lip}\\{{0},{1}\\}$. We say that $f:[a,b]\to X$ is bounded on $[a,b]$ if there exists $y^{*}\in X$ such that $M(f,y^{*})=\sup_{x\in[a,b]}\rho(f(x),y^{*})<\infty.$ The following lemmas deal with relations between the above notions. ###### Lemma 2.8. If $f\in\mathrm{Lip}\\{{x},{\mathcal{L}}\\}$ and $f$ is bounded on $[a,b]$, then $f$ is also globally Lipschitz at $x$. ###### Proof. By Definition 2.6 (b), there is $\delta>0$ such that for $|z-x|<\frac{\delta}{2}$ we have $\rho(f(z),f(x))\leq\mathcal{L}|z-x|$. Also, since $f$ is bounded there is $y^{*}$ such that $\rho(f(z),f(x))\leq\rho(f(z),y^{*})+\rho(y^{*},f(x))\leq 2M(f,y^{*}).$ If $|z-x|\geq\frac{\delta}{2}$ we obtain from the above inequality the estimate $\rho(f(z),f(x))\leq\frac{4M(f,y^{*})}{\delta}|z-x|$. Altogether (8) holds with $\widetilde{\mathcal{L}}=\max\left(\mathcal{L},\frac{4M(f,y^{*})}{\delta}\right)$. ∎ ###### Lemma 2.9. If $f\in\mathrm{BV}[a,b]$ is locally Lipschitz around a point $x$ with a constant $\mathcal{L}>0$, then 1. (i) $v_{f}$ is locally Lipschitz around the point $x$ with the same constant $\mathcal{L}$, 2. (ii) $v_{f}\in\mathrm{Lip}\\{{x},{\mathcal{L}}\\}$. ###### Proof. Assume that $f$ satisfies (6) with some $\delta>0$. To prove (i), let $x-\frac{\delta}{2}<y<z<x+\frac{\delta}{2}$, then $v_{f}(z)-v_{f}(y)=V_{y}^{z}(f)=\sup_{\chi}\sum_{j=0}^{n-1}\rho(f(x_{j}),f(x_{j+1}))\leq\sup_{\chi}\sum_{j=0}^{n-1}\mathcal{L}|x_{j+1}-x_{j}|=\mathcal{L}|y-z|,$ where the supremum is taken over all partitions ${\chi=\\{y=x_{0}<\cdots<x_{n}=z\\}}$ of $[y,z]$. By (i) and Remark 2.7 we obtain (ii). ∎ ###### Lemma 2.10. Let $f\in\mathrm{BV}[a,b]$. 1. (i) If $v_{f}$ is locally Lipschitz around a point $x$ with some $\mathcal{L}>0$ and $\delta>0$, then $f$ is locally Lipschitz around $x$ with the same $\mathcal{L}$ and $\delta$. 2. (ii) If $v_{f}\in\mathrm{Lip}\\{{x},{\mathcal{L}}\\}$, then $f\in\mathrm{Lip}\\{{x},{\mathcal{L}}\\}$. ###### Proof. (i) For all $z,y$ such that ${x-\frac{\delta}{2}<y<z<x+\frac{\delta}{2}}$ we have $\rho(f(z),f(y))\leq V_{y}^{z}(f)=v_{f}(z)-v_{f}(y)\leq\mathcal{L}|z-y|\,.$ (9) To prove (ii) we replace either $y$ or $z$ in (9) by $x$. ∎ ###### Remark 2.11. The inverse implication of Lemma 2.10 (ii) does not hold. Consider, for example, the function $f(x)=x\sin(1/x)$ for $x\neq 0$ and $f(0)=0$. Clearly, $f\in\mathrm{Lip}\\{{0},{\mathcal{L}}\\}$ with $\mathcal{L}=1$, but $v_{f}\notin\mathrm{Lip}\\{{0},{\mathcal{L}}\\}$ for any $\mathcal{L}>0$. In the following we consider ${f:[a,b]\rightarrow{{\mathbb{R}}}^{d}}$, $f=\begin{pmatrix}f_{1}\\\ \vdots\\\ f_{d}\end{pmatrix}$. We recall that $|\cdot|$ is a fixed norm on ${{\mathbb{R}}}^{d}$. ###### Lemma 2.12. Let ${f:[a,b]\rightarrow{{\mathbb{R}}}^{d}}$ be Riemann integrable. Then $\left|\int_{a}^{b}f(x)dx\right|\leq\int_{a}^{b}|f(x)|dx.$ ###### Proof. Consider a sequence of partitions $\chi_{n}=\\{a=x_{0}^{n}<\cdots<x_{n}^{n}=b\\}$ with $\displaystyle\lim_{n\to\infty}|\chi_{n}|=0$. Take some $c_{i}^{n}\in[x_{i-1}^{n},x_{i}^{n}]$, $i=1,\ldots,n$. Using the triangle inequality in the estimate for the Riemann sums we obtain $\left|\int_{a}^{b}f(x)dx\right|=\left|\lim_{n\to\infty}\sum_{i=1}^{n}f(c_{i}^{n})(x_{i}^{n}-x_{i-1}^{n})\right|\leq\lim_{n\to\infty}\sum_{i=1}^{n}|f(c_{i}^{n})|(x_{i}^{n}-x_{i-1}^{n})=\int_{a}^{b}|f(x)|dx.$ ∎ We recall the notion of integral moduli of continuity for functions with values in ${{\mathbb{R}}}^{d}$. We extend a function ${f:[a,b]\rightarrow{{\mathbb{R}}}^{d}}$ outside $[a,b]$ in a simple way preserving its variation on $[a,b]$ $f(x)=\left\\{\begin{array}[]{ll}f(a),&x<a,\\\ f(x),&a\leq x\leq b,\\\ f(b),&x>b.\end{array}\right.$ For ${f:[a,b]\rightarrow{{\mathbb{R}}}^{d}}$, $f=\begin{pmatrix}f_{1}\\\ \vdots\\\ f_{d}\end{pmatrix}$, we write $f\in L^{1}[a,b]$ if all its components $f_{i}$, $i=1,\ldots,d$, are in $L^{1}[a,b]$. The distance between two functions $f,g\in L^{1}[a,b]$ is given by $\|f-g\|_{L^{1}}=\int_{a}^{b}|f(x)-g(x)|dx.$ We denote $\displaystyle\|f\|_{\infty}=\sup_{x\in[a,b]}|f(x)|$. The first order integral modulus of continuity of ${f:[a,b]\rightarrow{{\mathbb{R}}}^{d}}$, $f\in L^{1}[a,b]$ is $\vartheta(f,\delta)=\vartheta_{1}(f,\delta)=\sup_{0<h\leq\delta}\int_{a}^{b}|f(x+h)-f(x)|dx,\quad\delta>0.$ The second order integral modulus of continuity is $\vartheta_{2}(f,\delta)=\sup_{0<h\leq\delta}\int_{a}^{b}|f(x+h)-2f(x)+f(x-h)|dx,\quad\delta>0.$ It is easy to see that $\vartheta_{2}(f,\delta)\leq 2\vartheta(f,\delta).$ (10) Using (2) one can easily obtain that for $f\in\mathrm{BV}[a,b]$ $\vartheta\big{(}{f},{\delta}\big{)}\leq\delta V_{a}^{b}(f).$ (11) ### 2.3 Metric selections and the weighted metric integral of multifunctions We consider set-valued functions (SVFs, multifunctions) mapping a compact interval $[a,b]\subset{\mathbb{R}}$ to $\mathrm{K}({{\mathbb{R}}}^{d})$. The graph of a multifunction $F$ is the set of points in ${\mathbb{R}}^{d+1}$ ${\mathrm{Graph}}(F)=\left\\{(x,y)\ :\ y\in F(x),\;x\in[a,b]\right\\}.$ It is easy to see that if $F\in\mathrm{BV}[a,b]$ then ${\mathrm{Graph}}(F)$ is a bounded set and $\|F\|_{\infty}=\sup_{x\in[a,b]}|F(x)|<\infty.$ We denote the class of SVFs of bounded variation with compact graphs by $\mathcal{F}[a,b]$. For a set-valued function $F:[a,b]\to\mathrm{K}({{\mathbb{R}}}^{d})$, a single-valued function ${s:[a,b]\to{{\mathbb{R}}}^{d}}$ such that $s(x)\in F(x)$ for all $x\in[a,b]$ is called a selection of $F$. The notions of the metric selections and of the weighted metric integral are central in our work. We recall their definitions. Given a multifunction $F:[a,b]\to\mathrm{K}({{\mathbb{R}}}^{d})$, a partition $\chi=\\{x_{0},\ldots,x_{n}\\}\subset[a,b]$, $a=x_{0}<\cdots<x_{n}=b$, and a corresponding metric chain $\phi=(y_{0},\ldots,y_{n})\in{\mathrm{CH}}\left(F(x_{0}),\ldots,F(x_{n})\right)$ (see Definition 2.1), the chain function based on $\chi$ and $\phi$ is $c_{\chi,\phi}(x)=\left\\{\begin{array}[]{ll}y_{i},&x\in[x_{i},x_{i+1}),\quad i=0,\ldots,n-1,\\\ y_{n},&x=x_{n}.\end{array}\right.$ (12) A selection $s$ of $F$ is called a metric selection, if there is a sequence of chain functions $\\{c_{\chi_{k},\phi_{k}}\\}_{k\in{\mathbb{N}}}$ of $F$ with ${\lim_{k\to\infty}|\chi_{k}|=0}$ such that $s(x)=\lim_{k\to\infty}c_{\chi_{k},\phi_{k}}(x)\quad\mbox{pointwisely on}\ [a,b].$ We denote the set of all metric selections of $F$ by $\mathcal{S}(F)$. Note that the definitions of chain functions and metric selections imply that a metric selection $s$ of a multifunction $F$ is constant in any open interval where the graph of $s$ stays in the interior of ${\mathrm{Graph}}(F)$. Below we quote some results from [19] and [7] which are used in this paper. ###### Result 2.13. [19, Theorem 3.6] Let $s$ be a metric selection of $F\in\mathcal{F}[a,b]$. Then $V_{a}^{b}(s)\leq V_{a}^{b}(F)$ and $\|s\|_{\infty}\leq\|F\|_{\infty}$. ###### Result 2.14. [19, Corollary 3.7] Let $F\in\mathcal{F}[a,b]$. Through any point $\alpha\in{\mathrm{Graph}}(F)$ there exists a metric selection which we denote by ${\,s_{\alpha}}$. Moreover, $F$ has a representation by metric selections, namely $F(x)=\\{s_{\alpha}(x)\ :\ \alpha\in{\mathrm{Graph}}(F)\\}.$ ###### Result 2.15. [7, Theorem 4.9] Let $F\in\mathcal{F}[a,b]$, $s$ be a metric selection of $F$ and $x^{*}\in[a,b]$. Then $\omega\big{(}{s},{x^{*}},{\delta}\big{)}\leq\omega\big{(}{v_{F}},{x^{*}},{2\delta}\big{)},\quad\delta>0.$ In particular, if $F$ is continuous at $x^{*}$, then $s$ is continuous at $x^{*}$. ###### Result 2.16. [7, Theorem 4.13] For $F\in\mathcal{F}[a,b]$, the pointwise limit (if exists) of a sequence of metric selections of $F$ is a metric selection of $F$. ###### Result 2.17. [7, Lemmas 6.7, 6.8] Let $F\in\mathcal{F}[a,b]$ and $\delta>0$. Then any metric selection $s\in\mathcal{S}(F)$ satisfies $\varpi^{-}\big{(}{v_{s}},{x^{*}},{\delta}\big{)}\leq\varpi^{-}\big{(}{v_{F}},{x^{*}},{2\delta}\big{)},\;x^{*}\in(a,b],\qquad\varpi^{+}\big{(}{v_{s}},{x^{*}},{\delta}\big{)}\leq\varpi^{+}\big{(}{v_{F}},{x^{*}},{\delta}\big{)},\;x^{*}\in[a,b).$ A direct consequence of Lemma 2.5 and Result 2.17 is ###### Corollary 2.18. Let $F\in\mathcal{F}[a,b]$, $s\in\mathcal{S}(F)$ and $\delta>0$. Then $\varpi^{-}\big{(}{s},{x^{*}},{\delta}\big{)}\leq\varpi^{-}\big{(}{v_{F}},{x^{*}},{2\delta}\big{)},\;x^{*}\in(a,b],\qquad\varpi^{+}\big{(}{s},{x^{*}},{\delta}\big{)}\leq\varpi^{+}\big{(}{v_{F}},{x^{*}},{\delta}\big{)},\;x^{*}\in[a,b).$ Now we show that the metric selections of $F$ inherit the Lipschitz regularity property at $x$ from $v_{F}$. ###### Lemma 2.19. Let $F\in\mathcal{F}[a,b]$. If $v_{F}\in\mathrm{Lip}\\{{x},{\mathcal{L}}\\}$ then $s\in\mathrm{Lip}\\{{x},{4\mathcal{L}}\\}$ for all metric selections $s\in\mathcal{S}(F)$. ###### Proof. Since $v_{F}\in\mathrm{Lip}\\{{x},{\mathcal{L}}\\}$, there is $\delta>0$ such that ${|v_{F}(x)-v_{F}(z)|\leq{\mathcal{L}}|x-z|}$, ${\forall z\in\left(x-\frac{\delta}{2},x+\frac{\delta}{2}\right)\cap[a,b]}$, implying that $\omega(v_{F},x,\eta)\leq\mathcal{L}\eta$ for $\eta\leq\delta$. Thus by Result 2.15 we have for all $z\in\left(x-\frac{\delta}{4},x+\frac{\delta}{4}\right)\cap[a,b]$ $|s(x)-s(z)|\leq\omega(s,x,2|x-z|)\leq\omega(v_{F},x,4|x-z|)\leq 4\mathcal{L}|x-z|.$ ∎ The metric integral of SVFs is introduced in [19] and extended to the weighted metric integral in [7]. We recall its definition. For a multifunction $F:[a,b]\rightarrow\mathrm{K}({{\mathbb{R}}}^{d})$, a weight function $\kappa:[a,b]\to{\mathbb{R}}$ and for a partition ${\chi=\\{x_{0},\ldots,x_{n}\\}}$, ${a=x_{0}<x_{1}<\cdots<x_{n}=b}$, we define the weighted metric Riemann sum of $F$ by $\displaystyle{\scriptstyle(\mathcal{M}_{\kappa})}S_{\chi}F$ $\displaystyle=\left\\{\sum_{i=0}^{n-1}(x_{i+1}-x_{i})\kappa(x_{i})y_{i}:\ (y_{0},\ldots,y_{n-1})\in{\mathrm{CH}}(F(x_{0}),\ldots,F(x_{n-1}))\right\\}$ $\displaystyle=\bigoplus_{i=0}^{n-1}(x_{i+1}-x_{i})\kappa(x_{i})F(x_{i}).$ The weighted metric integral of $F$ with the weight function $\kappa$ is defined as the upper Kuratowski limit of weighted metric Riemann sums ${\scriptstyle(\mathcal{M}_{\kappa})}\int_{a}^{b}\kappa(x)F(x)dx=\limsup_{|\chi|\to 0}{\scriptstyle(\mathcal{M}_{\kappa})}S_{\chi}F.$ The set ${\scriptstyle(\mathcal{M}_{\kappa})}\int_{a}^{b}\kappa(x)F(x)dx$ is non-empty whenever the set-valued function $\kappa F$ has a bounded range. ###### Result 2.20. [7] Let $F\in\mathcal{F}[a,b]$ and $\kappa\in\mathrm{BV}[a,b]$. Then the set ${\scriptstyle(\mathcal{M}_{\kappa})}\int_{a}^{b}\kappa(x)F(x)dx$ is compact and ${\scriptstyle(\mathcal{M}_{\kappa})}\int_{a}^{b}\kappa(x)F(x)dx=\left\\{\int_{a}^{b}\kappa(x)s(x)dx\ :\ s\in\mathcal{S}(F)\right\\}.$ Here and in such occasions below we understand that the integral is applied to each component of $s=\begin{pmatrix}s_{1}\\\ \vdots\\\ s_{d}\end{pmatrix}$. ## 3 Rate of pointwise convergence of integral operators for real-valued functions Let $\\{\mathcal{K}_{n}(x,t)\\}_{n\in{\mathbb{N}}}$, ${\mathcal{K}}_{n}:[a,b]\times[a,b]\to{\mathbb{R}}$, be a sequence of functions that are integrable with respect to $t$ for each $x$. We term the functions $\mathcal{K}_{n}$ kernels. With the help of the sequence of the kernels we define a sequence of linear integral operators $\\{T_{n}\\}_{n\in{\mathbb{N}}}$ on real-valued functions in $L^{\infty}[a,b]$ by $T_{n}f(x)=\int_{a}^{b}\mathcal{K}_{n}(x,t)f(t)dt,\quad x\in[a,b],\quad n\in{\mathbb{N}}.$ (13) ###### Remark 3.1. If $f:[a,b]\to{{\mathbb{R}}}^{d}$, then $T_{n}f$ is obtained by the operation of $T_{n}$ on each component of $f$. Next we introduce the following notation. For $x\in[a,b]$ let $\alpha_{n}(x)=\left|\int_{a}^{b}\mathcal{K}_{n}(x,t)dt-1\right|$ and $\beta_{n}(x,\delta)=\int_{|x-t|\geq\delta}|\mathcal{K}_{n}(x,t)|dt,\quad\delta>0.$ Furthermore, for each $x\in[a,b]$ let $M(x)\in[0,\infty]$ be such that $\int_{a}^{b}|\mathcal{K}_{n}(x,t)|dt\leq M(x),\quad n\in{\mathbb{N}}.$ (14) ### 3.1 The case of continuity points The theorem below can be considered as a refinement of Theorem 2.1 in [12, Chapter 1]. For that we need the following result [22, page 12] ###### Lemma 3.2. If the function $f$ is defined on $[a,b]$ and $I\subseteq[a,b]$ is a closed interval such that $f$ is continuous at all the points of $I$, then for any $\epsilon>0$ there exists $\delta>0$ such that $|f(y)-f(x)|<\epsilon$ for all $x\in I$, $y\in[a,b]$ and $|y-x|<\delta$. ###### Theorem 3.3. Let $f:[a,b]\rightarrow{\mathbb{R}}$ be bounded and measurable on $[a,b]$, $x\in[a,b]$ and let $M(x)$ be as in (14). Then (i) For all $\delta>0$ and $n\in{\mathbb{N}}$ $|T_{n}f(x)-f(x)|\leq\omega(f,x,2\delta)M(x)+\|f\|_{\infty}2\beta_{n}(x,\delta)+|f(x)|\alpha_{n}(x).$ (15) (ii) If $\displaystyle\lim_{n\to\infty}\alpha_{n}(x)=0$, $\displaystyle\lim_{n\to\infty}\beta_{n}(x,\delta)=0$ for any sufficiently small $\delta>0$, $M(x)<\infty$, and if $x$ is a point of continuity of $f$, then $\lim_{n\to\infty}{|T_{n}f(x)-f(x)|}=0.$ (16) (iii) If $f$ is continuous at all points of a closed interval $I\subseteq[a,b]$, if $\displaystyle\lim_{n\to\infty}\alpha_{n}(x)=0$ and ${\displaystyle\lim_{n\to\infty}\beta_{n}(x,\delta)=0}$ uniformly in $x\in I$ for any sufficiently small $\delta>0$, and if $M(x)$ is bounded on $I$, then the convergence is uniform in $I$. ###### Proof. (i) We have $\displaystyle|T_{n}f(x)-f(x)|$ $\displaystyle=\left|\int_{a}^{b}\mathcal{K}_{n}(x,t)f(t)dt-\int_{a}^{b}\mathcal{K}_{n}(x,t)f(x)dt+f(x)\int_{a}^{b}\mathcal{K}_{n}(x,t)dt-f(x)\right|$ $\displaystyle\leq\int_{a}^{b}|\mathcal{K}_{n}(x,t)||f(t)-f(x)|dt+|f(x)|\left|\int_{a}^{b}\mathcal{K}_{n}(x,t)dt-1\right|$ $\displaystyle\leq\int_{|x-t|\leq\delta}|\mathcal{K}_{n}(x,t)||f(t)-f(x)|dt+\int_{|x-t|>\delta}|\mathcal{K}_{n}(x,t)||f(t)-f(x)|dt+|f(x)|\alpha_{n}(x).$ Using the fact that $|f(t)-f(x)|\leq\omega(f,x,2\delta)$ in the first integral and that $|f(t)-f(x)|\leq 2\|f\|_{\infty}$, we obtain $\displaystyle|T_{n}f(x)-f(x)|$ $\displaystyle\leq\omega(f,x,2\delta)\int_{|x-t|\leq\delta}|\mathcal{K}_{n}(x,t)|dt+2\|f\|_{\infty}\int_{|x-t|>\delta}|\mathcal{K}_{n}(x,t)|dt+|f(x)|\alpha_{n}(x)$ $\displaystyle\leq\omega(f,x,2\delta)M(x)+2\|f\|_{\infty}\beta_{n}(x,\delta)+|f(x)|\alpha_{n}(x)$ which gives (15). (ii) Fix $\delta>0$ and let $n\to\infty$ in (15). With the assumption of (ii), it follows from (15) that $\limsup_{n\to\infty}{|T_{n}f(x)-f(x)|}\leq\omega(f,x,2\delta)M(x).$ Since this is valid for each $\delta>0$ and since $\displaystyle\lim_{\delta\to 0+}\omega(f,x,2\delta)=0$ when $x$ is a point of continuity, we obtain (16). (iii) By Lemma 3.2, $\omega(f,x,2\delta)$ tends to zero uniformly in $x\in I$ when $\delta\to 0+$ and it follows from (15) that the convergence is uniform in $I$. ∎ In view of Lemma 2.12, estimates in the proof of Theorem 3.3 can be repeated verbatim for $f:[a,b]\to{{\mathbb{R}}}^{d}$, $f\in\mathrm{BV}[a,b]$, since such $f$ is Riemann integrable. We obtain ###### Corollary 3.4. If $f:[a,b]\to{{\mathbb{R}}}^{d}$, $f\in\mathrm{BV}[a,b]$ and $\mathcal{K}_{n}(x,\cdot)\in\mathrm{BV}[a,b]$ for each $x\in[a,b]$, then the estimates in Theorem 3.3 hold. ###### Remark 3.5. It is easy to see that if $f:[a,b]\to{{\mathbb{R}}}^{d}$ is bounded and measurable, then (15) becomes $|T_{n}f(x)-f(x)|\leq C{\omega}(f,x,2\delta)M(x)+C\|f\|_{\infty}2\beta_{n}(x,\delta)+C|f(x)|\alpha_{n}(x),\quad x\in[a,b],\;n\in{\mathbb{N}},\;\delta>0,$ where $C$ depends only on the underlying norm $|\cdot|$ in ${{\mathbb{R}}}^{d}$. ### 3.2 The case of discontinuity points In this section we follow and refine the analysis in [21] for real-valued functions. Let $f:[a,b]\to{\mathbb{R}}$, $f\in\mathrm{BV}[a,b]$. Fix $x\in(a,b)$, following [21] we introduce the function $g_{x}:[a,b]\to{\mathbb{R}}$, $g_{x}(t)=\begin{cases}f(t)-f(x-),&t\in[a,x),\\\ f(t)-f(x)=0,&t=x,\\\ f(t)-f(x+),&t\in(x,b].\end{cases}$ Note that $f\in\mathrm{BV}[a,b]$ implies that $g_{x}$ is well defined and is continuous at $x$. It is easy to check that $f(t)=g_{x}(t)+f(x-)\chi_{[a,x)}(t)+f(x+)\chi_{(x,b]}(t)+f(x)\chi_{\\{x\\}}(t),\quad t\in[a,b],$ where $\chi_{A}$ denotes the characteristic function of a set $A$ $\chi_{A}(t)=\begin{cases}1,&t\in A,\\\ 0,&t\not\in A.\end{cases}$ A simple computation shows that for $t\in[a,b]$ $\displaystyle f(t)$ $\displaystyle-\frac{1}{2}\big{[}f(x+)+f(x-)\big{]}=g_{x}(t)+\frac{1}{2}\big{[}f(x+)-f(x-)\big{]}\big{(}\chi_{(x,b]}(t)-\chi_{[a,x)}(t)\big{)}$ $\displaystyle+\left(f(x)-\frac{1}{2}\big{[}f(x+)+f(x-)\big{]}\right)\chi_{\\{x\\}}(t)$ $\displaystyle=g_{x}(t)+\frac{1}{2}\big{[}f(x+)-f(x-)\big{]}{\mathrm{sign}}(t-x)+\left(f(x)-\frac{1}{2}\big{[}f(x+)+f(x-)\big{]}\right)\chi_{\\{x\\}}(t),$ where ${\mathrm{sign}}(t)=\begin{cases}1,&t>0,\\\ 0,&t=0,\\\ -1,&t<0.\end{cases}$ Inserting this representation into the operator (13), we obtain $T_{n}f(x)-\frac{1}{2}[f(x+)+f(x-)]\int_{a}^{b}\mathcal{K}_{n}(x,t)dt=T_{n}g_{x}(x)+\frac{1}{2}[f(x+)-f(x-)]T_{n}({\mathrm{sign}}(\cdot-x))(x).$ (17) Taking into account that $g_{x}(x)=0$, the estimate (15) for $T_{n}g_{x}(x)$ takes the form $|T_{n}g_{x}(x)|=|T_{n}g_{x}(x)-g_{x}(x)|\leq\omega(g_{x},x,2\delta)M(x)+2\|g_{x}\|_{\infty}\beta_{n}(x,\delta).$ (18) By the definition of the function $g_{x}$ we easily see that $\|g_{x}\|_{\infty}\leq 2\|f\|_{\infty}$. By the definition of the local modulus of continuity and the local quasi-moduli (see Section 2.2), we get $\omega(g_{x},x,2\delta)\leq 2\max{\\{\varpi^{-}\big{(}{f},{x},{\delta}\big{)},\varpi^{+}\big{(}{f},{x},{\delta}\big{)}\\}}=2\varpi\big{(}{f},{x},{\delta}\big{)}.$ (19) Clearly, $\frac{1}{2}\big{|}f(x+)-f(x-)\big{|}\leq\|f\|_{\infty}$, since $f\in\mathrm{BV}[a,b]$. Combining this with (17), (18), (19) we arrive at $\begin{split}&\left|T_{n}f(x)-\frac{1}{2}\big{[}f(x+)+f(x-)\big{]}\int_{a}^{b}\mathcal{K}_{n}(x,t)dt\right|\\\ &\leq 2\varpi\big{(}{f},{x},{\delta}\big{)}M(x)+\|f\|_{\infty}\left(4\beta_{n}(x,\delta)+\big{|}T_{n}({\mathrm{sign}}(\cdot-x))(x)\big{|}\right).\end{split}$ (20) Thus we obtain the following result, ###### Theorem 3.6. Let $f:[a,b]\to{\mathbb{R}}$, $f\in\mathrm{BV}[a,b]$ and $x\in(a,b)$. (i) For all $\delta>0$ and $n\in{\mathbb{N}}$ we have $\displaystyle\left|T_{n}f(x)-\frac{1}{2}[f(x+)+f(x-)]\right|$ $\displaystyle\leq 2\varpi\big{(}{f},{x},{\delta}\big{)}M(x)+\|f\|_{\infty}\left(4\beta_{n}(x,\delta)+\alpha_{n}(x)+\big{|}T_{n}({\mathrm{sign}}(\cdot-x))(x)\big{|}\right).$ (ii) If $\displaystyle\lim_{n\to\infty}\alpha_{n}(x)=0$, $\displaystyle\lim_{n\to\infty}\beta_{n}(x,\delta)=0$ for any sufficiently small $\delta>0$, ${\displaystyle\lim_{n\to\infty}T_{n}({\mathrm{sign}}(\cdot-x))(x)=0}$, and $M(x)<\infty$, then $\lim_{n\to\infty}{\left|T_{n}f(x)-\frac{1}{2}[f(x+)+f(x-)]\right|}=0.$ ###### Proof. By the triangle inequality and (4.2) we have $\displaystyle\left|T_{n}f(x)-\frac{1}{2}[f(x+)+f(x-)]\right|$ $\displaystyle\leq\left|T_{n}f(x)-\frac{1}{2}[f(x+)+f(x-)]\int_{a}^{b}\mathcal{K}_{n}(x,t)dt\right|+\frac{1}{2}|f(x+)+f(x-)|\left|\int_{a}^{b}\mathcal{K}_{n}(x,t)dt-1\right|$ $\displaystyle\leq 2\varpi\big{(}{f},{x},{\delta}\big{)}M(x)+\|f\|_{\infty}\left(4\beta_{n}(x,\delta)+\big{|}T_{n}({\mathrm{sign}}(\cdot-x))(x)\,\big{|}\,\right)\,+\|f\|_{\infty}\alpha_{n}(x),$ which leads to the first claim. The second claim follows directly from it by applying (5). ∎ Similarly to Corollary 3.4 for vector-valued functions we obtain ###### Corollary 3.7. Let $f:[a,b]\to{{\mathbb{R}}}^{d}$, $f\in\mathrm{BV}[a,b]$, $\mathcal{K}_{n}(z,\cdot)\in\mathrm{BV}[a,b]$ for each $z\in[a,b]$, then for $x\in(a,b)$ $\displaystyle\left|T_{n}f(x)-\frac{1}{2}[f(x+)+f(x-)]\right|$ $\displaystyle\leq 2\varpi\big{(}{f},{x},{\delta}\big{)}M(x)+\|f\|_{\infty}\big{(}4\beta_{n}(x,\delta)+\alpha_{n}(x)+\big{|}T_{n}({\mathrm{sign}}(\cdot-x))(x)\big{|}\,\big{)}.$ ## 4 Rate of pointwise convergence of integral operators for set-valued functions Let $F\in\mathcal{F}[a,b]$ and $\mathcal{K}_{n}(x,\cdot)\in\mathrm{BV}[a,b]$ for any $x\in[a,b]$. Denoting $\kappa_{n,x}(t)=\mathcal{K}_{n}(x,t)$ and using the concept of the weighted metric integral (see Section 2.3), we define $T_{n}F(x)={\scriptstyle(\mathcal{M}_{\kappa_{n,x}})}\int_{a}^{b}\kappa_{n,x}(t)F(t)dt,\quad x\in[a,b],\quad n\in{\mathbb{N}}.$ (21) By Result 2.20 and since $\kappa_{n,x}(t)=\mathcal{K}_{n}(x,t)$ we have for $F\in\mathcal{F}[a,b]$ $T_{n}F(x)=\left\\{\int_{a}^{b}\mathcal{K}_{n}(x,t)s(t)dt\ :\ s\in\mathcal{S}(F)\right\\}=\left\\{T_{n}s(x)\ :\ s\in\mathcal{S}(F)\right\\},\quad x\in[a,b],$ (22) where $\mathcal{S}(F)$ is the set of metric selections of $F$. By Result 2.14 we have as well $F(x)=\left\\{s(x)\ :\ s\in\mathcal{S}(F)\right\\},\quad x\in[a,b].$ (23) It is easy to obtain from (22) and (23) that $\mathrm{haus}\left(T_{n}F(x),F(x)\right)\leq\sup{\left\\{\left|T_{n}s(x)-s(x)\right|\ :\ s\in\mathcal{S}(F)\right\\}}.$ (24) The arguments leading to (24) are similar to those in the proof of Lemma 6.3 below. ###### Remark 4.1. Any bounded linear operator defined for single-valued functions can be extended to set-valued functions from the class $\mathcal{F}[a,b]$ by $TF(x)=\\{Ts(x)\ :\ s\in\mathcal{S}(F)\\},\quad x\in[a,b],\quad F\in\mathcal{F}[a,b].$ This approach was used for operators $T$ defined on continuous real-valued functions in [18, Section 8.2]. ### 4.1 The case of continuity points Recall that by Results 2.13 and 2.15 for each selection $s\in\mathcal{S}(F)$ we have $\|s\|_{\infty}\leq\|F\|_{\infty}\quad\mbox{and}\quad\omega(s,x,\delta)\leq\omega(v_{F},x,2\delta)\,,\;\,x\in[a,b]\,,\;\delta>0.$ Thus for each $s\in\mathcal{S}(F)$ the estimate in Corollary 3.4 turns into $|T_{n}s(x)-s(x)|\leq\omega(v_{F},x,4\delta)M(x)+\|F\|_{\infty}(2\beta_{n}(x,\delta)+\alpha_{n}(x)).$ (25) We arrive at ###### Theorem 4.2. Let $F\in\mathcal{F}[a,b]$, $x\in[a,b]$, $\\{\mathcal{K}_{n}(x,t)\\}_{n\in{\mathbb{N}}}$ be a sequence of kernels on $[a,b]\times[a,b]$ such that $\mathcal{K}_{n}(x,\cdot)\in\mathrm{BV}[a,b]$ for any $x\in[a,b]$ and let $\alpha_{n}(x)$, $\beta_{n}(x,\delta)$ and $M(x)$ be as in Section 3. Then (i) For all $\delta>0$ and $n\in{\mathbb{N}}$ $\mathrm{haus}(T_{n}F(x),F(x))\leq\omega(v_{F},x,4\delta)M(x)+\|F\|_{\infty}(2\beta_{n}(x,\delta)+\alpha_{n}(x))\,,\quad x\in[a,b].$ (ii) If $\displaystyle\lim_{n\to\infty}\alpha_{n}(x)=0$, $\displaystyle\lim_{n\to\infty}\beta_{n}(x,\delta)=0$ for any sufficiently small $\delta>0$, $M(x)<\infty$ and if $x$ is a point of continuity of $F$, then $\lim_{n\to\infty}{\mathrm{haus}\left(T_{n}F(x),F(x)\right)}=0.$ (iii) If $F$ is continuous at all points of a closed interval $I\subseteq[a,b]$, if ${\displaystyle\lim_{n\to\infty}\alpha_{n}(x)=0}$ and ${\displaystyle\lim_{n\to\infty}\beta_{n}(x,\delta)=0}$ uniformly in $x\in I$ for any sufficiently small $\delta>0$, and if $M(x)$ is bounded on $I$, then the convergence is uniform in $I$. ###### Proof. The statements (i) and (ii) follow from the first two claims of Theorem 3.3 combined with (24) and (25). The proof of the statements (iii) is based on Result 2.3, which implies that $v_{F}$ is uniformly continuous on compact intervals, and therefore $\omega(v_{F},x,4\delta)\to 0$ uniformly for $x\in I$. ∎ ### 4.2 The case of discontinuity points For $x\in(a,b)$ we define the set (see also [7, Section 6.3]) $A_{F}(x)=\left\\{\frac{1}{2}\left(s(x+)+s(x-)\right)\ :\ s\in\mathcal{S}(F)\right\\}.$ Similarly to (24) we have $\mathrm{haus}\left(T_{n}F(x),A_{F}(x)\right)\leq\sup{\left\\{\left|T_{n}s(x)-\frac{1}{2}(s(x+)+s(x-))\right|\ :\ s\in\mathcal{S}(F)\right\\}}.$ (26) Recall that by Result 2.13, $\|s\|_{\infty}\leq\|F\|_{\infty}$ and $\frac{1}{2}|s(x+)+s(x-)|\leq\|F\|_{\infty}$. In view of Corollary 2.18 we have for each $s\in\mathcal{S}(F)$ $\varpi\big{(}{s},{x},{\delta}\big{)}\leq\varpi\big{(}{v_{F}},{x},{2\delta}\big{)}.$ By Corollary 3.7 and the above inequality we have for any $s\in\mathcal{S}(F)$ $\displaystyle\left|T_{n}s(x)-\frac{1}{2}[s(x+)+s(x-)]\right|$ $\displaystyle\leq 2\varpi\big{(}{v_{F}},{x},{2\delta}\big{)}M(x)+\|F\|_{\infty}\big{(}4\beta_{n}(x,\delta)+\alpha_{n}(x)+\big{|}T_{n}({\mathrm{sign}}(\cdot-x))(x)\,\big{|}\,\big{)}.$ This together with (26), by arguments as in the proof of (ii) of Theorem 3.3, leads to the following result ###### Theorem 4.3. Let $F\in\mathcal{F}[a,b]$ and let $\\{\mathcal{K}_{n}(x,t)\\}_{n\in{\mathbb{N}}}$ be a sequence of kernels on $[a,b]\times[a,b]$ such that $\mathcal{K}_{n}(x,\cdot)\in\mathrm{BV}[a,b]$ for any $x\in[a,b]$. Then for any $x\in(a,b)$ the following statements hold. (i) For all $\delta>0$ and $n\in{\mathbb{N}}$ $\displaystyle\mathrm{haus}(T_{n}F(x),A_{F}(x))\leq 2\varpi\big{(}{v_{F}},{x},{2\delta}\big{)}M(x)+\|F\|_{\infty}\big{(}4\beta_{n}(x,\delta)+\alpha_{n}(x)+\big{|}T_{n}({\mathrm{sign}}(\cdot-x))(x)\big{|}\big{)}.$ (ii) If $\displaystyle\lim_{n\to\infty}\alpha_{n}(x)=0$, $\displaystyle\lim_{n\to\infty}\beta_{n}(x,\delta)=0$ for any sufficiently small $\delta>0$, ${\displaystyle\lim_{n\to\infty}T_{n}({\mathrm{sign}}(\cdot-x))(x)=0}$, and $M(x)<\infty$, then $\lim_{n\to\infty}{\mathrm{haus}\left(T_{n}F(x),A_{F}(x)\right)}=0.$ ## 5 Specific operators Here we apply the results of the previous sections to two specific integral operators. In particular, at points of discontinuity we obtain error estimates that combine ideas from [21, 29] with the local quasi-moduli of continuity (3), (4). ### 5.1 Bernstein-Durrmeyer operators For $x\in[0,1]$, $n\in{\mathbb{N}}$ the Bernstein basis polynomials are defined as $p_{n,k}(x)={n\choose k}x^{k}(1-x)^{n-k}\,,\quad k=0,1,\ldots,n.$ (27) Note that $\displaystyle\sum_{k=0}^{n}p_{n,k}(x)=1$ and $\displaystyle\int_{0}^{1}p_{n,k}(x)dx=\frac{1}{n+1}$. The Bernstein-Durrmeyer operator is defined for $f\in L^{1}[0,1]$ by $M_{n}f(x)=(n+1)\sum_{k=0}^{n}p_{n,k}(x)\int_{0}^{1}p_{n,k}(t)f(t)dt=\int_{0}^{1}\mathcal{K}_{n}(x,t)f(t)dt\,,\quad x\in[0,1]\,,\;n\in{\mathbb{N}},$ where $\mathcal{K}_{n}(x,t)=(n+1)\sum_{k=0}^{n}p_{n,k}(x)p_{n,k}(t).$ (28) The Bernstein-Durrmeyer operator for a set-valued function $F\in\mathcal{F}[0,1]$ is $M_{n}F(x)={\scriptstyle(\mathcal{M}_{\kappa_{n,x}})}\int_{0}^{1}\kappa_{n,x}(t)F(t)dt,\quad x\in[0,1],\quad n\in{\mathbb{N}}\,,$ with $\kappa_{n,x}(t)=\mathcal{K}_{n}(x,t)$, where $\mathcal{K}_{n}(x,t)$ is given in (28). The properties of the Bernstein basis polynomials $p_{n,k}$ yield $\mathcal{K}_{n}(x,t)\geq 0,\quad\int_{0}^{1}\mathcal{K}_{n}(x,t)dt=1,$ so that $\alpha_{n}(x)=0$ and $M(x)=1$, $x\in[0,1]$. Denote $e_{i}(t)=t^{i}$, $i=0,1,2$. A direct calculation shows that (see e.g. [21]) $M_{n}e_{0}(x)=1\,,\qquad M_{n}e_{1}(x)=\frac{nx+1}{n+2}\,,\qquad M_{n}e_{2}(x)=\frac{n(n-1)x^{2}+4nx+2}{(n+2)(n+3)}.$ This implies $M_{n}((\cdot-x)^{2})(x)=\frac{2[(n-3)x(1-x)+1]}{(n+2)(n+3)}.$ (29) Following S. Guo (see Lemma 6 in [21]), we estimate $\beta_{n}(x,\delta)$ as follows: $\displaystyle\beta_{n}(x,\delta)$ $\displaystyle=\int_{|x-t|\geq\delta}\mathcal{K}_{n}(x,t)dt\leq\int_{|x-t|\geq\delta}\left(\frac{x-t}{\delta}\right)^{2}\mathcal{K}_{n}(x,t)dt$ $\displaystyle\leq\frac{1}{\delta^{2}}\int_{0}^{1}(x-t)^{2}\mathcal{K}_{n}(x,t)dt=\frac{1}{\delta^{2}}M_{n}((\cdot-x)^{2})(x)=\frac{1}{\delta^{2}}\frac{2[(n-3)x(1-x)+1]}{(n+2)(n+3)}.$ Using the inequality $\displaystyle x(1-x)\leq\frac{1}{4}$, $x\in[0,1]$, we arrive at the estimate $\displaystyle\beta_{n}(x,\delta)\leq\frac{n+1}{2\delta^{2}(n+2)(n+3)}\leq\frac{1}{2n\delta^{2}}.$ Thus, for the Bernstein-Durrmeyer operator we have $M(x)=1\;,\;\alpha_{n}(x)=0\;,\;\beta_{n}(x,\delta)\leq\frac{1}{2n\delta^{2}}\,,\quad x\in[0,1].$ The case of continuity points The assumptions of Theorem 3.3 are fulfilled for the Bernstein-Durrmeyer operator. Thus we obtain the following result, where part (i), which provides rate of convergence, is new in this form, while parts (ii) and (iii) are already known. ###### Corollary 5.1. Let $f\in L^{1}[0,1]$ be bounded on $[0,1]$, $x\in[0,1]$. Then (i) For all $\delta>0$ and $n\in{\mathbb{N}}$ $|M_{n}f(x)-f(x)|\leq\omega(f,x,2\delta)+\|f\|_{\infty}\frac{1}{n\delta^{2}}.$ (ii) If $x$ is a point of continuity of $f$, then $\lim_{n\to\infty}{|M_{n}f(x)-f(x)|}=0$. (iii) If $f$ is continuous at all points of a closed interval $I\subseteq[0,1]$, then the convergence is uniform in $I$. From Theorem 4.2 for the Bernstein-Durrmeyer operator we get ###### Corollary 5.2. Let $F\in\mathcal{F}[0,1]$, $x\in[0,1]$. Then (i) For all $\delta>0$ and $n\in{\mathbb{N}}$ $\mathrm{haus}(M_{n}F(x),F(x))\leq\omega(v_{F},x,4\delta)+\|F\|_{\infty}\frac{1}{n\delta^{2}}.$ (ii) If $x$ is a point of continuity of $F$, then $\lim_{n\to\infty}{\mathrm{haus}\left(M_{n}F(x),F(x)\right)}=0$. (iii) If $F$ is continuous at all points of a closed interval $I\subseteq[0,1]$, then the convergence is uniform in $I$. The case of discontinuity points Theorem 3.6 for the Bernstein-Durrmeyer operator leads to $\displaystyle\left|M_{n}f(x)-\frac{1}{2}[f(x+)+f(x-)]\right|\leq 2\varpi\big{(}{f},{x},{\delta}\big{)}+\|f\|_{\infty}\left(\frac{2}{n\delta^{2}}+\big{|}M_{n}({\mathrm{sign}}(\cdot-x))(x)\big{|}\right)$ for $x\in(0,1)$. S. Guo proved in [21] (see Proof of the Theorem) that $|M_{n}({\mathrm{sign}}(\cdot-x))(x)|\leq\frac{13}{2\sqrt{nx(1-x)}}$ for $x\in(0,1)$ and $n$ large enough. Thus we get ###### Corollary 5.3. Let $f:[0,1]\to{\mathbb{R}}$, $f\in\mathrm{BV}[a,b]$ and $x\in(0,1)$. Then (i) For all $\delta>0$ and $n\in{\mathbb{N}}$ large enough $\left|M_{n}f(x)-\frac{1}{2}[f(x+)+f(x-)]\right|\leq 2\varpi\big{(}{f},{x},{\delta}\big{)}+\|f\|_{\infty}\left(\frac{2}{n\delta^{2}}+\frac{13}{2\sqrt{nx(1-x)}}\right),$ (ii) $\displaystyle\lim_{n\to\infty}{\left|M_{n}f(x)-\frac{1}{2}[f(x+)+f(x-)]\right|}=0$. Note that (ii) and an estimate for the order of convergence are obtained by S. Guo in [21]. Theorem 4.3 for the Bernstein-Durrmeyer operator results in ###### Corollary 5.4. Let $F\in\mathcal{F}[0,1]$, $x\in(0,1)$. Then (i) For all $\delta>0$ and large enough $n\in{\mathbb{N}}$ $\mathrm{haus}(M_{n}F(x),A_{F}(x))\leq 2\varpi\big{(}{v_{F}},{x},{2\delta}\big{)}+\|F\|_{\infty}\left(\frac{2}{n\delta^{2}}+\frac{13}{2\sqrt{nx(1-x)}}\right).$ (ii) $\displaystyle\lim_{n\to\infty}{\mathrm{haus}\left(M_{n}F(x),A_{F}(x)\right)}=0$. ### 5.2 Kantorovich operators For $f\in L^{1}[0,1]$ the Kantorovich operator is defined by $K_{n}f(x)=(n+1)\sum_{k=0}^{n}p_{n,k}(x)\int_{\frac{k}{n+1}}^{\frac{k+1}{n+1}}f(t)dt=\int_{0}^{1}\mathcal{K}_{n}(x,t)f(t)dt,\quad x\in[0,1],\quad n\in{\mathbb{N}},$ where $\mathcal{K}_{n}(x,t)=(n+1)\sum_{k=0}^{n}p_{n,k}(x)\chi_{\left[\frac{k}{n+1},\frac{k+1}{n+1}\right]}(t)$ and $p_{n,k}(x)$ are defined in (27). By formula (21) we can extend this operator to set-valued functions $F\in\mathcal{F}[0,1]$. Using the properties of $p_{n,k}(x)$, we obtain, as in the case of the Bernstein-Durrmeyer operator, that $\mathcal{K}_{n}(x,t)\geq 0$, ${\int_{0}^{1}\mathcal{K}_{n}(x,t)dt=1}$, leading to $\alpha_{n}(x)=0$ and $M(x)=1$, $x\in[0,1]$. Calculating $K_{n}e_{i}(x)$, $i=0,1,2$ one can get (see e.g. [29]) $K_{n}e_{0}(x)=1,\qquad K_{n}e_{1}(x)=\frac{2nx+1}{2(n+1)},\qquad K_{n}e_{2}(x)=\frac{3n(n-1)x^{2}+6nx+1}{3(n+1)^{2}},$ and $K_{n}((\cdot-x)^{2})(x)=\frac{3(n-1)x(1-x)+1}{3(n+1)^{2}}.$ (30) Thus, estimating $\beta_{n}(x,\delta)$ as in the case of the Bernstein- Durrmeyer operator, we obtain $\beta_{n}(x,\delta)\leq\frac{1}{\delta^{2}}K_{n}((\cdot-x)^{2})(x)=\frac{1}{\delta^{2}}\frac{3(n-1)x(1-x)+1}{3(n+1)^{2}}\leq\frac{1}{4n\delta^{2}}.$ The case of continuity points Theorem 3.3 takes for the Kantorovich operator the following form, where part (i) is new in this form (providing rate of convergence), while parts (ii) and (iii) are well-known. ###### Corollary 5.5. Let $f\in L^{1}[0,1]$ be bounded on $[0,1]$, $x\in[0,1]$. Then (i) For all $\delta>0$ and $n\in{\mathbb{N}}$ $|K_{n}f(x)-f(x)|\leq\omega(f,x,2\delta)+\|f\|_{\infty}\frac{1}{2n\delta^{2}}.$ (ii) If $x$ is a point of continuity of $f$, then $\lim_{n\to\infty}{|K_{n}f(x)-f(x)|}=0$. (iii) If $f$ is continuous at all points of a closed interval $I\subseteq[0,1]$, then the convergence is uniform in $I$. Applying Theorem 4.2, we get ###### Corollary 5.6. Let $F\in\mathcal{F}[0,1]$, $x\in[0,1]$. Then (i) For all $\delta>0$ and $n\in{\mathbb{N}}$ $\mathrm{haus}(K_{n}F(x),F(x))\leq\omega(v_{F},x,4\delta)+\|F\|_{\infty}\frac{1}{2n\delta^{2}}.$ (ii) If $x$ is a point of continuity of $F$, then $\lim_{n\to\infty}{\mathrm{haus}\left(K_{n}F(x),F(x)\right)}=0$. (iii) If $F$ is continuous at all points of a closed interval $I\subseteq[0,1]$, then the convergence is uniform in $I$. The case of discontinuity points Theorem 3.6 for the Kantorovich operator gives the estimate $\displaystyle\left|K_{n}f(x)-\frac{1}{2}[f(x+)+f(x-)]\right|\leq 2\varpi\big{(}{f},{x},{\delta}\big{)}+\|f\|_{\infty}\left(\frac{1}{n\delta^{2}}+\big{|}K_{n}({\mathrm{sign}}(\cdot-x))(x)\big{|}\right)$ for $x\in(0,1)$. An estimate for $\big{|}K_{n}({\mathrm{sign}}(\cdot-x))(x)\big{|}$ is given in [29]. Closely following the consideration of Zeng and Piriou in [29], we derive here a very similar estimate in a slightly different form. We also replace the estimate $p_{n,k}(x)\leq\frac{1}{\sqrt{2e}}\frac{1}{\sqrt{nx(1-x)}}$ used in [29] and quoted there from an unpublished paper, by the estimate $p_{n,k}(x)\leq\frac{5}{2\sqrt{nx(1-x)}}$ (31) of Guo, which is published in [21] with a full proof. Let $x\in\left[\frac{\ell}{n+1},\frac{\ell+1}{n+1}\right)$ with a certain $\ell\in\\{0,1,\ldots,n\\}$, then $\displaystyle K_{n}({\mathrm{sign}}(\cdot-x))(x)$ $\displaystyle=(n+1)\sum_{k=0}^{n}p_{n,k}(x)\int_{\frac{k}{n+1}}^{\frac{k+1}{n+1}}{\mathrm{sign}}(t-x)dt$ $\displaystyle=-\sum_{k=0}^{\ell-1}p_{n,k}(x)+\sum_{k=\ell+1}^{n}p_{n,k}(x)+p_{n,\ell}(x)(n+1)\left(\frac{\ell+1}{n+1}-x-\left(x-\frac{\ell}{n+1}\right)\right)$ $\displaystyle=2\sum_{k=\ell+1}^{n}p_{n,k}(x)-1+2p_{n,\ell}(x)[(\ell+1)-(n+1)x],$ and, since $\ell\leq(n+1)x<\ell+1$, $\big{|}K_{n}({\mathrm{sign}}(\cdot-x))(x)\big{|}\leq 2\left|\sum_{(n+1)x<k\leq n}p_{n,k}(x)-\frac{1}{2}\right|+2p_{n,\ell}(x).$ It was proved in [29] (see proof of Lemma 2) that $\left|\sum_{nx<k\leq n}p_{n,k}(x)-\frac{1}{2}\right|\leq\frac{0.8(2x^{2}-2x+1)}{\sqrt{nx(1-x)}}<\frac{1}{\sqrt{nx(1-x)}},$ where the last inequality is easy to check. Combining this with (31), we obtain $\left|\sum_{(n+1)x<k\leq n}p_{n,k}(x)-\frac{1}{2}\right|\leq\frac{7}{2\sqrt{nx(1-x)}}$ and finally $\big{|}K_{n}({\mathrm{sign}}(\cdot-x))(x)\big{|}\leq\frac{12}{\sqrt{nx(1-x)}}$ for $x\in(0,1)$. Thus we obtain ###### Corollary 5.7. Let $f:[0,1]\to{\mathbb{R}}$, $f\in\mathrm{BV}[a,b]$ and $x\in(0,1)$. Then (i) For all $\delta>0$ and $n\in{\mathbb{N}}$ $\left|K_{n}f(x)-\frac{1}{2}[f(x+)+f(x-)]\right|\leq 2\varpi\big{(}{f},{x},{\delta}\big{)}+\|f\|_{\infty}\left(\frac{1}{n\delta^{2}}+\frac{12}{\sqrt{nx(1-x)}}\right),$ (ii) $\displaystyle\lim_{n\to\infty}{\left|K_{n}f(x)-\frac{1}{2}[f(x+)+f(x-)]\right|}=0$. Note that (ii) is known (see [29] and the bibliography therein), this is also the case for the order of convergence $1/\sqrt{n}$. Theorem 4.3 for the Kantorovich operator takes the following form. ###### Corollary 5.8. Let $F\in\mathcal{F}[0,1]$, $x\in(0,1)$. Then (i) For all $\delta>0$ and $n\in{\mathbb{N}}$ $\mathrm{haus}(K_{n}F(x),A_{F}(x))\leq 2\varpi\big{(}{v_{F}},{x},{2\delta}\big{)}+\|F\|_{\infty}\left(\frac{1}{n\delta^{2}}+\frac{12}{\sqrt{nx(1-x)}}\right).$ (ii) $\displaystyle\lim_{n\to\infty}{\mathrm{haus}\left(K_{n}F(x),A_{F}(x)\right)}=0$. ### 5.3 Rate of convergence of the Bernstein-Durrmeyer operators and the Kantorovich operators for locally Lipschitz functions In this section the sequence $\\{T_{n}\\}$ is the sequence of Bernstein- Durrmeyer operators $\\{M_{n}\\}$ or the sequence of Kantorovich operators $\\{K_{n}\\}$. Using the notion of locally Lipschitz functions we can state the following result. ###### Theorem 5.9. Let $F\in\mathcal{F}[0,1]$ be locally Lipschitz around a point $x\in[0,1]$. Then $\mathrm{haus}\left(T_{n}F(x),F(x)\right)=O\left(\frac{1}{\sqrt{n}}\right),\quad n\to\infty.$ ###### Proof. By (24) $\mathrm{haus}\left(T_{n}F(x),F(x)\right)\leq\sup{\left\\{\left|T_{n}s(x)-s(x)\right|\ :\ s\in\mathcal{S}(F)\right\\}}.$ Since $F$ is Lipschitz around the point $x$ with some $\mathcal{L}>0$, by (ii) of Lemma 2.9, $v_{F}\in\mathrm{Lip}\\{{x},{\mathcal{L}}\\}$ and by Lemma 2.19 each metric selection $s\in\mathcal{S}(F)$ satisfies $s\in\mathrm{Lip}\\{{x},{4\mathcal{L}}\\}$. By Lemma 2.8 there exists $\widetilde{\mathcal{L}}>0$ such that $|s(z)-s(x)|\leq\widetilde{\mathcal{L}}|z-x|,\quad\forall z\in[0,1].$ (32) Using Lemma 2.12, (32), the Cauchy-Schwarz inequality and (29) for $\\{M_{n}\\}$ or (30) for $\\{K_{n}\\}$, we get $\displaystyle\left|T_{n}s(x)-s(x)\right|$ $\displaystyle\leq\int_{0}^{1}|s(t)-s(x)|\mathcal{K}_{n}(x,t)dt\leq\widetilde{\mathcal{L}}\int_{0}^{1}|t-x|\mathcal{K}_{n}(x,t)dt$ $\displaystyle\leq\widetilde{\mathcal{L}}\left(\int_{0}^{1}|t-x|^{2}\mathcal{K}_{n}(x,t)dt\right)^{\frac{1}{2}}\left(\int_{0}^{1}\mathcal{K}_{n}(x,t)dt\right)^{\frac{1}{2}}=O\left(\frac{1}{\sqrt{n}}\right),\quad n\to\infty,$ and the statement follows. ∎ ## 6 Approximation of the set of metric selections in $\mathbb{L^{1}[a,b]}$ In this section we consider the sequence of operators $\\{T_{n}\\}$ defined in (13) and study two sets of functions, $\mathcal{S}(F)$ and the set ${T_{n}\mathcal{S}(F)}=\\{T_{n}s:\,s\in\mathcal{S}(F)\\}.$ Note that for $F\in\mathcal{F}[a,b]$, the set $\mathcal{S}(F)\subset L^{1}[a,b]$, since it consists of functions of bounded variation. We assume that $\mathcal{K}_{n}(\cdot,\cdot)\in C([a,b]^{2})$. This condition guaranties that $T_{n}:\mathcal{S}(F)\to C[a,b]\subset L^{1}[a,b]$. Motivated by (22), we regard the set of functions $T_{n}\mathcal{S}(F)$ as an approximant to $F$, represented, in view of Result 2.14, by the set of functions $\mathcal{S}(F)$. We show below that $\mathcal{S}(F)$ and $T_{n}\mathcal{S}(F)$ are elements of the metric space $\widetilde{H}$ of compact non-empty subsets of $L^{1}[a,b]$, endowed with the Hausdorff metric. We use this metric to measure the approximation error of $\mathcal{S}(F)$ by $T_{n}\mathcal{S}(F)$. ### 6.1 Two compact sets of functions In the the next two lemmas we prove that the sets $\mathcal{S}(F)$, $T_{n}\mathcal{S}(F)$ are compact in $L^{1}{[a,b]}$, thus they are elements of $\widetilde{H}$. It is enough to show that they are sequentially compact, i.e. that every sequence in $\mathcal{S}(F)$ (in $T_{n}\mathcal{S}(F)$, respectively) has a convergent subsequence with a limit in $\mathcal{S}(F)$ (in $T_{n}\mathcal{S}(F)$, respectively). ###### Lemma 6.1. For $F\in\mathcal{F}[a,b]$ the set $\mathcal{S}(F)$ is compact in $L^{1}{[a,b]}$. ###### Proof. By Result 2.13 any metric selection $s\in\mathcal{S}(F)$ satisfies $\|s\|_{\infty}\leq\|F\|_{\infty}$ and $V_{a}^{b}(s)\leq V_{a}^{b}(F)$. Applying Helly’s selection principle, we conclude that for any sequence of metric selections there is a subsequence converging pointwisely at all points $x\in[a,b]$. By Result 2.16 the pointwise limit function of such a subsequence is a metric selection. Thus by Lebesgue Dominated Convergence Theorem this subsequence converges in the $L^{1}$-norm to the same limit metric selection. ∎ ###### Lemma 6.2. Let $F\in\mathcal{F}[a,b]$ and assume that $\mathcal{K}_{n}(\cdot,\cdot)\in C([a,b]^{2})$. Then the set $T_{n}\mathcal{S}(F)$ is compact in $L^{1}[a,b]$. ###### Proof. We have to show that any sequence $\\{\sigma_{k}\\}_{k=1}^{\infty}\subset T_{n}\mathcal{S}(F)$ has a subsequence converging in the $L^{1}$-norm and that its $L^{1}$-limit is in $T_{n}\mathcal{S}(F)$. By definition $\sigma_{k}(x)=T_{n}s_{k}(x)=\int_{a}^{b}\mathcal{K}_{n}(x,t)s_{k}(t)dt,$ for some $s_{k}\in\mathcal{S}(F)$. By Helly’s selection principle there exists a subsequence $\\{s_{k_{\ell}}\\}_{\ell=1}^{\infty}$ such that $\displaystyle\lim_{\ell\to\infty}s_{k_{\ell}}(t)=s^{\infty}(t)$ pointwisely for all $t\in[a,b]$. By Result 2.16, $s^{\infty}\in\mathcal{S}(F)$. Since $\mathcal{K}_{n}$ is bounded, we have $\displaystyle\lim_{\ell\to\infty}\mathcal{K}_{n}(x,t)s_{k_{\ell}}(t)=\mathcal{K}_{n}(x,t)s^{\infty}(t)$, for all $x,t\in[a,b]$. Clearly, the functions $\mathcal{K}_{n}(x,t)s_{k_{\ell}}(t)$, $\ell\in{\mathbb{N}}$ are dominated by $\|\mathcal{K}_{n}\|_{\infty}\|F\|_{\infty}$. Applying Lebesgue Dominated Convergence Theorem we get $\lim_{\ell\to\infty}\sigma_{k_{\ell}}(x)=\lim_{\ell\to\infty}\int_{a}^{b}\mathcal{K}_{n}(x,t)s_{k_{\ell}}(t)dt=\int_{a}^{b}\mathcal{K}_{n}(x,t)s^{\infty}(t)dt=\sigma^{\infty}(x),\qquad x\in[a,b].$ (33) Thus, we have that $\\{\sigma_{k_{\ell}}(x)\\}_{\ell=1}^{\infty}$ converges pointwisely to $\sigma^{\infty}(x)$ for all $x\in[a,b]$. Also the sequence $\\{\sigma_{k_{\ell}}\\}_{\ell=1}^{\infty}$ is dominated by $\|F\|_{\infty}\|\mathcal{K}_{n}\|_{\infty}(b-a)$, indeed $|\sigma_{k_{\ell}}(x)|\leq\int_{a}^{b}|\mathcal{K}_{n}(x,t)|\,|s_{k_{\ell}}(t)|dt\leq\|F\|_{\infty}\int_{a}^{b}|\mathcal{K}_{n}(x,t)|dt\leq\|F\|_{\infty}\|\mathcal{K}_{n}\|_{\infty}(b-a),\quad x\in[a,b].$ It follows that the sequence of functions $\\{\sigma_{k_{\ell}}(x)\\}$, ${\ell\in{\mathbb{N}}}$, $x\in[a,b]$ satisfies the assumptions of the Lebesgue Dominated Convergence Theorem and we get $\lim_{\ell\to\infty}\int_{a}^{b}|\sigma_{k_{\ell}}(x)-\sigma^{\infty}(x)|dx=0.$ To finish the proof, note that by (33) $\sigma^{\infty}\in T_{n}\mathcal{S}(F)$, since $s^{\infty}\in\mathcal{S}(F)$. ∎ The Hausdorff distance between the sets $\mathcal{S}(F)$ and $T_{n}\mathcal{S}(F)$ is $\mathrm{haus}(\mathcal{S}(F),T_{n}\mathcal{S}(F))=\max\left\\{\sup_{\sigma\in T_{n}\mathcal{S}(F)}\,\inf_{s\in\mathcal{S}(F)}\|s-\sigma\|_{L^{1}}\,,\,\sup_{s\in\mathcal{S}(F)}\inf_{\sigma\in T_{n}\mathcal{S}(F)}\|s-\sigma\|_{L^{1}}\right\\}.$ ###### Lemma 6.3. Let $F\in\mathcal{F}[a,b]$. For the sets $\mathcal{S}(F)$ and $T_{n}\mathcal{S}(F)$ $\mathrm{haus}(\mathcal{S}(F),T_{n}\mathcal{S}(F))\leq\sup_{s\in\mathcal{S}(F)}\|s-T_{n}s\|_{L^{1}}.$ ###### Proof. For any fixed $s^{*}\in\mathcal{S}(F)$ we have $\displaystyle\inf_{\sigma\in T_{n}\mathcal{S}(F)}\|s^{*}-\sigma\|_{L^{1}}\leq\|s^{*}-T_{n}s^{*}\|_{L^{1}}\leq\sup_{s\in\mathcal{S}(F)}\|s-T_{n}s\|_{L^{1}}.$ Similarly, for any fixed $\sigma^{*}\in T_{n}\mathcal{S}(F)$ there is $s^{*}\in\mathcal{S}(F)$ with $T_{n}s^{*}=\sigma^{*}$ and we get $\displaystyle\inf_{s\in\mathcal{S}(F)}\|s-\sigma^{*}\|_{L^{1}}\leq\|s^{*}-T_{n}s^{*}\|_{L^{1}}\leq\sup_{s\in\mathcal{S}(F)}\|s-T_{n}s\|_{L^{1}}.$ Now the statement follows easily. ∎ ### 6.2 Error estimates For $e_{i}(x)=x^{i}$, denote ${\displaystyle\lambda_{n}=(\max_{i=0,1}\|T_{n}e_{i}-e_{i}\|_{L^{1}})^{1/2}}$. Here we use the integral moduli $\vartheta$ and $\vartheta_{2}$ defined in Section 2.2. In light of the Theorem in [8] we get for real-valued functions (as a special case of the statement in [8] for $p=1$) ###### Result 6.4. Let $T_{n}$ be a positive linear operator satisfying $\|T_{n}g\|_{L^{1}}\leq\|g\|_{L^{1}}$ for any $g\in L^{1}[a,b]$. Then $\|T_{n}f-f\|_{L^{1}}\leq C_{1}\left(\lambda_{n}^{2}\|f\|_{L^{1}}+\vartheta_{2}(f,\lambda_{n})\right),\quad f\in L^{1}[a,b].$ (34) If, in addition $T_{n}e_{0}\equiv e_{0}$, then $\|T_{n}f-f\|_{L^{1}}\leq C_{2}\left(\lambda_{n}^{2}\,\vartheta\big{(}{f},{b-a}\big{)}+\vartheta_{2}\big{(}f,\lambda_{n}\big{)}\right),\quad f\in L^{1}[a,b].$ (35) Here the constants $C_{i}$, $i=0,1$ depend only on the interval $[a,b]$. Applying (34) and (35) on a metric selection $s\in\mathcal{S}(F)$ and using (10) and (11) we get $\|T_{n}s-s\|_{L^{1}}\leq\widetilde{C}_{1}\left(\lambda_{n}^{2}\,\|s\|_{L^{1}}+2\vartheta\big{(}{s},{\lambda_{n}}\big{)}\right)\leq\widetilde{C}_{1}\left(\lambda_{n}^{2}\,\|s\|_{L^{1}}+2\lambda_{n}V_{a}^{b}(s)\right),$ and in the case when $T_{n}e_{0}\equiv e_{0}$ $\|T_{n}s-s\|_{L^{1}}\leq\widetilde{C}_{2}\left(\lambda_{n}^{2}\,\vartheta\big{(}{s},{b-a}\big{)}+2\vartheta\big{(}{s},{\lambda_{n}}\big{)}\right)\leq\widetilde{C}_{2}\left(\lambda_{n}^{2}\,(b-a)V_{a}^{b}(s)+2\lambda_{n}V_{a}^{b}(s)\right),$ where $\widetilde{C}_{i}$, $i=0,1$ depend only on $[a,b]$ and on the underlying norm in ${{\mathbb{R}}}^{d}$. Since by Result 2.13 $V_{a}^{b}(s)\leq V_{a}^{b}(F)$, we obtain for any $s\in\mathcal{S}(F)$ $\sup_{s\in\mathcal{S}(F)}\|T_{n}s-s\|_{L^{1}}\leq\widetilde{C}_{1}\left(\lambda_{n}^{2}\,\sup_{s\in\mathcal{S}(F)}\|s\|_{L^{1}}+2\lambda_{n}V_{a}^{b}(F)\right),$ (36) and if $T_{n}e_{0}\equiv e_{0}$ $\sup_{s\in\mathcal{S}(F)}\|T_{n}s-s\|_{L^{1}}\leq\widetilde{C}_{2}\left(\lambda_{n}^{2}\,(b-a)+2\lambda_{n}\right)V_{a}^{b}(F).$ (37) Thus in view of Lemma 6.3, (36) and (37) we arrive at ###### Theorem 6.5. Let $T_{n}$ be a positive linear operator satisfying $\|T_{n}g\|_{L^{1}}\leq\|g\|_{L^{1}}$ for any $g\in L^{1}[a,b]$ and let $F\in\mathcal{F}[a,b]$. Then $\mathrm{haus}(\mathcal{S}(F),T_{n}\mathcal{S}(F))\leq\widetilde{C}_{1}\left(\lambda_{n}^{2}\,\sup_{s\in\mathcal{S}(F)}\|s\|_{L^{1}}+2\lambda_{n}V_{a}^{b}(F)\right)\leq\widetilde{C}_{1}\left(\lambda_{n}^{2}\,\|F\|_{\infty}(b-a)+2\lambda_{n}V_{a}^{b}(F)\right).$ If, in addition, $T_{n}e_{0}\equiv e_{0}$, then $\mathrm{haus}(\mathcal{S}(F),T_{n}\mathcal{S}(F))\leq\widetilde{C}_{2}\left(\lambda_{n}^{2}\,(b-a)+2\lambda_{n}\right)V_{a}^{b}(F).$ Here $\widetilde{C}_{i}$, $i=0,1$ depend only on $[a,b]$ and on the underlying norm in ${{\mathbb{R}}}^{d}$. ### 6.3 Examples We apply the results of Section 6.2 to the Bernstein-Durrmeyer operator $M_{n}$ and the Kantorovich operator $K_{n}$. As was shown in Section 5, these operators satisfy $M_{n}e_{0}\equiv e_{0}$ , $K_{n}e_{0}\equiv e_{0}$. Moreover, $M_{n}e_{1}(x)-e_{1}(x)=\frac{nx+1}{n+2}-x=\frac{1-2x}{n+2},\quad K_{n}e_{1}(x)-e_{1}(x)=\frac{2nx+1}{2(n+1)}-x=\frac{1-2x}{2(n+1)}\,.$ Combining this with the second claim of Theorem 6.5 we end up with the following result. ###### Theorem 6.6. Let $F\in\mathcal{F}[0,1]$ and let $T_{n}$ be the Bernstein-Durrmeyer operator or the Kantorovich operator. Then $\mathrm{haus}(\mathcal{S}(F),T_{n}\mathcal{S}(F))\leq CV_{a}^{b}(F)\left(\lambda_{n}^{2}\,(b-a)+2\lambda_{n}\right)=O\left(\frac{1}{\sqrt{n}}\right),\quad n\to\infty,$ with $C$ depending only on the underlying norm in ${{\mathbb{R}}}^{d}$. Note that the rate $O\left(\frac{1}{\sqrt{n}}\right)$ here is obtained for any $F\in\mathcal{F}[0,1]$, while in the pointwise estimates of Section 5.3 we obtain this rate under the assumption of local Lipschitz continuity of $F\in\mathcal{F}[0,1]$. ## References * [1] Z. Artstein, Piecewise linear approximations of set-valued maps, J. Approx. Theory 56 (1989), 41–47. * [2] J.-P. Aubin, H. Frankowska, Set-valued analysis, Birkhäuser Boston Inc., Boston (1990). * [3] R. J. Aumann, Integrals of set-valued functions, J. Math. Anal. Appl. 12 (1965), 1–12. * [4] V. F. Babenko, V. V. Babenko, M. V. Polishchuk, Approximation of some classes of set-valued periodic functions by generalized trigonometric polynomials, Ukrainian Math. J. 68(4) (2016), 502–514. * [5] R. Baier, G. Perria, Set-valued Hermite interpolation, J. Approx. Theory 163 (2011), 1349–1372. * [6] E. E. Berdysheva, N. Dyn, E. Farkhi, A. Mokhov, Metric approximation of set-valued functions of bounded variation, J. Comput. Appl. Math. 349 (2019), 251–264. * [7] E. E. Berdysheva, N. Dyn, E. Farkhi, A. Mokhov, Metric Fourier approximation of set-valued functions of bounded variation, J. Fourier Anal. Appl. (2021) 27:17. * [8] H. Berens, R. A. DeVore, Quantitative Korovkin theorems for $L_{p}$-spaces, In: Approximation Theory II, Proc. Internat. Sympos., Univ. Texas, Austin, Tex., Academic Press, New York (1976), 289–298. * [9] M. Campiti, Korovkin-type approximation in spaces of vector-valued and set-valued functions, Appl. Anal. 98 (2019), 2486–2496. * [10] M. Campiti, Korovkin approximation of set-valued integrable functions, manuscript (2022). * [11] V. V. Chistyakov, On mappings of bounded variation, J. Dyn. Control Syst. 3 (1997), 261–289. * [12] R. A. De Vore, G. G. Lorentz, Constructive Approximation, Springer-Verlag, Berlin Heidelberg (1993). * [13] N. Dyn, E. Farkhi, Spline subdivision schemes for convex compact sets, J. Comput. Appl. Math. 119(1-2) (2000), 133–144. * [14] N. Dyn, E. Farkhi, Spline subdivision schemes for compact sets with metric averages, In: Trends in approximation theory (Nashville, TN, 2000), Innov. Appl. Math., Vanderbilt Univ. Press, Nashville, TN, 2001, 93–102. * [15] N. Dyn, E. Farkhi, Set-valued approximations with Minkowski averages — convergence and convexification rates, Numer. Funct. Anal. Optim. 25 (2004), 363–377. * [16] N. Dyn, E. Farkhi, A. Mokhov, Approximations of set-valued functions by metric linear operators, Constr. Approx. 25 (2007), 193–209. * [17] N. Dyn, E. Farkhi, A. Mokhov, Approximation of univariate set-valued functions—an overview. Serdica Math. J. 33 (2007), no. 4, 495–514. * [18] N. Dyn, E. Farkhi, A. Mokhov, Approximations of Set-Valued Functions: Adaptation of Classical Approximation Operators, Imperial College Press, London (2014). * [19] N. Dyn, E. Farkhi, A. Mokhov, The metric integral of set-valued functions, Set-Valued Var. Anal. 26 (2018), 867–885. * [20] N. Dyn, A. Mokhov, Approximations of set-valued functions based on the metric average, Rend. Mat. Appl., VII. Ser. 26(3-4) (2006), 249–266. * [21] S. Guo, On the rate of convergence of the Durrmeyer operator for functions of bounded variation, J. Approx. Theory 51 (1987), 183–192. * [22] P.P. Korovkin, Linear Operators and Approximation Theory, Hindustan Publ. Corp., Delhi (1960). * [23] M. Mureşan, Set-valued approximation of multifunctions, Stud. Univ. Babeş-Bolyai, Math. 55 (1) (2010), 107–148. * [24] M. S. Nikolskiĭ, Approximation of continuous convex-valued multivalued mappings (Russian), Optimization 21(2) (1990), 209–214. * [25] R. T. Rockafellar, R. Wets, Variational Analysis, Springer-Verlag, Berlin (1998). * [26] B. Sendov, V. Popov, The Averaged Moduli of Smoothness: Applications in Numerical Methods and Approximation, Wiley (1988). * [27] R. Schneider, Convex Bodies: The Brunn–Minkowski Theory, Cambridge University Press, Cambridge (1993). * [28] R. A. Vitale, Approximations of convex set-valued functions, J. Approx. Theory 26 (1979), 301–316. * [29] X. Zeng, A. Piriou, On the rate of convergence of two Bernstein-Bézier type operators for bounded variation functions J. Approx. Theory 95 (1998), 369–387.
# Enveloping norms on the spaces of regularly $\cal{P}$-operators in Banach lattices ###### Abstract We introduce and study the enveloping norms of regularly $\cal{P}$-operators between Banach lattices $E$ and $F$, where $\cal{P}$ is a subspace of the space $\text{\rm L}(E,F)$ of continuous operators from $E$ to $F$. We prove that if $\cal{P}$ is closed in $\text{\rm L}(E,F)$ in the operator norm then the regularly $\cal{P}$-operators forms a Banach space under the enveloping norm. Conditions providing that regularly $\cal{P}$-operators forms a Banach lattice under the enveloping norm are given. Safak Alpay1, Eduard Emelyanov2, Svetlana Gorokhova 3 $1$ Department of Mathematics, Middle East Technical University, 06800 Ankara, Turkey<EMAIL_ADDRESS> $2$ Sobolev Institute of Mathematics, Acad. Koptyug avenue, 4, 630090, Novosibirsk, Russia<EMAIL_ADDRESS> $3$ Uznyj matematiceskij institut VNC RAN, Vatutin str., 53, 362025, Vladikavkaz, Russia<EMAIL_ADDRESS> Keywords: Banach lattice, regularly $\cal{P}$-operator, enveloping norm MSC2020: 46B25, 46B42, 46B50, 47B60 Throughout the paper, all vector spaces are real, all operators are linear, letters $E$ and $F$ denote Banach latices, $B_{E}$ denotes the closed unit ball of $X$, $\text{\rm L}(E,F)$ (K(E,F); W(E,F)) denotes the space of all continuous (resp. compact; weakly compact) operators from $E$ to $F$, and $E_{+}$ the positive cone of $E$. Let ${\cal P}\subseteq\text{\rm L}(E,F)$. We call elements of ${\cal P}$ by ${\cal P}$-operators and say that ${\cal P}$-operators satisfy the domination property if $0\leq S\leq T\in{\cal P}$ implies $S\in{\cal P}$. An operator $T\in\text{\rm L}(E,F)$ is called ${\cal P}$-dominated if $\pm T\leq U$ for some $U\in{\cal P}$. Under the assumption ${\cal P}\pm{\cal P}\subseteq{\cal P}\neq\emptyset$, ${\cal P}$-operators satisfy the domination property iff each ${\cal P}$-dominated operator lies in ${\cal P}$. We refer to [1, 10, 12] for unexplained terminology and notations. ## 0.1 In the present paper we continue the investigation of regularly ${\cal P}$-operators following [9, 2] and introduce enveloping norms on spaces of regularly ${\cal P}$-operators. An operator $T:E\to F$ is called regular if $T=T_{1}-T_{2}$ for some $T_{1},T_{2}\in\text{\rm L}_{+}(E,F)$. We denote by $\text{\rm L}_{r}(E,F)$ (resp. $\text{\rm L}_{ob}(E,F)$, $\text{\rm L}_{oc}(E,F)$) the ordered space of all regular (resp. order bounded, o-continuous) operators in $\text{\rm L}(E,F)$. The space $\text{\rm L}_{r}(E,F)$ is a Banach space under the regular norm [12, Prop.1.3.6] $\|T\|_{r}:=\inf\\{\|S\|:\pm T\leq S\in\text{\rm L}(E,F)\\}.$ (1) Furthermore, for every $T\in\text{\rm L}_{r}(E,F)$, $\|T\|_{r}=\inf\\{\|S\|:S\in\text{\rm L}(E,F),|Tx|\leq S|x|\ \forall x\in E\\}\geq\|T\|,$ (2) and if $F$ is Dedekind complete then $(\text{\rm L}_{r}(E,F),\|\cdot\|_{r})$ is a Banach lattice with $\|T\|_{r}=\|~{}|T|~{}\|$ for each $T\in\text{\rm L}_{r}(E,F)$ [12, Prop.1.3.6]. The following notion was introduced in [9, Def.2] (cf. also [2, Def.1.5.1]). ###### Definition 1. Let ${\cal P}\subseteq\text{\rm L}(E,F)$. An operator $T:E\to F$ is called a regularly ${\cal P}$-operator (shortly, an r-${\cal P}$-operator) if $T=T_{1}-T_{2}$ with $T_{1},T_{2}\in{\cal P}\cap\text{\rm L}_{+}(E,F)$. We denote by: 1. ${\cal P}(E,F):={\cal P}$ the set of all ${\cal P}$-operators in $\text{\rm L}(E,F)$; 2. ${\cal P}_{r}(E,F)$ the set of all regular operators in ${\cal P}(E,F)$; 3. $\text{\rm r-}{\cal P}(E,F)$ the set of all r-${\cal P}$-operators in $\text{\rm L}(E,F)$. ###### Assertion 2. ([2, Prop.1.5.2]) Let ${\cal P}\subseteq\text{\rm L}(E,F)$, ${\cal P}\pm{\cal P}\subseteq{\cal P}\neq\emptyset$, and $T\in\text{\rm L}(E,F)$. Then the following holds. 1. (i) $T$ is an r-${\cal P}$-operator iff $T$ is a ${\cal P}$-dominated ${\cal P}$-operator. 2. (ii) Suppose ${\cal P}$-operators satisfy the domination property and the modulus $|T|$ exists in $\text{\rm L}(E,F)$. Then $T$ is an r-${\cal P}$-operator iff $|T|\in{\cal P}$. The next fact was established in [2, Prop.1.5.3]. ###### Assertion 3. $\text{\rm r-}{\cal P}(E,F)$ is a Dedekind complete vector lattice if $F$ is Dedekind complete and ${\cal P}$ is a subspace of $\text{\rm L}(E,F)$ satisfying the domination property. ## 0.2 In the definition (1) of the regular norm, we replace $\text{\rm L}(E,F)$ by an arbitrary subspace ${\cal P}$ of $\text{\rm L}(E,F)$ as follows: $\|T\|_{\text{\rm r-}{\cal P}}:=\inf\\{\|S\|:\pm T\leq S\in{\cal P}\\}\ \ \ \ (T\in\text{\rm r-}{\cal P}(E,F)).$ (3) ###### Lemma 4. For a subspace ${\cal P}$ of $\text{\rm L}(E,F)$, the formula $(\ref{enveloping norm})$ defines a norm on $\text{\rm r-}{\cal P}(E,F)$, called the enveloping norm. Moreover, $\|T\|_{\text{\rm r-}{\cal P}}=\inf\\{\|S\|:S\in{\cal P}\ \&\ (\forall x\in E)\ |Tx|\leq S|x|\\}\ \ \ (T\in\text{\rm r-}{\cal P}(E,F)).$ (4) If ${\cal P}_{1}$ is a subspace of ${\cal P}$ then $\|T\|_{\text{\rm r-}{\cal P}_{1}}\geq\|T\|_{\text{\rm r-}{\cal P}}\geq\|T\|_{r}\geq\|T\|\ \ \ \ \ (\forall\ T\in\text{\rm r-}{\cal P}_{1}(E,F)).$ (5) ###### Proof. Only the triangle inequality and the formula (4) require explanations. (A) Let $T_{1},T_{2}\in\text{\rm r-}{\cal P}(E,F)$ and $\varepsilon>0$. Pick $S_{1},S_{2}\in{\cal P}$ with $\pm T_{1}\leq S_{1}$, $\pm T_{2}\leq S_{2}$, $\|S_{1}\|\leq\|T_{1}\|_{\text{\rm r-}{\cal P}}+\varepsilon$, and $\|S_{2}\|\leq\|T_{2}\|_{\text{\rm r-}{\cal P}}+\varepsilon$. Then $\pm(T_{1}+T_{2})\leq S_{1}+S_{2}\in{\cal P}$, and $\|T_{1}+T_{2}\|_{\text{\rm r-}{\cal P}}\leq\|S_{1}+S_{2}\|\leq\|S_{1}\|+\|S_{2}\|\leq\|T_{1}\|_{\text{\rm r-}{\cal P}}+\|T_{2}\|_{\text{\rm r-}{\cal P}}+2\varepsilon.$ Since $\varepsilon>0$ is arbitrary, $\|T_{1}+T_{2}\|_{\text{\rm r-}{\cal P}}\leq\|T_{1}\|_{\text{\rm r-}{\cal P}}+\|T_{2}\|_{\text{\rm r-}{\cal P}}$. (B) Denote the right side of (4) by $R(T)$. If $\pm T\leq S\in{\cal P}$ then $\pm Tx=\pm(T(x_{+})-T(x_{-}))=\pm T(x_{+})\mp T(x_{-})\leq S(x_{+})+S(x_{-})=S|x|$ for all $x\in E$. Then $|Tx|\leq S|x|$ for all $x\in E$, and hence $\|T\|_{\text{\rm r-}{\cal P}}\geq R(T)$. If $S\in{\cal P}$ satisfies $|Tx|\leq S|x|$ for all $x\in E$ then, for all $y\in E_{+}$, $|Ty|\leq Sy$ and consequently $\pm Ty\leq Sy$. Therefore $\pm T\leq S$, and hence $R(T)\geq\|T\|_{\text{\rm r-}{\cal P}}$. ∎ Let ${\cal P}$ be a subspace of $\text{\rm L}(E,F)$. What are the conditions for ${\cal P}_{r}(E,F)=\text{\rm r-}{\cal P}(E,F)$? Trivially it happens for ${\cal P}=\text{\rm L}(E,F)$. The following proposition shows that if $\text{\rm L}(E,F)$ is non-trival then we can always find a closed subspace ${\cal P}$ such that $\text{\rm r-}{\cal P}(E,F)$ is a proper subspace of ${\cal P}_{r}(E,F)$. More interesting case of the question is included in Example 20 below. ###### Proposition 5. Let $\dim(\text{\rm L}(E,F))>1$. Then there exists a one-dimensional subspace ${\cal P}$ of $\text{\rm L}(E,F)$ such that $\text{\rm r-}{\cal P}(E,F)\subsetneqq{\cal P}_{r}(E,F)$. ###### Proof. If $\dim(E)>1$ then $\dim(E^{\prime})>1$ and hence there are two linearly independent functionals $g_{1},g_{2}\in E^{\prime}_{+}$. We take a nonzero element $y\in F_{+}$ and set ${\cal P}:=\text{\rm span}((g_{1}-g_{2})\otimes y)$. If $\dim(F)>1$, there are two linearly independent elements $y_{1},y_{2}\in F_{+}$. We take a nonzero functional $g\in E^{\prime}_{+}$ and set ${\cal P}:=\text{\rm span}(g\otimes(y_{1}-y_{2}))$. In the both cases ${\cal P}$ is one-dimensional subspace of $\text{\rm L}(E,F)$. Moreover, $\text{\rm r-}{\cal P}(E,F)=\\{0\\}\neq{\cal P}={\cal P}_{r}(E,F)$. ∎ ## 0.3 The following result is an extension of [12, Prop.1.3.6], [9, Lm.1] (see also [6, Prop.2.2] and [7, Thm.2.3] for particular cases). Its proof is a modification of the proof of [12, Prop.1.3.6]. ###### Theorem 6. Let ${\cal P}$ be a subspace of $\text{\rm L}(E,F)$ closed in the operator norm. Then $\text{\rm r-}{\cal P}(E,F)$ is a Banach space under the enveloping norm. ###### Proof. Let $(T_{n})$ be a $\|\cdot\|_{\text{\rm r-}{\cal P}}$-Cauchy sequence in $\text{\rm r-}{\cal P}(E,F)$, say $T_{n}=P_{n}-R_{n}$ for $P_{n},R_{n}\in{\cal P}\cap{\cal L}_{+}(E,F)$. Without lost of generality, we can assume $\|T_{n+1}-T_{n}\|_{\text{\rm r-}{\cal P}}<2^{-n}\ \ \ (\forall n\in\mathbb{N}).$ Since $\|\cdot\|_{\text{\rm r-}{\cal P}}\geq\|\cdot\|$, there exists a $T\in\text{\rm L}(E,F)$ with $\|T-T_{n}\|\to 0$. We obtain $T\in{\cal P}$ because $T_{n}\in{\cal P}$ and ${\cal P}$ is closed in the operator norm. Pick $S_{n}\in{\cal P}$ with $\|S_{n}\|<2^{-n}$ and $\pm(T_{n+1}-T_{n})\leq S_{n}$. By (4), $|(T_{n+1}-T_{n})x|\leq S_{n}|x|\ \ \ \ \ (\forall x\in E)(\forall n\in\mathbb{N}).$ As ${\cal P}$ is closed, $Q_{n}:=\|\cdot\|\text{\rm-}\sum\limits_{k=n}^{\infty}S_{k}\in{\cal P}$ for each $n\in\mathbb{N}$. Since $|(T-T_{n})x|=\lim\limits_{k\to\infty}|(T_{k}-T_{n})x|\leq\sum\limits_{k=n}^{\infty}|(T_{k+1}-T_{n})x|\leq Q_{n}|x|\ \ \ \ \ (\forall x\in E),$ then $\pm(T-T_{n})\leq Q_{n}$ for all $n\in\mathbb{N}$. Thus $-Q_{n}\leq(T-T_{n})\leq Q_{n}$ and hence $0\leq(T-T_{n})+Q_{n}$ for all $n\in\mathbb{N}$. In particular, $T=[(T-T_{n})+Q_{n}]+[T_{n}-Q_{n}]=[(T-T_{n})+Q_{n}+P_{n}]-[R_{n}+Q_{n}]\in\text{\rm r-}{\cal P}(E,F),$ and hence $(T-T_{n})\in\text{\rm r-}{\cal P}(E,F)$ for all $n\in\mathbb{N}$. Now, $\|T-T_{n}\|_{\text{\rm r-}{\cal P}}\leq\|Q_{n}\|<2^{1-n}$ implies $(T_{n})\stackrel{{\scriptstyle\|\cdot\|_{\text{\rm r-}{\cal P}}}}{{\to}}T$. ∎ ###### Theorem 7. Let $F$ be Dedekind complete and let ${\cal P}$ be a closed in the operator norm subspace of $\text{\rm L}(E,F)$ satisfying the domination property. Then $\text{\rm r-}{\cal P}(E,F)$ is a Dedekind complete Banach lattice under the enveloping norm. ###### Proof. By Assertion 3, $\text{\rm r-}{\cal P}(E,F)$ is a Dedekind complete vector lattice. Theorem 6 implies that $\text{\rm r-}{\cal P}(E,F)$ is $\|\cdot\|_{\text{\rm r-}{\cal P}}$-complete. Let $|S|\leq|T|$ for some $S,T\in\text{\rm r-}{\cal P}(E,F)$. By Assertion 2 (ii), $|S|,|T|\in{\cal P}$, and hence $\|S\|_{\text{\rm r-}{\cal P}}=\|~{}|S|~{}\|=\sup\limits_{x\in E_{+}\cap B_{E}}\||S|x\|\leq\sup\limits_{x\in E_{+}\cap B_{E}}\||T|x\|=\|~{}|T|~{}\|=\|T\|_{\text{\rm r-}{\cal P}}.$ Therefore, $(\text{\rm r-}{\cal P}(E,F),\|\cdot\|_{\text{\rm r-}{\cal P}})$ is a Banach lattice. ∎ ## 0.4 In general, $\text{\rm r-}{\cal P}(E,F)\subsetneqq\text{\rm r-}{\overline{\cal P}}(E,F)$, where ${\overline{\cal P}}$ is the norm-closure of ${\cal P}$ in $\text{\rm L}(E,F)$. From the other hand, for ${\cal P}:=\text{\rm r-L}_{ob}(E,F)$ (which is almost never closed in $\text{\rm L}(E,F)$ in the operator norm), $\text{\rm r-}{\cal P}(E,F)=\text{\rm r-}{\overline{\cal P}}(E,F)=\text{\rm L}_{r}(E,F)$. The enveloping norms on these three spaces agree with the regular norm. The next result coupled with Example 9 shows that the enveloping norm on $\text{\rm r-}{\cal P}(E,F)$ can be complete even if ${\cal P}(E,F)$ is not complete in the operator norm. ###### Proposition 8. Let the norm in $F$ be o-continuous. Then $\text{\rm r-L}_{oc}(E,F)$ is a Banach space under the enveloping norm. ###### Proof. Let $(T_{n})$ be a Cauchy sequence in $\text{\rm r-L}_{oc}(E,F)$ in the enveloping norm. We can assume that $\|T_{n+1}-T_{n}\|_{\text{\rm r-L}_{oc}(E,F)}<2^{-n}$ for all $n\in\mathbb{N}$. Let $T\in\text{\rm L}(E,F)$ satisfy $\|T-T_{n}\|\to 0$. Pick $S_{n}\in\text{\rm L}_{oc}(E,F)$ with $\|S_{n}\|<2^{-n}$ and $\pm(T_{n+1}-T_{n})\leq S_{n}$. First, we claim $Q_{n}:=\|\cdot\|\text{\rm-}\sum\limits_{k=n}^{\infty}S_{k}\in\text{\rm L}_{oc}(E,F)$ for all $n\in\mathbb{N}$. To prove the claim, it suffices to show that $Q_{1}\in\text{\rm L}_{oc}(E,F)$. So, let $x_{\alpha}\downarrow 0$ in $E$. Passing to a tail we can assume that $\|x_{\alpha}\|\leq M\in\mathbb{R}$ for all $\alpha$. Since $Q_{1}\geq 0$ then $Q_{1}x_{\alpha}\downarrow\geq 0$ and hence in order to show that $Q_{1}x_{\alpha}\downarrow 0$ it is enough to prove that $\|Q_{1}x_{\alpha}\|\to 0$. Let $\varepsilon>0$. Fix an $m\in\mathbb{N}$ with $\|Q_{m+1}\|\leq\varepsilon$. Since the positive operators $S_{1},...,S_{m}$ are all o-continuous, and since the norm in $F$ is o-continuous, there exists an $\alpha_{1}$ such that $\sum\limits_{k=1}^{m}\|S_{k}x_{\alpha}\|\leq\varepsilon$ for all $\alpha\geq\alpha_{1}$. Since $\varepsilon>0$ is arbitrary, it follows from $\|Q_{1}x_{\alpha}\|\leq\|\sum\limits_{k=1}^{m}S_{k}x_{\alpha}\|+\|Q_{m+1}x_{\alpha}\|\leq\varepsilon+M\|Q_{m+1}\|\leq 2\varepsilon\ \ \ \ (\forall\alpha\geq\alpha_{1})$ that $\|Q_{1}x_{\alpha}\|\to 0$, which proves our claim that $Q_{n}\in\text{\rm L}_{oc}(E,F)$ for all $n\in\mathbb{N}$. Since $\pm(T_{n+1}-T_{n})\leq S_{n}$ then by formula (4), $|(T-T_{n})x|=\lim\limits_{k\to\infty}|(T_{k}-T_{n})x|\leq\sum\limits_{k=n}^{\infty}|(T_{k+1}-T_{n})x|\leq\sum\limits_{k=n}^{\infty}S_{n}|x|=Q_{n}|x|$ for all $x\in E$. In particular, $|T-T_{1}|\leq Q_{1}\in\text{\rm L}_{oc}(E,F)$ and, since $\text{\rm L}_{oc}(E,F)$ is an order ideal in $\text{\rm L}_{r}(E,F)$, then $(T-T_{1})\in\text{\rm L}_{oc}(E,F)$. Since $T_{1}\in\text{\rm L}_{oc}(E,F)$, it follows $T\in\text{\rm L}_{oc}(E,F)$. Now, $\|T-T_{n}\|_{\text{\rm r-L}_{oc}(E,F)}\leq\|Q_{n}\|<2^{1-n}$ implies $(T_{n})\stackrel{{\scriptstyle\|\cdot\|_{\text{\rm r-L}_{oc}(E,F)}}}{{\longrightarrow}}T$. ∎ ###### Example 9. Consider the following modification of Krengel’s example [11] (cf. [1, Ex.5.6]$)$. Define a sequence $(S_{n})$ in $\text{\rm L}_{oc}((\oplus_{n=1}^{\infty}\ell^{2}_{2^{n}})_{0})$ by $S_{n}x:=(2^{-\frac{1}{3}}T_{1}x_{1},2^{-\frac{2}{3}}T_{2}x_{2},\dots 2^{-\frac{n}{3}}T_{n}x_{n},0,0,\dots),$ where $T_{n}:\ell^{2}_{2^{n}}\to\ell^{2}_{2^{n}}$ is an isometry as in [1, Ex.5.6]. Then $\|S_{n}-S\|\to 0$ for an operator $S\in\text{\rm L}(E,F)$, defined by $Sx:=(2^{-\frac{1}{3}}T_{1}x_{1},2^{-\frac{2}{3}}T_{2}x_{2},\dots 2^{-\frac{n}{3}}T_{n}x_{n},\dots).$ Notice that $(\oplus_{n=1}^{\infty}\ell^{2}_{2^{n}})_{0}$ has o-continuous norm, and hence is Dedekind complete. However $|S|$ does not exist. Thus, $S$ is not order bounded and hence $S\not\in\text{\rm L}_{oc}((\oplus_{n=1}^{\infty}\ell^{2}_{2^{n}})_{0})$. ## 0.5 We include several applications of Theorem 6 and Assertion 2. We begin with recalling the following definition (see [5, 8]). ###### Definition 10. A continuous operator $T:E\to F$ is called limited (resp. almost limited) if $T^{\prime}$ takes $\text{\rm w}^{\ast}$-null (resp. disjoint $\text{\rm w}^{\ast}$-null) sequences of $F^{\prime}$ to norm null sequences of $E^{\prime}$. By $\text{\rm Lm}(E,F)$ (by $\text{\rm a-Lm}(E,F)$) we denote the space of limited (resp. almost limited) operators from $E$ to $F$. ###### Lemma 11. The spaces $\text{\rm a-Lm}(E,F)$ and $\text{\rm Lm}(E,F)$ are both closed in $\text{\rm L}(E,F)$ under the operator norm. ###### Proof. Suppose $\text{\rm a-Lm}(E,F)\ni T_{k}\stackrel{{\scriptstyle\|\cdot\|}}{{\to}}T\in\text{\rm L}(E,F)$. Let $(f_{n})$ be disjoint $\text{\rm w}^{\ast}$-null in $F^{\prime}$. We need to show that $(T^{\prime}f_{n})$ is norm null $X^{\prime}$. Let $\varepsilon>0$. Choose any $k$ with $\|T^{\prime}_{k}-T^{\prime}\|\leq\varepsilon$. Since $T_{k}\in\text{\rm a-Lm}(X,F)$, there exists $n_{0}$ such that $\|T_{k}^{\prime}f_{n}\|\leq\varepsilon$ whenever $n\geq n_{0}$. As $(f_{n})$ is $\text{\rm w}^{\ast}$-null, there exists $M\in\mathbb{R}$ such that $\|f_{n}\|\leq M$ for all $n\in\mathbb{N}$. Then $\|T^{\prime}f_{n}\|\leq\|T^{\prime}_{k}f_{n}\|+\|T^{\prime}_{k}f_{n}-T^{\prime}f_{n}\|\leq\varepsilon+\|T^{\prime}_{k}-T^{\prime}\|\|f_{n}\|\leq\varepsilon+M\varepsilon$ for $n\geq n_{0}$. Since $\varepsilon>0$ is arbitrary then $(T^{\prime}f_{n})$ is norm null, as desired. Similarly, the space $\text{\rm Lm}(E,F)$ is closed. ∎ The next result follows from Theorem 6 and Lemma 11. ###### Corollary 12. For arbitrary Banach lattices $E$ and $F$, $\text{\rm r-Lm}(E,F)$ and $\text{\rm r-a-Lm}(E,F)$ are both Banach spaces (each under its enveloping norm). ###### Definition 13. (see [10]) A continuous operator $T:E\to F$ is called Grothendieck (resp. almost Grothendieck) if $T^{\prime}$ takes $\text{\rm w}^{\ast}$-null (resp. disjoint $\text{\rm w}^{\ast}$-null) sequences of $F^{\prime}$ to w-null sequences of $E^{\prime}$. By $\text{\rm G}(E,F)$ (by $\text{\rm a-G}(E,F)$) we denote the space of Grothendieck (resp. almost Grothendieck) operators from $E$ to $F$. By Definitions 10 and 13, $\text{\rm K}(E,F)\subseteq\text{\rm Lm}(E,F)\subseteq\text{\rm a-Lm}(E,F)\subseteq\text{\rm a-G}(E,F),$ (6) and $\text{\rm Lm}(E,F)\subseteq\text{\rm G}(E,F)\subseteq\text{\rm a-G}(E,F).$ (7) ###### Lemma 14. The spaces $\text{\rm a-G}(E,F)$ and $\text{\rm G}(E,F)$ are both closed in $\text{\rm L}(E,F)$ under the operator norm. ###### Proof. Suppose $\text{\rm a-G}(E,F)\ni T_{k}\stackrel{{\scriptstyle\|\cdot\|}}{{\to}}T\in\text{\rm L}(E,F)$. Let $(f_{n})$ be disjoint $\text{\rm w}^{\ast}$-null in $F^{\prime}$. We need to show that $(T^{\prime}f_{n})$ is w-null in $E^{\prime}$. So, take any $g\in F^{\prime\prime}$. Let $\varepsilon>0$. Pick a $k$ with $\|T^{\prime}_{k}-T^{\prime}\|\leq\varepsilon$. Since $T_{k}\in\text{\rm a-G}(E,F)$, there exists $n_{0}$ such that $|g(T_{k}^{\prime}f_{n})|\leq\varepsilon$ whenever $n\geq n_{0}$. Note that $\|f_{n}\|\leq M$ for some $M\in\mathbb{R}$ and for all $n\in\mathbb{N}$. Since $\varepsilon>0$ is arbitrary, it follows from $|g(T^{\prime}f_{n})|\leq|g(T_{k}^{\prime}f_{n}-T^{\prime}f_{n})|+|g(T^{\prime}f_{n})|\leq\|g\|\|T_{k}^{\prime}-T^{\prime}\|\|f_{n}\|+\varepsilon\leq(\|g\|M+1)\varepsilon$ for $n\geq n_{0}$, that $g(T^{\prime}f_{n})\to 0$. Since $g\in F^{\prime\prime}$ is arbitrary, $T\in\text{\rm a-G}(E,F)$. Similarly, the space $\text{\rm G}(E,F)$ is closed. ∎ The next result follows from Theorem 6 and Lemma 14. ###### Corollary 15. For arbitrary Banach lattices $E$ and $F$, $\text{\rm r-G}(E,F)$ and $\text{\rm r-a-G}(E,F)$ are both Banach spaces (each under its enveloping norm). We continue with the following definition. ###### Definition 16. A continuous operator $T:E\to F$ is called: 1. a) Dunford–Pettis (shortly, $T\in\text{\rm DP}(E,F)$) if $T$ takes w-null sequences to norm null ones (cf. [1, p.340]). 2. b) weak Dunford–Pettis (shortly, $T\in\text{\rm wDP}(E,F)$) if $f_{n}(Tx_{n})\to 0$ whenever $(f_{n})$ is w-null in $F^{\prime}$ and $(x_{n})$ is w-null in $E$ [1, p.349]. 3. c) almost Dunford–Pettis (shortly, $T\in\text{\rm a-DP}(E,F)$) if $T$ takes disjoint w-null sequences to norm null ones [13]; 4. d) almost weak Dunford–Pettis (shortly, $T\in\text{\rm a-wDP}(E,F)$) if $f_{n}(Tx_{n})\to 0$ whenever $(f_{n})$ is w-null in $F^{\prime}$ and $(x_{n})$ is disjoint w-null in $E$. Clearly, $\text{\rm K}(E,F)\subseteq\text{\rm DP}(E,F)\subseteq\text{\rm wDP}(E,F)\bigcap\text{\rm a-DP}(E,F),$ (8) and $\text{\rm wDP}(E,F)\subseteq\text{\rm a-wDP}(E,F).$ (9) The operator $L^{1}[0,1]\stackrel{{\scriptstyle T}}{{\to}}\ell^{\infty}$, $Tf=\left(\int_{0}^{1}f(t)r^{+}_{k}(t)\,dt\right)_{k=1}^{\infty}$ is wDP yet not DP, as $r_{n}\stackrel{{\scriptstyle\text{\rm w}}}{{\to}}0$ and $\|Tr_{n}\|\geq\int_{0}^{1}r_{n}(t)r^{+}_{n}(t)\,dt\equiv\frac{1}{2}$. By [3, Thm.4.1], $\text{\rm a-DP}(E,F)=\text{\rm DP}(E,F)$ for all $F$ iff the lattice operations in $E$ are sequentially w-continuous. The identity operator: 1. $I:L^{1}[0,1]\to L^{1}[0,1]$ is a-DP yet not DP; 2. $I:c\to c$ is a-wDP yet neither wDP nor a-DP. The domination property for a-DP-operators was established in [4, Cor.2.3]. We omit the proof of the next proposition as it is a modification of the proof of the Kalton–Saab domination theorem (cf. [1, Thm.5.101]. ###### Proposition 17. Any positive operator dominated by an a-wDP-operator is likewise an a-wDP- operator. We also omit the straightforward proof of the following fact. ###### Lemma 18. The spaces $\text{\rm a-DP}(E,F)$ and $\text{\rm a-wDP}(E,F)$ are both closed in $\text{\rm L}(E,F)$ under the operator norm. By [4, Cor.2.3], Proposition 17, and Lemma 18, the next result follows from Theorem 7. ###### Corollary 19. If $F$ is Dedekind complete then $\text{\rm r-a-DP}(E,F)$ and $\text{\rm r-a-wDP}(E,F)$ are both Banach lattices under their enveloping norms. ## 0.6 Finally, we discuss once more the Krengel example (cf. [1, Ex.5.6]. ###### Example 20. Let $A_{n}$ be a $2^{n}\times 2^{n}$ matrix constructed inductively: $A_{1}=\left[\begin{matrix}1&1\\\ 1&-1\end{matrix}\right]\quad\text{\rm and}\quad A_{n+1}=\left[\begin{matrix}A_{n}&A_{n}\\\ A_{n}&-A_{n}\end{matrix}\right].$ For each $n$, let $T_{n}:\ell^{2}_{2^{n}}\to\ell^{2}_{2^{n}}$ be an isometry defined by the orthogonal matrix $2^{-\frac{n}{2}}A_{n}$. Then $|T_{n}|\in\text{\rm L}(\ell^{2}_{2^{n}})$ is the operator, whose $2^{n}\times 2^{n}$ coordinate matrix has all entries equal to $2^{-\frac{n}{2}}$, in particular $\||T_{n}|\|=2^{\frac{n}{2}}$ for all $n$. Consider the $c_{0}$-direct sum $E:=(\oplus_{n=1}^{\infty}\ell^{2}_{2^{n}})_{0}$. Note that $E$ has o-continuous norm and hence is Dedekind complete. Define $T:E\to E$, by $Tx:=(2^{-\frac{n}{2}}T_{n}x_{n})\ \ \ \ (x=(x_{1},x_{2},\dots,x_{n},\dots)\in E).$ (10) It is straightforward to see that $T\in\text{\rm K}(E)$, and then $T\in\text{\rm a-G}(E)$ by (6), and $T\in\text{\rm a-DP}(E)$ by (8). The modulus $|T|$ exists and is given by $|T|x=(2^{-\frac{n}{2}}|T_{n}|x_{n})\ \ \ \ (x\in E).$ (11) However, $|T|\not\in\text{\rm a-G}(E)$. Indeed, consider a disjoint $\text{\rm w}^{\ast}$-null sequence $(f_{n})$ in $E^{\prime}=(\oplus_{n=1}^{\infty}\ell^{2}_{2^{n}})_{1}$, where all terms in $f_{n}$ are zero, except the $n$-th term which is equal to $(2^{-\frac{n}{2}},2^{-\frac{n}{2}},\dots,2^{-\frac{n}{2}})\in\ell^{2}_{2^{n}}$. Since $|T|^{\prime}\in\text{\rm L}((\oplus_{n=1}^{\infty}\ell^{2}_{2^{n}})_{1})$ satisfies $|T|^{\prime}y=(2^{-\frac{n}{2}}|T_{n}|^{\prime}y_{n})$ for all $y\in E^{\prime}$, where the coordinate matrix of $|T_{n}|^{\prime}$ has all entries equal to $2^{-\frac{n}{2}}$, then all terms of $|T|^{\prime}f_{n}$ are zero, except the $n$-th term which is equal to $(2^{-\frac{n}{2}},2^{-\frac{n}{2}},\dots,2^{-\frac{n}{2}})\in\ell^{2}_{2^{n}}$. Take $g\in E^{\prime\prime}=(\oplus_{k=1}^{\infty}\ell^{2}_{2^{k}})_{\infty}$ by letting each $k$-th term of $g$ equals to $(2^{-\frac{k}{2}},2^{-\frac{k}{2}},\dots,2^{-\frac{k}{2}})\in\ell^{2}_{2^{k}}$, then $g(|T|^{\prime}f_{n})=\sum\limits_{i=1}^{2^{n}}2^{-\frac{n}{2}}\cdot 2^{-\frac{n}{2}}=\sum\limits_{i=1}^{2^{n}}2^{-n}=1\not\to 0,$ and hence $T\not\in\text{\rm a-G}(E)$. In particular, $T\not\in\text{\rm a-Lm}(E)$ by (6). Now, take a disjoint w-null sequence $(x_{n})$ in $E=(\oplus_{n=1}^{\infty}\ell^{2}_{2^{n}})_{0}$, whose terms are zero, except the $n$-th one which is equals to $(2^{-\frac{n}{2}},2^{-\frac{n}{2}},\dots,2^{-\frac{n}{2}})\in\ell^{2}_{2^{n}}$. Then $|T|x_{n}$ has all the terms zero, except the $n$-th term which is equal to $(2^{-\frac{n}{2}},2^{-\frac{n}{2}},\dots,2^{-\frac{n}{2}})$. Thus $\||T|x_{n}\|=1$ for all $n$, and hence $|T|\not\in\text{\rm a-DP}(E)$. Since a-G\- and a-DP-operators satisfy the domination property, it follows from Assertion 2 (ii) that $T\in\text{\rm a-G}_{r}((\oplus_{n=1}^{\infty}\ell^{2}_{2^{n}})_{0})\setminus\text{\rm r-a-G}((\oplus_{n=1}^{\infty}\ell^{2}_{2^{n}})_{0})$ and $T\in\text{\rm a-DP}_{r}((\oplus_{n=1}^{\infty}\ell^{2}_{2^{n}})_{0})\setminus\text{\rm r-a- DP}((\oplus_{n=1}^{\infty}\ell^{2}_{2^{n}})_{0})$. ## References * [1] C. D. Aliprantis, O. Burkinshaw, Positive Operators. Springer, Dordrecht, 2006. * [2] S. Alpay, E. Emelyanov, S. Gorokhova, Generalizations of L- and M-weakly compact operators. https://arxiv.org/abs/2206.02718v4 * [3] B. Aqzzouz, K. Bouras, (L) sets and almost (L) sets in Banach lattices. Quaest. Math. 36, no.1 (2013), 107–118. * [4] B. Aqzzouz, A. Elbour, Some characterizations of almost Dunford–Pettis operators and applications. Positivity 15 (2011), 369–380. * [5] J. Bourgain, J. Diestel, Limited operators and strict cosingularity. Math. Nachr. 119 (1984), 55–58. * [6] Z. L. Chen, A. W. Wickstead, Incompleteness of the linear span of the positive compact operators. Proc. Amer. Math. Soc. 125 (1997), 3381–3389. * [7] N. Cheng, Dedekind $\sigma$-complete vector lattice of b-AM-compact operators. Quaest. Math. 40 (2017), no. 3, 313–318. * [8] A. Elbour, Some characterizations of almost limited operators. Positivity 21 (2017), 865–874. * [9] E. Emelyanov, Algebras of Lebesgue and KB regular operators on Banach lattices. https://arxiv.org/abs/2203.08326v3 * [10] P. Galindo, V. C. C. Miranda, Grothendieck-type subsets of Banach lattices. J. Math. Anal. Appl. 506 (2022) * [11] U. Krengel, Remark on the modulus of compact operators. Bull. Amer. Math. Soc. 72 (1966), 132–133. * [12] P. Meyer-Nieberg, Banach Lattices. Universitext, Springer-Verlag, Berlin 1991. * [13] J. A. Sanchez, Operators on Banach lattices. Ph. D. Thesis, Complutense University, Madrid, 1985.
# Nontrivial lower bounds for the $p$-adic valuations of some type of rational numbers Bakir FARHI Laboratoire de Mathématiques appliquées Faculté des Sciences Exactes Université de Bejaia, 06000 Bejaia, Algeria <EMAIL_ADDRESS> http://farhi.bakir.free.fr/ ###### Abstract In this paper, we will show that the $p$-adic valuation (where $p$ is a given prime number) of some type of rational numbers is unusually large. This generalizes the very recent results by the author and by A. Dubickas, which are both related to the special case $p=2$. The crucial point for obtaining our main result is the fact that the $p$-adic valuation of the rational numbers in question is unbounded from above. We will confirm this fact by three different methods; the first two are elementary while the third one leans on the $p$-adic analysis. MSC 2010: Primary 11B83; Secondary 11A41, 05A10, 05A19, 11B65. Keywords: $p$-adic valuations, binomial coefficients, combinatorial identities, $p$-adic analysis. ## 1 Introduction and Notation Throughout this paper, we let ${\mathbb{N}}$ denote the set of positive integers and ${\mathbb{N}}_{0}:={\mathbb{N}}\cup\\{0\\}$ denote the set of non-negative integers. For $x\in{\mathbb{R}}$, we let $\lfloor x\rfloor$ denote the integer part of $x$. For a given prime number $p$ and a given non- zero rational number $r$, we let $\vartheta_{p}(r)$ denote the usual $p$-adic valuation of $r$; if in addition $r$ is positive then we let $\log_{p}(r)$ denote its logarithm to the base $p$ (i.e., $\log_{p}(r):=\frac{\log{r}}{\log{p}}$). Next, the least common multiple of given positive integers $u_{1},u_{2},\dots,u_{n}$ ($n\in{\mathbb{N}}$) is denoted by $\mathrm{lcm}(u_{1},u_{2},\dots,u_{n})$ or by $\mathrm{lcm}\\{u_{1},u_{2},\dots,u_{n}\\}$ if this is more convenient. In several places of this paper, we will use the immediate estimate $\vartheta_{p}(n)\leq\log_{p}(n)$ (for any prime $p$ and any $n\in{\mathbb{N}}$). We also often use the immediate formula $\vartheta_{p}\left(\mathrm{lcm}(1,2,\dots,n)\right)=\left\lfloor\log_{p}(n)\right\rfloor$ (for any prime $p$ and any $n\in{\mathbb{N}}$). At the end of the paper, we need to use the $p$-adic logarithm function which we denote by $L_{p}$ (to differentiate from the notation $\log_{p}$, which is reserved to denote the logarithm to the base $p$). With the usual notation ${\mathbb{Q}}_{p}$ for the field of $p$-adic numbers, ${\mathbb{C}}_{p}$ for the field of the $p$-adic complex numbers, and ${|\cdot|}_{p}$ for the usual $p$-adic absolute value on ${\mathbb{C}}_{p}$, recall that $L_{p}$ can be defined by: $-L_{p}(1-x):=\sum_{k=1}^{+\infty}\frac{x^{k}}{k}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\forall x\in{\mathbb{C}}_{p},{|x|}_{p}<1).$ (See [5]). The fundamental property of $L_{p}$ is that it satisfies the functional equation: $L_{p}(uv)=L_{p}(u)+L_{p}(v)$ (for all $u,v\in{\mathbb{C}}_{p}$, with ${|u-1|}_{p}<1$ and ${|v-1|}_{p}<1$). In [2, 3], the author have obtained nontrivial lower bounds for the $2$-adic valuation of the rational numbers of the form $\sum_{k=1}^{n}\frac{2^{k}}{k}$ ($n\in{\mathbb{N}}$). The stronger one is $\vartheta_{2}\left(\sum_{k=1}^{n}\frac{2^{k}}{k}\right)\geq n-\left\lfloor\log_{2}(n)\right\rfloor~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\forall n\in{\mathbb{N}}).$ (1.1) In [3], the author also posed the problem of generalizing (1.1) to other prime numbers $p$ other than $p=2$. In [1], Dubickas has found arguments to improving and optimizing (1.1) by leaning only on the fact that the sequence $\left\\{\vartheta_{2}\left(\sum_{k=1}^{n}\frac{2^{k}}{k}\right)\right\\}_{n\geq 1}$ is unbounded from above. However, he does not established any way to prove this fact without using (1.1). The main result in [1] states that we have for any $n\in{\mathbb{N}}$: $\vartheta_{2}\left(\sum_{k=1}^{n}\frac{2^{k}}{k}\right)\geq(n+1)-\log_{2}(n+1),$ (1.2) with equality if and only if $n$ has the form $n=2^{\alpha}-1$ ($\alpha\in{\mathbb{N}}$). The goal of this paper is twofold. On the one hand, we expand and improve the arguments in [1] to establish a general result providing to us nontrivial lower bounds for the $p$-adic valuation of a sum of rational numbers under some conditions (see Theorem 2.2). On the other hand, we solve the problem posed in [3] by generalizing (1.1) and (1.2) to other prime numbers. Precisely, we show (in different ways) that for any prime number $p$ and any non-multiple integer $a$ of $p$, the sequence $\left\\{\vartheta_{p}\left(\sum_{k=1}^{n}\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}\right)\right\\}_{n\geq 1}$ (1.3) is unbounded from above. Then, by using our general theorem 2.2, we derive an optimal lower bound for the sequence in (1.3). It must be noted that the crucial point of the non-boundness from above of the sequence in (1.3) is established by three methods. The first two are elementary and effective while the third one leans on the $p$-adic analysis and it is ineffective; precisely, it uses the function $L_{p}$ described above. Personally, we consider that the deep reason why the sequence in (1.3) is unbounded from above is rather given by the third method. ## 2 The results and the proofs Our main result is the following: ###### Theorem 2.1. Let $p$ be a prime number and $a$ be an integer not multiple of $p$. Then we have for all positive integer $n$: $\vartheta_{p}\left(\sum_{k=1}^{n}\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}\right)\geq(n+1)-\log_{p}\left(\frac{n+1}{2}\right).$ (2.1) In addition, this inequality becomes an equality if and only if $n$ has the form $n=2p^{\alpha}-1$ ($\alpha\in{\mathbb{N}}_{0}$). Note that Theorem 2.1 generalizes the recent results of the author [2, 3] and Dubickas [1], which are both related to the particular case $p=2$. Especially, if we take $p=2$ and $a=1$ in Theorem 2.1, we exactly obtain (after some obvious simplifications) the main result of [1], stating that: $\vartheta_{2}\left(\sum_{k=1}^{n}\frac{2^{k}}{k}\right)\geq(n+1)-\log_{2}(n+1)~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\forall n\in{\mathbb{N}}),$ with equality if and only if $n$ has the form $(2^{\alpha}-1)$ ($\alpha\in{\mathbb{N}}$). The proof of Theorem 2.1 is based in part on the following result which can serve us well in other situations for bounding from below the $p$-adic valuation of a sum of rational numbers when it is unbounded from above. It must be also noted that the result below is obtained by generalizing the arguments in [1]. ###### Theorem 2.2. Let $p$ be a fixed prime number and ${(r_{n})}_{n\geq 1}$ be a sequence of rational numbers such that the sequence $\left\\{\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right)\right\\}_{n\geq 1}$ is unbounded from above. Let also ${(\ell_{k})}_{k\geq 2}$ be an increasing real sequence satisfying the property: $\ell_{k}\leq\vartheta_{p}(r_{k})~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\forall k\geq 2).$ (2.2) Then we have for any positive integer $n$: $\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right)\geq\min_{k\geq n+1}\vartheta_{p}(r_{k})\geq\ell_{n+1}.$ (2.3) In addition, the inequality $\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right)\geq\ell_{n+1}$ becomes an equality if and only if we have $\min_{k\geq n+1}\vartheta_{p}(r_{k})=\ell_{n+1}.$ (2.4) Our main result (i.e., Theorem 2.1) is proved in two steeps: in the first one, we suppose (in the situation of Theorem 2.1) that the sequence $\left\\{\vartheta_{p}\left(\sum_{k=1}^{n}\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}\right)\right\\}_{n\geq 1}$ is unbounded from above and we apply for it Theorem 2.2 to prove the lower bound (2.1) and to characterize the $n$’s for which it is attained. In the second one, we return to prove the non-boundness from above of the considered sequence. We do this by three different methods: the first one is based on two identities, one is combinatorial and the other is arithmetic. The second one uses a certain functional equation and the Taylor polynomials. The third one uses the $p$-adic analysis; precisely the $p$-adic logarithm function. Let us begin by proving Theorem 2.2. ###### Proof of Theorem 2.2. Let $n$ be a fixed positive integer. Let us show the first inequality of (2.3). Since, by hypothesis, the sequence $\left\\{\vartheta_{p}\left(\sum_{k=1}^{N}r_{k}\right)\right\\}_{N\geq 1}$ is unbounded from above then there exists $m\in{\mathbb{N}}$, with $m>n$, such that: $\vartheta_{p}\left(\sum_{k=1}^{m}r_{k}\right)>\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right).$ Then, by using the elementary properties of the $p$-adic valuation, we have on the one hand: $\vartheta_{p}\left(\sum_{k=n+1}^{m}r_{k}\right)=\vartheta_{p}\left(\sum_{k=1}^{m}r_{k}-\sum_{k=1}^{n}r_{k}\right)=\min\left(\vartheta_{p}\left(\sum_{k=1}^{m}r_{k}\right),\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right)\right)=\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right)$ and on the other hand: $\vartheta_{p}\left(\sum_{k=n+1}^{m}r_{k}\right)\geq\min_{n+1\leq k\leq m}\vartheta_{p}(r_{k})\geq\min_{k\geq n+1}\vartheta_{p}(r_{k}).$ By comparing these two results, we deduce that: $\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right)\geq\min_{k\geq n+1}\vartheta_{p}(r_{k}),$ which is nothing else the first inequality of (2.3). The second inequality of (2.3) is immediately derived from its first one together with the properties of the sequence ${(\ell_{k})}_{k\geq 2}$. Indeed, we have $\displaystyle\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right)$ $\displaystyle\geq\min_{k\geq n+1}\vartheta_{p}(r_{k})~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\text{by the first inequality of \eqref{eq2}})$ $\displaystyle\geq\min_{k\geq n+1}\ell_{k}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\text{by using \eqref{eq1}})$ $\displaystyle=\ell_{n+1}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\text{since }{(\ell_{k})}_{k}\text{ is increasing by hypothesis}),$ confirming the second inequality of (2.3). Now, let us prove the second part of Theorem 2.2. If $\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right)=\ell_{n+1}$ then we have (according to (2.3), proved above): $\min_{k\geq n+1}\vartheta_{p}(r_{k})=\ell_{n+1}$, as required. Conversely, suppose that $\min_{k\geq n+1}\vartheta_{p}(r_{k})=\ell_{n+1}$ and let us show that $\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right)=\ell_{n+1}$. To do so, we first show that: $\vartheta_{p}(r_{n+1})<\vartheta_{p}(r_{k})~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\forall k>n+1).$ (2.5) To prove (2.5), let us argue by contradiction. So, suppose that there is an integer $k_{0}>n+1$ which satisfies $\vartheta_{p}(r_{n+1})\geq\vartheta_{p}(k_{0})$. So we have $\min_{k\geq n+1}\vartheta_{p}(r_{k})=\min_{k\geq n+2}\vartheta_{p}(r_{k})\geq\min_{k\geq n+2}\ell_{k}=\ell_{n+2}>\ell_{n+1}$ (according to (2.2) and the increase of ${(\ell_{k})}_{k}$), contradicting the supposition $\min_{k\geq n+1}\vartheta_{p}(r_{k})=\ell_{n+1}$. This contradiction confirms (2.5). Now, we shall use (2.5) to prove the desired equality $\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right)=\ell_{n+1}$. On the one hand, we have (according to the first part of this proof): $\vartheta_{p}\left(\sum_{k=n+1}^{m}r_{k}\right)=\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right).$ But on the other hand, we have (according to (2.5) and the elementary properties of the $p$-adic valuation): $\vartheta_{p}\left(\sum_{k=n+1}^{m}r_{k}\right)=\min_{n+1\leq k\leq m}\vartheta_{p}(r_{k})=\vartheta_{p}(r_{n+1})=\min_{k\geq n+1}\vartheta_{p}(r_{k})=\ell_{n+1}.$ Comparing the two results, we derive the required equality: $\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right)=\ell_{n+1}$. This confirms the second part of Theorem 2.2 and completes this proof. ∎ Next, we have the following fundamental result: ###### Theorem 2.3. Let $p$ be a prime number and $a$ be an integer not multiple of $p$. Then the sequence $\left\\{\vartheta_{p}\left(\sum_{k=1}^{n}\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}\right)\right\\}_{n\geq 1}$ is unbounded from above. Admitting for the moment Theorem 2.3, our main result is obtained as an application of Theorem 2.2. ###### Proof of Theorem 2.1 by admitting Theorem 2.3. Let us put ourselves in the situation of Theorem 2.1. We apply Theorem 2.2 with $r_{k}:=\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}$ ($\forall k\in{\mathbb{N}}$) and $\ell_{k}:=k-\log_{p}\left(\frac{k}{2}\right)$ ($\forall k\geq 2$). The non- boundness from above of the sequence $\left\\{\vartheta_{p}\left(\sum_{k=1}^{n}r_{k}\right)\right\\}_{n\geq 1}$ is guaranteed by Theorem 2.3 (admitted for the moment). Next, the increase of the sequence ${(\ell_{k})}_{k\geq 2}$ can be derived from the increase of the function $x\mapsto x-\log_{p}\left(\frac{x}{2}\right)$ in the interval $[2,+\infty)$. Finally, we have for any integer $k\geq 2$: $\displaystyle\vartheta_{p}(r_{k})$ $\displaystyle=\vartheta_{p}\left(\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}\right)$ $\displaystyle=\vartheta_{p}\left(\frac{a^{k}+(p-a)^{k}}{\left(a(p-a)\right)^{k}}\cdot\frac{p^{k}}{k}\right)$ $\displaystyle=\vartheta_{p}\left(a^{k}+(p-a)^{k}\right)+k-\vartheta_{p}(k)~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\text{since }a\text{ is coprime with }p).$ If $k$ is even, we use $\vartheta_{p}\left(a^{k}+(p-a)^{k}\right)\geq\vartheta_{p}(2)$ (for $p>2$, this is obvious and for $p=2$, observe that $a^{k}+(p-a)^{k}$ is even). So, we obtain $\vartheta_{p}(r_{k})\geq k-\vartheta_{p}\left(\frac{k}{2}\right)\geq k-\log_{p}\left(\frac{k}{2}\right)=\ell_{k}$ (because $k/2$ is a positive integer if $k$ is even). However, if $k$ is odd, we use $\vartheta_{p}\left(a^{k}+(p-a)^{k}\right)\geq 1$ (since $a^{k}+(p-a)^{k}\equiv a^{k}+(-a)^{k}\ (\mathrm{mod}\ p)\equiv 0\ (\mathrm{mod}\ p)$). So, we obtain again: $\vartheta_{p}(r_{k})\geq 1+k-\vartheta_{p}(k)\geq\log_{p}(2)+k-\log_{p}(k)=k-\log_{p}\left(\frac{k}{2}\right)=\ell_{k}.$ Consequently, we have for any integer $k\geq 2$: $\vartheta_{p}(r_{k})\geq\ell_{k}$. So, all the hypothesis of Theorem 2.2 are satisfied; thus we can apply it for our situation. Applying the first part of Theorem 2.2, we get for any positive integer $n$: $\vartheta_{p}\left(\sum_{k=1}^{n}\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}\right)\geq\min_{k\geq n+1}\vartheta_{p}\left(\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}\right)\geq(n+1)-\log_{p}\left(\frac{n+1}{2}\right),$ confirming Inequality (2.1) of Theorem 2.1. Next, for a given positive integer $n$, the second part of Theorem 2.2 tells us that (2.1) becomes an equality if and only if we have $\min_{k\geq n+1}\vartheta_{p}\left(\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}\right)=(n+1)-\log_{p}\left(\frac{n+1}{2}\right),$ that is $\min_{k\geq n+1}\vartheta_{p}\left(\left(a^{k}+(p-a)^{k}\right)\frac{p^{k}}{k}\right)=(n+1)-\log_{p}\left(\frac{n+1}{2}\right).$ (2.6) So, it remains to prove that (2.6) holds if and only if $n$ has the form $n=2p^{\alpha}-1$ ($\alpha\in{\mathbb{N}}_{0}$). Let us prove this last fact. • Suppose that (2.6) holds. Then, we have $\log_{p}\left(\frac{n+1}{2}\right)=(n+1)-\min_{k\geq n+1}\vartheta_{p}\left(\left(a^{k}+(p-a)^{k}\right)\frac{p^{k}}{k}\right)\in{\mathbb{Z}}.$ But since $\log_{p}\left(\frac{n+1}{2}\right)\geq 0$, we have even $\log_{p}\left(\frac{n+1}{2}\right)\in{\mathbb{N}}_{0}$. By setting $\alpha:=\log_{p}\left(\frac{n+1}{2}\right)\in{\mathbb{N}}_{0}$, we get $n=2p^{\alpha}-1$, as required. • Conversely, suppose that $n=2p^{\alpha}-1$ for some $\alpha\in{\mathbb{N}}_{0}$. We first claim that we have $\vartheta_{p}\left(a^{n+1}+(p-a)^{n+1}\right)=\vartheta_{p}(2).$ (2.7) To confirm (2.7), we distinguish two cases: — 1st case: (If $p=2$). In this case, because $a$ is coprime with $p$ then $a$ and $(p-a)$ are both odd, implying that $a^{2}\equiv 1\ (\mathrm{mod}\ 4)$ and $(p-a)^{2}\equiv 1\ (\mathrm{mod}\ 4)$. Then, because $n+1=2p^{\alpha}$ is even, we have also $a^{n+1}\equiv 1\ (\mathrm{mod}\ 4)$ and $(p-a)^{n+1}\equiv 1\ (\mathrm{mod}\ 4)$; thus $a^{n+1}+(p-a)^{n+1}\equiv 2\ (\mathrm{mod}\ 4)$, implying that $\vartheta_{p}\left(a^{n+1}+(p-a)^{n+1}\right)=1=\vartheta_{p}(2)$. — 2nd case: (If $p$ is odd). In this case, because $n+1=2p^{\alpha}$ is even, we have $a^{n+1}+(p-a)^{n+1}\equiv a^{n+1}+(-a)^{n+1}\ (\mathrm{mod}\ p)\equiv 2a^{n+1}\ (\mathrm{mod}\ p)\not\equiv 0\ (\mathrm{mod}\ p)$ (since $p$ is assumed odd and $a$ is coprime with $p$). Thus $\vartheta_{p}\left(a^{n+1}+(p-a)^{n+1}\right)=0=\vartheta_{p}(2)$. Our claim (2.7) is proved. Now, using (2.7), we have $\vartheta_{p}\left(\left(a^{n+1}+(p-a)^{n+1}\right)\frac{p^{n+1}}{n+1}\right)=\vartheta_{p}(2)+(n+1)-\vartheta_{p}(n+1)=(n+1)-\vartheta_{p}\left(\frac{n+1}{2}\right)\\\ =n+1-\vartheta_{p}\left(p^{\alpha}\right)=n+1-\alpha=n+1-\log_{p}\left(\frac{n+1}{2}\right).$ This shows that (2.6) is equivalent to: $\vartheta_{p}\left(\left(a^{k}+(p-a)^{k}\right)\frac{p^{k}}{k}\right)\geq(n+1)-\log_{p}\left(\frac{n+1}{2}\right)~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\forall k\geq n+2),$ which is weaker than: $k-(n+1)\geq\vartheta_{p}(k)-\log_{p}\left(\frac{n+1}{2}\right)~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\forall k\geq n+2).$ (2.8) Let us prove (2.8) for a given integer $k\geq n+2$. To do so, we distinguish two cases according to whether $\vartheta_{p}(k)\leq\alpha+1$ or not. — 1st case: (If $\vartheta_{p}(k)\leq\alpha+1$). In this case, we have $k-(n+1)\geq 1~{}~{}~{}~{}\text{and}~{}~{}~{}~{}\vartheta_{p}(k)-\log_{p}\left(\frac{n+1}{2}\right)=\vartheta_{p}(k)-\alpha\leq 1.$ Thus (2.8) is true. — 2nd case: (If $\vartheta_{p}(k)\geq\alpha+2$). In this case, we have $k\geq p^{\alpha+2}\geq 2p^{\alpha+1}$. Hence $k-(n+1)=k-2p^{\alpha}\geq\frac{p-1}{p}k\geq\frac{k}{p}\geq p^{\vartheta_{p}(k)-1}\geq 2^{\vartheta_{p}(k)-1}\geq\vartheta_{p}(k)\geq\vartheta_{p}(k)-\log_{p}\left(\frac{n+1}{2}\right),$ confirming (2.8) for this case also. Consequently, (2.8) is valid for any integer $k\geq n+2$. This completes the proof of the second part of Theorem 2.1 and achieves this proof. ∎ The rest of the paper is now devoted to prove Theorem 2.3. We achieve this by three different methods: ### The first method We lean on two identities. The first one (due to Mansour [6]) is combinatorial and states that: $\sum_{k=0}^{n}\frac{x^{k}y^{n-k}}{\binom{n}{k}}=\frac{n+1}{(x+y)\left(\frac{1}{x}+\frac{1}{y}\right)^{n+1}}\sum_{k=1}^{n+1}\frac{\left(x^{k}+y^{k}\right)\left(\frac{1}{x}+\frac{1}{y}\right)^{k}}{k}$ (2.9) (for any $x,y\in{\mathbb{R}}^{*}$, with $x+y\neq 0$, and any $n\in{\mathbb{N}}_{0}$). While the second one (due to the author [4]) is arithmetic and states that: $\mathrm{lcm}\left\\{\binom{n}{0},\binom{n}{1},\dots,\binom{n}{n}\right\\}=\frac{\mathrm{lcm}\left(1,2,\dots,n,n+1\right)}{n+1}$ (2.10) (for any $n\in{\mathbb{N}}_{0}$). Using (2.9) and (2.10), we are now ready to prove Theorem 2.3. Let $p$ be a prime number and $a$ be an integer non-multiple of $p$. By applying (2.9) for $x=a$ and $y=p-a$ and replacing $n$ by $(n-1)$ (where $n\in{\mathbb{N}}$), we get (after simplifying and rearranging) $\sum_{k=1}^{n}\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}=\frac{p^{n+1}}{n\left(a(p-a)\right)^{n}}\sum_{k=0}^{n-1}\frac{a^{k}(p-a)^{n-1-k}}{\binom{n-1}{k}}.$ (2.11) On the other hand, for any $n\in{\mathbb{N}}$, we have (according to (2.10)): $1=\frac{n}{\mathrm{lcm}(1,2,\dots,n)}\mathrm{lcm}\left\\{\binom{n-1}{0},\binom{n-1}{1},\dots,\binom{n-1}{n-1}\right\\}.$ (2.12) Then, for a given $n\in{\mathbb{N}}$, by multiplying side to side (2.11) and (2.12), we obtain $\sum_{k=1}^{n}\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}=\frac{p^{n+1}}{\left(a(p-a)\right)^{n}\mathrm{lcm}(1,2,\dots,n)}\\\ \times\mathrm{lcm}\left\\{\binom{n-1}{0},\binom{n-1}{1},\dots,\binom{n-1}{n-1}\right\\}\sum_{k=0}^{n-1}\frac{a^{k}(p-a)^{n-1-k}}{\binom{n-1}{k}}.$ But since the rational number $\mathrm{lcm}\left\\{\binom{n-1}{0},\binom{n-1}{1},\dots,\binom{n-1}{n-1}\right\\}\sum_{k=0}^{n-1}\frac{a^{k}(p-a)^{n-1-k}}{\binom{n-1}{k}}$ is obviously an integer, we derive from the last identity that: $\displaystyle\vartheta_{p}\left(\sum_{k=1}^{n}\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}\right)$ $\displaystyle\geq\vartheta_{p}\left(\frac{p^{n+1}}{\left(a(p-a)\right)^{n}\mathrm{lcm}(1,2,\dots,n)}\right)$ $\displaystyle=n+1-\vartheta_{p}\left(\mathrm{lcm}(1,2,\dots,n)\right)~{}~{}~{}~{}~{}(\text{since }a\text{ is not a multiple of }p)$ $\displaystyle=n+1-\left\lfloor\log_{p}(n)\right\rfloor,$ confirming the non-boundness from above of the sequence $\left\\{\vartheta_{p}\left(\sum_{k=1}^{n}\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}\right)\right\\}_{n\geq 1}$ (since $n+1-\left\lfloor\log_{p}(n)\right\rfloor\rightarrow+\infty$ as $n\rightarrow+\infty$). $\square$ ### The second method Let $p$ be a prime number and $a$ be an integer non-multiple of $p$. For a given $n\in{\mathbb{N}}$, consider the rational function $R_{n}$ defined by: $R_{n}(X):=\sum_{k=1}^{n}\left(\frac{1}{a^{k}}+\frac{1}{(X-a)^{k}}\right)\frac{X^{k}}{k}=\sum_{k=1}^{n}\dfrac{\left(\frac{X}{a}\right)^{k}}{k}+\sum_{k=1}^{n}\dfrac{\left(\frac{X}{X-a}\right)^{k}}{k}.$ Consider also the real function $f$ defined at the neighborhood of $0$ by: $f(X):=-\log\left(1-X\right),$ which satisfies the functional equation: $f\left(\frac{X}{a}\right)+f\left(\frac{X}{X-a}\right)=0$ (2.13) and whose the $n$th degree Taylor polynomial at $0$ is $\sum_{k=1}^{n}\frac{X^{k}}{k}$. On the one hand, according to the well-known properties of Taylor polynomials, the $n$th degree Taylor polynomial of the function $X\stackrel{{\scriptstyle g}}{{\mapsto}}f\left(\frac{X}{a}\right)+f\left(\frac{X}{X-a}\right)$ at $0$ is the same with the $n$th degree Taylor polynomial of $\sum_{k=1}^{n}\dfrac{\left(\frac{X}{a}\right)^{k}}{k}+\sum_{k=1}^{n}\dfrac{\left(\frac{X}{X-a}\right)^{k}}{k}=R_{n}(X).$ But on the other hand, in view of (2.13), this $n$th degree Taylor polynomial of $g$ at $0$ is zero. Comparing these two results, we deduce that the multiplicity of $0$ in $R_{n}$ is at least $(n+1)$. Consequently, $R_{n}(X)$ can be written as: $R_{n}(X)=X^{n+1}\cdot\frac{U_{n}(X)}{a^{n}(X-a)^{n}\mathrm{lcm}(1,2,\dots,n)},$ where $U_{n}\in{\mathbb{Z}}[X]$. In particular, we have $R_{n}(p)=p^{n+1}\cdot\frac{U_{n}(p)}{a^{n}(p-a)^{n}\mathrm{lcm}(1,2,\dots,n)}.$ Next, because $U_{n}(p)\in{\mathbb{Z}}$ (since $U_{n}\in{\mathbb{Z}}[X]$) and $a$ is not a multiple of $p$, then by taking the $p$-adic valuations in the two sides of the last identity, we derive that: $\vartheta_{p}\left(R_{n}(p)\right)\geq n+1-\vartheta_{p}\left(\mathrm{lcm}(1,2,\dots,n)\right)=n+1-\left\lfloor\log_{p}(n)\right\rfloor,$ implying that the sequence $\left\\{\vartheta_{p}\left(R_{n}(p)\right)\right\\}_{n\geq 1}$ is unbounded from above, as required by Theorem 2.3. $\square$ ###### Remark 2.4. Curiously, the two previous methods give the same upper bound $\vartheta_{p}\left(\sum_{k=1}^{n}\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}\right)\geq n+1-\left\lfloor\log_{p}(n)\right\rfloor.$ Furthermore, this last estimate is remarkably very close to the optimal one of Theorem 2.1. In the third method below, we will show the non-boundness of the sequence in Theorem 2.3 without providing any estimate! ### The third method Let $p$ be a prime number and $a$ be an integer non-multiple of $p$. For all $n\in{\mathbb{N}}$, set $r_{n}:=\left(\frac{1}{a^{n}}+\frac{1}{(p-a)^{n}}\right)\frac{p^{n}}{n}~{}~{}\text{and}~{}~{}s_{n}:=\sum_{k=1}^{n}r_{k}.$ The property we have to show is that the sequence ${\left(\vartheta_{p}(s_{n})\right)}_{n\geq 1}$ is unbounded from above; in other words, we have that $\limsup_{n\rightarrow+\infty}\vartheta_{p}(s_{n})=+\infty$. So, if we show the stronger property $\lim_{n\rightarrow+\infty}\vartheta_{p}(s_{n})=+\infty$, then we are done. To do so, observe that: $\displaystyle\lim_{n\rightarrow+\infty}\vartheta_{p}(s_{n})=+\infty$ $\displaystyle\Longleftrightarrow\lim_{n\rightarrow+\infty}{\left|s_{n}\right|}_{p}=0$ $\displaystyle\Longleftrightarrow\lim_{n\rightarrow+\infty}s_{n}=0~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\text{in the }p\text{-adic sense})$ $\displaystyle\Longleftrightarrow\sum_{k=1}^{+\infty}r_{k}=0~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}~{}(\text{in the }p\text{-adic sense}).$ Consequently, it suffices to show that: $\sum_{k=1}^{+\infty}\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}=0$ (2.14) (in the $p$-adic sense). Let us show (2.14). By using the $p$-adic logarithm function (recalled in §1), we have $\displaystyle\sum_{k=1}^{+\infty}\left(\frac{1}{a^{k}}+\frac{1}{(p-a)^{k}}\right)\frac{p^{k}}{k}$ $\displaystyle=\sum_{k=1}^{+\infty}\frac{\left(\frac{p}{a}\right)^{k}}{k}+\sum_{k=1}^{+\infty}\frac{\left(\frac{p}{p-a}\right)^{k}}{k}$ $\displaystyle=-L_{p}\left(1-\frac{p}{a}\right)-L_{p}\left(1-\frac{p}{p-a}\right)$ $\displaystyle=-\left[L_{p}\left(\frac{a-p}{a}\right)+L_{p}\left(\frac{-a}{p-a}\right)\right]$ $\displaystyle=-L_{p}\left(\frac{a-p}{a}\cdot\frac{-a}{p-a}\right)$ $\displaystyle=-L_{p}(1)=0,$ as required. The non-boundness from above of the sequence ${\left(\vartheta_{p}(s_{n})\right)}_{n\geq 1}$ follows. $\square$ ## References * [1] A. Dubickas. On a sequence of integers with unusual divisibility by a power of $2$, to appear in Miskolc Math. Notes. * [2] B. Farhi. On a curious integer sequence, preprint 2022, available at https://arxiv.org/abs/2204.10136. * [3] B. Farhi. The integrality of the Genocchi numbers obtained through a new identity and other results, Notes Number Theory Discrete Math., 28, n°4 (2022), p. 749-757. * [4] B. Farhi. An identity involving the least common multiple of binomial coefficients and its application, Amer. Math. Monthly, 116 (2009), p. 836-839. * [5] N. Koblitz. $p$-adic numbers, $p$-adic analysis, and zeta-functions, 2nd edition, Springer-Verlag, New York, 1984. * [6] T. Mansour. Combinatorial identities and inverse binomial coefficients, Adv. Appl. Math, 28 (2002), p. 196-202.
# Density Functional Theory for two-dimensional homogeneous materials with magnetic fields David Gontier CEREMADE, University of Paris-Dauphine, PSL University, 75016 Paris, France & ENS/PSL University, Département de Mathématiques et Applications, F-75005, Paris, France<EMAIL_ADDRESS>, Salma Lahbabi EMAMI, LRI, ENSEM, UHIIC, 7 Route d’El Jadida, B.P. 8118 Oasis, Casablanca; MSDA, Mohammed VI Polytechnic University, Lot 660, Hay Moulay Rachid Ben Guerir, 43150, Morocco<EMAIL_ADDRESS>and Abdallah Maichine Department of Mathematics, Faculty of Sciences, Mohammed V University in Rabat, 4 Avenue Ibn Battouta P.B. 1014 RP, Rabat, Morocco & Ecole Centrale Casablanca, Bouskoura, Ville Verte, P.B. 27182, Morocco <EMAIL_ADDRESS> ###### Abstract. This paper studies DFT models for homogeneous 2D materials in 3D space, under a constant perpendicular magnetic field. We show how to reduce the three–dimensional energy functional to a one–dimensional one, similarly as in our previous work. This is done by minimizing over states invariant under magnetic translations and that commute with the Landau operator. In the reduced model, the Pauli principle no longer appears. It is replaced by a penalization term in the energy. ###### Contents 1. 1 Introduction 2. 2 States commuting with magnetic translations 1. 2.1 Two dimensional Landau operator 2. 2.2 Magnetic translations 3. 2.3 Diagonalisation of states commuting with magnetic translations 3. 3 Reduction of the kinetic energy, and applications 1. 3.1 Reduction of the kinetic energy 2. 3.2 Some properties of the function F 3. 3.3 Reduced DFT models 4. 3.4 The reduced Hartree–Fock case 4. 4 Models with spin ## 1\. Introduction The analysis of quantum properties of two dimensional materials is an active research area in physics and material science. Some 2D materials such as graphene or phosphorene exhibits many interesting physical properties [16, 3, 10, 11] which has many applications such as High Electron Mobility Transistors [12]. Some of these properties are not yet fully understood. This is the main motivation to revisit Density Functional Theory (DFT) when applied to quantum two dimensional systems (see [1, 9] for previous works). As in our previous work [9], we study homogeneous two–dimensional slabs, when embedded in three dimensional space, but this time, we include a constant perpendicular magnetic field. We consider a charge distribution $\mu$ which is equidistributed in the first two dimensions: $\mu(x_{1},x_{2},x_{3})=\mu(x_{3})$, and with a constant perpendicular magnetic field ${\mathbb{B}}=b{\mathbb{e}}_{3}$. One key result of our previous work was an inequality for the kinetic energy per unit surface for translationally invariant states. Let us quickly summarize the result. Let ${\mathcal{P}}:=\left\\{\gamma\in\underline{\mathfrak{S}}^{1}(L^{2}(\mathbb{R}^{3})),\quad 0\leq\gamma\leq 1,\quad\forall{\mathbb{R}}\in\mathbb{R}^{2},\quad\tau_{\mathbb{R}}\gamma=\gamma\tau_{\mathbb{R}}\right\\}$ denote the set of one-body density matrices which commute with all $\mathbb{R}^{2}$ translations. Here, $\underline{\mathfrak{S}}^{1}(L^{2}(\mathbb{R}^{3}))$ stands for locally trace class self–adjoint operators with finite trace per unit surface $\underline{\rm Tr}(\gamma)<{\infty}$ (see Section 2.3). For ${\mathbb{R}}\in\mathbb{R}^{2}\subset\mathbb{R}^{3}$, we have denoted by $\tau_{\mathbb{R}}f(x_{1},x_{2},x_{3}):=f(x_{1}-R_{1},x_{2}-R_{2},x_{3})$ the usual translation along the first two dimensions. Let us also introduce the set of reduced states ${\mathcal{G}}:=\left\\{G\in\mathfrak{S}^{1}(L^{2}(\mathbb{R})),\quad G\geq 0\right\\},$ where $\mathfrak{S}^{1}(L^{2}(\mathbb{R}))$ refers to the space of trace class self-adjoint operators on $L^{2}(\mathbb{R})$. We have proved in [9] that, for any (representable) density $\rho=\rho(x_{3})$ depending only on the third variable, we have (1.1) $\inf_{\gamma\in{\mathcal{P}}\atop\rho_{\gamma}=\rho}\left\\{\frac{1}{2}\underline{\rm Tr}(-\Delta_{3}\gamma)\right\\}=\inf_{G\in{\mathcal{G}}\atop\rho_{G}=\rho}\left\\{\frac{1}{2}{\rm Tr}(-\Delta_{1}G)+\pi{\rm Tr}(G^{2})\right\\},$ where $\Delta_{d}$ denotes the Laplacian operator in $d$–space dimension. The energy appearing in the right hand side leads to one–dimensional reduced models for homogeneous semi-infinite slabs in the context of DFT. One typically obtains a minimization problem of the form (1.2) $\inf\left\\{\frac{1}{2}{\rm Tr}(-\Delta_{1}G)+\pi{\rm Tr}(G^{2})+\frac{1}{2}{\mathcal{D}}_{1}(\rho_{G}-\mu)+E^{\rm xc}(\rho_{G}),\quad G\in{\mathcal{G}}\right\\},$ where ${\mathcal{D}}_{1}(\cdot)$ is the one–dimensional Coulomb interaction energy (see [9, Section 3.1] for a discussion about this term), and $E^{\rm xc}$ some exchange-correlation energy per unit surface. Note that there is no Pauli principle for the operator $G$; it has been replaced by the penalization term $\pi{\rm Tr}(G^{2})$ in the energy, which prevents $G$ from having large eigenvalues. We refer to [9] for details, where we also studied the reduced Hartree–Fock model, which corresponds to the case $E^{\rm xc}=0$. The scope of this paper is to apply a similar reduction when taking into account magnetic effects. Without considering the spin (we refer to Section 4 for the case with spin), the kinetic energy per unit surface is of the form $\frac{1}{2}\underline{\rm Tr}\left((-{\mathrm{i}}\nabla_{3}+{\mathbb{A}})^{2}\gamma\right),$ where ${\mathbb{A}}=b(0,x_{1},0)$ is a vector potential so that ${\bf curl}\,{\mathbb{A}}={\mathbb{B}}=b{\mathbb{e}}_{3}$. We chose a gauge which is not symmetric, but which will simplify some computations. The Laplacian operator $-\Delta_{3}$ has been replaced by the Landau operator ${\mathbb{L}}^{\mathbb{A}}_{3}:=(-{\mathrm{i}}\nabla_{3}+{\mathbb{A}})^{2}={\mathbb{L}}^{\mathbb{A}}_{2}-\partial_{x_{3}x_{3}}^{2},\quad\text{with}\quad{\mathbb{L}}^{\mathbb{A}}_{2}=-\partial_{x_{1}x_{1}}^{2}+(-{\mathrm{i}}\partial_{x_{2}}+bx_{1})^{2}.$ In analogy with [9], we only consider states commuting with the so–called magnetic translations ${\mathfrak{m}}_{\mathbb{R}}$ and with the Landau operator. We refer to Section 2 for the definition of these operators, and for the justification of this choice. We denote the set of such states by (1.3) ${\mathcal{P}}^{\mathbb{A}}:=\left\\{\gamma\in\underline{\mathfrak{S}}^{1}(L^{2}(\mathbb{R}^{3})),\ 0\leq\gamma\leq 1,\ \forall{\mathbb{R}}\in\mathbb{R}^{2},\ {\mathfrak{m}}_{\mathbb{R}}\gamma=\gamma{\mathfrak{m}}_{\mathbb{R}},\;{\mathbb{L}}^{\mathbb{A}}_{2}\gamma=\gamma{\mathbb{L}}^{\mathbb{A}}_{2}\right\\}.$ In Theorem 2.7, we prove that any $\gamma\in{\mathcal{P}}^{\mathbb{A}}$ has a simple decomposition, in terms of the different projectors on the Landau levels. Using this decomposition, we prove in Theorem 3.1 that, in the magnetic case, we have an equality similar to (1.1), which reads (1.4) $\boxed{\inf_{\gamma\in{\mathcal{P}}^{\mathbb{A}}\atop\rho_{\gamma}=\rho}\left\\{\frac{1}{2}\underline{\rm Tr}({\mathbb{L}}^{\mathbb{A}}_{3}\gamma)\right\\}=\inf_{G\in{\mathcal{G}}\atop\rho_{G}=\rho}\left\\{\frac{1}{2}{\rm Tr}(-\Delta_{1}G)+{\rm Tr}\left(F(b,G)\right)\right\\}}$ with the penalization term $F$ defined by $F(b,g):=\pi g^{2}+\frac{b^{2}}{4\pi}\left\\{\frac{2\pi g}{b}\right\\}\left(1-\left\\{\frac{2\pi g}{b}\right\\}\right),$ where $\\{x\\}$ denotes the fractional part of $x$. The function $F$ is studied in Proposition 3.2. It is a piece-wise linear function, reflecting the contribution of the different Landau levels. The function $F$ is not new, and already appears in the context of the two-dimensional Thomas Fermi (TF) theory under constant magnetic fields (see [14] and related references). The (spinless) TF kinetic energy takes the form $E_{\rm kin}^{\rm TF}(b,\rho):=\int_{\mathbb{R}^{2}}F(b,\rho({\mathbf{x}})){\mathrm{d}}{\mathbf{x}}.$ For $b=0$, we recover the usual two–dimensional TF kinetic energy $\int\pi\rho^{2}$. It is different from the three–dimensional TF kinetic energy of a gas under a constant magnetic field, which has been derived and studied in [6, 19], and is obtained by assuming that the electron density is constant, hence also invariant under the third–direction translation. Equation (1.4) allows the reduction of DFT models for two-dimensional homogeneous slabs under constant magnetic field. In fact, one obtains a one- dimensional problem of the form (compare with (1.2)) $\inf\left\\{\frac{1}{2}{\rm Tr}(-\Delta_{1}G)+{\rm Tr}\left(F(b,G)\right)+\frac{1}{2}{\mathcal{D}}_{1}(\rho_{G}-\mu)+E^{\rm xc}(b,\rho_{G}),\quad G\in{\mathcal{G}}\right\\}.$ Note that the exchange-correlation function may depend on the external magnetic field $b$ (see [6, Eqn. (4.1)-(4.2)] for the expression of the exchange energy for the Landau gas). In Section 3.3, we study the corresponding reduced Hartree-Fock model, where $E^{\rm xc}=0$. This article is structured as follows. In Section 2, we start by introducing the Landau operator and studying its spectral decomposition, then we define magnetic translations $\\{{\mathfrak{m}}_{\mathbb{R}}\\}_{{\mathbb{R}}\in\mathbb{R}^{2}}$ with some of their properties, and we characterize the states in ${\mathcal{P}}^{\mathbb{A}}$. Using this characterization, we explain in section 3 how to reduce the kinetic energy, we give some properties of the penalization term $F$ and we study the corresponding reduced Hartree-Fock model. Finally, we show in Section 4 how to extend our results to systems with spin. ### Acknowledgments The research leading to these results has received funding from OCP grant AS70 “Towards phosphorene based materials and devices”. ## 2\. States commuting with magnetic translations In this section, we prove that states $\gamma\in{\mathcal{P}}^{\mathbb{A}}$ have a particular structure. ### 2.1. Two dimensional Landau operator We start by recalling some classical facts about the (two–dimensional) Landau operator ${\mathbb{L}}^{\mathbb{A}}_{2}=-\partial_{x_{1}x_{1}}^{2}+(-{\mathrm{i}}\partial_{x_{2}}+bx_{1})^{2}$. In what follows, we assume $b\neq 0$. For $n\in\mathbb{N}_{0}=\\{0,1,2,\cdots\\}$, we introduce the function $\varphi_{n}:\mathbb{R}\to\mathbb{R}$ defined by (2.1) $\varphi_{n}(x):=a_{n}|b|^{1/4}{\mathcal{H}}_{n}(\sqrt{|b|}x){\mathrm{e}}^{-\tfrac{1}{2}|b|x^{2}},$ where ${\mathcal{H}}_{n}(x)=(-1)^{n}{\mathrm{e}}^{x^{2}}\frac{{\mathrm{d}}^{n}}{{\mathrm{d}}x^{n}}{\mathrm{e}}^{-x^{2}}$ refers to the $n$-th Hermite polynomial and $a_{n}=(2^{n}n!)^{-1/2}/\pi^{1/4}$ is a normalization constant so that $\|\varphi_{n}\|_{L^{2}(\mathbb{R})}=1$. ###### Proposition 2.1. The operator ${\mathbb{L}}^{A}_{2}$ has purely discrete spectrum (2.2) $\sigma({\mathbb{L}}^{A}_{2})=b(2\mathbb{N}_{0}+1).$ The eigenvalue $\varepsilon_{n}:=b(2n+1)$ is of infinite multiplicity, with eigenspace (2.3) $E_{n}:=\ker({\mathbb{L}}^{\mathbb{A}}_{2}-\varepsilon_{n})=\\{{\mathcal{W}}(\varphi_{n},g),\;g\in L^{2}(\mathbb{R})\\},$ where ${\mathcal{W}}$ is a Wigner type transform defined on $L^{2}(\mathbb{R})\times L^{2}(\mathbb{R})$ by (2.4) ${\mathcal{W}}(\varphi,g)({\mathbf{x}}):=\frac{1}{\sqrt{2\pi}}\int_{\mathbb{R}}{\mathrm{e}}^{-{\mathrm{i}}kx_{2}}\varphi\left(x_{1}-\frac{k}{b}\right)g(k){\mathrm{d}}k.$ In particular, the spectral projector $\mathbf{P}_{n}$ onto $\ker({\mathbb{L}}^{\mathbb{A}}_{2}-\varepsilon_{n})$ has kernel (2.5) $\mathbf{P}_{n}({\mathbf{x}};{\mathbf{y}})=\frac{1}{2\pi}\int_{\mathbb{R}}{\mathrm{e}}^{-{\mathrm{i}}k(x_{2}-y_{2})}\varphi_{n}\left(x_{1}-\frac{k}{b}\right)\varphi_{n}\left(y_{1}-\frac{k}{b}\right){\mathrm{d}}k.$ Its density $\rho_{\mathbf{P}_{n}}({\mathbf{x}}):=\mathbf{P}_{n}({\mathbf{x}};{\mathbf{x}})=\frac{b}{2\pi}$ is constant and independent of $n$. ###### Remark 2.2. The definition (2.4) is slightly different from the classical Wigner transform (see for example [18, Chapter 2]) which is rather adapted to study Landau operator with the gauge $\tilde{{\mathbb{A}}}=\frac{b}{2}\begin{pmatrix}-x_{2}\\\ x_{1}\end{pmatrix}$, for $b=1$. A gauge transformation links the two transforms. ###### Proof. First, we remark that ${\mathbb{L}}^{\mathbb{A}}_{2}$ commutes with all translations in the $x_{2}$–direction. We introduce the Fourier transform with respect to the $x_{2}$–variable (2.6) ${\mathcal{F}}[f](x_{1},k):=\frac{1}{\sqrt{2\pi}}\int_{\mathbb{R}}f(x_{1},x_{2}){\mathrm{e}}^{{\mathrm{i}}kx_{2}}{\mathrm{d}}x_{2}$ and its inverse ${\mathcal{F}}^{-1}[\phi](x_{1},x_{2}):=\frac{1}{\sqrt{2\pi}}\int_{\mathbb{R}}\phi(x_{1},k){\mathrm{e}}^{-{\mathrm{i}}kx_{2}}{\mathrm{d}}k.$ We have (2.7) ${\mathcal{F}}{\mathbb{L}}^{A}_{2}{\mathcal{F}}^{-1}=\int_{\mathbb{R}}^{\oplus}{\mathfrak{h}}_{b,k}{\mathrm{d}}k,\quad\text{with}\quad{\mathfrak{h}}_{b,k}:=-\partial_{xx}^{2}+(bx-k)^{2}.$ The operator ${\mathfrak{h}}_{b,k}$ is a translation of the harmonic oscillator ${\mathfrak{h}}:=-\partial_{xx}^{2}+b^{2}x^{2}$, whose spectral decomposition is ${\mathfrak{h}}=\sum_{n=0}^{\infty}\varepsilon_{n}|\varphi_{n}\rangle\langle\varphi_{n}|,$ with $\varepsilon_{n}=b(2n+1)$ and $\varphi_{n}$ as defined in (2.1). We deduce that the spectral decomposition of ${\mathfrak{h}}_{b,k}$ is ${\mathfrak{h}}_{b,k}:=\sum_{n=0}^{\infty}\varepsilon_{n}|\varphi_{n}(\cdot-\tfrac{k}{b})\rangle\langle\varphi_{n}(\cdot-\tfrac{k}{b})|,$ which proves (2.2). Using (2.7), we see that $W$ is an eigenfunction of ${\mathbb{L}}^{{\mathbb{A}}}_{2}$, corresponding to eigenvalue $\varepsilon_{n}$, if and only if it is of the form $W={\mathcal{F}}^{-1}\left[(x_{1},k)\mapsto g(k)\varphi_{n}\left(x_{1}-\frac{k}{b}\right)\right]={\mathcal{W}}(\varphi_{n},g),$ where $g\in L^{2}(\mathbb{R})$. Thus $E_{n}:=\ker({\mathbb{L}}^{\mathbb{A}}_{2}-\varepsilon_{n})=\left\\{{\mathcal{W}}(\varphi_{n},g),\;g\in L^{2}(\mathbb{R})\right\\}.$ To compute the kernel $\mathbf{P}_{n}({\mathbf{x}},{\mathbf{y}})$ of the projector on $E_{n}$, we use the Moyal identity, that we recall here. ###### Proposition 2.3. Let $f_{1},g_{1},f_{2},g_{2}$ in $L^{2}(\mathbb{R})$. Then, ${\mathcal{W}}(f_{1},g_{1}),{\mathcal{W}}(f_{2},g_{2})\in L^{2}(\mathbb{R}^{2})$ and (Moyal identity) (2.8) $\langle{\mathcal{W}}(f_{1},g_{1}),{\mathcal{W}}(f_{2},g_{2})\rangle_{L^{2}(\mathbb{R}^{2})}=\langle f_{1},f_{2}\rangle_{L^{2}(\mathbb{R})}\,\langle g_{1},g_{2}\rangle_{L^{2}(\mathbb{R})}.$ ###### Proof. We first prove the result for $f_{1},g_{1},f_{2},g_{2}\in C^{\infty}_{0}(\mathbb{R})$ and conclude by density. Parseval identity gives $\displaystyle\langle{\mathcal{W}}\left(f_{1},g_{1}\right),{\mathcal{W}}\left(f_{2},g_{2}\right)\rangle_{L^{2}(\mathbb{R}^{2})}$ $\displaystyle=\iint_{\mathbb{R}^{2}}(\overline{f_{1}}f_{2})\left(x_{1}-\frac{k}{b}\right)(\overline{g_{1}}g_{2})(k){\mathrm{d}}x_{1}{\mathrm{d}}k$ $\displaystyle=\langle f_{1},f_{2}\rangle_{L^{2}(\mathbb{R})}\,\langle g_{1},g_{2}\rangle_{L^{2}(\mathbb{R})}.$ ∎ In particular, if $(\psi_{m})_{m\in\mathbb{N}}$ is any basis of $L^{2}(\mathbb{R})$, then $\\{{\mathcal{W}}(\varphi_{n},\psi_{m})\\}_{m\in\mathbb{N}}$ is a basis of $E_{n}$. Notice that, for any fixed ${\mathbf{x}}\in\mathbb{R}^{2}$, we have ${\mathcal{W}}(\varphi,g)({\mathbf{x}})=\langle\varphi_{{\mathbf{x}}},g\rangle_{L^{2}(\mathbb{R})},\quad\text{with}\quad\varphi_{\mathbf{x}}(k):=\frac{1}{\sqrt{2\pi}}{\mathrm{e}}^{{\mathrm{i}}kx_{2}}\overline{\varphi}\left(x_{1}-\frac{k}{b}\right).$ Hence $\displaystyle\mathbf{P}_{n}({\mathbf{x}};{\mathbf{y}})$ $\displaystyle=\sum_{m\in\mathbb{N}}{\mathcal{W}}(\varphi_{n},\psi_{m})({\mathbf{x}})\overline{{\mathcal{W}}(\varphi_{n},\psi_{m})}({\mathbf{y}})=\sum_{m\in\mathbb{N}}\langle\varphi_{n,{\mathbf{x}}},\psi_{m}\rangle\langle\psi_{m},\varphi_{n,{\mathbf{y}}}\rangle.$ Using that $\sum_{m}|\psi_{m}\rangle\langle\psi_{m}|=\mathbb{I}_{L^{2}}$, we obtain $\mathbf{P}_{n}({\mathbf{x}};{\mathbf{y}})=\langle\varphi_{n,{\mathbf{x}}},\varphi_{n,{\mathbf{y}}}\rangle$, which, given that $\varphi_{n}$ is real-valued, gives (2.5). The density of $\mathbf{P}_{n}$ is thus $\rho_{\mathbf{P}_{n}}({\mathbf{x}})=\mathbf{P}_{n}({\mathbf{x}};{\mathbf{x}})=\frac{1}{2\pi}\int_{\mathbb{R}}\left|\varphi_{n}\left(x_{1}-\frac{k}{b}\right)\right|^{2}{\mathrm{d}}k=\frac{b}{2\pi}.$ ∎ ### 2.2. Magnetic translations The Landau operator does not commute with the usual translations, however it commutes with the magnetic translations, that we define now. We write ${\mathbb{L}}_{2}^{\mathbb{A}}:=p_{{\mathbb{A}},1}^{2}+p_{{\mathbb{A}},2}^{2},\quad\text{with}\quad p_{{\mathbb{A}},1}:=-{\mathrm{i}}\partial_{x_{1}},\quad p_{{\mathbb{A}},2}:=-{\mathrm{i}}\partial_{x_{2}}+bx_{1}.$ The operators $p_{{\mathbb{A}},1}$ and $p_{{\mathbb{A}},2}$ do not commute, and do not commute with $L_{2}^{\mathbb{A}}$. Actually, we have $\left[p_{{\mathbb{A}},1},p_{{\mathbb{A}},2}\right]=-{\mathrm{i}}b,\quad\left[p_{{\mathbb{A}},1},L_{2}^{\mathbb{A}}\right]=-2{\mathrm{i}}bp_{{\mathbb{A}},2},\quad\left[p_{{\mathbb{A}},2},L_{2}^{\mathbb{A}}\right]=2{\mathrm{i}}bp_{{\mathbb{A}},1}.$ However, introducing the dual momentum operators $\widetilde{p}_{{\mathbb{A}},1}:=-{\mathrm{i}}\partial_{x_{1}}+bx_{2},\quad\widetilde{p}_{{\mathbb{A}},2}:=-{\mathrm{i}}\partial_{x_{2}},$ we can check that $\left[\widetilde{p}_{{\mathbb{A}},1},L_{2}^{\mathbb{A}}\right]=\left[\widetilde{p}_{{\mathbb{A}},1},L_{2}^{\mathbb{A}}\right]=0$. The magnetic translation ${\mathfrak{m}}_{\mathbb{R}}$, ${\mathbb{R}}\in\mathbb{R}^{2}$, is the unitary operator $\displaystyle{\mathfrak{m}}_{\mathbb{R}}$ $\displaystyle=\exp(-\tfrac{{\mathrm{i}}}{2}bR_{1}R_{2})\exp\left(-{\mathrm{i}}\widetilde{{\mathbb{p}}}_{\mathbb{A}}\cdot{\mathbb{R}}\right)$ $\displaystyle=\exp(-\tfrac{{\mathrm{i}}}{2}bR_{1}R_{2})\exp\left(-{\mathrm{i}}\left(\widetilde{p}_{{\mathbb{A}},1}R_{1}+\widetilde{p}_{{\mathbb{A}},2}R_{2}\right)\right).$ Note that we have added a phase factor in order to match the usual convention. Using the Baker–Campbell–Haussdorf formula and the fact that $\left[\widetilde{p}_{{\mathbb{A}},1},\widetilde{p}_{{\mathbb{A}},2}\right]={\mathrm{i}}b$ commutes with all operators, we obtain the explicit expression ${\mathfrak{m}}_{\mathbb{R}}=\exp(-{\mathrm{i}}bR_{1}x_{2})\tau_{\mathbb{R}},\quad\text{that is}\quad\left({\mathfrak{m}}_{\mathbb{R}}f\right)({\mathbf{x}})=\exp(-{\mathrm{i}}bR_{1}x_{2})f({\mathbf{x}}-{\mathbb{R}}),$ where $\tau_{\mathbb{R}}f({\mathbf{x}}):=f({\mathbf{x}}-{\mathbb{R}})$ is the usual translation operator. By construction, the magnetic translations commute with $L^{\mathbb{A}}_{2}$ and $\mathbf{P}_{n}$, but they do not commute among them. Actually, we have ${\mathfrak{m}}_{\mathbb{R}}{\mathfrak{m}}_{\tilde{{\mathbb{R}}}}={\mathrm{e}}^{{\mathrm{i}}bR_{2}\tilde{R}_{1}}{\mathfrak{m}}_{{\mathbb{R}}+\tilde{{\mathbb{R}}}}\quad\text{and}\quad{\mathfrak{m}}_{\mathbb{R}}^{*}={\mathfrak{m}}_{\mathbb{R}}^{-1}={\mathrm{e}}^{{\mathrm{i}}bR_{1}R_{2}}{\mathfrak{m}}_{-{\mathbb{R}}}.$ An important feature of magnetic translations is that they form an irreducible family on each eigenspace $E_{n}$, in the sense of [2, Definition 2.3.7]. ###### Proposition 2.4. The set of magnetic translation operators $({\mathfrak{m}}_{\mathbb{R}})_{\mathbb{R}}$ is an irreducible family of operators on each $E_{n}$, in the sense that (2.9) $\forall\Psi\in E_{n}\setminus\left\\{0\right\\},\quad E_{n}={\mathrm{span}}\\{{\mathfrak{m}}_{\mathbb{R}}\Psi:{\mathbb{R}}\in\mathbb{R}^{2}\\}.$ ###### Proof. Assume otherwise, and let $\Psi$ so that $\widetilde{E}_{n}(\Psi):={\mathrm{span}}\\{{\mathfrak{m}}_{\mathbb{R}}\Psi:{\mathbb{R}}\in\mathbb{R}^{2}\\}\subsetneq E_{n}.$ Then there is $\Phi\in E_{n}\setminus\\{0\\}$ so that $\Phi\perp\widetilde{E}_{n}(\Psi)$. Let $f,g\in L^{2}(\mathbb{R})$ so that $\Psi={\mathcal{W}}(\varphi_{n},g)$ and $\Phi={\mathcal{W}}(\varphi_{n},f)$. Using the Moyal identity, and the fact that ${\mathfrak{m}}_{\mathbb{R}}{\mathcal{W}}(\varphi_{n},g)={\mathcal{W}}(\varphi_{n},{\mathfrak{t}}_{\mathbb{R}}g),\quad\text{with}\quad{\mathfrak{t}}_{\mathbb{R}}g:k\mapsto{\mathrm{e}}^{-{\mathrm{i}}bR_{1}R_{2}}{\mathrm{e}}^{{\mathrm{i}}kR_{2}}g(k-bR_{1}),$ the condition $\langle\Phi,{\mathbb{m}}_{\mathbb{R}}\Psi\rangle=0$ for all ${\mathbb{R}}\in\mathbb{R}^{2}$ reads $\forall{\mathbb{R}}\in\mathbb{R}^{2},\quad\langle f,{\mathfrak{t}}_{\mathbb{R}}g\rangle=0,\quad\text{hence}\quad\int_{\mathbb{R}}\overline{f}(k){\mathrm{e}}^{{\mathrm{i}}kR_{2}}g(k-bR_{1}){\mathrm{d}}k=0.$ Applying the inverse Fourier transform to $k\mapsto\overline{f}(k)g(k-R_{1})=0$ shows that $\overline{f}(k)g(k-R_{1})=0$ a.e. for all $R_{1}\in\mathbb{R}$. Squaring and integrating in $R_{1}$ gives $f=0$, a contradiction. ∎ ### 2.3. Diagonalisation of states commuting with magnetic translations In what follows, we are interested in one-body density matrices which commute with all magnetic translations. In the case without magnetic field, if a state commutes with all usual translations $\tau_{\mathbb{R}}$, then it commutes with the Laplacian operator. In the magnetic case, there are operators which commute with all magnetic translations $({\mathfrak{m}}_{\mathbb{R}})_{{\mathbb{R}}\in\mathbb{R}^{2}}$, but which do not commute with the Landau operator (we give an example of such an operator in Remark 2.6 below). So, we rather consider one-body density matrices which commute with ${\mathfrak{m}}_{\mathbb{R}}$, and with the Landau operator. It turns out that such operators have an explicit and simple characterization. We first enunciate our result in dimension two before turning to the three dimensional case. ###### Proposition 2.5. Let $\eta\in{\mathcal{S}}(L^{2}(\mathbb{R}^{2}))$ be such that $\eta{\mathfrak{m}}_{\mathbb{R}}={\mathfrak{m}}_{\mathbb{R}}\eta$ for all ${\mathbb{R}}\in\mathbb{R}^{2}$ and $\eta{\mathbb{L}}_{\mathbb{A}}={\mathbb{L}}_{\mathbb{A}}\eta$. Then, there is a family of real numbers $(\lambda_{n})_{n\in\mathbb{N}_{0}}$ so that (2.10) $\eta=\sum_{n\in\mathbb{N}_{0}}\lambda_{n}\mathbf{P}_{n}.$ If $\eta$ is a locally trace class operator, then its density is constant $\rho_{\eta}=\frac{b}{2\pi}\sum_{n\in\mathbb{N}_{0}}\lambda_{n}.$ ###### Proof. Since $\eta$ commutes with ${\mathbb{L}}^{\mathbb{A}}_{2}$, it commutes with any spectral projector $\mathbf{P}_{n}$, hence leaves invariant $E_{n}={\rm Ran}(\mathbf{P}_{n})$, for all $n\in\mathbb{N}$. The operator $\eta_{n}:=\mathbf{P}_{n}\eta\mathbf{P}_{n}\in{\mathcal{S}}(E_{n})$ commutes with all magnetic translations. Since the family $\\{{\mathbb{m}}_{\mathbb{R}}\\}_{{\mathbb{R}}}$ is irreducible, it implies that $\eta_{n}$ is proportional to the identity on $E_{n}$, hence is of the form $\eta_{n}=\lambda_{n}\mathbb{I}_{E_{n}}$. This is a kind of Schur’s Lemma, see [2, Proposition 2.3.8]. ∎ ###### Remark 2.6. The hypothesis $\eta{\mathbb{L}}^{\mathbb{A}}_{2}={\mathbb{L}}^{\mathbb{A}}_{2}\eta$ is not a consequence of the commutation with the operators $({\mathfrak{m}}_{\mathbb{R}})_{{\mathbb{R}}}$. Indeed, consider for a normalized $\zeta\in L^{2}(\mathbb{R})$, the projector $P_{\zeta}$ onto the vectorial space $E_{\zeta}:=\left\\{{\mathcal{W}}(\zeta,f),\ f\in L^{2}(\mathbb{R})\right\\}.$ Since ${\mathfrak{m}}_{\mathbb{R}}{\mathcal{W}}(\zeta,f)={\mathcal{W}}(\zeta,{\mathfrak{t}}_{\mathbb{R}}f)\in E_{\zeta}$, the set $E_{\zeta}$ is invariant by ${\mathfrak{m}}_{\mathbb{R}}$, hence $P_{\zeta}$ commutes with all magnetic translations. However, we have, using the decomposition (2.7) that $\displaystyle{\mathbb{L}}^{\mathbb{A}}_{2}\left[{\mathcal{W}}(\zeta,f)\right]$ $\displaystyle=\frac{1}{\sqrt{2\pi}}\int_{\mathbb{R}}{\mathrm{e}}^{-{\mathrm{i}}kx_{2}}\left[(-\partial_{x_{1}x_{1}}^{2}+\left(bx_{1}-k\right)^{2})\zeta\left(x_{1}-\frac{k}{b}\right)\right]f(k){\mathrm{d}}k$ $\displaystyle={\mathcal{W}}(\widetilde{\zeta},f)$ with $\widetilde{\zeta}:=\left(-\partial_{xx}^{2}+b^{2}x^{2}\right)\zeta$. So ${\mathbb{L}}^{{\mathbb{A}}}_{2}E_{\zeta}=E_{\widetilde{\zeta}}$, and ${\mathbb{L}}^{\mathbb{A}}_{2}$ leaves $E_{\zeta}$ invariant iff $\widetilde{\zeta}$ is collinear to $\zeta$. This happens if and only if $\zeta$ is an eigenstate of the harmonic oscillator, that is $\zeta=\varphi_{n}$ for some $n\in\mathbb{N}_{0}$. The three–dimensional analogue of the previous Proposition reads as follows. ###### Theorem 2.7. Let $\gamma$ be a bounded operator on $L^{2}(\mathbb{R}^{3})$ commuting with all magnetic translations ${\mathfrak{m}}_{\mathbb{R}}$ and with ${\mathbb{L}}^{\mathbb{A}}_{2}\otimes\mathbb{I}$. Then, there exists a family $(\gamma_{n})_{n\in\mathbb{N}}$ of bounded operators on $L^{2}(\mathbb{R})$, with $\|\gamma_{n}\|\leq\|\gamma\|$, and so that (2.11) $\gamma=\sum_{n=0}^{\infty}\mathbf{P}_{n}\otimes\gamma_{n}.$ If $\gamma$ is a locally trace class operator, then its density depends only on $x_{3}$ $\rho_{\gamma}({\mathbf{x}})=\rho_{\gamma}(x_{3})=\frac{b}{2\pi}\sum_{n=0}^{\infty}\rho_{\gamma_{n}}(x_{3}).$ ###### Proof. Let us consider two fixed test functions $\phi,\psi\in L^{2}(\mathbb{R})$, and define the operator $\eta_{\phi,\psi}$ on $L^{2}(\mathbb{R}^{2})$ by $\forall f,g\in L^{2}(\mathbb{R}^{2}),\quad\langle f,\eta_{\phi,\psi},g\rangle_{L^{2}(\mathbb{R}^{2})}:=\langle f\otimes\phi,\gamma(g\otimes\psi)\rangle_{L^{2}(\mathbb{R}^{3})}.$ The conditions on $\gamma$ imply that $\eta_{\phi,\psi}$ is a bounded self–adjoint operator that commutes with all ${\mathfrak{m}}_{\mathbb{R}}$ and with ${\mathbb{L}}_{\mathbb{A}}$. Thus, using Proposition 2.5, $\eta_{\phi,\psi}$ can be decomposed as $\eta_{\phi,\psi}=\sum_{n\in\mathbb{N}_{0}}\lambda_{n}(\phi,\psi)\mathbf{P}_{n}.$ Since $\eta$ is a bounded operator, we have for any normalized $\Phi_{n}\in L^{2}(\mathbb{R}^{2})$ in the range of $\mathbf{P}_{n}$, $\left|\lambda_{n}(\phi,\psi)\right|=\left|\langle\Phi_{n}\otimes\phi,\gamma(\Phi_{n}\otimes\psi)\rangle_{L^{2}(\mathbb{R}^{3})}\right|\leq\|\gamma\|_{\rm op}\|\psi\|_{L^{2}(\mathbb{R})}\left\|\phi\right\|_{L^{2}(\mathbb{R})}.$ We deduce that the map $(\phi,\psi)\mapsto\lambda_{n}(\phi,\psi)$ is sesquilinear and bounded, with bound smaller than $\|\eta\|_{\rm op}$. The result then follows by taking $\gamma_{n}$ the bounded self-adjoint operator on $L^{2}(\mathbb{R})$ defined by $\langle\phi,\gamma_{n}\psi\rangle:=\lambda_{n}(\phi,\psi).$ Finally, for $\eta$ of the form (2.11), we obtain, using $\rho_{\mathbf{P}_{n}}=\frac{b}{2\pi}$, $\rho_{\eta}({\mathbf{x}})=\eta({\mathbf{x}},{\mathbf{x}})=\sum_{n\in\mathbb{N}}\rho_{\mathbf{P}_{n}}(x_{1},x_{2})\rho_{\gamma_{n}}(x_{3})=\frac{b}{2\pi}\sum_{n\in\mathbb{N}}\rho_{\gamma_{n}}(x_{3}).$ ∎ For an operator $\eta$ of the form (2.11), the trace per unit-surface, defined as the limit $\underline{\rm Tr}(\eta)=\lim_{L\to\infty}\frac{1}{L^{2}}{\rm Tr}\left({\mathds{1}}_{\Gamma_{L}}\eta{\mathds{1}}_{\Gamma_{L}}\right),\quad\Gamma_{L}=[-\tfrac{L}{2},\tfrac{L}{2}]^{2}\times\mathbb{R}$ takes the simpler form $\underline{\rm Tr}(\eta)=\frac{b}{2\pi}\sum_{n=0}^{\infty}{\rm Tr}_{1}(\eta_{n}),$ where we have used that the density of any Landau level is $\rho_{\mathbf{P}_{n}}=\frac{b}{2\pi}$. ## 3\. Reduction of the kinetic energy, and applications We now exploit the particular structure of states $\gamma\in{\mathcal{P}}^{\mathbb{A}}$ to deduce their kinetic energy. ### 3.1. Reduction of the kinetic energy Recall that the set ${\mathcal{P}}^{\mathbb{A}}$ has been defined in (1.3) as the set of one-body density matrices $\gamma$, satisfying the Pauli principle $0\leq\gamma\leq 1$, which commute with all magnetic translations, and with the Landau operator ${\mathbb{L}}^{\mathbb{A}}_{2}\otimes\mathbb{I}$. We also recall that the set of reduced states ${\mathcal{G}}$ is defined by ${\mathcal{G}}:=\left\\{G\in{\mathcal{S}}(L^{2}(\mathbb{R})),\quad G\geq 0\right\\}.$ The main result of this section is the following. ###### Theorem 3.1. For any $\gamma\in{\mathcal{P}}^{\mathbb{A}}$, there is an operator $G\in{\mathcal{G}}$ satisfying $\rho_{G}=\rho_{\gamma}$ and (3.1) $\frac{1}{2}\underline{\rm Tr}({\mathbb{L}}^{\mathbb{A}}_{3}\gamma)\geq\frac{1}{2}{\rm Tr}(-\Delta_{1}G)+{\rm Tr}(F(b,G)),$ where (3.2) $F(b,g):=\pi g^{2}+\frac{b^{2}}{4\pi}\left\\{\frac{2\pi g}{b}\right\\}\left(1-\left\\{\frac{2\pi g}{b}\right\\}\right).$ Conversely, for any $G\in{\mathcal{G}}$, there is $\gamma\in{\mathcal{P}}^{\mathbb{A}}$ so that $\rho_{\gamma}=\rho_{G}$, and for which there is equality in (3.1). In particular, for any (representable) density $\rho$, (3.3) $\inf_{\gamma\in{\mathcal{P}}^{\mathbb{A}}\atop\rho_{\gamma}=\rho}\left\\{\frac{1}{2}\underline{\rm Tr}({\mathbb{L}}^{\mathbb{A}}_{3}\gamma)\right\\}=\inf_{G\in{\mathcal{G}}\atop\rho_{G}=\rho}\left\\{\frac{1}{2}{\rm Tr}(-\Delta_{1}G)+{\rm Tr}(F(b,G))\right\\}.$ ###### Proof. According to Theorem (2.7), any $\gamma\in{\mathcal{P}}^{\mathbb{A}}$ can be decomposed as (3.4) $\gamma=\sum_{n=0}^{\infty}\mathbf{P}_{n}\otimes\gamma_{n}\quad\text{with}\quad\gamma_{n}\in{\mathcal{S}}(L^{2}(\mathbb{R})),\quad 0\leq\gamma_{n}\leq 1.$ For a state of the form (3.4), we define the operator $G_{\gamma}\in{\mathcal{S}}(L^{2}(\mathbb{R}))$ by $G_{\gamma}:=\frac{b}{2\pi}\sum_{n=0}^{\infty}\gamma_{n}.$ Since $\gamma_{n}\geq 0$, we have $G_{\gamma}\geq 0$ as well, so $G_{\gamma}\in{\mathcal{G}}$. Also, since $\rho_{\mathbf{P}_{n}}({\mathbf{x}})=\frac{b}{2\pi}$, we deduce that $\rho_{G}=\rho_{\gamma}$. Recalling that ${\mathbb{L}}^{\mathbb{A}}_{3}:={\mathbb{L}}^{\mathbb{A}}_{2}\otimes\mathbb{I}_{L^{2}(\mathbb{R})}+\mathbb{I}_{L^{2}(\mathbb{R}^{2})}\otimes(-\Delta_{1})$, and using Proposition 2.1, we obtain that $\displaystyle\frac{1}{2}\underline{\rm Tr}\left({\mathbb{L}}^{\mathbb{A}}_{3}\gamma\right)$ $\displaystyle=\frac{1}{2}\sum_{n=0}^{\infty}\underline{\rm Tr}\left(\varepsilon_{n}\mathbf{P}_{n}\otimes\gamma_{n}\right)+\frac{1}{2}\sum_{n=0}^{\infty}\underline{\rm Tr}\left(\mathbf{P}_{n}\otimes(-\Delta_{1}\gamma_{n})\right)$ $\displaystyle=\frac{b}{4\pi}\sum_{n=0}^{\infty}\varepsilon_{n}{\rm Tr}(\gamma_{n})+\frac{b}{4\pi}\sum_{n=0}^{\infty}{\rm Tr}\left(-\Delta_{1}\gamma_{n}\right)$ $\displaystyle=\frac{b}{4\pi}\sum_{n=0}^{\infty}\varepsilon_{n}{\rm Tr}(\gamma_{n})+\frac{1}{2}{\rm Tr}\left(-\Delta_{1}G\right).$ The first term cannot be expressed directly as a function of $G$, but we have an inequality for this term. Since $G$ is a positive operator with finite trace, it is compact, and admits a spectral decomposition of the form $G=\sum_{j}g_{j}|\psi_{j}\rangle\langle\psi_{j}|$ with $g_{j}>0$ and $\sum_{j}g_{j}<{\infty}$. Evaluating the trace of $\gamma_{n}$ in the $\\{\psi_{j}\\}$ basis, and changing the order of the sums (all terms are positive), we obtain $\sum_{n=0}^{\infty}\varepsilon_{n}{\rm Tr}(\gamma_{n})=\sum_{j=1}^{\infty}\left(\sum_{n=0}^{\infty}\varepsilon_{n}\left\langle\psi_{j},\gamma_{n}\psi_{j}\right\rangle\right).$ Since $0\leq\gamma_{n}\leq 1$, the quantity $m_{j}(n):=\left\langle\psi_{j},\gamma_{n}\psi_{j}\right\rangle$ satisfies $0\leq m_{j}(n)\leq 1$. In addition, we have $\frac{b}{2\pi}\sum_{n}m_{j}(n)=\frac{b}{2\pi}\langle\psi_{j},\sum_{n}\gamma_{n}\psi_{j}\rangle=\langle\psi_{j},G\psi_{j}\rangle=g_{j}$. So we have the inequality (3.5) $\sum_{n=0}^{\infty}\varepsilon_{n}{\rm Tr}(\gamma_{n})\geq\sum_{j=1}^{\infty}\inf_{m}\left\\{\sum_{n=0}^{\infty}\varepsilon_{n}m(n),\ 0\leq m(n)\leq 1,\ \sum_{n=0}^{\infty}m(n)=\frac{2\pi g_{j}}{b}\right\\}.$ Since the $\varepsilon_{n}$ are ranked in increasing order, we can apply the bathtub principle [13, Theorem 1.4]. The optimal $m$ for the above minimization is given by (3.6) $m_{j}^{*}(n)=\begin{cases}1&\text{for all}\quad 0\leq n\leq\lfloor\frac{2\pi g_{j}}{b}\rfloor-1\\\ \\{\frac{2\pi g_{j}}{b}\\}&\text{for}\quad n=\lfloor\frac{2\pi g_{j}}{b}\rfloor\\\ 0&\text{otherwise}.\end{cases}.$ We now calculate the infimum in the RHS of (3.5) using the explicit formula of the optimal function $m^{*}$. Recalling that $\varepsilon_{n}=b(2n+1)$ and denoting by $x:=\frac{2\pi g_{j}}{b}$, we obtain $\displaystyle\sum_{n=0}^{\infty}\varepsilon_{n}m_{j}^{*}(n)$ $\displaystyle=b\sum_{n=0}^{\lfloor x\rfloor-1}(2n+1)+b\left(2\left\lfloor x\right\rfloor+1\right)\left\\{x\right\\}$ $\displaystyle=b\left(x^{2}+\\{x\\}(1-\\{x\\})\right)=\frac{4\pi}{b}F(b,g_{j}),$ with the function $F$ defined in (3.2). Summing in $j$ and gathering the terms gives the inequality $\frac{1}{2}\underline{\rm Tr}\left({\mathbb{L}}^{\mathbb{A}}_{3}\gamma\right)\geq{\rm Tr}\left(F(b,G)\right)+\frac{1}{2}{\rm Tr}\left(-\Delta_{1}G\right),$ which proves the first part of the Theorem. Conversely, given $G=\sum_{j}g_{j}|\psi_{j}\rangle\langle\psi_{j}|\in{\mathcal{G}}$, we consider the state $\gamma^{*}:=\sum_{n=0}^{\infty}\mathbf{P}_{n}\otimes\gamma_{n}^{*},\quad\text{with}\quad\gamma_{n}^{*}:=\sum_{j=1}^{\infty}m_{j}^{*}(n)|\psi_{j}\rangle\langle\psi_{j}|$ and $m_{j}^{*}$ defined as in (3.6). The operator $\gamma^{*}$ belongs to ${\mathcal{P}}^{\mathbb{A}}$, satisfies $G_{\gamma^{*}}=G$, and gives an equality in (3.1). ∎ ### 3.2. Some properties of the function F Let us collect some useful properties of the function $F$. A plot of $F$ is displayed in Figure 1 below. ###### Proposition 3.2. The function $F$ in (3.2) is continuous and satisfies (3.7) $\pi g^{2}\leq F(b,g)\leq\pi g^{2}+\frac{b^{2}}{16\pi},$ with equality in the left for $\frac{2\pi g}{b}\in\mathbb{N}$, and equality in the right for $\frac{2\pi g}{b}\in\mathbb{N}+\frac{1}{2}$, and $F(b,g)\to\pi g^{2}$ as $b\to 0$. For any $b\geq 0$, the map $g\mapsto F(b,g)-\pi g^{2}$ is $\frac{b}{2\pi}$ periodic and the map $g\in\mathbb{R}_{+}\mapsto F(b,g)$ is piece-wise linear, increasing and convex. Finally, for all $0\leq g<\frac{b}{2\pi}$, we have $F(b,g)=\frac{1}{2}bg$ ###### Proof. The first part is straightforward from the definition (3.2). To see that it is convex, piece-wise linear and increasing, we use the alternative form (3.8) $F(b,g)=\frac{b^{2}}{4\pi}\left(x(1+2\lfloor x\rfloor)-\lfloor x\rfloor-\lfloor x\rfloor^{2}\right),$ where we have denoted by $x:=\frac{2\pi g}{b}$. When $0\leq g<\frac{b}{2\pi}$, which corresponds to $0\leq x<1$, $F(b,g)=\frac{1}{2}bg$. ∎ ###### Remark 3.3. The left inequality of (3.7) implies ${\rm Tr}(F(b,G))\geq\pi{\rm Tr}(G^{2})$, hence, together with (1.1), that $\inf_{\gamma\in{\mathcal{P}}^{\mathbb{A}}\atop\rho_{\gamma}=\rho}\left\\{\frac{1}{2}\underline{\rm Tr}\left({\mathbb{L}}^{\mathbb{A}}_{3}\gamma\right)\right\\}\geq\inf_{\gamma\in{\mathcal{P}}\atop\rho_{\gamma}=\rho}\left\\{\frac{1}{2}\underline{\rm Tr}\left(-\Delta_{3}\gamma\right)\right\\}.$ In particular, the kinetic energy is higher with the magnetic field. This is a kind of diamagnetic inequality for 2D materials. ###### Remark 3.4. The fact that $F(b,g)\to\pi g^{2}$ as $b\to 0$, for all $g\in\mathbb{R}^{+}$, means that, our reduction approach in this manuscript is coherent with the one without magnetic field already treated in [9]. ###### Remark 3.5. Splitting $F$ into $F(b,G)=\pi g^{2}+\tilde{F}(b,g)$, we see that the effect of adding a magnetic field ${\mathbb{B}}=(0,0,b)$ is a periodic perturbation of the energy with no magnetic field. ### 3.3. Reduced DFT models In the context of DFT, the previous result suggests modelling the electronic state in a homogeneous slab of charge distribution $\mu({\mathbf{x}})=\mu(x_{3})$ under a constant magnetic field ${\mathbb{B}}=b(0,0,x_{3})$ by a reduced state $G\in{\mathcal{G}}$ whose energy per unit surface is given by (3.9) ${\mathcal{E}}(G)=\frac{1}{2}{\rm Tr}(-\Delta_{1}G)+{\rm Tr}\left(F_{b}(G)\right)+\frac{1}{2}{\mathcal{D}}_{1}(\rho_{G}-\mu)+E^{\rm xc}(\rho_{G}).$ Here, $E^{\rm xc}$ models is an exchange-correlation energy per unit surface, and ${\mathcal{D}}_{1}$ is the one–dimensional Hartree term. This last term has been extensively studied in our previous work [9, Section 3.1], and is defined as follows. For $f\in{\mathcal{C}}:=\left\\{f\in L^{1}(\mathbb{R}),\ W_{f}\in L^{2}(\mathbb{R})\right\\}$, where $W_{f}(x):=\int_{-\infty}^{x}f$ is a primitive of $f$, we have ${\mathcal{D}}_{1}(f):=4\pi\int_{\mathbb{R}}|W_{f}|^{2}(x){\mathrm{d}}x.$ We have proved in [9, Proposition 3.3] that the elements $f\in{\mathcal{C}}$ have null integral $\int f=0$, that the map ${\mathcal{C}}\ni f\mapsto{\mathcal{D}}_{1}(f)$ is convex, and that ${\mathcal{D}}_{1}(f)=4\pi\iint_{(\mathbb{R}_{+})^{2}\times(\mathbb{R}_{-})^{2}}\min\\{|x|,|y|\\}f(x)f(y){\mathrm{d}}x{\mathrm{d}}y=\int_{\mathbb{R}}\Phi_{f}(x)f(x){\mathrm{d}}x,$ with the mean-field potential $\Phi_{f}(x):=4\pi\int_{\mathbb{R}^{\pm}}\min\\{|x|,|y|\\}f(y){\mathrm{d}}y,\quad x\in\mathbb{R}^{\pm}.$ The function $\Phi_{f}$ is continuous, and is the (unique) solution to $-\Phi_{f}^{\prime\prime}(x)=4\pi f,\quad\Phi_{f}^{\prime}(x)\xrightarrow[x\to\pm\infty]{}0,\quad\Phi_{f}(0)=0.$ In practice, we restrict the minimization problem to the $G$ so that $\rho_{G}-\mu\in{\mathcal{C}}$. This implies in particular the neutrality condition ${\rm Tr}(G)=\int\rho_{G}=\int\mu=\nu$. On the other hand, if $G$ is a trace class operator, then $\rho_{G}\in L^{1}$ (we assume that $\mu\in L^{1}$ as well), and if $\rho_{G}-\mu$ has null integral, then ${\mathcal{D}}_{1}(\rho_{G}-\mu)<\infty$ iff $\rho_{G}-\mu\in{\mathcal{C}}$. Note that there is no Pauli principle on $G$ for admissible states ${\mathcal{G}}$. It has been replaced by a penalization term $+F(b,G)$ in the energy. ###### Remark 3.6. The energy (3.9) is obtained when minimizing a 3-dimensional DFT model over transitionally invariant states. In particular, this model does not include possible spatial symmetry breaking along the first $2$ variables. Such phenomena are known to exist in two-dimensional electron gas under magnetic field due to the de Haas–van Alphen effect [5]. In some real-life systems e.g. Br2, magnetic domains form, sometimes called Condon domains [4, 15]. Our simple model is unable to capture these effects. ### 3.4. The reduced Hartree–Fock case Let us illustrate the previous discussion in the particular case of the reduced Hartree-Fock (rHF) model, in which $E^{\rm xc}=0$ in (3.9). We let $0\leq\mu\in L^{1}(\mathbb{R})$ be a nuclear density describing a homogeneous 2D material and denote by $\nu=\int_{\mathbb{R}}\mu$ the total charge per unit surface. We denote by (3.10) ${\mathcal{E}}_{b}^{\rm rHF}(G):=\frac{1}{2}{\rm Tr}(-\Delta_{1}G)+{\rm Tr}(F(b,G))+\frac{1}{2}{\mathcal{D}}_{1}(\rho_{G}-\mu)$ the corresponding rHF energy per unit surface, and study the minimization problem ${\mathcal{I}}_{b}^{\rm rHF}:=\inf\left\\{{\mathcal{E}}_{b}^{\rm rHF}(G),\;G\in{\mathcal{G}}^{\nu}\right\\},\quad\text{with}\quad{\mathcal{G}}^{\nu}:=\left\\{G\in{\mathcal{G}},\ {\rm Tr}(G)=\nu\right\\}.$ Following the exact same lines as [9, Theorem 2.7], one has the following. ###### Theorem 3.7. The problem ${\mathcal{I}}_{b}^{\rm rHF}$ admits a minimizer, and all minimizers share the same density. We skip the proof for brevity. The uniqueness of the density comes from the fact that the problem is strictly convex in $\rho_{G}$. However, unlike the case without magnetic field, $F(b,\cdot)$ is not strictly convex for $b>0$. It is unclear to us whether the minimizer of ${\mathcal{I}}_{b}^{\rm rHF}$ is unique. We would like to write the Euler-Lagrange equations for a minimizer $G_{*}$. Recall that $g\mapsto F(b,g)$ is continuous and convex (but is not smooth). We denote by $f_{b}:=\partial_{g}F(b,\cdot)$ its subdifferential, a set-valued function defined by $f_{b}(g):=\left\\{a\in\mathbb{R},\quad\forall g^{\prime}\in\mathbb{R},\quad F(b,g^{\prime})-F(b,g)\geq a(g^{\prime}-g)\right\\}.$ In our case, since the function $F_{b}$ is piece-wise linear, $f_{b}$ is explicit. From (3.8), we obtain (in the following lines, $\\{a\\}$ denotes the singleton $a$) $f_{b}(g)=\begin{cases}\left\\{\frac{b^{2}}{4\pi}(2n+1)\right\\}\quad\text{if }\exists n\in\mathbb{N}_{0},\;\quad n<\frac{2\pi g}{b}<n+1\\\ \left[\frac{b^{2}}{4\pi}(2n-1),\frac{b^{2}}{4\pi}(2n+1)\right]\quad\text{if}\quad\frac{2\pi g}{b}=n\in\mathbb{N}_{0}.\end{cases}$ Its inverse map, noted $h_{b}$, is the set-valued function so that $y\in f_{b}(x)$ iff $x\in h_{b}(y)$. One finds, for $y>0$, $h_{b}(y)=\begin{cases}\left[n\frac{b}{2\pi},(n+1)\frac{b}{2\pi}\right]\quad\text{if}\quad n:=\frac{1}{2}\left(\frac{4\pi}{b^{2}}y-1\right)\in\mathbb{N}_{0},\\\ \left\\{n\frac{b}{2\pi}\right\\}\quad\text{if there is $n\in\mathbb{N}_{0}$ so that}\quad n-1<\frac{1}{2}\left(\frac{4\pi}{b^{2}}y-1\right)<n.\end{cases}$ We extend the definition of $h_{b}$ by setting $h_{b}(y)=0$ for $y<0$. In order to work with functions, it is useful to introduce the maps (3.11) $f_{b}^{\pm}(g):=\lim_{t\to 0^{\pm}}\frac{1}{t}\left(F(b,g+t)-F(b,g)\right)$ so that $f_{b}(g)=[f_{b}^{-}(g),f_{b}^{+}(g)]$ for all $g\in\mathbb{R}^{+}$. Of course, if $g\notin\frac{b}{2\pi}\mathbb{N}_{0}$ is a regular point, then $f_{b}(y)=f_{b}^{+}(y)=f_{b}^{-}(y)$. We define the maps $h_{b}^{\pm}$ similarly, so that $h_{b}(y)=[h_{b}^{-}(y),h_{b}^{+}(y)]$ for all $y\in\mathbb{R}$. The Euler–Lagrange equations for $G_{*}$ takes the following form (see end of the section for the proof). ###### Proposition 3.8 (Euler-Lagrange equations). Let $G_{*}$ be a minimizer of ${\mathcal{I}}^{{\rm rHF}}_{b}$. Then there is $\lambda\in\mathbb{R}$ so that (3.12) $\begin{cases}h_{b}^{-}(\lambda-H_{*})\leq G_{*}\leq h_{b}^{+}(\lambda- H_{*})\\\ H_{*}:=-\frac{1}{2}\Delta_{1}+\Phi_{*}\\\ -\Phi_{*}^{\prime\prime}=4\pi(\rho_{*}-\mu),\quad\Phi_{*}^{\prime}(x)\xrightarrow[x\to\pm\infty]{}0,\quad\Phi_{*}(0)=0,\end{cases}$ where $\rho_{*}=\rho_{G_{*}}$ is the associated density of $G_{*}$, and $\Phi_{*}$ is the mean-field potential, defined as the unique solution of the last equation. The first equation can also be written as $G_{*}\in h_{b}(\lambda-H_{*}),$ and means that if $G_{*}=\sum_{j}g_{j}|\psi_{j}\rangle\langle\psi_{j}|$ is the spectral decomposition of the optimizer, then $\psi_{j}$ is also an eigenfunction of $H_{*}$ for an eigenvalue $\varepsilon_{j}$ so that $g_{j}\in h_{b}(\lambda-\varepsilon_{j}),\quad\text{or equivalently}\quad\varepsilon_{j}\in\lambda-f_{g}(g_{j}).$ Conversely, if $\varepsilon<\lambda$ is an eigenvalue of $H_{*}$, then $\varepsilon=\varepsilon_{j}$ for some $j$. In practice, for numerical purpose, one rather considers an approximation $F^{\delta}_{b}$ of $F$, which is smooth, strictly convex, and so that $\|F^{\delta}_{n}-F\|_{\infty}<\delta$. In this case, one can repeat the arguments in [9, Theorem 2.7], and the first line of (3.12) becomes $\left(F_{b}^{\delta}\right)^{\prime}(G_{*})=\lambda-H_{*},\quad\text{or, equivalently}\quad G_{*}=\left[\left(F_{b}^{\delta}\right)^{\prime}\right]^{-1}(\lambda-H_{*}).$ ###### Remark 3.9 (Strong magnetic fields). In the case where $b>2\pi\nu$, any $G\in{\mathcal{G}}^{\nu}$are positive and satisfies ${\rm Tr}(G)=\nu$, hence all eigenvalues of $G$ are smaller than $\nu$. In particular, $F_{b}(G)=\frac{1}{2}bG,\quad\text{hence}\quad{\rm Tr}\left(F_{b}(G)\right)=\frac{1}{2}b\nu$ is a constant, independent of $G\in{\mathcal{G}}^{\nu}$. In this case, $G_{*}$ is also the minimizer of $\inf\left\\{\frac{1}{2}{\rm Tr}(-\Delta_{1}G)+\frac{1}{2}{\mathcal{D}}_{1}(\rho_{G}-\mu),\quad G\in{\mathcal{G}}^{\nu}\right\\}.$ This minimizer is therefore independent of $b>2\pi\nu$, reflecting the fact that all electrons lie in the lowest Landau level. Following the previous lines, we deduce that $G_{*}$ is a rank-$1$ operator, of the form $G_{*}=\nu|\psi_{*}\rangle\langle\psi_{*}|$, with $\psi_{*}$ minimizing $\inf\left\\{\frac{\nu}{2}\int_{\mathbb{R}}|\nabla\psi_{*}|^{2}+\frac{1}{2}{\mathcal{D}}_{1}\left(\nu|\psi_{*}|^{2}-\mu\right),\quad\psi_{*}\in L^{2}(\mathbb{R}),\ \|\psi_{*}\|=1\right\\}.$ ###### Proof of Proposition 3.8. Let $G_{*}$ be a minimizer of ${\mathcal{I}}^{\rm rHF}_{b}$, and let $\rho_{*}$ and $\Phi_{*}$ be the corresponding density and mean-field potential, and set $H_{*}:=-\frac{1}{2}\Delta_{1}+\Phi_{*}$. Recall that $\rho_{*}$ (hence $\Phi_{*}$ and $H_{*}$) is uniquely defined. First, we claim that $G_{*}$ commutes with $H_{*}$. This is a standard result in the case where the map $F$ is smooth (say of class $C^{1}$), using that ${\rm Tr}(F(G_{*}+H))={\rm Tr}(F(G_{*}))+{\rm Tr}(F^{\prime}(G_{*})H)+o(H).$ In our case however, the map $F$ is only piece-wise smooth, and we need a direct proof. Let $A$ be a finite-rank symmetric operator on $L^{2}(\mathbb{R})$, and set $G_{t,A}:={\mathrm{e}}^{-{\mathrm{i}}tA}G_{*}{\mathrm{e}}^{{\mathrm{i}}tA}$. Since $G_{t,A}$ is a unitary transformation of $G_{*}$, we have ${\rm Tr}(G_{t,A})={\rm Tr}(G_{*})=\nu$, that is $G\in{\mathcal{G}}^{\nu}$ and ${\rm Tr}(F(b,G_{t,A}))={\rm Tr}(F(b,G_{*}))$. In particular, $\displaystyle{\mathcal{E}}^{\rm rHF}_{b}(G_{t,A})$ $\displaystyle={\mathcal{E}}^{\rm rHF}_{b}(G_{*})$ $\displaystyle+\frac{1}{2}\left({\rm Tr}(-\Delta_{1}[G_{t,A}-G_{*}])+{\mathcal{D}}_{1}\left(\rho_{G_{t,A}}-\mu\right)-{\mathcal{D}}_{1}\left(\rho_{*}-\mu\right)\right).$ Together with the fact that $G_{t,A}=G_{*}+{\mathrm{i}}t\left[G_{*},A\right]+o(t),$ and the definition of $H_{*}$, we deduce that ${\mathcal{E}}^{\rm rHF}_{b}(G_{t,A})={\mathcal{E}}^{\rm rHF}_{b}(G_{*})+{\mathrm{i}}t{\rm Tr}\left(H_{*}[G_{*},A]\right)+o(t).$ Since the minimum of ${\mathcal{E}}^{\rm rHF}_{b}$ is obtained for $t=0$, the linear term in $t$ must vanish, that is: $\displaystyle 0$ $\displaystyle={\rm Tr}\left(H_{*}[G_{*},A]\right)={\rm Tr}\left(H_{*}G_{*}A-H_{*}AG_{*}\right)$ $\displaystyle={\rm Tr}\left(H_{*}G_{*}A-G_{*}H_{*}A\right)={\rm Tr}\left([H_{*},G_{*}]A\right),$ where we have used cyclicity of the trace, and the fact that $A$ is finite rank (so all operators are trace-class). Since this is true for all finite rank symmetric operators $A$, we deduce as wanted that $[H_{*},G_{*}]=0$. Recall that $G_{*}$ is positive compact (even trace class). So there is $M:={\rm rank}(G_{*})\in\mathbb{N}\cup\\{\infty\\}$ and an orthonormal family $\\{\psi_{j}\\}_{j\in[1,M]}$ so that $G_{*}=\sum_{j=1}^{M}g_{j}|\psi_{j}\rangle\langle\psi_{j}|,\quad\text{with}\quad H\psi_{j}=\varepsilon_{j}\psi_{j},$ and with $g_{1}\geq g_{2}\geq\cdots\geq g_{M}>0$. The orthonormal family $\\{\psi_{j}\\}_{1\leq j\leq M}$ spans ${\rm Ran}(G_{*})$. For two indices $(i,j)\in[1,M]$, we consider the operator $G_{t}^{(i,j)}:=G_{*}+t\left(|\psi_{i}\rangle\langle\psi_{i}|-|\psi_{j}\rangle\langle\psi_{j}|\right).$ For $t$ small enough ($|t|<\min\\{g_{i},g_{j}\\}$), the operator $G_{t}^{(i,j)}$ is positive, and with ${\rm Tr}(G_{t}^{(i,j)})={\rm Tr}(G_{*})=\nu$, hence $G_{t}^{(i,j)}\in{\mathcal{G}}^{\nu}$. In addition, we have $\displaystyle{\mathcal{E}}^{\rm rHF}_{b}(G_{t}^{(i,j)})-{\mathcal{E}}^{\rm rHF}_{b}(G_{*})=$ $\displaystyle\left(F(b,g_{i}+t)-F(b,g_{i})\right)+\left(F(b,g_{j}-t)-F(b,g_{j})\right)$ $\displaystyle+t\underbrace{{\rm Tr}(H_{*}(|\psi_{i}\rangle\langle\psi_{i}|-|\psi_{j}\rangle\langle\psi_{j}|))}_{=\varepsilon_{i}-\varepsilon_{j}}+o(t).$ By optimality of $G_{*}$, this quantity is always positive. Taking the limit $t\to 0^{+}$, and recalling the definition of $f_{b}^{\pm}$ in (3.11), we deduce that $f_{b}^{+}(g_{i})+\varepsilon_{i}\geq f_{b}^{-}(g_{j})+\varepsilon_{j}.$ This inequality is valid for all $1\leq i,j\leq M$, so $\inf_{1\leq i\leq M}\left(f_{b}^{+}(g_{i})+\varepsilon_{i}\right)\geq\sup_{1\leq j\leq M}\left(f_{b}^{-}(g_{j})+\varepsilon_{j}\right)=:\lambda$ For all $1\leq j\leq M$, we have $f_{b}^{-}(g_{j})+\varepsilon_{j}\leq\lambda\leq f_{b}^{+}(g_{j})+\varepsilon_{j},\quad\text{hence}\quad\lambda-\varepsilon_{j}\in[f_{b}^{-}(g_{j}),f_{b}^{+}(g_{j})].$ This is also $\lambda-\varepsilon_{j}\in f_{b}(g_{j})$, or equivalently $g_{j}\in h_{b}(\lambda-\varepsilon_{j})$. Note that if there is an eigenvalue $g_{j}\notin\frac{b}{2\pi}\mathbb{Z}$, then $f_{b}^{-}(g_{j})=f_{b}^{+}(g_{j})$ for this eigenvalue, and there is equality. In other words, we have proved that $G_{*}\in h_{b}(\lambda-H_{*})\quad\text{on}\quad{\rm Ran}(G_{*}).$ It remains to prove the result on ${\rm Ker}(G_{*})$. Let $\psi\in{\rm Ker}(G_{*})$. This time, we consider the perturbed state $G_{t}^{(j)}:=G_{*}+t\left(|\psi\rangle\langle\psi|-|\psi_{j}\rangle\langle\psi_{j}|\right),$ which is in ${\mathcal{G}}^{\nu}$ for all $0\leq t\leq g_{j}$. Taking the limit $t\to 0^{+}$, and reasoning as before, we get $f_{b}^{+}(0)+\langle\psi,H_{*}\psi\rangle\geq f_{b}^{-}(g_{j})+\varepsilon_{j},\quad\text{hence}\quad f_{b}^{+}(0)+\langle\psi,H_{*}\psi\rangle\geq\lambda,$ where we took the supremum in $1\leq j\leq M$ in the last inequality. We deduce that, for all $\psi\in{\rm Ker}(G_{*})$, $f_{b}^{+}(\langle\psi,G_{*}\psi\rangle)=f_{b}^{+}(0)\geq\lambda-\langle\psi,H_{*}\psi\rangle,$ so $f_{b}^{+}(G_{*})\geq\lambda-H_{*},\quad\text{on}\quad{\rm Ker}(G_{*}).$ Together with the fact that $f_{b}^{-}(G_{*})=0$ on ${\rm Ker}(G_{*})$, we obtain $G_{*}\in h_{b}(\lambda-H_{*})$ on ${\rm Ker}(G_{*})$ as well. ∎ ## 4\. Models with spin In this section, we explain how to extend our results to the case where the spin is taken into account. In this case, the density matrix is an operator $\gamma\in{\mathcal{S}}(L^{2}(\mathbb{R}^{3},\mathbb{C}^{2}))$ satisfying the Pauli-principle $0\leq\gamma\leq 1$. Such an operator can be decomposed as a $2\times 2$ matrix of the form ${\gamma}=\begin{pmatrix}\gamma^{\uparrow\uparrow}&\gamma^{\uparrow\downarrow}\\\ \gamma^{\downarrow\uparrow}&\gamma^{\downarrow\downarrow}\end{pmatrix}.$ The spin–density $2\times 2$ matrix of $\gamma$ is $R_{\gamma}({\mathbf{x}})=\gamma({\mathbf{x}},{\mathbf{x}})$ (see [7] for details), and the total density is $\rho_{\gamma}({\mathbf{x}})=R_{\gamma}^{\uparrow\uparrow}({\mathbf{x}})+R_{\gamma}^{\downarrow\downarrow}({\mathbf{x}})=\gamma^{\uparrow\uparrow}({\mathbf{x}},{\mathbf{x}})+\gamma^{\downarrow\downarrow}({\mathbf{x}},{\mathbf{x}})$. The kinetic energy operator is now the Pauli operator $\mathbf{P}^{\mathbb{A}}_{3}:=\left[\sigma\cdot\left(-{\mathrm{i}}\nabla+{\mathbb{A}}\right)\right]^{2},$ where $\sigma$ contains the Pauli matrices. In the case of constant magnetic field ${\mathbb{B}}=(0,0,b)$ with the gauge ${\mathbb{A}}=b(0,x_{1},0)$, this operator becomes $\mathbf{P}^{\mathbb{A}}_{3}={\mathbb{L}}^{\mathbb{A}}_{3}\mathbb{I}_{2}+\begin{pmatrix}b&0\\\ 0&-b\end{pmatrix}={\mathbb{L}}^{\mathbb{A}}_{2}\mathbb{I}_{2}+\begin{pmatrix}b&0\\\ 0&-b\end{pmatrix}-\partial_{x_{3}x_{3}}^{2}\mathbb{I}_{2}.$ In what follows, we denote by ${\mathcal{B}}:=\begin{pmatrix}b&0\\\ 0&-b\end{pmatrix}$. This term corresponds to the Zeeman term. There are several ways to read this operator. Indeed, we have $\displaystyle L^{2}(\mathbb{R}^{3},\mathbb{C}^{2})$ $\displaystyle\simeq L^{2}(\mathbb{R}^{3})\otimes\mathbb{C}^{2}\simeq L^{2}(\mathbb{R}^{2})\otimes L^{2}(\mathbb{R})\otimes\mathbb{C}^{2}$ (4.1) $\displaystyle\simeq L^{2}(\mathbb{R}^{2})\otimes L^{2}(\mathbb{R},\mathbb{C}^{2})$ (4.2) $\displaystyle\simeq L^{2}(\mathbb{R}^{2},\mathbb{C}^{2})\otimes L^{2}(\mathbb{R}).$ In the decomposition (4.1) (resp. (4.2)), we split the Pauli operator as $\mathbf{P}^{\mathbb{A}}_{3}={\mathbb{L}}^{\mathbb{A}}_{2}\otimes\mathbb{I}+\mathbb{I}\otimes({\mathcal{B}}-\partial_{x_{3}x_{3}}^{2}),\quad(\text{resp. }\quad\mathbf{P}^{\mathbb{A}}_{3}=({\mathbb{L}}^{\mathbb{A}}_{2}+{\mathcal{B}})\otimes\mathbb{I}+\mathbb{I}\otimes(-\partial_{x_{3}x_{3}}^{2}).)$ First splitting: keeping the spin structure. This splitting is useful if one wants to keep track of the spin–density structure of $R_{\gamma}$ as a function of $x_{3}$. This happens for instance when using exchange–correlations functionals which depend explicitly on the spin, as in the LSDA model [17, 8]. In this case, we consider one-body density matrices $\gamma$ of the form (4.3) $\gamma=\sum_{n=0}^{\infty}\mathbf{P}_{n}\otimes\gamma_{n},\quad\gamma_{n}\in{\mathcal{S}}(L^{2}(\mathbb{R},\mathbb{C}^{2})),\ 0\leq\gamma_{n}\leq 1.$ This is similar to what we have studied in the present article, but the $\gamma_{n}$ operators now act on $L^{2}(\mathbb{R},\mathbb{C}^{2})$. Introducing the operator $G:=\frac{b}{2\pi}\sum\gamma_{n}$ and the set ${\mathcal{G}}^{\rm spin}:=\left\\{G\in L^{2}(\mathbb{R},\mathbb{C}^{2}),\ G\geq 0\right\\},$ we obtain as before that, for all representable spin–density $2\times 2$ matrix $R$, we have $\displaystyle\inf_{\gamma\ \text{of the form\leavevmode\nobreak\ \eqref{eq:gamma-spin1}}\atop R_{\gamma}=R}\left\\{\frac{1}{2}\underline{\rm Tr}(\mathbf{P}^{\mathbb{A}}_{3}\gamma)\right\\}$ $\displaystyle\qquad=\inf_{G\in{\mathcal{G}}^{\rm spin}\atop R_{G}=R}\left\\{\frac{1}{2}{\rm Tr}(-\Delta_{1}G)+{\rm Tr}(F(b,G))\right\\}+b\int_{\mathbb{R}}(R^{\uparrow\uparrow}-R^{\downarrow\downarrow}).$ The last term is the Zeeman term, where we have used that ${\rm Tr}({\mathcal{B}}G)=b{\rm Tr}(G^{\uparrow\uparrow}-G^{\downarrow\downarrow})=b\int(R^{\uparrow\uparrow}-R^{\downarrow\downarrow})$. Second splitting: loosing the spin structure. If one is only interested in keeping the total density $\rho_{\gamma}$ instead of the full spin–density $2\times 2$ matrix $R_{\gamma}$, one should instead use the spectral decomposition of the operator $\mathbf{P}^{\mathbb{A}}_{2}:={\mathbb{L}}^{\mathbb{A}}_{2}+{\mathcal{B}}$. This one is easily deduced from the one of ${\mathbb{L}}^{\mathbb{A}}_{2}$ in Proposition 2.1. Using that $\varepsilon_{n}=b(2n+1)$, we obtain $\mathbf{P}^{\mathbb{A}}_{2}=\sum_{n=0}^{\infty}\widetilde{\varepsilon}_{n}\widetilde{\mathbf{P}}_{n}$ with $\widetilde{\varepsilon}_{n}=2nb,\quad\widetilde{\mathbf{P}}_{0}=\begin{pmatrix}0&0\\\ 0&\mathbf{P}_{0}\end{pmatrix},\quad\widetilde{\mathbf{P}}_{n}=\begin{pmatrix}\mathbf{P}_{n-1}&0\\\ 0&\mathbf{P}_{n}.\end{pmatrix}\text{ for }n\geq 1.$ The lowest Landau level has now energy $\widetilde{\varepsilon}_{0}=0$, and is only occupied by spin-down electrons. The corresponding eigenspace has density $\rho_{\widetilde{\mathbf{P}_{0}}}=\frac{b}{2\pi}$. The other Landau levels are <<doubly>> occupied, with density $\rho_{\widetilde{\mathbf{P}_{n}}}=\frac{b}{\pi}$. In this case, we consider density matrices $\gamma$ of the form (4.4) $\gamma=\sum_{n=0}^{\infty}\widetilde{\mathbf{P}_{n}}\otimes\widetilde{\gamma}_{n},\quad\widetilde{\gamma}_{n}\in{\mathcal{S}}(L^{2}(\mathbb{R},\mathbb{C})),\ 0\leq\widetilde{\gamma}_{n}\leq 1.$ ###### Remark 4.1. One can prove that the set of such states is smaller than the set of states $\gamma$ commuting with $\mathbf{P}^{\mathbb{A}}_{2}$ and the magnetic translations ${\mathfrak{m}}_{\mathbb{R}}$. This comes from the fact that the family $({\mathfrak{m}}_{\mathbb{R}})$ is not irreducible on $\widetilde{E}_{n}:={\rm Ran}\widetilde{P_{n}}$. The conclusion of Theorem 3.1 still holds in the spin setting, and the proof is similar with only minor modifications: we now consider the $G$ matrix defined by $G_{\gamma}=\frac{b}{2\pi}\left(\widetilde{\gamma}_{0}+2\sum_{n=1}^{\infty}\widetilde{\gamma}_{n}\right)\qquad\in{\mathcal{G}},$ and the optimal $m_{j}^{*}$ functions are given by $m_{j}^{*}(0)=\\{\frac{2\pi g_{j}}{b}\\}$ and $m_{j}^{*}(n)=0$ if $\frac{2\pi g_{j}}{b}<1$, and, if $\frac{2\pi g_{j}}{b}\geq 1$, $m_{j}^{*}(n)=\begin{cases}1&\quad\text{for}\quad n\leq\lfloor\frac{\pi}{b}g_{j}+\frac{1}{2}\rfloor-1\\\ \left\\{\frac{\pi}{b}g_{j}+\frac{1}{2}\right\\}&\quad\text{for}\quad n=\lfloor\frac{\pi}{b}g_{j}+\frac{1}{2}\rfloor\\\ 0&\quad\text{otherwise}.\end{cases}$ Denoting by $y=\frac{\pi}{b}g_{j}+\frac{1}{2}$, this gives the energy $\displaystyle\frac{b}{2\pi}\sum_{n=0}^{\infty}\widetilde{\varepsilon}_{n}m_{j}^{*}(n)$ $\displaystyle=\frac{b^{2}}{2\pi}\sum_{n=1}^{\lfloor y\rfloor-1}2n+\frac{b^{2}}{\pi}\lfloor y\rfloor\left\\{y\right\\}=\frac{b^{2}}{2\pi}\left(y^{2}-y+\\{y\\}(1-\\{y\\})\right).$ We obtain that for any representable density $\rho$. $\inf_{\gamma\ \text{of the form\leavevmode\nobreak\ \eqref{eq:gamma- spin2}}\atop\rho_{\gamma}=\rho}\left\\{\frac{1}{2}\underline{\rm Tr}(\mathbf{P}^{\mathbb{A}}_{3}\gamma)\right\\}=\inf_{G\in{\mathcal{G}}\atop\rho_{G}=\rho}\left\\{\frac{1}{2}{\rm Tr}(-\Delta_{1}G)+{\rm Tr}(F^{\rm spin}(b,G))\right\\},$ with the new functional $F^{\rm spin}(b,g)=\frac{}{}\frac{\pi}{2}g^{2}-\frac{b^{2}}{8\pi}+\frac{b^{2}}{2\pi}\left\\{\frac{\pi g}{b}+\frac{1}{2}\right\\}\left(1-\left\\{\frac{\pi g}{b}+\frac{1}{2}\right\\}\right).$ A plot of $F^{\rm spin}(b=1,g)$ is displayed in Figure 1. The constant $\frac{\pi}{2}$ in the $\frac{\pi}{2}g^{2}$ term is the two-dimensional Thomas-Fermi constant, when the spin of the electron is included. Again, we see that magnetic effects gives a correction to the Thomas-Fermi approximation, but this time, including the Zeeman term, the magnetic energy is always lower than the no-magnetic one, and we have (4.5) $\frac{\pi}{2}g^{2}-\frac{b^{2}}{8\pi}\leq F^{\rm spin}(b,g)\leq\frac{\pi}{2}g^{2}.$ Figure 1. (Left). The map $g\mapsto F(b=1,g)$ (red) together with its lower/upper bounds in (3.7). (Right) The map $g\mapsto F^{\rm spin}(b=1,g)$ (red) and its bounds (4.5). ## References * [1] X. Blanc and C. Le Bris. Thomas-Fermi type theories for polymers and thin films. Advances in Differential Equations, 5(7-9):977 – 1032, 2000. * [2] O. Bratteli and D.W. Robinson. Operator Algebras and Quantum Statistical Mechanics: C*- and W*-algebras, symmetry groups, decomposition of states. Texts and monographs in physics. Springer-Verlag, 1987. * [3] Vivek Chaudhary, Petr Neugebauer, Omar Mounkachi, Salma Lahbabi, and A. EL Fatimy. Phosphorene - an emerging two-dimensional material: recent advances in synthesis, functionalization, and applications. 2D Materials, 9, 05 2022. * [4] J.H. Condon. Nonlinear de Haas-van Alphen effect and magnetic domains in Beryllium. Phys. Rev., 145(2):526, 1966. * [5] W.J. De Haas and P.M. Van Alphen. The dependence of the susceptibility of diamagnetic metals upon the field. Proc. Netherlands Roy. Acad. Sci, 33(1106):170, 1930. * [6] I. Fushiki, E.H. Gudmundsson, C.J. Pethick, and J. Yngvason. Matter in a magnetic field in the Thomas-Fermi and related theories. Annals of Physics, 216(1):29–72, 1992. * [7] D. Gontier. N-representability in noncollinear spin-polarized density-functional theory. Phys. Rev. Lett., 111:153001, 2013. * [8] D. Gontier. Existence of minimizers for Kohn–Sham within the local spin density approximation. Nonlinearity, 28(1):57, 2014. * [9] D. Gontier, S. Lahbabi, and Maichine A. Density functional theory for homogeneous two dimensional materials. Comm. Math. Phys., 388:1475–1505, 2021. * [10] D. Gupta, V. Chauhan, and R. Kumar. A comprehensive review on synthesis and applications of molybdenum disulfide (mos2) material: Past and recent developments. Inorganic Chemistry Communications, 121:108200, 2020. * [11] Y. Kaddar, G. Tiouitchi, A. El Kenz, A . Benyoussef, P. Neugebauer, S. Lahbabi, A. El Fatimy, and O. Mounkachi. Black phosphorus synthesized by the chemical vapor transport: Magnetic properties induced by oxygen atoms. Submitted for publication, 2022. * [12] L.L. Lev, I.O. Maiboroda, M.A. Husanu, and et al. k-space imaging of anisotropic 2D electron gas in GaN/GaAlN high-electron-mobility transistor heterostructures. Nature Communication, 9(2653), 2018. * [13] E.H. Lieb and M. Loss. Analysis, volume 14. American Mathematical Soc., 2001. * [14] E.H. Lieb, J.P. Solovej, and J. Yngvason. Ground states of large quantum dots in magnetic fields. Phys Rev B Condens Matter, 51(16):10646–10665, 1995. * [15] R.S. Markiewicz, M. Meskoob, and C. Zahopoulos. Giant magnetic interaction (Condon domains) in two dimensions. Phys. Rev. Lett., 54(13):1436, 1985. * [16] J. Saha and A. Dutta. A review of graphene: Material synthesis from biomass sources. Waste and Biomass Valorization, 13, 03 2022. * [17] U. Von Barth and L. Hedin. A local exchange-correlation potential for the spin polarized case. I. Journal of Physics C: Solid State Physics, 5(13):1629, 1972. * [18] M. W. Wong. Wigner Transforms. Universitext. Springer, New York, 1998. * [19] Y. Yngvason. Thomas-Fermi theory for matter in a magnetic field as a limit of quantum mechanics. Letters in Mathemattcal Physics, 22:107–117, 1991.
XX 2022 ACML 2022 # Graph annotation generative adversarial networks Yoann Boget<EMAIL_ADDRESS> University of Geneva and Geneva School for Business administration HES-SO Rue de la Tambourine 17 Carouge Switzerland Magda Gregorova<EMAIL_ADDRESS> Center for Artificial Intelligence and Robotics (CAIRO) FHWS Franz-Horn-Strasse 2 Würzburg-Schweinfurt Germany Alexandros Kalousis<EMAIL_ADDRESS> Geneva School for Business administration HES-SO Rue de la Tambourine 17 Carouge Switzerland ###### Abstract We consider the problem of modelling high-dimensional distributions and generating new examples of data with complex relational feature structure coherent with a graph skeleton. The model we propose tackles the problem of generating the data features constrained by the specific graph structure of each data point by splitting the task into two phases. In the first it models the distribution of features associated with the nodes of the given graph, in the second it complements the edge features conditionally on the node features. We follow the strategy of implicit distribution modelling via generative adversarial network (GAN) combined with permutation equivariant message passing architecture operating over the sets of nodes and edges. This enables generating the feature vectors of all the graph objects in one go (in 2 phases) as opposed to a much slower one-by-one generations of sequential models, prevents the need for expensive graph matching procedures usually needed for likelihood-based generative models, and uses efficiently the network capacity by being insensitive to the particular node ordering in the graph representation. To the best of our knowledge, this is the first method that models the feature distribution along the graph skeleton allowing for generations of annotated graphs with user specified structures. Our experiments demonstrate the ability of our model to learn complex structured distributions through quantitative evaluation over three annotated graph datasets. ###### keywords: Graph, Annotation, Generative Model, GAN, adversarial. ††editors: Emtiyaz Khan and Mehmet Gonen ## 1 Introduction Modern deep learning approaches for learning high-dimensional data distributions and synthesizing new data examples have achieved great successes in a number of domains. We focus here on the particularly challenging problem of generating new data examples ${\mathbf{x}}_{i}\in\mathbb{R}^{m_{i}}$ with different dimensionalities $m_{i}$ between individual instances and complex relational structures in the feature spaces organized in graphs. More concretely, we focus on learning the distributions and generating new examples of features associated with nodes and edges of annotated graphs conditioned on the graph structure. We formulate the problem as that of modeling a conditional distribution, where we learn to generate the node and edge features conditionally on the given graph skeleton (non-annotated graph). In this respect, our paper complements the existing tool-set of models for non-annotated graph generations (e.g. You et al. (2018); Liao et al. (2019)). The non-annotated graph can be generated by one of those or, more interestingly, can be provided by the user based on the needs of a downstream task that requires a specific graph structure (e.g. basis for scaffold-based _de novo_ drug discovery or molecular docking, types and interactions of particles in high energy physic, individual features and types of relations in a social network). The graph annotation generative adversarial networks (GrannGAN) method that we propose follows the strategy of implicit distribution modeling via adversarial training Goodfellow et al. (2014), allowing us to sample new graphs from a distribution $p$ approximating the true $p^{*}$ without explicitly formulating the $p$ distribution function. In principle, the method can generate node of edge features independently. In practice, we generate the node and edge features in two phases depicted in Figure 1. In the first phase, our model samples the node features conditioning on the graph skeleton. In the second phase, it uses the generated node features as an additional conditional variable to sample the edge features. The entire model architecture relies on the permutation equivariant transformations of message passing neural networks (MPNN) Gilmer et al. . Thanks to these, we circumvent the difficulties related to the ordering of graph objects in their representation. This critical property of the method allows for efficient use of the model capacity, which needs to learn neither a particular heuristic for unique representation ordering (as in the case of linearized representations) nor the complete set of equivalent permutations. In the following, we describe the newly proposed GrannGAN method and provide some details of the technical implementation in sections 2. We then position the method within the existing state of research and document its performance competitive with the best of the well-established methods for graph generations on a set of experiments in section 3 and 4. We conclude with a discussion of possible future directions in section 5. ## 2 GrannGAN Let ${\mathcal{G}}=\\{{\mathcal{V}},{\mathcal{E}},V,E\\}$ be an undirected graph with a set of nodes (vertices) ${\mathcal{V}}$ and a set of edges ${\mathcal{E}}$ between pairs of nodes in the graph, $V$ and $E$ are the corresponding node and edge features. Let $\nu_{i}$ denote a node $i$ and ${\mathbf{v}}_{i}\in\mathbb{R}^{d}$ the features associated with that node. An edge $\epsilon_{ij}$ is connecting the pair of nodes $\nu_{i}$ and $\nu_{j}$ and ${\mathbf{e}}_{ij}\in\mathbb{R}^{c}$ are features corresponding to that edge. We consider only the case where the edges are undirected so that $\epsilon_{ij}=\epsilon_{ji}$ and ${\mathbf{e}}_{ij}={\mathbf{e}}_{ji}$, but the model can be easily extended to the case with directed edges. We further use the term _skeleton_ and the letter ${\mathcal{S}}$ to refer to the non- annotated graph ${\mathcal{S}}=\\{{\mathcal{V}},{\mathcal{E}}\\}$ corresponding to ${\mathcal{G}}$. ### 2.1 Model factorization In our approach, we model the underlying graph distribution in the following factorization $\displaystyle p({\mathcal{G}})=p({\mathcal{S}},V,E)=p({\mathcal{S}})p(V\rvert{\mathcal{S}})p(E\rvert{\mathcal{S}},V)\enspace.$ (1) Our model consists of modeling $p(V\rvert{\mathcal{S}})$ and $p(E\rvert{\mathcal{S}},V)$ with conditional GAN. Note that we could equivalently reverse the conditioning order and model $p(E\rvert{\mathcal{S}})$ and $p(V\rvert{\mathcal{S}},E)$. In our experiments, we found the first option yielding better results. So, we keep this ordering for the following. During training, we sample the graph skeleton from the data distribution $p^{*}({\mathcal{S}})$. At inference, we can either sample the graph skeleton or keep it fixed depending on the task. ### 2.2 Implicit data generation We use the Wasserstein-GAN (WGAN) Arjovsky et al. (2017) formulation of adversarial training with spectral normalization Miyato et al. (2018) in all linear layers of the critic to enforce 1-lipschitzness. To sample new examples of data ${\mathbf{x}}$ from a model distribution $p_{\theta}({\mathbf{x}})$ WGAN uses a generator $g_{\theta}$ mapping from a random latent variable ${\mathbf{z}}$ to the output ${\mathbf{x}}=g_{\theta}({\mathbf{x}})$. The generator is learned to minimize the Wasserstein-1 (or Earth-Mover) distance $W(p_{\theta}({\mathbf{x}}),p^{*}({\mathbf{x}}))$ between the implicit model distribution and the true generative distribution through a min-max optimization. $\displaystyle\min_{\theta}\max_{\varphi}\ \mathbb{E}_{(y)\sim p^{*}}f_{\varphi}({\mathbf{y}})-\mathbb{E}_{z\sim p(z)}f_{\varphi}\big{(}g_{\theta}({\mathbf{z}})\big{)}\enspace,$ (2) where $f_{\varphi}$ is the K-Lipschitz critic function mapping from the data ${\mathbf{y}}$ (real or generated) to a real-valued score $f_{\varphi}({\mathbf{y}})\in\mathbb{R}$. We use the conditional WGAN formulation twice in the GrannGAN pipeline sketched out in figure 1. Following the factorization in equation (1) we first model the conditional distribution $p(V\rvert{\mathcal{S}})$ in a _node- annotation_ phase. Here, the method generates the node features $\hat{V}=g_{V,\theta}({\mathcal{Z}},{\mathcal{S}})$ from the set of latent noise variables ${\mathcal{Z}}$ and the skeleton ${\mathcal{S}}$. So, equation 2 becomes $\displaystyle\min_{\theta}\max_{\varphi}\ \mathbb{E}_{(V,{\mathcal{S}})\sim p^{*}}f_{V,\varphi}(V,{\mathcal{S}})-\mathbb{E}_{z\sim p(z),{\mathcal{S}}\sim p^{*}}f_{V,\varphi}\big{(}g_{V,\theta}({\mathcal{Z}},{\mathcal{S}}),{\mathcal{S}}\big{)}\enspace.$ (3) Similarly in the following _edge-annotation_ phase, we model the conditional $p(E\rvert{\mathcal{S}},V)$. The edge features $\hat{E}=g_{E,\theta}({\mathcal{Z}},V,{\mathcal{S}})$ are generated conditionally on the previously generated node features by including these as an additional input variable to the generator and critic. In this phase, equation 2 becomes $\displaystyle\min_{\theta}\max_{\varphi}\ \mathbb{E}_{(E,V,{\mathcal{S}})\sim p^{*}}f_{E,\varphi}(E,V,{\mathcal{S}})-\mathbb{E}_{z\sim p(z),(V,{\mathcal{S}})\sim p^{*}}f_{E,\varphi}\big{(}g_{E,\theta}({\mathcal{Z}},V,{\mathcal{S}}),V,{\mathcal{S}}\big{)}\enspace,$ (4) Note that the conditioning on the skeleton ${\mathcal{S}}=({\mathcal{V}},{\mathcal{E}})$ acts on the computational graph of the critic $f$ and of the generator $g$, i.e. on the structure of the Message Passing Neural Network (MPNN, explained in detail in the section 2.4). This special way of conditioning through the computational graph is one of the novelties of our method. ### 2.3 Implementation In practice, we start by sampling the skeleton ${\mathcal{S}}=\\{{\mathcal{V}},{\mathcal{E}}\\}$ from the data. For all node $\nu_{i}\in{\mathcal{V}}$, we sample a random noise vector ${\mathbf{z}}_{i}$. So, we get ${\mathcal{Z}}=\\{{\mathbf{z}}_{i}\\}_{1}^{n}$ and, by construction $|{\mathcal{V}}|=|{\mathcal{Z}}|$. Doing so, we obtain a graph with random noise vectors as initial latent representation of the node features ${\mathcal{G}}_{0}=\\{{\mathcal{V}},{\mathcal{E}},{\mathcal{Z}}\\}$ (and without edge feature). This graph is the input of the _node-annotation_ generator. Similarly, for the edge generation, we sample a graph from the data without edge feature. We also sample ${\mathcal{Z}}=\\{{\mathbf{z}}_{i}\\}_{1}^{m}$, where $m=|{\mathcal{E}}|$. The input of the _edge-annotation_ generator is the graph ${\mathcal{G}}_{0}=\\{{\mathcal{V}},{\mathcal{E}},V,{\mathcal{Z}}\\}$ with latent random noise vectors as initial latent edge representation. The generators are permutation-equivarient functions. They output node or edge features following the order of the input. The critics receive alternatively real or generated graphs (without edge feature during the annotation step). The critics are permutation-invariant functions. We provide details the generators and the critics in the next subsection. The two steps (the two conditional GANs) are trained independently. Using teacher forcing, we use real data for the conditioning during training. For inference, we generate the edges features by conditioning on the previously generated node features. The Figure 1 illustrates the model architecture. The software implementation of our method together with instructions for replicating our experiments are available at https://github.com/yoboget/GrannGAN. In this section we provide some details of the implementation to help the reader understand important design decisions. Figure 1: GrannGAN architecture: grayscale boxes represent random noise, colored boxes represent node and edge features. ### 2.4 Message passing neural network In both the node- and the edge-annotation phases described in section 2.2 all GrannGAN generator $g$ and critic $f$ functions are message passing neural networks (MPNN) Gilmer et al. . As mentioned here-above, the node- neighbourhood structure of the MPNNs is derived from the conditioning on the graph skeleton $\\{{\mathcal{V}}{\mathcal{E}}\\}$. Therefore, the skeleton acts on the $i$ and $j$ indices. We use the following equations to perform $L$ update steps over the hidden states ${\mathbf{h}}_{i}^{(l)}$ and ${\mathbf{r}}_{ij}^{(l)}$ of each node and edge in the graph respectively $\displaystyle{\mathbf{r}}_{ij}^{(l+1)}$ $\displaystyle=\phi_{r}^{(l)}([{\mathbf{h}}_{i}^{(l)},{\mathbf{h}}_{j}^{(l)},{\mathbf{r}}_{ij}^{(l)}])$ (5) $\displaystyle\widetilde{{\mathbf{h}}}_{i}^{(l+1)}$ $\displaystyle=\sum_{j\in{\mathcal{N}}_{i}}\frac{1}{\sqrt{d_{i}}{\sqrt{d_{j}}}}{\mathbf{r}}_{ij}^{(l+1)}$ (6) $\displaystyle{\mathbf{h}}_{i}^{(l+1)}$ $\displaystyle=\phi_{h}^{(l)}([{\mathbf{h}}_{i}^{(l)},\widetilde{{\mathbf{h}}}_{i}^{(l+1)}])\enspace.$ (7) Here $\phi$ are learned differentiable functions (we use small feedforward networks), ${\mathcal{N}}_{i}$ is the first order neighbourhood of node ${\mathbf{v}}_{i}$ as given by the graph skeleton ${\mathcal{S}}$, $d_{i}$ is its degree, and $[{\bm{a}},{\bm{b}}]$ is the concatenation of vectors ${\bm{a}}$ and ${\bm{b}}$. We outline the whole generative pipeline in figure 1. In the generator of the node-annotation phase $g_{\theta_{\mathcal{V}}}$ we initiate the node hidden states with a latent random noise vectors ${\mathbf{h}}_{i}^{(0)}={\mathbf{z}}_{i},\,\forall{\mathbf{v}}_{i}\in{\mathcal{V}}$ (and drop ${\mathbf{r}}_{ij}^{(0)}$ from inputs to $\phi_{r}^{(0)}$). The generated node features are the hidden states of the last update step $\widehat{{\mathbf{v}}}_{i}={\mathbf{h}}_{i}^{(L)}$. In contrast, in the edge- annotation generator $g_{\theta_{\mathcal{E}}}$ the latent random noise is used for initiating the edge hidden states ${\mathbf{r}}_{ij}^{(0)}={\mathbf{z}}_{ij},\,\forall{\mathbf{e}}_{ij}\in{\mathcal{E}}$ while the node hidden states are the node features generated from the previous phase ${\mathbf{h}}_{i}^{(0)}=\widehat{{\mathbf{v}}}_{i}$. The generated edge features are the hidden edge states of the last update step $\widehat{\mathbf{e}}_{ij}={\mathbf{r}}_{ij}^{(L)}$. The critic functions MPNNs are initiated from the real data examples as ${\mathbf{h}}_{i}^{(0)}={\mathbf{v}}_{i}$ and ${\mathbf{r}}_{ij}^{(0)}={\mathbf{e}}_{ij}$ or from the synthetic examples of the generators respectively. To produce the critic scores, the last-step update functions ($\phi_{h}^{(L)}$ for the node critic and $\phi_{r}^{(L)}$ for the edge critic) have scalar outputs that are averaged to enter the loss in equation (2). Therefore, the critic evaluates each node, aggregating information from all the nodes included in a radius of $2(L-1)$. While this receptive field may not cover the whole graph, we assume that the node and edge features can be evaluated locally. It has been shown Arvind et al. (2020) that MPNN cannot capture graph substructures other than forests of stars. To alleviate this issue, we further embody the graph topology into the MPNNs by extending the node representation by a set of skeleton-related features. Similarly to Bouritsas et al. (2020), we complement the node hidden states at the initial step ${\mathbf{h}}_{i}^{(0)}$ of all generators and critics by the node degree $d_{i}$ and the number of $k$-cycles (cycles of length $k$ the node is part of) extracted from the graph skeleton. ### 2.5 Graph representations An important property of graph representation ${\mathcal{G}}$ are their non- uniqueness. There are in general $n!$ possible permutations $\pi$ determining the ordering of the nodes. The particular choice of the representation ordering from the complete set $\Pi=\\{\pi_{i}\\}_{i=1}^{n!}$ is therefore another source of stochasticity so that $\displaystyle p({\mathcal{G}})=p\big{(}\bigcup_{\pi\in\Pi}\pi({\mathcal{G}})\big{)}=\sum_{\pi\in\Pi}p\big{(}\pi({\mathcal{G}})\big{)}\enspace.$ (8) From equation (8) we observe that when relying on the ordered graph representations ${\mathcal{G}}$, one shall in principle model the complete set of distributions $p(\pi({\mathcal{G}}))$ for all the $n!$ permutations $\pi$ to capture the unordered-set graph distribution $p({\mathcal{G}})$. However, operating over the individual permutations $\pi({\mathcal{G}})$ would lead to inefficient use of the model capacity. As an alternative, previous methods often fallback to operating over a single representation of each graph using some heuristic to fix the canonical ordering (such as the minimum depth-first- search Goyal et al. (2020), the unique breadth-first-search You et al. (2018) or the further sequentiallized domain driven SMILES Daylight Chemical Information Systems (2011)). In result, the model has to learn not only the distribution of the graphs but also the ordering heuristic which is again inefficient from the perspective of generating the unordered graph sets ${\mathcal{G}}$. A fundamental property of GrannGAN is that it is completely insensitive to the ordering $\pi({\mathcal{G}})$ that can in result be chosen arbitrarily. This permuation equivariance111A function $f$ is equivariant with respect to permutation $\pi$ if $f(\pi({\bm{x}}))=\pi(f({\bm{x}})).$ of the message passing operations is pivotal for our method as it solves the problem of the graph representation non-uniqueness while allowing for efficient use of the model capacity. ## 3 Related work GrannGAN is closely related to the broad category of graph generative models, an area which has attracted significant attention of the research community resulting in a flurry of papers in the last several years. A systematic review of the major advancements in the field can be found for example in the excellent survey Faez et al. (2021). Unlike previous work on that field, our method focuses on the particular problem of generating node and edge features conditionally over a given graph skeleton. To the best of our knowledge, we are the first to investigate such structural-based conditional graph feature generation. The closest to our settings are methods modelling the distribution of annotated graphs. These can be categorized into two large families: the ones generating graphs sequentially and those generating in the whole graph in one go. ### 3.1 Sequential graph generation The first are methods generating the graphs sequentially, starting from small structures to which they gradually connect new graph components (nodes, edges, or complete sub-structures). Most of these focus on generating chemical molecules and adopt various measures to improve the performance on this particular generative problem. CharacterVAE Gómez-Bombarelli et al. (2018) and GrammarVAE Kusner et al. (2017) rely directly on sequential SMILES representation of the molecules. JT-VAE Jin et al. (2018) operates over hand- crafted vocabulary of chemically valid sub-structures. MolecularRNN Popova et al. (2019) and GraphAF Shi et al. (2019) generate the graphs by successive node and edge sampling steps and ensure validity of the generated molecules by valency checking and the possibility for resampling at each of these steps. Thanks to these domain-motivated components, the sequential methods often achieve excellent performance on molecular graphs datasets. However, those directly relying on chemistry-specific data representations (e.g SMILES) cannot be easily extended to beyond the molecular problems. On the other hand, methods not using the SMILES representation need to establish their own heuristic for the sequential traversal of the graph such as the breadth-first- search in GraphRNN You et al. (2018) which in turn needs to be learned by the generative model together with the graph data distribution. As any other sequential models, the graph autoregressive generators need to be particularly careful about capturing long-term dependencies, and suffer from slow sampling process. Different from the previous models NetGAN Bojchevski et al. (2018) uses GANs to learn the distribution of random walks over one big graph. ### 3.2 Annotated graph generation in one go The other large family of models, into which GrannGAN can also be related, are those generating the graphs in one go, that is all the nodes and edges and their features together. GraphVAE Simonovsky and Komodakis (2018) proposes to sidestep the problem of graph linearization characteristic for sequential models by relying on the variational auto-encoder framework Kingma and Welling (2014). Due to modelling the graph in the ordered representation of annotation and adjacency matrix, it needs to employ an expensive (though inexact) graph- matching procedure in the training loss calculation, which significantly hampers its scalability. The flow-based graphNVP Madhawa et al. (2019) adopts the coupling strategy of Dinh et al. (2017) applying it to the rows of the graph annotation and adjacency matrices and thus preserving the ordering of the nodes and edges through the flow. GraphNVP is the only model for annotated graph using permutation equivarient generative function. In general, annotated graph generation in one go is still an open field of research. Our contribution can be seen as a proposition to tackle this challenging issue. Misc-GAN Zhou et al. (2019) aims to translate graphs from a source-domain graph into a target-domain graph using a multiscale cycle-GAN. It operates on unannotated graphs. The closest to our GrannGAN is the MolGAN De Cao and Kipf (2018) model, which uses GAN for molecule generation. Unlike in GrannGAN, the MolGAN generator is a feed-forward network sampling the ordered annotation and adjacency matrices of the graph representation. The authors of MolGAN observe that the model tends to suffer from mode collapse resulting in insufficient variation in the generated samples. We suppose that this issue comes from the loss of capacity by learning to generate various permutations of the same graph and from an additional Reinforcement Learning module encouraging the generator to produce graphs with some specific properties. We never experienced mode collapse during our experiments. As we show it in the experiment section, our model presents excellent uniqueness and novelty rates. We compare to MolGAN and other methods discussed here in our experiments in section 4. ## 4 Experiments In this section, we present experiments using our model as a graph generative model. There currently exists no other models conditioning on the graph skeleton and, therefore, no other method to directly compare with. Instead, we use generative models presented in the previous section as baselines. However, we underline that these models do not have the same purpose. In the second part of this section, we also present the results of the conditional generation by fixing the skeleton. We evaluate our method in a set of experiments over three datasets containing (node- and edge-) annotated graphs222Unfortunately, we did not found other public dataset with enough node- and edge-annotated graphs.. Two of the datasets, QM9 Ramakrishnan et al. (2014) and ZINC Sterling and Irwin (2015), are from the chemical domain that is frequently used as a test bed for annotated graph modeling. The third is the fingerprint dataset Riesen and Bunke (2008) included in the TUDataset collection Morris et al. (2020). As it has recently been discussed, for example in O’Bray et al. (2021) and Thompson et al. (2021), evaluating generative models of graph structures is particularly challenging due to the impracticality of visual or perceptual comparisons of the data examples. The comparison with existing graph generative models is therefore difficult and can provide only crude indication of the method capabilities. For the chemical datasets the most common evaluation metrics are the following * • _validity_ is the proportion of chemically valid molecules in the total generated examples and measures the ability of the method to understand the chemical constraints differentiating general graphs from chemically valid molecules * • _uniqueness_ is the proportion of unique samples within the _valid_ generated samples and measures the variability of the generated data * • _novelty_ is the proportion of examples within the _valid_ and _unique_ set that do not exist in the training dataset and measures the ability of the method to go beyond data memorization We complement these by an overall score of _valid-unique-novel_ molecules calculated simply as the product of the three above metrics and measuring the overall quality of the generated examples as a proportion of samples with all three of these desirable properties in the set of generated data. So, the overall score gives the rate of valid molecules that are neither in the dataset nor already generated. Note that in most cases the overall score is the metric of interest. While these evaluation metrics may be useful indicators for chemical downstream tasks, they do not provide any measure of the distance between the distribution of reference and the generated distributions. Various indicators calling on the framework of maximum mean discrepancy (MMD) Gretton et al. (2012) have been proposed in the literature. These are used very inconsistently and, as demonstrated in O’Bray et al. (2021), they are highly sensitive to the specific choice of the graph statistics, the kernel and the hyperparameters used for the MMD calculation. Instead of calculating the MMDs on graphs statistics, we report directly the distance between the generated feature distributions, using the Jenson-Shannon Distance for each feature. We show that with respect to these metrics, our model outperforms by far GraphAF, model considered as the state-of-the-art. ### 4.1 QM9 QM9 is one of the most commonly used datasets for testing models for annotated graph generation. It consists of $\sim$134k stable organic molecules with up to 9 atoms of four types. The 4 atom types and 4 bond types are encoded as one-hot vectors in the node feature descriptions ${\mathbf{v}}$. In table 1 we present an overview of the generative performance of our GrannGAN method in comparison with a set of well-established graph generative methods: MolGAN De Cao and Kipf (2018), GraphVAE Simonovsky and Komodakis (2018), GraphNVP Madhawa et al. (2019), GraphAF Shi et al. (2019), CharacterVAE Gómez-Bombarelli et al. (2018), GrammarVAE Kusner et al. (2017), and JT-VAE Jin et al. (2018). The GrannGAN results are calculated from 1000 new data examples generated by conditioning on skeletons randomly sampled from the training data. The results of the baseline methods are those reported by the authors in the original papers333CharacterVAE and GrammarVAE results (not reported in the original papers) are taken over from the method replications in GraphNVP.. Despite the frailty of such comparisons due to inconsistencies in the original experimental protocols (for example, some of the methods, such as GraphAF, work over the _kekulized_ versions of the molecules reducing the number of edge categories to three), our GrannGAN achieves excellent results. It outperforms all one-go generative models and is on par with the best auto- regressive model. Table 1: QM9: performance comparison | Model | valid | unique | novel | overall ---|---|---|---|---|--- no chemistry | GrannGAN | 82.5 | 99.9 | 64.9 | 53.4 MolGAN (wo. RL) | 87.7 | 2.9 | 97.7 | 2.5 GraphVAE | 55.7 | 76.0 | 61.6 | 26.1 GraphNVP | 83.1 | 58.2 | 99.2 | 46.8 GraphAF (wo. validity) | 67.0 | 94.5 | 88.8 | 56.3 chemistry | MolGAN-RL | 99.8 | 2.3 | 97.9 | 2.2 CharacterVAE | 10.3 | 90.0 | 67.5 | 11.9 GrammarVAE | 60.2 | 80.9 | 9.3 | 11.9 GraphAF | 100 | 94.5 | 88.8 | 83.9 As explained above, the metrics in the table are very much domain specific. Methods in the lower part of table 1 expressly focus on these introducing into the models various expert-designed modules promoting chemical validity of the generated molecular graphs (e.g., valid sub-substructures vocabulary, valency constraints, rejection sampling, etc.). While these improve the targeted metrics in table 1, they also bias the generative process and, therefore, the model distribution towards the validity metric in exchange for the distribution approximation goal. We list these methods in table 1 for completeness despite them not being really comparable to our approach which focuses on the distribution approximation of general graphs. Figure 2: QM9: empirical distributions of real and generated features (samples of 1000 graphs) Table 2: Jenson-Shannon distances between real and generated distributions for node and edge features of the QM9 and ZINC datasets (the lower the better) Model | nodes - QM9 | edges - QM9 | nodes - ZINC | edges - ZINC ---|---|---|---|--- GrannGAN | $1.87\cdot 10^{-3}$ | $0.04\cdot 10^{-4}$ | $1.50\cdot 10^{-3}$ | $0.60\cdot 10^{-3}$ GraphAF | $15.80\cdot 10^{-3}$ | $11.5\cdot 10^{-4}$ | $15.5\cdot 10^{-3}$ | $3.39\cdot 10^{-3}$ Figure 2 documents the ability of the GrannGAN method to learn the feature distribution by comparing the empirical real data distribution with the distribution of the generated examples. We have replicated the results of the GraphAF method, which is the most competitive with GrannGAN according to table 1), to extract similar statistics. GrannGAN matches the feature distributions of the reference dataset almost exactly while we observe a less closer match in the GraphAF444GraphAF uses somewhat different training set than we. For example, by kekulizing the molecules it removes all aromatic edges and thus reduces the number of edge types to 3. This is true also for the ZINC dataset. generated data. This observation is numerically confirmed in table 2 listing the Jensen-Shannon distances (JSD) between the real and generated data samples, where GrannGAN is systematically better by one order of magnitude. ### 4.2 ZINC We use here the ZINC250k version used previously for generative modeling evaluations. It consists of 250k randomly selected molecules from the complete ZINC set with up to 38 atoms of 9 types and 4 bond types. Table 3 summarizes the generative performance of GrannGAN over 1000 newly synthesized examples compared to results reported in the literature. MolGAN and GraphVAE did not experiment (or did not provide the results) on this more challenging dataset of larger graphs. GrannGAN scales to this dataset easily and performs competitively with respect to the chemically motivated metrics (keeping in mind the difficulties of such comparisons related to differences in experimental protocols and evaluation procedures). Figure 3 documents the excellent performance of GrannGAN in learning the feature distributions, matching it closely for both the node and edge features. The differences in the true and GraphAF generated data distributions are more pronounced as is also clear from the numerical evaluation via the JSD presented in table 2. Table 3: ZINC: performance comparison | Model | valid | unique | novel | overall ---|---|---|---|---|--- no chemistry | GrannGAN | 56.5 | 100 | 100 | 56.5 MolGAN (wo. RL) | — | — | — | — GraphVAE | 13.5 | — | — | — GraphNVP | 42.6 | 94.8 | 100 | 40.4 GraphAF (wo. validity) | 68.0 | 99.1 | 100 | 67.3 chemistry | MolGAN-RL | — | — | — | — CharacterVAE | 7.2 | 9.0 | 100 | 0.6 GrammarVAE | 0.7 | 67.5 | 100 | 0.5 JT-VAE | 100 | 100 | 100 | 100 GraphAF | 100 | 99.1 | 100 | 99.1 Figure 3: ZINC: empirical distributions of real and generated features (samples of 1000 graphs) ### 4.3 Fingerprints Fingerprint is a dataset of graphs representing the skeletonized regions of interest in human fingerprints Riesen and Bunke (2008). The graphs consist of up to 26 nodes. Both nodes and edges are described by 2 continuous features related to their position and orientation in the 2d fingerprint image. We have linearly re-scaled both node and edge features between -1 and 1. We use the non-annotated graphs as the skeletons for new feature generations. Each row displays graphs arranged according to the new features generated by GrannGAN. Figure 4: Fingerprints: empirical distributions ${\mathbf{v}}_{1}$, ${\mathbf{v}}_{2}$ and ${\mathbf{d}}$ sampled from our GrannGAN and from the dataset (samples of 1000 graphs) Figure 5: Fingerprints: empirical distributions ${\mathbf{e}}_{1}$ and ${\mathbf{e}}_{2}$ sampled from our GrannGAN and from the dataset (samples of 1000 graphs) In table 4 we list the JS distance (with the logarithm in base 2) between the empirical distributions of the training data and newly generated features starting from 500 randomly selected skeletons. These have been calculated by discretizing the features into 200 even sized bins. The columns in the table correspond to the two node ${\mathbf{v}}$ and edge ${\mathbf{e}}$ features respectively. Following the same methodology, we also report the JS distance in the distribution of the euclidean distances between connected nodes ${\mathbf{d}}$. Unlike the feature distribution, this statistic consider the relation between node features. In figures 4 and 5, we show the node and edge features distribution for the data and newly generated examples from our model as well as the distribution of distances between node ${\mathbf{d}}$555We assume that the oscillation of the curve is just an artifact coming from the relatively small data points by bin. Table 4: Jenson-Shannon distances between real and generated distributions for node and edge features of the Fingerprint dataset (the lower the better) | v1 | v2 | d | e1 | e2 ---|---|---|---|---|--- GrannGAN | $5.02\cdot 10^{-2}$ | $4.55\cdot 10^{-2}$ | $32.84\cdot 10^{-2}$ | $16.78\cdot 10^{-2}$ | $9.74\cdot 10^{-2}$ Data samples | $1.13\cdot 10^{-2}$ | $1.27\cdot 10^{-2}$ | $1.49\cdot 10^{-2}$ | $1.37\cdot 10^{-2}$ | $1.07\cdot 10^{-2}$ Again, our model shows excellent performance in closely matching the node and edge features and their relations. These experiments show the quality of our GrannGAN for the generative task. It outperforms all previous one-go generative models. The sequential GraphAF is the only model presenting slightly better results using the molecular metrics. However, we show that our model is much better at capturing the feature distributions. ### 4.4 Conditional generation In this section, we show that our model can be used for conditional generation by fixing the skeleton and still producing a high level of diversity. We want to guarantee that the sample diversity does not come mainly from the skeleton. To test the generative capacity of our GrannGAN, we randomly sampled 100 skeletons from the ZINC dataset. We generate 1000 new instances conditioned on each sampled skeleton. We compute the average rate of uniqueness, i.e. the rate of unique instances over all the valid molecules. Table 5 reports the results of the experiment. As expected, the validity rate is similar to the one by sampling a new skeleton for each instance. Even by fixing the skeleton, we reached a rate of 92.5% unique instances on average. On average, the conditional generation of 1000 instances over the same skeleton produces 527.0 unique molecules. So, we show that the performance of our model does not depend critically on the sampling of multiple skeletons and can be properly used as a conditional generative model. Table 5: Conditional generation of 1000 instances from the same skeleton (average) Model | Validity (%) | Uniqueness (%) | Valid and unique (‰) ---|---|---|--- GrannGAN | $57.0$ | $92.5$ | $527.0$ We have demonstrated that the GrannGAN is competitive with state-of-the-art model for graph generation even though it is first meant for conditional generation. GraphAF is slightly better on the molecular metric, but much worst at matching the feature distribution. In addition, our model generates instances in one go and is therefore much faster during inference. More fundamentally, our model preserves its quality while generating new instances conditioned on the skeleton, a task that no other model can do. ## 5 Conclusions We have presented here a new method, GrannGAN, for implicit distribution learning and generating features of nodes and edges conditioned on graph skeletons. The promising results of the method, as documented in our experiments, indicate that the method can learn the complex high-dimensional distributions and generate new data examples with features coherent with the underlying graph structure. When the generation of new skeletons is not crucial, GrannGAN can be used as a full generative model by sampling the skeleton from the data. It is also the first method for graph generation conditioned on the skeleton. These favorable results provide a solid starting point for multiple possible extensions in future work. For example, in some applications (e.g. bio- chemistry) a reasonable starting point for the feature generations may be a partly annotated graph (in place of a completely non-annotated graph). Our model, which is already conditional, could be easily adapted to this setting by re-using the ideas of our second edge-annotation phase. Another interesting extension is that of property optimization. An additional conditioning variable could be included in the generators and critics to guide the model in generating property-conditioned features. The contribution of Magda Gregorova has been supported by High-Tech Agenda Bayern. ## References * Arjovsky et al. (2017) Martin Arjovsky, Soumith Chintala, and Léon Bottou. Wasserstein Generative Adversarial Networks. In _Proceedings of the 34th International Conference on Machine Learning_ , pages 214–223. PMLR, July 2017. URL https://proceedings.mlr.press/v70/arjovsky17a.html. ISSN: 2640-3498. * Arvind et al. (2020) V. Arvind, Frank Fuhlbrück, Johannes Köbler, and Oleg Verbitsky. On Weisfeiler-Leman invariance: Subgraph counts and related graph properties. _Journal of Computer and System Sciences_ , 113:42–59, nov 2020. ISSN 10902724. * Bojchevski et al. (2018) Aleksandar Bojchevski, Oleksandr Shchur, Daniel Zügner, and Stephan Günnemann. NetGAN: Generating graphs via random walks. In Jennifer Dy and Andreas Krause, editors, _Proceedings of the 35th International Conference on Machine Learning_ , volume 80 of _Proceedings of Machine Learning Research_ , pages 610–619. PMLR, 10–15 Jul 2018. URL https://proceedings.mlr.press/v80/bojchevski18a.html. * Bouritsas et al. (2020) Giorgos Bouritsas, Fabrizio Frasca, Stefanos Zafeiriou, and Michael M. Bronstein. Improving Graph Neural Network Expressivity via Subgraph Isomorphism Counting. _IEEE Transactions on Pattern Analysis and Machine Intelligence_ , jun 2020. URL http://arxiv.org/abs/2006.09252. * Daylight Chemical Information Systems (2011) Inc. Daylight Chemical Information Systems. Daylight Theory Manual, August 2011. URL https://www.daylight.com/dayhtml/doc/theory/. * De Cao and Kipf (2018) Nicola De Cao and Thomas Kipf. MolGAN: An implicit generative model for small molecular graphs. _arXiv:1805.11973 [cs, stat]_ , May 2018. URL http://arxiv.org/abs/1805.11973. arXiv: 1805.11973. * Dinh et al. (2017) Laurent Dinh, Jascha Sohl-Dickstein, and Samy Bengio. Density estimation using real NVP. In _5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings_. OpenReview.net, 2017. URL https://openreview.net/forum?id=HkpbnH9lx. * Faez et al. (2021) Faezeh Faez, Yassaman Ommi, Mahdieh Soleymani Baghshah, and Hamid R. Rabiee. Deep Graph Generators: A Survey. _IEEE Access_ , 9:106675–106702, 2021. ISSN 2169-3536. Conference Name: IEEE Access. * (9) Justin Gilmer, Samuel S. Schoenholz, Patrick F. Riley, Oriol Vinyals, and George E. Dahl. Neural message passing for quantum chemistry. In _Proceedings of the 34th International Conference on Machine Learning_ , pages 1263–1272. PMLR. URL https://proceedings.mlr.press/v70/gilmer17a.html. ISSN: 2640-3498. * Goodfellow et al. (2014) Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative Adversarial Nets. In _Advances in Neural Information Processing Systems_ , volume 27. Curran Associates, Inc., 2014. URL https://papers.nips.cc/paper/2014/hash/5ca3e9b122f61f8f06494c97b1afccf3-Abstract.html. * Goyal et al. (2020) Nikhil Goyal, Harsh Vardhan Jain, and Sayan Ranu. GraphGen: A Scalable Approach to Domain-agnostic Labeled Graph Generation. 2020\. URL https://doi.org/10.1145/3366423.3380201. * Gretton et al. (2012) Arthur Gretton, Karsten M. Borgwardt, Malte J. Rasch, Bernhard Schölkopf, and Alexander Smola. A Kernel Two-Sample Test. _Journal of Machine Learning Research_ , 13(25):723–773, 2012. ISSN 1533-7928. URL http://jmlr.org/papers/v13/gretton12a.html. * Gómez-Bombarelli et al. (2018) Rafael Gómez-Bombarelli, Jennifer N. Wei, David Duvenaud, José Miguel Hernández-Lobato, Benjamín Sánchez-Lengeling, Dennis Sheberla, Jorge Aguilera-Iparraguirre, Timothy D. Hirzel, Ryan P. Adams, and Alán Aspuru-Guzik. Automatic Chemical Design Using a Data-Driven Continuous Representation of Molecules. _ACS Central Science_ , 4(2):268–276, February 2018. ISSN 2374-7943. URL https://doi.org/10.1021/acscentsci.7b00572. Publisher: American Chemical Society. * Jin et al. (2018) Wengong Jin, Regina Barzilay, and Tommi Jaakkola. Junction Tree Variational Autoencoder for Molecular Graph Generation. In _Proceedings of the 35th International Conference on Machine Learning_ , pages 2323–2332. PMLR, July 2018. URL https://proceedings.mlr.press/v80/jin18a.html. ISSN: 2640-3498. * Kingma and Welling (2014) Diederik P. Kingma and Max Welling. Auto-encoding variational bayes. In Yoshua Bengio and Yann LeCun, editors, _2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference Track Proceedings_ , 2014. URL http://arxiv.org/abs/1312.6114. * Kusner et al. (2017) Matt J. Kusner, Brooks Paige, and José Miguel Hernández-Lobato. Grammar Variational Autoencoder. In _Proceedings of the 34th International Conference on Machine Learning_ , pages 1945–1954. PMLR, July 2017. URL https://proceedings.mlr.press/v70/kusner17a.html. ISSN: 2640-3498. * Liao et al. (2019) Renjie Liao, Yujia Li, Yang Song, Shenlong Wang, Will Hamilton, David K Duvenaud, Raquel Urtasun, and Richard Zemel. Efficient Graph Generation with Graph Recurrent Attention Networks. In _Advances in Neural Information Processing Systems_ , volume 32. Curran Associates, Inc., 2019. URL https://proceedings.neurips.cc/paper/2019/hash/d0921d442ee91b896ad95059d13df618-Abstract.html. * Madhawa et al. (2019) Kaushalya Madhawa, Katsuhiko Ishiguro, Kosuke Nakago, and Motoki Abe. GraphNVP: an Invertible Flow-based Model for Generating Molecular Graphs. September 2019. URL https://openreview.net/forum?id=ryxQ6T4YwB. * Miyato et al. (2018) Takeru Miyato, Toshiki Kataoka, Masanori Koyama, and Yuichi Yoshida. Spectral Normalization for Generative Adversarial Networks. February 2018. URL https://openreview.net/forum?id=B1QRgziT-&foo=bar. * Morris et al. (2020) Christopher Morris, Nils M. Kriege, Franka Bause, Kristian Kersting, Petra Mutzel, and Marion Neumann. Tudataset: A collection of benchmark datasets for learning with graphs. In _ICML 2020 Workshop on Graph Representation Learning and Beyond (GRL+ 2020)_ , 2020. URL www.graphlearning.io. * O’Bray et al. (2021) Leslie O’Bray, Max Horn, Bastian Rieck, and Karsten Borgwardt. Evaluation Metrics for Graph Generative Models: Problems, Pitfalls, and Practical Solutions. September 2021. URL https://openreview.net/forum?id=tBtoZYKd9n. * Popova et al. (2019) Mariya Popova, Mykhailo Shvets, Junier Oliva, and Olexandr Isayev. MolecularRNN: Generating realistic molecular graphs with optimized properties. _arXiv:1905.13372 [cs, q-bio, stat]_ , May 2019. URL http://arxiv.org/abs/1905.13372. arXiv: 1905.13372. * Ramakrishnan et al. (2014) Raghunathan Ramakrishnan, Pavlo O. Dral, Matthias Rupp, and O. Anatole von Lilienfeld. Quantum chemistry structures and properties of 134 kilo molecules. _Scientific Data_ , 1(1):140022, August 2014. ISSN 2052-4463. URL https://www.nature.com/articles/sdata201422. Number: 1 Publisher: Nature Publishing Group. * Riesen and Bunke (2008) Kaspar Riesen and Horst Bunke. IAM graph database repository for graph based pattern recognition and machine learning. In _Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)_ , volume 5342 LNCS, pages 287–297. Springer, Berlin, Heidelberg, 2008. ISBN 3540896880. URL http://www.iam.unibe.ch/fki/databases/iam-graph-database. * Shi et al. (2019) Chence Shi, Minkai Xu, Zhaocheng Zhu, Weinan Zhang, Ming Zhang, and Jian Tang. GraphAF: a Flow-based Autoregressive Model for Molecular Graph Generation. September 2019. URL https://openreview.net/forum?id=S1esMkHYPr. * Simonovsky and Komodakis (2018) Martin Simonovsky and Nikos Komodakis. GraphVAE: Towards Generation of Small Graphs Using Variational Autoencoders. _arXiv:1802.03480 [cs]_ , February 2018. URL http://arxiv.org/abs/1802.03480. arXiv: 1802.03480. * Sterling and Irwin (2015) Teague Sterling and John J. Irwin. ZINC 15 – Ligand Discovery for Everyone. _Journal of Chemical Information and Modeling_ , 55(11):2324–2337, November 2015. ISSN 1549-9596. URL https://doi.org/10.1021/acs.jcim.5b00559. Publisher: American Chemical Society. * Thompson et al. (2021) Rylee Thompson, Boris Knyazev, Elahe Ghalebi, Jungtaek Kim, and Graham W. Taylor. On Evaluation Metrics for Graph Generative Models. September 2021. URL https://openreview.net/forum?id=EnwCZixjSh. * You et al. (2018) Jiaxuan You, Rex Ying, Xiang Ren, William Hamilton, and Jure Leskovec. GraphRNN: Generating Realistic Graphs with Deep Auto-regressive Models. In _Proceedings of the 35th International Conference on Machine Learning_ , pages 5708–5717. PMLR, July 2018. URL https://proceedings.mlr.press/v80/you18a.html. ISSN: 2640-3498. * Zhou et al. (2019) Dawei Zhou, Lecheng Zheng, Jiejun Xu, and Jingrui He. Misc-gan: A multi-scale generative model for graphs. _Frontiers in Big Data_ , 2:3, 4 2019. ISSN 2624909X.
# The Canonical BV Laplacian on Half-Densities Alberto S. Cattaneo Institut für Mathematik, Universität Zürich Winterthurerstrasse 190, CH-8057 Zürich, Switzerland<EMAIL_ADDRESS> ###### Abstract. This is a didactical review on the canonical BV Laplacian on half-densities. I acknowledge partial support of the SNF Grant No. 200020_192080, of the Simons Collaboration on Global Categorical Symmetries, and of the COST Action 21109 (CaLISTA). This research was (partly) supported by the NCCR SwissMAP, funded by the Swiss National Science Foundation. > _Das Bekannte überhaupt ist darum, weil es bekannt ist, nicht erkannt_ > > GWFH ###### Contents 1. 1 Introduction 1. 1.1 Overview 2. 2 The standard BV Laplacian 3. 3 The BV bracket 4. 4 The BV Laplacian on functions 1. 4.1 Digression: further properties of the BV Laplacian 5. 5 The BV Laplacian on densities 1. 5.1 Digression: Darboux expression of $\Delta_{\mu}^{(s)}$ 6. 6 Odd cotangent bundles 7. 7 Conclusion of the argument 8. A Some background details 1. A.1 The Lie derivative 2. A.2 Densities 9. B Applications of the BV formalism 10. C Historical remarks ## 1\. Introduction The BV Laplacian on functions is a central ingredient of the Batalin–Vilkovisky (BV) formalism. Its global definition is not canonical, yet a remarkable fact observed by Khudaverdian [7, 8] is that its extension to half-densities is so, and there are various ways to show it. In this review we give a short, self-contained presentation which requires just a few simple computations. It gives a special emphasis on the fact that half-densities have a natural inner product. This note has no pretence of originality. In Appendix C, we give a short historical overview with the relevant references. In Appendix A, we collect some technical background material. In Appendix B, we recall why the BV formalism is important in applications. ### 1.1. Overview We recall definitions and properties of BV Laplacians (on functions and densities) on an oriented odd symplectic manifold with the goal of proving that on half-densities it is canonically defined (whereas on other kinds of densities, including functions, it requires the choice of a compatible density). This goes as follows: 1. (1) On functions we define $\Delta_{\mu}f=\frac{1}{2}{\mathrm{div}}_{\mu}X_{f}$, where $\mu$ is an even, nowhere vanishing111Throughout the paper, nowhere vanishing means nonvanishing at every point of the body after the nilpotents are set to zero. An even, nowhere vanishing density $\mu$ is the same as a basis for the module of densities. The $k$-density $\mu^{k}$ is then a basis for the module of $k$-densities. density, and $X_{f}$ denotes the hamiltonian vector field of $f$. 2. (2) On half-densities we define $\Delta^{(\frac{1}{2})}_{\mu}\sigma=\Delta_{\mu}(\sigma/\mu^{1/2})\mu^{1/2}$. 3. (3) We check that the leading term is invariant under symplectomorphisms and does not depend on the choice of $\mu$. 4. (4) We check that $\Delta^{(\frac{1}{2})}_{\mu}$ is a symmetric operator. This implies that in any Darboux chart only the zeroth-order term may depend on $\mu$. 5. (5) We show that for a compatible density $\mu$, i.e., for $\mu$ satisfying $\Delta_{\mu}^{2}=0$, the zeroth-order term of the previous point must vanish. This shows that $\Delta^{(\frac{1}{2})}_{\mu}$ does not depend on the choice of the compatible density $\mu$. 6. (6) We show, via the presentation of the given odd symplectic manifold as a symplectomorphic odd cotangent bundle, that compatible densities always exist. 7. (7) We conclude that the BV Laplacian on half-densities, with the property of squaring to zero, is canonically defined. 8. (8) As an aside, we also show how to define this canonical operator, denoted by $\Delta$, directly on Darboux charts without any choice involved. Moreover, we prove that the compatibility condition for a density $\mu$ can also be expressed as $\Delta\mu^{\frac{1}{2}}=0$ and that the BV Laplacian on functions can now be recovered as $\Delta_{\mu}=\Delta(f\mu^{\frac{1}{2}})/\mu^{\frac{1}{2}}$. We go through details, also checking signs carefully, to make this presentation useful also as a reference. However, it should be observed that, roughly, the above flow can be easily checked without going into sign details—the only crucial point being (4), which requires some care. ###### Acknowledgment. I thank A. Cabrera, D. Fiorenza, N. Moshayedi, M. Schiavina, and P. Ševera for useful remarks. I am especially grateful to T. Voronov for several extremely helpful exchanges. ## 2\. The standard BV Laplacian On functions of odd variables $p_{1},\dots,p_{n}$ and even variables $q^{1},\dots,q^{n}$, the standard BV Laplacian222This operator was introduced to show that certain integrals are invariant under deformations of the integration domain; see Appendix B. (a.k.a. BV Laplace operator or BV operator) is defined as $\triangle\coloneqq\partial_{i}\partial^{i},$ where we use Einstein’s summation convention and the shorthand notations $\partial_{i}=\frac{\partial}{\partial q^{i}}\quad\text{and}\quad\partial^{i}=\frac{\partial}{\partial p_{i}}.$ Another ingredient of the BV formalism is the BV bracket (1) $(f,g)\coloneqq f\overleftarrow{\partial{{}^{i}}}\,\overrightarrow{\partial{{}_{i}^{\phantom{i}}}}g-f\overleftarrow{\partial{{}_{i}^{\phantom{i}}}}\,\overrightarrow{\partial{{}^{i}}}g,$ where $f$ and $g$ are functions of homogeneous degree, and the arrows denote the directions the derivatives are applied from. Explicitly, $\overrightarrow{\partial{{}_{i}^{\phantom{i}}}}f=\partial_{i}f$, $\overrightarrow{\partial{{}^{i}}}f=-\partial^{i}f$, and $\begin{split}f\overleftarrow{\partial{{}_{i}^{\phantom{i}}}}&=\partial_{i}f,\\\ f\overleftarrow{\partial{{}^{i}}}&=-(-1)^{f}\partial^{i}f.\end{split}$ Here and in the following, when we put a function (or any other object) as an exponent of $(-1)$ we mean its degree.333A more precise but more cumbersome notation would be $(-1)^{|f|}$, where $|f|$ denotes the degree of $f$. The standard BV Laplacian has the following properties which are easily proved by direct computation: (2a) $\displaystyle\triangle^{2}=0,$ (2b) $\displaystyle\triangle(fg)=(\triangle f)g+(-1)^{f}f\triangle g-(-1)^{f}(f,g),$ (2c) $\displaystyle\triangle\mathrm{e}^{S}=\left(\triangle S-\frac{1}{2}(S,S)\right)\mathrm{e}^{S}.$ Here again $f$ and $g$ are functions of homogeneous degree, whereas the function $S$ is assumed to be even. As an exercise, just using (2a) and (2b), you can also show that (3) $\triangle(f,g)=(\triangle f,g)-(-1)^{f}(f,\triangle g).$ The geometric interpretation of the content of this section is that the $p_{i}$s and $q^{i}$s are Darboux coordinates for the odd symplectic form $\omega={\mathrm{d}}p_{i}{\mathrm{d}}q^{i}$ with $(\ ,\ )$ its associated odd Poisson bracket.444If the $q^{i}$s are coordinates on an open subset $U$ of ${\mathbb{R}}^{n}$, the functions of the $p$ and $q$ variables can be identified with multivector fields on $U$. The BV bracket then gets identified with the Schouten–Nijenhuis bracket. We will return to this, with a general basis manifold, in Section 6; see footnote 15 on page 15. The main problem with this, however, is that the standard BV Laplacian does not transform well under symplectomorphisms, so it cannot be used as such to define an operator on functions on an odd symplectic manifold. In the next sections, we will see how to obviate this problem. ## 3\. The BV bracket Let $(\mathcal{M},\omega)$ be an odd symplectic manifold. Given a function $f$, we denote by $X_{f}$ its hamiltonian vector field defined via $\iota_{X_{f}}\omega={\mathrm{d}}f.$ Given a second function $g$, we define the BV bracket as $(f,g)\coloneqq(-1)^{f+1}X_{f}(g).$ This is an odd Poisson bracket, which in local Darboux coordinates agrees with the one of the previous section:555This is a simple computation which however requires checking signs. It is just needed to make sure that the conventions we choose are compatible. ###### Lemma 3.1. In Darboux coordinates, $\omega={\mathrm{d}}p_{i}{\mathrm{d}}q^{i}$, the BV bracket agrees with the standard one defined in (1). ###### Proof. Writing $X=X^{i}\partial_{i}+X_{i}\partial^{i}$, we get666We use the convention of total degree, i.e., internal degree plus parity of the form degree. This means that ${\mathrm{d}}$ is odd and that $\iota_{X}$ has parity opposite to that of $X$. Moreover, $X^{i}$ has the same parity as $X$, whereas $X_{i}$ has opposite parity. Finally, $f$ and $X_{f}$ have opposite parity to each other. $\iota_{X}\omega={\mathrm{d}}p_{i}\,X^{i}+X_{i}\,{\mathrm{d}}q^{i}={\mathrm{d}}p_{i}\,X^{i}-(-1)^{X}{\mathrm{d}}q^{i}\,X_{i}.$ From ${\mathrm{d}}f={\mathrm{d}}p_{i}\,\partial^{i}f+{\mathrm{d}}q^{i}\,\partial_{i}f$ we then get (4a) $\displaystyle(X_{f})^{i}$ $\displaystyle=\partial^{i}f=(-1)^{f+1}f\overleftarrow{\partial{{}^{i}}}$ and (4b) $\displaystyle(X_{f})_{i}$ $\displaystyle=(-1)^{f}\partial_{i}f=(-1)^{f}f\overleftarrow{\partial{{}_{i}^{\phantom{i}}}},$ so $(-1)^{f+1}X_{f}(g)$ is given by (1). ∎ The BV bracket satisfies several properties. ###### Proposition 3.2. For all functions $f$, $g$, and $h$, we have $\displaystyle[X_{f},X_{g}]$ $\displaystyle=X_{(f,g)},$ $\displaystyle(f,gh)$ $\displaystyle=(f,g)\,h+(-1)^{(f+1)g}g\,(f,h),$ $\displaystyle(g,f)$ $\displaystyle=-(-1)^{(f+1)(g+1)}(f,g),$ $\displaystyle(f,(g,h))$ $\displaystyle=((f,g),h)+(-1)^{(f+1)(g+1)}(g,(f,h)).$ The last three properties say that $(\ ,\ )$ is an odd Poisson bracket,777In particular this means that $(\ ,\ )$ is a super Lie bracket with respect to the opposite parity of its arguments. whereas the first says the the map $f\mapsto X_{f}$ is a morphism of super Lie algebras. ###### Proof. We prove only the first identity. The others are also easily obtained from the definitions (or recovered from the identical identities for the standard BV bracket which one obtains in each Darboux chart). We have $\displaystyle\iota_{[X_{f},X_{g}]}\omega=[\mathsf{L}_{X_{f}},\iota_{X_{g}}]\omega=\mathsf{L}_{X_{f}}\iota_{X_{g}}\omega=\mathsf{L}_{X_{f}}{\mathrm{d}}g$ $\displaystyle=(-1)^{f+1}{\mathrm{d}}\mathsf{L}_{X_{f}}g=(-1)^{f+1}{\mathrm{d}}X_{f}(g)={\mathrm{d}}(f,g)=\iota_{X_{(f,g)}}\omega,$ where $\mathsf{L}_{X_{f}}$ denotes the Lie derivative with respect to $X_{f}$. ∎ ## 4\. The BV Laplacian on functions Let $(\mathcal{M},\omega)$ be an odd symplectic manifold and $\mu$ an even, nowhere vanishing density on $\mathcal{M}$, which we assume to be orientable. We then define the $\mu$-dependent BV Laplacian as (5) $\Delta_{\mu}f\coloneqq\frac{1}{2}{\mathrm{div}}_{\mu}X_{f}.$ Recall (or see Appendix A.2) that the divergence operator of a vector field $X$ with respect to an even, nowhere vanishing density $\mu$ is defined via ${\mathrm{div}}_{\mu}X\,\mu=\mathsf{L}_{X}\mu,$ where $\mathsf{L}_{X}$ denotes the Lie derivative. Again we have to make sure that this definition of the BV Laplacian agrees with the standard one in the appropriate case: ###### Proposition 4.1. In Darboux coordinates, $\omega={\mathrm{d}}p_{i}{\mathrm{d}}q^{i}$, with standard density $\mu_{\text{stand}}\coloneqq{\mathrm{d}}^{n}q\,{\mathrm{d}}^{n}p$, we have $\Delta_{\mu_{\text{stand}}}=\triangle.$ ###### Proof. With the standard density the divergence of a vector field $X=X^{i}\partial_{i}+X_{i}\partial^{i}$ is given by (14a), i.e., ${\mathrm{div}}_{\mu_{\text{stand}}}X=\partial_{i}X^{i}-(-1)^{X}\partial^{i}X_{i}.$ If we now insert $X_{f}$ as calculated in (4), we get ${\mathrm{div}}_{\mu_{\text{stand}}}X_{f}=2\partial_{i}\partial^{i}f.$ ∎ Next we check how the operator depends on the choice of density. Note that, given two even, nowhere vanishing densities $\mu$ and $\widetilde{\mu}$, there is a unique even, nowhere vanishing function $h$ such that $\widetilde{\mu}=h\mu$. ###### Proposition 4.2. We have $\Delta_{h\mu}=\Delta_{\mu}+\frac{1}{2}\frac{1}{h}X_{h}.$ ###### Proof. The divergence operator depends on the choice of density as ${\mathrm{div}}_{h\mu}X_{f}={\mathrm{div}}_{\mu}X_{f}+\frac{1}{h}X_{f}(h)$ (see point (2) in Proposition A.4). However, $X_{f}(h)=(-1)^{f+1}(f,h)=-(h,f)=X_{h}(f)$, since $h$ is even. ∎ We can always write $h=\pm\mathrm{e}^{S}$ with $S$ an even function. This way the formula simplifies to $\Delta_{h\mu}=\Delta_{\mu}+\frac{1}{2}X_{S},$ since $X_{\mathrm{e}^{S}}=\mathrm{e}^{S}X_{S}$. In particular, we have the ###### Lemma 4.3. In Darboux coordinates with $\mu=\pm\mathrm{e}^{S}\mu_{\text{stand}}$, we have (6) $\Delta_{\mu}=\triangle+\frac{1}{2}X_{S}=\triangle-\frac{1}{2}(S,\ ).$ This, together with Proposition 4.2, gives the ###### Proposition 4.4. The BV Laplacian on functions is a second-order differential operator, and its leading term is independent of the choice of density. ###### Remark 4.5. Under the assumptions of Lemma 4.3, from (6), also using (3) and the Jacobi identity, we get $\Delta_{\mu}^{2}f=-\frac{1}{2}(F_{S},f)$ with (7) $F_{S}\coloneqq\triangle S-\frac{1}{4}(S,S).$ We then see that $\Delta_{\mu}^{2}=0$ if and only if $F_{S}$ is constant. As $F_{S}$ is odd, this happens if and only if $F_{S}=0$.888Note that this condition can also be rephrased as $\triangle\mathrm{e}^{\frac{1}{2}S}=0$. The appearance of the factor $\frac{1}{2}$ in the exponent will become clear when we study half-densities (see Lemma 5.13). Therefore, in general the property of equation (2a) does not extend to the global BV Laplacian. ###### Definition 4.6. We say that an even, nowhere vanishing density $\mu$ is compatible with the odd symplectic structure if $\Delta_{\mu}^{2}=0$. ###### Remark 4.7. We will see that every odd symplectic manifold admits a compatible density. ###### Remark 4.8. As we have seen in Remark 4.5, the property of equation (2a) does not extend in general. One can try to obviate it by modifying the definition of the BV operator. For example, in Darboux coordinates with $\mu=\pm\mathrm{e}^{S}\mu_{\text{stand}}$, we could define $\widetilde{\Delta}_{\mu}\coloneqq\Delta_{\mu}+F_{S}=\triangle+\frac{1}{2}X_{S}+\frac{1}{2}F_{S}$ with $F_{S}$ as in (7). It is an easy exercise (show first that $\Delta_{\mu}F_{S}=0$ for every $S$) to show that $\widetilde{\Delta}_{\mu}^{2}=0$ for every $\mu$. However, since $\widetilde{\Delta}_{\mu}$ is not of the form in Lemma 4.3 (unless, of course, $F_{S}=0$), it is not a BV Laplacian. If $\Psi\colon\mathcal{M}\to\mathcal{N}$ is a diffeomorphism, we have the pushforward $\Psi_{*}\colon\operatorname{End}(C^{\infty}(\mathcal{M}))\to\operatorname{End}(C^{\infty}(\mathcal{N}))$ defined by $\Psi_{*}P=\Psi_{*}\circ P\circ\Psi_{*}^{-1}$, where on the right- hand side $\Psi_{*}$ denotes the pushforward of functions. The BV Laplacian transforms nicely under symplectomorphisms. ###### Proposition 4.9. For every symplectomorphism $\Psi$, $\Psi_{*}\Delta_{\mu}=\Delta_{\Psi_{*}\mu}.$ ###### Proof. The divergence operator changes under a diffeomorphism as $\Psi_{*}{\mathrm{div}}_{\mu}X_{f}={\mathrm{div}}_{\Psi_{*}\mu}\Psi_{*}X_{f}$ (see point (1) in Proposition A.4). If $\Psi$ is a symplectomorphism, we also have $\Psi_{*}X_{f}=X_{\Psi_{*}f}$. ∎ ###### Remark 4.10. In conjunction with Proposition 4.4, this implies that under a symplectomorphism $\Psi\colon\mathcal{M}\to\mathcal{M}$, the leading term of the BV Laplacian on functions is invariant under symplectomorphisms. ###### Remark 4.11. The proposition also implies that a symplectomorphism sends compatible densities to compatible densities. ### 4.1. Digression: further properties of the BV Laplacian Even though we are not going to use this in the following, it is good to know that the properties stated in equations (2b) and (2c) also hold for the global BV Laplacian. ###### Proposition 4.12. For every even, nowhere vanishing density $\mu$, every functions $f$ and $g$, and every even function $S$, we have (8a) $\displaystyle\Delta_{\mu}(fg)=(\Delta_{\mu}f)g+(-1)^{f}f\Delta_{\mu}g-(-1)^{f}(f,g),$ (8b) $\displaystyle\Delta_{\mu}\mathrm{e}^{S}=\left(\Delta_{\mu}S-\frac{1}{2}(S,S)\right)\mathrm{e}^{S}.$ ###### Proof. The identities may be proved by observing, thanks to (6), that in every Darboux chart the BV Laplacian differs from $\triangle$ by a vector field. They may also be proved directly using properties of the divergence operator (see point (3) in Proposition A.4) and of hamiltonian vector fields. To illustrate this, we prove the second identity. Since $X_{\mathrm{e}^{S}}=\mathrm{e}^{S}X_{S}$, we have ${\mathrm{div}}_{\mu}X_{\mathrm{e}^{S}}={\mathrm{div}}_{\mu}(\mathrm{e}^{S}X_{S})=\mathrm{e}^{S}{\mathrm{div}}_{\mu}X_{S}+X_{S}(\mathrm{e}^{S}),$ which proves the identity once we observe that $X_{S}(\mathrm{e}^{S})=-\mathrm{e}^{S}(S,S)$. ∎ ###### Remark 4.13. As we have seen in Remark 4.8, it is possible to define an operator $\widetilde{\Delta}_{\mu}$ that squares to zero, no matter what $\mu$ is, so as to satisfy the extension of the property of equation (2a). However, since this is obtained by adding to $\Delta_{\mu}$ a multiplication operator, we see that now (8a) is no longer satisfied (unless, of course, $F_{S}=0$). ## 5\. The BV Laplacian on densities A differential operator $P$ defined on functions on a (super)manifold $\mathcal{M}$ can be extended to sections of a trivial line bundle $L$ over $\mathcal{M}$ once an even, nowhere vanishing section $\lambda$ of $L$ has been chosen. For $\sigma\in\Gamma(L)$ there is a uniquely determined function $f$ such that $\sigma=f\lambda$, and we set $P^{(\lambda)}\sigma\coloneqq P(f)\lambda.$ ###### Lemma 5.1. The leading term of the differential operator $P^{(\lambda)}$ does not depend on the choice of $\lambda$. ###### Proof. If $\widetilde{\lambda}$ is also a nowhere vanishing section, then there is a uniquely determined nowhere vanishing function $h$ with $\widetilde{\lambda}=h\lambda$. For $\sigma=f\widetilde{\lambda}=fh\lambda$, we have $P^{(\widetilde{\lambda})}\sigma=P(f)\widetilde{\lambda}=P(f)h\lambda.$ On the other hand, we have $P^{(\lambda)}\sigma=P(fh)\lambda=((P_{\text{leading}}f)\,h+\cdots)\lambda,$ where the dots denote terms where less than the maximum number of derivatives hit $f$. ∎ We now want to extend the BV Laplacian $\Delta_{\mu}$ to $s$-densities (see Appendix A.2 for a review). Since we already have an even, nowhere vanishing section, $\mu$, of the density bundle, we can use it to get our reference section, $\mu^{s}$, of the $s$-density bundle.999Since we are interested also in nonintegral $s$, in particular $s=\frac{1}{2}$, it is essential to require that $\mathcal{M}$ is oriented. That is, ###### Definition 5.2. For an $s$-density $\sigma$, which we can uniquely write as $\sigma=f\mu^{s}$, we set101010 We avoid the more pedantic notation $\Delta_{\mu}^{(\mu^{s})}$ on the left-hand side. $\Delta_{\mu}^{(s)}\sigma\coloneqq\Delta_{\mu}(f)\mu^{s}.$ Note that $\Delta_{\mu}^{(0)}$ is the same as $\Delta_{\mu}$. ###### Remark 5.3. An immediate consequence of this definition is that $\Delta_{\mu}^{(s)}$ squares to zero on $s$-densities if and only if $\Delta_{\mu}$ does so on functions (i.e., $\mu$ is compatible with the odd symplectic structure). Proposition 4.4 and Lemma 5.1 immediately imply ###### Proposition 5.4. The BV Laplacian on $s$-densities is a second-order differential operator, and its leading term is independent of the choice of reference density. Proposition 4.9 immediately implies ###### Proposition 5.5. For every symplectomorphism $\Psi$ and for every $s$, $\Psi_{*}\Delta^{(s)}_{\mu}=\Delta^{(s)}_{\Psi_{*}\mu}.$ ###### Remark 5.6. In particular, this implies that, for every $s$, under a symplectomorphism $\Psi\colon\mathcal{M}\to\mathcal{M}$, the leading term of the BV Laplacian on $s$-densities is invariant under symplectomorphisms. Note that the product of an $s$-density $\sigma$ and a $(1-s)$-density $\tau$ yields a density, which can be integrated (if it has compact support). Given a differential operator $P$ on $s$-densities, we define its transpose $P^{t}$ as the differential operator on $(1-s)$-densities that satisfies $\int_{\mathcal{M}}P\sigma\,\tau=(-1)^{P\,\sigma}\int_{\mathcal{M}}\sigma\,P^{t}\tau$ for all $s$-densities $\sigma$ with compact support (the upper index “$P\,\sigma$” denotes the product of the degrees; see footnote 3 on page 3). The case $s=\frac{1}{2}$ is special, since transposition becomes an endomorphism on the space of differential operators. ###### Proposition 5.7. We have $(\Delta_{\mu}^{(s)})^{t}=\Delta_{\mu}^{(1-s)}$ In particular, $\Delta_{\mu}^{(\frac{1}{2})}$ is symmetric. ###### Proof. Write $\sigma=f\mu^{s}$ and $\tau=g\mu^{1-s}$. Then $\Delta_{\mu}^{(s)}\sigma\,\tau=\Delta_{\mu}f\,g\,\mu.$ But $\displaystyle 2\Delta_{\mu}f\,g={\mathrm{div}}_{\mu}X_{f}\,g=(-1)^{(f+1)g}g\,{\mathrm{div}}_{\mu}X_{f}$ $\displaystyle=(-1)^{(f+1)g}({\mathrm{div}}_{\mu}(gX_{f})-(-1)^{(f+1)g}X_{f}(g))$ $\displaystyle=(-1)^{(f+1)g}{\mathrm{div}}_{\mu}(gX_{f})+(-1)^{f}(f,g),$ where we have used part (3) of Proposition A.4. Similarly, $\sigma\,\Delta_{\mu}^{(1-s)}\tau=f\,\Delta_{\mu}g\,\mu,$ and $\displaystyle 2f\,\Delta_{\mu}g=f\,{\mathrm{div}}_{\mu}X_{g}={\mathrm{div}}_{\mu}(fX_{g})-(-1)^{f(g+1)}X_{g}(f)$ $\displaystyle={\mathrm{div}}_{\mu}(fX_{g})+(-1)^{fg+f+g}(g,f)={\mathrm{div}}_{\mu}(fX_{g})+(f,g).$ Therefore, $2\Delta_{\mu}f\,g-(-1)^{f}2f\,\Delta_{\mu}g=(-1)^{(f+1)g}{\mathrm{div}}_{\mu}(gX_{f})-(-1)^{f}{\mathrm{div}}_{\mu}(fX_{g}),$ which implies, thanks to part (4) of Proposition A.4, that111111Note the different sign in (8a), which instead yields $\int_{\mathcal{M}}(\Delta_{\mu}^{(s)}\sigma\,\tau+(-1)^{\sigma}\sigma\,\Delta_{\mu}^{(1-s)}\tau)=\int_{\mathcal{M}}(\Delta_{\mu}f\,g+(-1)^{f}f\,\Delta_{\mu}g)\,\mu=(-1)^{f}\int_{\mathcal{M}}(f,g)\,\mu.$ Subtracting the two identities yields the interesting formula $\int_{\mathcal{M}}(f,g)\,\mu=2\int_{\mathcal{M}}f\,\Delta_{\mu}g\,\mu,$ which in particular shows that the integral of the BV bracket of two functions (at least one of which has compact support) against the density $\mu$ vanishes if one of the two functions is in the kernel of $\Delta_{\mu}$. $\int_{\mathcal{M}}(\Delta_{\mu}^{(s)}\sigma\,\tau-(-1)^{\sigma}\sigma\,\Delta_{\mu}^{(1-s)}\tau)=0.$ ∎ ###### Remark 5.8. By inspection in the proof one sees that it is essential that we have used the same density $\mu$ to define the BV Laplacian on functions and to extend it to $s$-densities. We now come to the fundamental consequence of this proposition. ###### Lemma 5.9. In each Darboux chart, there is an odd function $G_{\mu}$, depending on $\mu$, such that $\Delta_{\mu}^{(\frac{1}{2})}\sigma=(\triangle f+G_{\mu}f)\,\mu_{\text{stand}}^{\frac{1}{2}},$ where we have written $\sigma=f\,\mu_{\text{stand}}^{\frac{1}{2}}$. ###### Proof. Since $\Delta_{\mu}^{(\frac{1}{2})}$ is a second-order differential operator and its leading term is independent of $\mu$, we have $\Delta_{\mu}^{(\frac{1}{2})}\sigma=(\triangle f+Y_{\mu}(f)+G_{\mu}\,f)\,\mu_{\text{stand}}^{\frac{1}{2}},$ for some vector field $Y_{\mu}$ and some function $G_{\mu}$. Since $\triangle$ has constant coefficients, it is symmetric. Moroever, since $Y_{\mu}$ is a vector field, we have $Y_{\mu}^{t}=-Y_{\mu}+$some function. Since $\Delta_{\mu}^{(\frac{1}{2})}$ is symmetric, we get $Y_{\mu}=0$. ∎ This implies the following ###### Corollary 5.10. For every compatible $\mu$, the BV Laplacian on half-densities has a canonical form in every Darboux chart: $\Delta_{\mu}^{(\frac{1}{2})}\sigma=\triangle f\,\mu_{\text{stand}}^{\frac{1}{2}},$ where we have written $\sigma=f\,\mu_{\text{stand}}^{\frac{1}{2}}$. In particular, $\Delta_{\mu}^{(\frac{1}{2})}$ is independent of the choice of the compatible density $\mu$. ###### Proof. Using $\triangle^{2}=0$, $G_{\mu}^{2}=0$, and (2b), we get $0=(\Delta_{\mu}^{(\frac{1}{2})})^{2}\sigma=(\triangle G_{\mu}\,f+(G_{\mu},f))\mu_{\text{stand}}^{\frac{1}{2}}.$ Therefore, for every function $f$, we have $\triangle G_{\mu}\,f+(G_{\mu},f)=0.$ In particular, for $f=1$, we get $\triangle G_{\mu}=0$, so the condition simplifies to $(G_{\mu},f)=0$ for every $f$. This implies that $G_{\mu}$ is constant, but then it must vanish because it is odd.121212This argument fails if one wants to work in families, as in this case odd constants are allowed. ∎ In every Darboux chart, we can consider the BV Laplacian $\Delta_{\mu_{\text{stand}}}^{(\frac{1}{2})}$, which sends $f\,\mu_{\text{stand}}^{\frac{1}{2}}$ to $\triangle f\,\mu_{\text{stand}}^{\frac{1}{2}}$. If we move from this Darboux chart to another one via the transition map $\phi$, the operator goes to $\phi_{*}\Delta_{\mu_{\text{stand}}}^{(\frac{1}{2})}$, which, by Proposition 5.5, is $\Delta_{\phi_{*}\mu_{\text{stand}}}^{(\frac{1}{2})}$. However, since $\mu_{\text{stand}}$ is a compatible density (and therefore, by Remark 4.11, also $\phi_{*}\mu_{\text{stand}}$ is a compatible density), by the corollary this is again $\Delta_{\mu_{\text{stand}}}^{(\frac{1}{2})}$ (in the new chart). Therefore, we have the ###### Theorem/Definition 5.11. There is a canonical BV operator $\Delta$ acting on half-densities, defined as $\Delta_{\mu_{\text{stand}}}^{(\frac{1}{2})}$ in every Darboux chart. Note that, by its very definition, the canonical BV operator is invariant under symplectomorphisms and satisfies $\Delta^{2}=0$. Corollary 5.10 may now be reformulated as ###### Theorem 5.12. For every compatible $\mu$, the BV Laplacian on half-densities is equal to the canonical BV operator: $\Delta_{\mu}^{(\frac{1}{2})}=\Delta.$ By this theorem we have that the canonical BV operator is a BV Laplacian on half-densities whenever the odd symplectic manifold admits compatible densities,131313Unlike in the even symplectic case, where a canonical choice of density is provided by the top exterior power of the symplectic form, there is no canonical density in the odd case. which we have not proved yet. On the other hand, what is remarkable is that $\Delta$ is canonically defined (by Theorem/Definition 5.11) without choosing any density (in particular it would exist even if there were no compatible densities). The canonical BV operator gives another way of defining the BV Laplacian on functions. Namely, given an even, nowhere vanishing density $\mu$, we define a differential operator $\widehat{\Delta}_{\mu}$ on functions via (9) $\widehat{\Delta}_{\mu}f=\Delta(f\mu^{\frac{1}{2}})\,\mu^{-\frac{1}{2}}.$ Note that $\widehat{\Delta}_{\mu}^{2}=0$ for every $\mu$. However, it is not a BV Laplacian in general. We actually have the ###### Lemma 5.13. For every $\mu$, $\widehat{\Delta}_{\mu}=\widetilde{\Delta}_{\mu}$ in every Darboux chart, where $\widetilde{\Delta}_{\mu}$ is the operator defined in Remark 4.8. ###### Proof. Writing $\mu=\mathrm{e}^{S}\mu_{\text{stand}}$, we have $\widehat{\Delta}_{\mu}f\,\mathrm{e}^{\frac{S}{2}}\mu_{\text{stand}}^{\frac{1}{2}}=\Delta(f\mathrm{e}^{\frac{S}{2}}\mu_{\text{stand}}^{\frac{1}{2}})=\triangle(f\mathrm{e}^{\frac{S}{2}})\,\mu_{\text{stand}}^{\frac{1}{2}}.$ One can easily compute $\triangle(f\mathrm{e}^{\frac{S}{2}})=\left(\triangle f-\frac{1}{2}(S,f)+\frac{1}{2}F_{S}\right)\mathrm{e}^{\frac{S}{2}}=\widetilde{\Delta}_{\mu}f\,\mathrm{e}^{\frac{S}{2}},$ with $F_{S}$ as in (7). ∎ This yields the following ###### Corollary 5.14. Let $\mu$ be an even, nowhere vanishing density. Then 1. (1) $\widetilde{\Delta}_{\mu}$ is globally defined; 2. (2) $\mu$ is compatible if and only if $\widehat{\Delta}_{\mu}=\Delta_{\mu}$; 3. (3) $\mu$ is compatible if and only if $\Delta\mu^{\frac{1}{2}}=0$. ###### Proof. The first two properties are now trivial. We prove the third. Setting $f=1$ in (9) yields $\Delta(\mu^{\frac{1}{2}})=\widehat{\Delta}_{\mu}1\,\mu^{\frac{1}{2}}$. Lemma 5.13 then implies that $\Delta(\mu^{\frac{1}{2}})=0$ if and only if in every Darboux chart $F_{S}=\widetilde{\Delta}_{\mu}1=0$. By Remark 4.5 this happens if and only if $\mu$ is compatible. ∎ ### 5.1. Digression: Darboux expression of $\Delta_{\mu}^{(s)}$ In a Darboux chart we have $\mu=\mathrm{e}^{S}\mu_{\text{stand}}$. If $\sigma=f\,\mu_{\text{stand}}^{s}$, we then have $\sigma=\mathrm{e}^{-sS}f\,\mu^{s}$, so $\Delta_{\mu}^{(s)}\sigma=\Delta_{\mu}(\mathrm{e}^{-sS}f)\,\mu^{s}.$ By using (6) and the properties of $\triangle$ in (2), we get $\Delta_{\mu}^{(s)}\sigma=\left(\triangle f+\left(s-\frac{1}{2}\right)(S,f)-\left(s\triangle S+\frac{1}{2}s(s-1)(S,S)\right)f\right)\mu_{\text{stand}}^{s}.$ One can now explictly check that $(\Delta_{\mu}^{(s)})^{t}=\Delta_{\mu}^{(1-s)}$. Moreover, one explicitly sees that the first-order term vanishes for every $S$ if and only if $s=\frac{1}{2}$. Finally, for $\mu$ compatible, i.e., $\Delta S=\frac{1}{4}(S,S)$, the zeroth-order term becomes $\frac{1}{4}(2s-1)s(S,S)$, which vanishes for $s=\frac{1}{2}$ or $s=0$. This formula, for $s=\frac{1}{2}$, yields a different way to prove Corollary 5.10 and therefore Theorem/Definition 5.11 and Theorem 5.12. However, proving this formula is a bit more laborious than proving Proposition 5.7. ## 6\. Odd cotangent bundles We now focus on an odd cotangent bundle $\Pi T^{*}M$,141414Here $\Pi$ denotes the change-of-parity functor. It simply means that the fiber coordinates are now regarded as odd. where $M$ is an ordinary manifold. We will see that it is easy to show the existence of compatible densities in this case. Since every odd symplectic manifold is symplectomorphic to the odd cotangent bundle of its body [13], this implies that every odd symplectic manifold has a compatible density and therefore that the canonical BV operator is indeed a BV Laplacian on half-densities. The first observation is that functions on $\Pi T^{*}M$ are the same as multivector fields on $M$: $C^{\infty}(\Pi T^{*}M)=\Gamma(\Lambda^{\bullet}TM)$ as super algebras.151515The interested reader might also appreciate that $(C^{\infty}(\Pi T^{*}M),(\ ,\ ))$ and $(\Gamma(\Lambda^{\bullet}TM),[\ ,\ ]_{\text{SN}})$ are naturally isomorphic odd Poisson algebras, where $(\ ,\ )$ is the canonical BV bracket, associated to the canonical symplectic form, and $[\ ,\ ]_{\text{SN}}$ is the Schouten–Nijenhuis bracket. The second observation is about the density bundle on $\Pi T^{*}M$. If $\phi_{\alpha\beta}$ denotes a transition function on $M$, then the corresponding transition function on $T^{*}M$ has a fiber part given by $(({\mathrm{d}}\phi_{\alpha\beta})^{*})^{-1}$. Therefore, the Berezinian of the corresponding transition function for $\Pi T^{*}M$ is $({\mathrm{d}}\phi_{\alpha\beta})^{2}$. This means that densities on $\Pi T^{*}M$ transform like $2$-densities on $M$. Since we assume $M$ to be oriented, we also have that half-densities on $\Pi T^{*}M$ transform like top forms on $M$. More precisely, we have a canonical isomorphism $\Gamma(\operatorname{Dens}^{\frac{1}{2}}(\Pi T^{*}M))\simeq\Gamma(\Lambda^{\bullet}TM)\otimes_{C^{\infty}(M)}\Omega^{n}(M),$ where $n$ is the dimension of $M$. The next observation is that we have a canonical isomorphism (of $C^{\infty}(M)$-modules) $\phi\colon\begin{array}[t]{ccc}\Gamma(\Lambda^{\bullet}TM)\otimes_{C^{\infty}(M)}\Omega^{n}(M)&\to&\Omega^{n-\bullet}(M)\\\ X\otimes v&\mapsto&\iota_{X}v\end{array}$ We can use this isomorphism to define an operator on half-densities: $\mathcal{D}\coloneqq\phi^{-1}\circ{\mathrm{d}}\circ\phi,$ where ${\mathrm{d}}$ is the de Rham differential. One immediately sees that $\mathcal{D}^{2}=0$ and that $\mathcal{D}v=0$ for every top form $v$ on $M$. ###### Proposition 6.1. The canonical operator $\mathcal{D}$ is the same as the canonical BV operator $\Delta$ on half-densities. ###### Proof. We just have to check how $\mathcal{D}$ acts in Darboux coordinates. Let $q^{1},\dots,q^{n}$ denote coordinates on a chart of $M$ and let $p_{1},\dots,p_{n}$ be the corresponding odd fiber coordinates. By linearity it is enough to consider a half-density $\sigma=g\,\mu_{\text{stand}}$ with $g(p,q)$ of the form $f(q)\,p_{i_{1}}\cdots\,p_{i_{k}}$, with pairwise distinct indices $i_{j}$. In the identification described above this corresponds to $\sigma=f(q)\,\partial_{i_{1}}\wedge\dots\wedge\partial_{i_{k}}\otimes{\mathrm{d}}^{n}q$, so $\phi(\sigma)=f(q)\,\iota_{\partial_{i_{1}}}\cdots\,\iota_{\partial_{i_{k}}}{\mathrm{d}}^{n}q.$ We then have ${\mathrm{d}}\phi(\sigma)={\mathrm{d}}f(q)\,\iota_{\partial_{i_{1}}}\cdots\,\iota_{\partial_{i_{k}}}{\mathrm{d}}^{n}q=\partial_{i}f(q)\,{\mathrm{d}}q^{i}\wedge\iota_{\partial_{i_{1}}}\cdots\,\iota_{\partial_{i_{k}}}{\mathrm{d}}^{n}q,$ since the form $\iota_{\partial_{i_{1}}}\cdots\,\iota_{\partial_{i_{k}}}{\mathrm{d}}^{n}q$ has a constant coefficient, so it is closed. If the index $i$ is different from all the $i_{j}$s, the corresponding term vanishes, since ${\mathrm{d}}q^{i}$ is already contained in ${\mathrm{d}}^{n}q$. Otherwise, we have $\begin{split}{\mathrm{d}}q^{i_{j}}\wedge\iota_{\partial_{i_{1}}}\cdots\,\iota_{\partial_{i_{k}}}{\mathrm{d}}^{n}q&=(-1)^{j-1}{\mathrm{d}}q^{i_{j}}\wedge\iota_{\partial_{i_{j}}}\iota_{\partial_{i_{1}}}\cdots\,\widehat{\iota}_{\partial_{i_{j}}}\cdots\,\iota_{\partial_{i_{k}}}{\mathrm{d}}^{n}q\\\ &=(-1)^{j-1}\iota_{\partial_{i_{1}}}\cdots\,\widehat{\iota}_{\partial_{i_{j}}}\cdots\,\iota_{\partial_{i_{k}}}{\mathrm{d}}^{n}q,\end{split}$ where the caret denotes omission. On the other hand, $\partial_{i_{j}}(p_{i_{1}}\cdots\,p_{i_{k}})=(-1)^{j-1}p_{i_{1}}\cdots\,\widehat{p}_{i_{j}}\cdots\,p_{i_{k}}.$ Therefore, $\phi(\triangle g\,\mu_{\text{stand}})={\mathrm{d}}\phi(g\,\mu_{\text{stand}}),$ which shows that $\Delta=\mathcal{D}$. ∎ We then have $\Delta v=0$ for every top form $v$ on $M$ (regarded as a half- density on $\Pi T^{*}M$). As a notation, we write $\mu_{v}$ for the even, nowhere vanishing density $v^{2}$ on $\Pi T^{*}M$ associated to a volume form $v$ on $M$. We then have $\Delta\mu_{v}^{\frac{1}{2}}=0,$ which shows, by point (3) of Corollary 5.14, that $\mu_{v}$ is a compatible density. We then have the ###### Proposition 6.2. An odd cotangent bundle possesses compatible densities. This implies, by Theorem 5.12, that $\Delta=\Delta^{(\frac{1}{2})}_{\mu_{v}}$. Given a volume form $v$, we can now use (9) to get $\Delta_{\mu_{v}}$ on functions. It is an easy exercise to see that $\Delta_{\mu_{v}}=\phi_{v}^{-1}\circ{\mathrm{d}}\circ\phi_{v},$ where $\phi_{v}$ is the ($v$-dependent!) isomorphism $\phi_{v}\colon\begin{array}[t]{ccc}C^{\infty}(\Pi T^{*}M)=\Gamma(\Lambda^{\bullet}TM)&\to&\Omega^{n-\bullet}(M)\\\ X&\mapsto&\iota_{X}v\end{array}$ Note that, for a vector field $X$, we have $\Delta_{\mu_{v}}X={\mathrm{div}}_{v}X$. For this reason, $\Delta_{\mu_{v}}$ is interpreted as the extension to multivector fields of the divergence operator with respect to $v$. ## 7\. Conclusion of the argument It is a theorem (see [13]) that every odd symplectic manifold $\mathcal{M}$ is, noncanonically, isomorphic to the odd cotangent bundle $\Pi T^{*}M$ of its body $M$.161616This is a global Darboux theorem. If one is familiar with the proof of Darboux’s theorem via Moser’s trick, the reason why this now works globally is that it turns out that one has to integrate a vector field in the odd directions, which has no problem of definition domain. On $\Pi T^{*}M$ we have compatible densities (Proposition 6.2), e.g., $\mu_{v}$ for a volume form $v$ on $M$. If $\Psi\colon\Pi T^{*}M\to\mathcal{M}$ is a symplectomorphism, which exists by the above mentioned theorem, then $\Psi_{*}\mu_{v}$ is a compatible density on $\mathcal{M}$ (Remark 4.11). We therefore have—also using Theorem 5.12, equation (9), and point (2) of Corollary 5.14—our final result: ###### Theorem 7.1. Every odd symplectic manifold possesses compatible densities. For each compatible density $\mu$, we have $\Delta_{\mu}^{(\frac{1}{2})}=\Delta,$ with $\Delta$ the canonical BV operator on half-densities, and the BV Laplacian $\Delta_{\mu}$ on functions may be recovered from $\Delta_{\mu}f\,\mu^{\frac{1}{2}}=\Delta(f\mu^{\frac{1}{2}}).$ ## Appendix A Some background details ### A.1. The Lie derivative If $\Xi$ is an object (e.g., function, vector field, density) for which a notion of pushforward $\Psi_{*}$ under a diffeomorphism $\Psi$ exists, we define the Lie derivative with respect to a vector field $X$ as $\mathsf{L}_{X}\Xi\coloneqq\left.\frac{\partial}{\partial t}\right\rvert_{t=0}(\Phi^{X}_{-t})_{*}\Xi.$ For an even vector field $X$, the diffeomorphism $\Phi^{X}_{t}$ is the flow of $X$ at time $t$. If $X$ is odd, the variable $t$ is also odd (so the evaluation at zero in the formula is redundant), and we define the morphism $\Phi^{X}\colon\Pi{\mathbb{R}}\times\mathcal{M}\to\mathcal{M}$ via171717This defines the flow of $X$ at odd time $t$ if and only if $[X,X]=0$. Otherwise, the flow property $\Phi^{X}_{t+s}=\Phi^{X}_{t}\circ\Phi^{X}_{s}$ is not satisfied, so in general $\Phi^{X}$ is not a flow. $(\Phi^{X})^{*}\colon\begin{array}[t]{ccc}C^{\infty}(\mathcal{M})&\to&C^{\infty}(\Pi{\mathbb{R}}\times\mathcal{M})\\\ f&\mapsto&f+tX(f)\end{array}$ where $t$ is the odd coordinate on $\Pi{\mathbb{R}}$.181818Here $\Pi$ denotes the change-of-parity functor: $\Pi{\mathbb{R}}$ is the superdomain with one odd coordinate. Note that in local coordinates $z^{\mu}$, with $X=X^{\mu}{\small\frac{\partial}{\partial z^{\mu}}}$, we have in both cases (10) $(\Phi^{X})^{\mu}\coloneqq(\Phi^{X})^{*}z^{\mu}=z^{\mu}+tX^{\mu}+O(t^{2}),$ where of course $O(t^{2})=0$ in the odd case. Also note that in both cases we have $\Psi_{*}(\Phi^{X})_{*}=(\Phi^{\Psi_{*}X})_{*}\Psi_{*}$ for every diffeomorphism $\Psi$.191919In the even case, one just observes how solutions of ODEs are mapped under diffeomorphisms. In the odd case, it is just an immediate application of the chain rule. This implies that $\mathsf{L}_{\Psi_{*}X}\Psi_{*}\Xi=\Psi_{*}\mathsf{L}_{X}\Xi.$ If two objects $\Xi_{1}$ and $\Xi_{2}$ as above can be multiplied, we immediately get (11) $\mathsf{L}_{X}(\Xi_{1}\Xi_{2})=\mathsf{L}_{X}\Xi_{1}\,\Xi_{2}+(-1)^{X\Xi_{1}}\Xi_{1}\,\mathsf{L}_{X}\Xi_{2}.$ On a function $f$ or on a vector field $Y$, the Lie derivative is readily computed as $\mathsf{L}_{X}f=X(f)\quad\text{and}\quad\mathsf{L}_{X}Y=[X,Y].$ ### A.2. Densities A density is an object that transforms in such a way that its integration may be defined. For this we have to recall that, if $\Psi$ is a diffeomorphism between two superdomains, then the change-of-variables formula for berezinian integration involves the Berezinian $\operatorname{Ber}(\Psi)$ of the jacobian matrix of $\Psi$ times the sign of the jacobian determinant of the reduction to the body of $\Psi$. Since we assume throughout that the body of our supermanifold is oriented, we will not have this factor. Therefore, given an oriented atlas for $\mathcal{M}$ that has transition functions $\phi_{\alpha\beta}$, we define the density bundle $\operatorname{Dens}\mathcal{M}$ as the line bundle with transition functions $\operatorname{Ber}(\phi_{\alpha\beta})^{-1}$. More generally, for any real number $s$, we define the line bundle $\operatorname{Dens}^{s}\mathcal{M}$ of $s$-densities with transition functions $\operatorname{Ber}(\phi_{\alpha\beta})^{-s}$. Note that $1$-densities are the same as densities, and $0$-densities are the same as functions. With this definition the integral of a compactly supported202020The support is defined in terms of the coefficient body functions. Namely, an object (function, vector field, density,…) is compactly supported when in its expansion in odd variables all the coefficients are compactly supported functions on the body. density $\mu$ over $\mathcal{M}$ is defined as usual (by choosing a partition of unity subordinated to the atlas and by berezinian integration in each chart). The pushforward of $s$-densities under diffeomorphisms is also defined. If we go to charts, the pushforward is given by the pushforward of the representing function times the Berezinian of the transformation to the power $-s$. Armed with the pushforward, we can define the Lie derivative of an $s$-density. In particular, we need the following ###### Lemma A.1. In local coordinates $p_{1},\dots,p_{m}$ and $q^{1},\dots,q^{n}$ (odd and even, respectively), the Lie derivative of the standard density $\mu_{\text{stand}}={\mathrm{d}}^{n}q\,{\mathrm{d}}^{m}p$ is $\mathsf{L}_{X}{\mu_{\text{stand}}}=(\partial_{i}X^{i}-(-1)^{X}\partial^{i}X_{i})\,{\mu_{\text{stand}}},$ where we used the expansion $X=X^{i}\partial_{i}+X_{i}\partial^{i}$. ###### Proof. Using (10), we see that212121By $\delta^{i}_{j}$ we mean the Kronecker delta, which is equal to $1$ for $i=j$ and to $0$ otherwise. $\begin{split}\operatorname{Ber}(\Phi^{X}_{-t})&=\operatorname{Ber}\begin{pmatrix}\delta_{i}^{j}-\partial_{i}(tX^{j})&0\\\ 0&\delta^{i}_{j}-\partial^{i}(tX_{j})\end{pmatrix}+O(t^{2})\\\ &=-\partial_{i}(tX^{i})-\partial^{i}(tX_{i})+O(t^{2})\\\ &=-t\partial_{i}(X^{i})-(-1)^{X}t\partial^{i}(X_{i})+O(t^{2}),\end{split}$ using that the paritiy of $t$ is the same as that of $X$. Applying the definitions of Lie derivative and of pushforward of a density yields the result. ∎ Applying (11) to the case of densities, we get (12) $\mathsf{L}_{X}(h\mu)=X(h)\,\mu+(-1)^{Xh}h\,\mathsf{L}_{X}\mu$ for every function $h$. In particular, we get the general formula for the Lie derivative in local coordinates: $\begin{split}\mathsf{L}_{X}\mu&=(X(h)+(-1)^{Xh}h(\partial_{i}X^{i}-(-1)^{X}\partial^{i}X_{i}))\,{\mu_{\text{stand}}}\\\ &=(X(h)+(\partial_{i}X^{i}-(-1)^{X}\partial^{i}X_{i})h)\,{\mu_{\text{stand}}}\\\ &=X(h)\mu_{\text{stand}}+(\partial_{i}X^{i}-(-1)^{X}\partial^{i}X_{i})\mu,\end{split}$ with $\mu=h\mu_{\text{stand}}$. This implies the following ###### Lemma A.2. For every function $f$, vector field $X$, and density $\mu$, we have $\mathsf{L}_{fX}\mu=f\mathsf{L}_{X}\mu+(-1)^{fX}X(f)\mu.$ ###### Proof. It is enough to check the formula in local coordinates. We have $\begin{split}\mathsf{L}_{fX}\mu&=fX(h)\mu_{\text{stand}}+(\partial_{i}(fX^{i})-(-1)^{f+X}\partial^{i}(fX_{i}))\mu\\\ &=fX(h)\mu_{\text{stand}}+f\,(\partial_{i}X^{i}-(-1)^{X}\partial^{i}X_{i})\mu+(\partial_{i}fX^{i}-(-1)^{f+X}\partial^{i}fX_{i})\mu\\\ &=f\mathsf{L}_{X}\mu+((-1)^{Xf}X^{i}\partial_{i}f-(-1)^{f+X}(-1)^{(f+1)(X+1)}X_{i}\partial^{i}f)\mu\\\ &=f\mathsf{L}_{X}\mu+(-1)^{fX}X(f)\mu.\end{split}$ ∎ In the case of densities, the pushforward is related to the change-of- variables formula. Namely, if $\Psi\colon\mathcal{M}\to\mathcal{N}$ is a diffeomorphism (preserving the orientation of the bodies), then (13) $\int_{\mathcal{N}}\Psi_{*}\mu=\int_{\mathcal{M}}\mu$ for every compactly supported density $\mu$. This leads to ###### Lemma A.3. For every vector field $X$ and density $\mu$, one of which is compactly supported, we have $\int_{\mathcal{M}}\mathsf{L}_{X}\mu=0.$ ###### Proof. First consider the case when $\mu$ is compactly supported. In this case, $(\Phi^{X}_{-t})_{*}\mu$ and hence $\mathsf{L}_{X}\mu$ are also compactly supported. We then have $\int_{\mathcal{M}}\mathsf{L}_{X}\mu=\int_{\mathcal{M}}\left.\frac{\partial}{\partial t}\right\rvert_{t=0}(\Phi^{X}_{-t})_{*}\mu=\left.\frac{\partial}{\partial t}\right\rvert_{t=0}\int_{\mathcal{M}}(\Phi^{X}_{-t})_{*}\mu=\left.\frac{\partial}{\partial t}\right\rvert_{t=0}\int_{\mathcal{M}}\mu=0,$ where we have also used (13). If on the other hand $\mu$ is not compactly supported, but $X$ is, we replace $\mu$ with $\widetilde{\mu}=\rho\mu$, where $\rho$ is a compactly supported bump function which is identically equal to $1$ on the support of $X$. Since $\mathsf{L}_{X}\mu=\mathsf{L}_{X}\widetilde{\mu}$, we get the result by the previous case. ∎ If we have an even, nowhere vanishing density $\mu$, then every other density can be written as $f\mu$ for a uniquely determined function $f$. Therefore, we can define the divergence operator via the following formula: ${\mathrm{div}}_{\mu}X\,\mu=\mathsf{L}_{X}\mu.$ The properties of the Lie derivative immediately imply properties for the divergence operator: ###### Proposition A.4. Let $\mu$ be an even, nowhere vanishing density $\mu$ and $X$ a vector field. Then 1. (1) for every diffeomorphism $\Psi$, $\Psi_{*}{\mathrm{div}}_{\mu}X={\mathrm{div}}_{\Psi_{*}\mu}\Psi_{*}X;$ 2. (2) for every even, nowhere vanishing function $h$, ${\mathrm{div}}_{h\mu}X={\mathrm{div}}_{\mu}X+\frac{1}{h}X(h);$ 3. (3) for every function $f$, ${\mathrm{div}}_{\mu}(fX)=f{\mathrm{div}}_{\mu}X+(-1)^{fX}X(f);$ 4. (4) under the additional assumption that $X$ is compactly supported, $\int_{\mathcal{M}}{\mathrm{div}}_{\mu}X\,\mu=0.$ Moreover, from the local coordinate expression for the Lie derivative, we get (14a) $\displaystyle{\mathrm{div}}_{\mu_{\text{stand}}}X$ $\displaystyle=\partial_{i}X^{i}-(-1)^{X}\partial^{i}X_{i},$ and, rearranging the terms, (14b) $\displaystyle{\mathrm{div}}_{\mu}X$ $\displaystyle=\frac{1}{h}(\partial_{i}(hX^{i})-(-1)^{X}\partial^{i}(hX_{i})),$ for $\mu=h\mu_{\text{stand}}$. ## Appendix B Applications of the BV formalism We recall here the main reason why the BV formalism was introduced: to study the invariance of certain integrals. We start considering the case of Section 2, where we have Darboux coordinates $p_{1},\dots,p_{n}$ and $q^{1},\dots,q^{n}$. For the sake of the argument, we rearrange each pair $(p_{i},q^{i})$ into a new pair $(x_{i},y^{i})$, with $x_{i}=p_{i}$ and $y^{i}=q^{i}$ for some $i$s and $x_{i}=q^{i}$ and $y^{i}=p_{i}$ for the other $i$s. Note that $x_{i}$ has opposite parity to $y^{i}$ (but we do not insist on $y^{i}$ being even). We now use the shorthand notation $\partial_{i}=\frac{\partial}{\partial y^{i}}\quad\text{and}\quad\partial^{i}=\frac{\partial}{\partial x_{i}}.$ Note that the BV Laplacian still reads $\triangle=\partial_{i}\partial^{i}$. If $\psi$ is an odd function of the variables $y$, it makes sense, in terms of parity, to set $x_{i}=(-1)^{x_{i}}\partial_{i}\psi$ (we will explain the reason for the signs in a moment). For a function $f$ in the $(x,y)$ variables, we define $\int_{\mathcal{L}_{\psi}}f\coloneqq\int f|_{x_{j}=(-1)^{x_{j}}\partial_{j}\psi}\;{\mathrm{d}}^{n}y,$ where $|_{x_{j}=(-1)^{x_{j}}\partial_{j}\psi}$ means that we set each $x$ variable to the corresponding right-hand-side expression. The reason for the notation is that $\mathcal{L}_{\psi}$ is a Lagrangian submanifold determined by $\psi$. In fact, $\omega={\mathrm{d}}x_{i}{\mathrm{d}}y^{i}={\mathrm{d}}(x_{i}{\mathrm{d}}y^{i})$ and $(x_{i}{\mathrm{d}}y^{i})|_{x_{i}=(-1)^{x_{i}}\partial_{i}\psi}={\mathrm{d}}\psi.$ The first result is the BV lemma. ###### Lemma B.1. Suppose that $f=\triangle g$ and $g$ is integrable. Then $\int_{\mathcal{L}_{\psi}}f=0$ for every $\psi$. ###### Proof. We have $\begin{split}\sum_{i}\partial_{i}(\partial^{i}g)|_{x_{j}=(-1)^{x_{j}}\partial_{j}\psi}&=\sum_{i}(\partial_{i}\partial^{i}g)|_{x_{j}=(-1)^{x_{j}}\partial_{j}\psi}\\\ &\phantom{=}\ +\sum_{ik}(\partial_{i}((-1)^{x_{k}}\partial_{k}\psi)\,\partial^{k}\partial^{i}g)|_{x_{j}=(-1)^{x_{j}}\partial_{j}\psi}.\end{split}$ Since $\sum_{ik}(-1)^{x_{k}}(\partial_{i}\partial_{k}\psi\,\partial^{k}\partial^{i}g)=0$, as one can easily see by exchanging the indices, we get $(\triangle g)|_{x_{j}=(-1)^{x_{j}}\partial_{j}\psi}=\sum_{i}\partial_{i}(\partial^{i}g)|_{x_{j}=(-1)^{x_{j}}\partial_{j}\psi},$ so $\int_{\mathcal{L}_{\psi}}f$ vanishes. ∎ This leads to the fundamental BV theorem. ###### Theorem B.2. Let $\psi_{t}$ be a family of odd functions in the $y$ variables depending smoothly on the even parameter $t$. If $f$ is integrable, on every $\mathcal{L}_{\psi_{t}}$, and $\triangle f=0$, then $I_{t}\coloneqq\int_{\mathcal{L}_{\psi_{t}}}f$ is constant. ###### Proof. We have $\dot{I}_{t}=\int\sum_{i}((-1)^{x_{i}}\partial_{i}\dot{\psi}\,\partial^{i}f)|_{x_{j}=(-1)^{x_{j}}\partial_{j}\psi}\;{\mathrm{d}}^{n}y,$ where the dot denotes derivative with respect to $t$. From $\triangle(\dot{\psi}\,f)=\sum_{i}\partial_{i}\partial^{i}(\dot{\psi}\,f)=\sum_{i}(-1)^{x_{i}}\partial_{i}(\dot{\psi}\,\partial^{i}f)=\sum_{i}(-1)^{x_{i}}\partial_{i}\dot{\psi}\,\partial^{i}f-\dot{\psi}\,\triangle f,$ we get $\dot{I}_{t}=\int\triangle(\dot{\psi}\,f)|_{x_{j}=(-1)^{x_{j}}\partial_{j}\psi}\;{\mathrm{d}}^{n}y=\int_{\mathcal{L}_{\psi_{t}}}\triangle(\dot{\psi}\,f)=0,$ where we have used the previous lemma. ∎ The BV theorem is used in quantum gauge theories,222222In this application, the odd symplectic manifold is actually infinite dimensional, so $\triangle$ is not defined. What one does is either to proceed formally or to regularize the theory, e.g., by replacing the space of fields with a finite-dimensional approximation. where the choice of $\psi$ (called the gauge-fixing fermion) corresponds to the choice of a gauge fixing and the invariance under deformations of this choice yields the independence of the theory from the gauge fixing, assuming of course that $\triangle f=0$. Using the results of Sections 4 and 5, one can extend the BV lemma and the BV theorem globally. The main observation is that the restriction of a half- density on an odd symplectic manifold to a Lagrangian submanifold defines a density there [13]. ###### Theorem B.3. Let $\mathcal{M}$ be an odd symplectic manifold and $\sigma$ a half-density. The following hold: 1. (1) If $\sigma=\Delta\tau$ for some $\tau$, then $\int_{\mathcal{L}}\sigma=0$ for every a Lagrangian submanifold $\mathcal{L}$ of $\mathcal{M}$ on which $\sigma$ is integrable. 2. (2) If $\Delta\sigma=0$ and $\mathcal{L}_{t}$ is a smooth family of Lagrangian submanifolds of $\mathcal{M}$ on which $\sigma$ is integrable, then $I_{t}\coloneqq\int_{\mathcal{L}_{t}}\sigma$ is constant. In the case of an odd cotangent bundle $\Pi T^{*}M$, one can show that every Lagrangian submanifold is a smooth deformation of an odd conormal bundle $\Pi N^{*}C$, where $C$ is a submanifold of $M$. Moreover, one can show that $\int_{\Pi N^{*}C}\sigma=\int_{C}\phi(\sigma)$ for every half-density $\sigma$. By Stokes’ theorem and the characterization of the canonical BV operator as in Proposition 6.1 (i.e., $\phi\circ\Delta={\mathrm{d}}\circ\phi$), one gets that $\int_{\Pi N^{*}C_{1}}\sigma=\int_{\Pi N^{*}C_{2}}\sigma$ if $C_{1}$ and $C_{2}$ are homologous and $\Delta\sigma=0$. By Schwarz’ theorem one can then generalize the global BV theorem a bit. ###### Theorem B.4. Let $\mathcal{M}$ be an odd symplectic manifold and $\sigma$ a half-density satisfying $\Delta\sigma=0$. Then $\int_{\mathcal{L}_{1}}\sigma=\int_{\mathcal{L}_{2}}\sigma$ whenever $\mathcal{L}_{2}$ can be obtained from $\mathcal{L}_{1}$ by a combination of smooth deformations and homologous changes of the body, assuming that $\sigma$ is integrable on each intermediate step. ## Appendix C Historical remarks The BV formalism was developed by Batalin and Vilkovisky [1, 2] as a generalization of the BRST formalism [3, 15], which in turn put gauge fixing and the Faddeev–Popov determinant [5] into a cohomological setting. They constructud the BV Laplacian (as in our Section 2) and proved their fundamental theorem (in our notes, Theorem B.2). In addition, they showed that, under suitable assumptions, a physical action can be extended to a BV action $S$ satisfying the classical master equation $(S,S)=0$. In [16] Witten recognized the relation between the BV operator on functions on an odd cotangent bundle and the divergence operator on multivector fields on the body (see the last paragraph of Section 6). The relation between the de Rham differential and the divergence operator on multivector fields on an ordinary manifold was already known to Soviet mathematicians in the ’80s (see, e.g., [12]). Also note that Bernshtein and Leites [4] had already introduced in 1977 the notion of “integral forms” on supermanifolds—where this notion differs from that of differential forms—and defined the exterior differential for them as the divergence operator. Khudaverdian [6] was the first to give the global definition of the BV Laplacian on functions, see our equation (5). Khudaverdian was also the first to observe that an odd symplectic manifold always admits global Darboux coordinates. This was used by Schwarz in [13], where he also observed how to globalize and extend the BV theorem (in our notes, Theorem B.3). Finally, in [7, 8],232323Note that O. M. Khudaverdian and H. M. Khudaverdian are just different spellings of the same name. Khudaverdian showed the existence of a canonical BV Laplacian on half-densities (in our notes, this corresponds to Theorem/Definition 5.11). In a series of papers (among others, [9, 10, 11]), Khudaverdian and Voronov further clarified the properties of the canonical BV Laplacian, essentially covering all the constructions we present in these notes (and more). In [9] they extended the construction of the odd Laplacian on half-densities to odd Poisson manifolds. In [10] they discussed the principal and subprincipal symbols of the BV Laplacians and their transposed operators (see our Section 5). In [11] they defined the canonical BV Laplacian on half-densities on an odd cotangent bundle in terms of the de Rham differential of the corresponding differential forms and then showed that this operator is invariant under all symplectomorphisms—not just those coming from the base. It is worth mentioning that Ševera [14] found a completely different construction for the canonical BV Laplacian. The main observation is that the complex of differential forms on $\mathcal{M}$ has two commuting coboundary operators: the de Rham differential ${\mathrm{d}}$ and the operator $\delta\coloneqq\omega\wedge\ $. It turns out that the associated spectral sequence lives up to the $E_{2}$-term. More precisely, Ševera shows that $E_{1}=H_{\delta}(\mathcal{M})$ is canonically isomorphic to $\operatorname{Dens}^{\frac{1}{2}}(\mathcal{M})$ and that the induced coboundary operator ${\mathrm{d}}_{1}$ vanishes, which implies that $E_{2}=E_{1}=\operatorname{Dens}^{\frac{1}{2}}(\mathcal{M})$. Finally, Ševera proves that the canonically induced coboundary operator ${\mathrm{d}}_{2}$ is the canonical BV Laplacian and that all higher coboundary operators vanish. ## References * [1] I. A. Batalin and G. A. Vilkovisky, “Relativistic S-matrix of dynamical systems with boson and fermion constraints,” Phys. Lett. B 69, 309–312 (1977). * [2] I. A. Batalin and G. A. Vilkovisky, “Gauge algebra and quantization,” Phys. Lett. B 102, 27–31 (1981). * [3] C. Becchi, A. Rouet, and R. Stora, “Renormalization of the abelian Higgs–Kibble model,” Commun. Math. Phys. 42, 127–162 (1975). * [4] I. N. Bernshtein and D. A. Leites, “Integral forms and the Stokes formula on supermanifolds,” Funkts. Anal. Prilozhen. 11, 55-56 (1977). * [5] L. D. Faddeev and V. Popov, “Feynman diagrams for the Yang–Mills field,” Phys. Lett. B 25, 29-30 (1967). * [6] O. M. Khudaverdian, “Geometry of superspace with even and odd brackets,” J. Math. Phys. 32, 1934–1937, (1991). * [7] O. M. Khudaverdian, “Delta-operator on semidensities and integral invariants in the Batalin–Vilkovisky geometry,” arXiv:math/9909117. * [8] H. M. Khudaverdian, “Semidensities on odd symplectic supermanifolds,” Commun. Math. Phys. 247, 353–390 (2004). * [9] H. M. Khudaverdian and T. Voronov, “On odd Laplace operators,” Lett. Math. Phys. 62, 127–142 (2002). * [10] H. M. Khudaverdian and T. Voronov, “On odd Laplace operators. II,” arXiv:math/0212311. * [11] H. M. Khudaverdian and T. Voronov, “Differential forms and odd symplectic geometry,” In: Geometry, Topology and Mathematical Physics. S. P. Novikov seminar: 2006-2007, (V. M. Buchstaber and I. M. Krichever, eds.) AMS Translations, Ser. 2, Vol. 224, Amer. Math. Soc., Providence, RI, 2008, 159–171, arXiv:math/0606560. * [12] A. A. Kirillov, “Invariant operators on geometric quantities,” J. of Soviet Mathematics 18, 1–21 (1982). * [13] A. Schwarz, “Geometry of Batalin–Vilkovisky quantization,”Commun. Math. Phys. 155, 249–260 (1993). * [14] P. Ševera, “On the origin of the BV operator on odd symplectic supermanifolds,” Lett. Math. Phys. 78, 55–59 (2006). * [15] I. V. Tyutin, “Gauge invariance in field theory and statistical physics in operator formalism,” Lebedev Physics Institute preprint 39 (1975), arXiv:0812.0580. * [16] E. Witten, “A note on the antibracket formalism,” Mod. Phys. Lett. A 5, 487–494 (1990).
# Active vector models generalizing 3D Euler and electron–MHD equations Dongho Chae Department of Mathematics, Chung-Ang University, Heuk-Seok Ro 84, Seoul, Republic of Korea 156-756. E-mail<EMAIL_ADDRESS>In-Jee Jeong Department of Mathematical Sciences and RIM, Seoul National University, Gwanak-gu, Gwanak 1, Seoul, Republic of Korea 08826. E-mail<EMAIL_ADDRESS> ###### Abstract We introduce an active vector system, which generalizes both the 3D Euler equations and the electron–magnetohydrodynamic equations (E–MHD). We may as well view the system as singularized systems for the 3D Euler equations, in which case the equations of (E–MHD) correspond to the order two more singular one than the 3D Euler equations. The generalized surface quasi-geostrophic equation (gSQG) can be also embedded into a special case of our system when the unknown functions are constant in one coordinate direction. We investigate some basic properties of this system as well as the conservation laws. In the case when the system corresponds up to order one more singular than the 3D Euler equations, we prove local well-posedness in the standard Sobolev spaces. The proof crucially depends on a sharp commutator estimate similar to the one used for (gSQG) in [11]. Since the system covers many areas of both physically and mathematically interesting cases, one can expect that there are various related problems to be investigated, parts of which are discussed here. ## 1 Introduction ### 1.1 The generalized SQG equations The generalized surface quasi-geostrophic (gSQG) equation is an active scalar system given in $\mathbb{R}^{2}$ as : $\left\\{\begin{aligned} &\partial_{t}\theta+v\cdot\nabla\theta=0,\\\ &v=\nabla^{\perp}\Gamma\theta,\end{aligned}\right.$ (gSQG) with $\theta(t,\cdot):\Omega\rightarrow\mathbb{R}$, where $\Gamma$ is a Fourier multiplier. In the recent years, the model (gSQG) together with its dissipative analogues has been intensively studied, with topics including local and global well-posedness ([14, 13, 12, 11, 85, 86, 28, 31, 31, 33, 58, 80, 89, 69, 70, 63, 62, 36, 84, 50, 10, 16, 17, 57, 78, 83]), finite and infinite time singularity formation ([73, 60, 74, 76, 46, 75, 92, 90, 47]), rotating and traveling-wave solutions ([40, 59, 39, 61, 8, 9, 41, 6, 7, 54, 55, 56]), non-uniqueness ([64, 24, 3, 88]), and so on. For any $\Gamma$, the system describes the transport of a scalar by an incompressible flow. The primary motivation for studying the system (gSQG) comes from the fact that (gSQG) interpolates the two physically important cases for $\Gamma=(-\Delta)^{-1}$ and $\Gamma=(-\Delta)^{-\frac{1}{2}}$, where (gSQG) corresponds to 2D Euler equations and the surface quasi-geostrophic (SQG) equations, respectively. The SQG equations were derived in [35, 34] from the system for non-homogeneous three-dimensional half-space rotating fluid under the so-called quasi-geostrophic approximation. Various interesting phenomena occur for solutions to (gSQG) with different choice of multipliers; see recent surveys [28, 77, 72] as well as aforementioned references. ### 1.2 The active vector system In this note, we introduce an active vector system defined in $\mathbb{R}^{3}$, which not only generalizes two physically important systems, namely the 3D Euler equations and the electron–MHD equations (also referred to as the Hall equations) but also contains the entire family of (gSQG) as sub- systems. The system we suggest is given by $\left\\{\begin{aligned} &\partial_{t}B+\nabla\times((\nabla\times\Gamma[B])\times B)=0,\\\ &\nabla\cdot B=0,\\\ \end{aligned}\right.$ (1.1) where $B(t,\cdot):\mathbb{R}^{3}\rightarrow\mathbb{R}^{3}$ and $\Gamma$ is a Fourier multiplier whose symbol will be denoted by $\gamma=\gamma(|\xi|)\geq 0$. Since $B$ is divergence-free, (1.1) can be written as an active vector form: $\left\\{\begin{aligned} &\partial_{t}B+V\cdot\nabla B=B\cdot\nabla V,\\\ &V=-\nabla\times\Gamma[B],\\\ &\nabla\cdot B=0.\end{aligned}\right.$ (1.2) The important special cases $\Gamma=(-\Delta)^{-1}$ and $\Gamma=1$ correspond to the 3D Euler equations (with $B=\nabla\times V$ being the vorticity) and the so-called electron magneto-hydrodynamics (E–MHD) system ($B$ being the magnetic field), which can be obtained from the Hall–MHD system by setting the fluid velocity to be zero ([1, 4, 81, 87]). That is, the equation for the magnetic field reads $\left\\{\begin{aligned} &\partial_{t}B+\nabla\times((\nabla\times B)\times B)=0,\\\ &\nabla\cdot B=0.\\\ \end{aligned}\right.$ (E–MHD) Recently, the Hall–MHD, E–MHD systems and their viscous counterparts have received a lot of attention in the PDE community ([15, 18, 19, 20, 21, 42, 43, 44, 45]), not only because these equations are important in various physical phenomena, but also they naturally appear in hydrodynamic limit problems ([1]) and raise new technical difficulties. Already at the level of proving well- posedness, (E–MHD) can be considered as a transport equation for $B$ with advecting velocity $V=\nabla\times B$, which is one order more singular than $B$. This is in stark contrast with the three-dimensional incompressible Euler equations ((1.2)), where $V=\nabla\times(-\Delta)^{-1}B$ is one order more regular than $B$. The 3D Euler equations have solutions which blow up in finite and infinite time ([52, 53, 23, 51, 25, 26]). Therefore, it will be interesting to investigate the dynamics of solutions for the interpolating systems, which was precisely the motivation for studying the generalized SQG equations. Somewhat surprisingly, it turns out that if we consider the special case of (1.2) with $B(t,x)=(0,0,\theta(t,x_{1},x_{2})),$ and $V=\nabla^{\perp}\Gamma[\theta]$, then we obtain precisely (gSQG), see 2.2 for details. That is, any gSQG equation can be embedded into our system (1.1), so that (1.1) contains all the phenomena observed in the gSQG case. In the current work, we shall focus on the multipliers $\Gamma$ which lie in between those above two cases and discuss well/ill-posedness of the Cauchy problem for (1.1). ###### Remark 1.1. The term “active vector” was used in [29] to denote transport systems for vector-valued functions associated with the incompressible Euler equations. We generalize the notion of an active vector to allow a stretching term on the right hand side. ### 1.3 Main result Our first main result provides local well-posedness of smooth solutions for (1.1) when the operator $\Gamma$ is up to one order more singular than the case of the 3D Euler equations. The precise assumptions we impose on $\Gamma$ are given as follows. Assumptions on $\Gamma$. Let $\Lambda=(-\Delta)^{\frac{1}{2}}$ be the Zygmund operator. We shall assume that $\Gamma$ satisfies * • $\Gamma\lesssim\max\\{\Lambda^{-1},\Lambda^{-2}\\}$, $\Gamma$ is decreasing with $\Lambda|\Gamma^{\prime}|\lesssim\Gamma$: to be precise, for any $|\xi|>0$, we have $\begin{split}\gamma(|\xi|)\leq{C(|\xi|^{-1}+|\xi|^{-2})}\end{split}$ (1.3) and $\begin{split}\gamma^{\prime}(|\xi|)\leq 0,\qquad|\xi||\gamma^{\prime}(|\xi|)|\leq C\gamma(|\xi|)\end{split}$ (1.4) for some $C>0$. * • $\Gamma\gtrsim\min\\{\Lambda^{-2},\Lambda^{-1}\\}$: that is, $\begin{split}\gamma(|\xi|)\geq c\,{\min\\{|\xi|^{-2},|\xi|^{-1}\\}}\end{split}$ (1.5) for all $|\xi|>0$ with some $c>0$. This assumption ensures that $\nabla\times\Gamma[B]$ is well-defined as an $L^{2}$ function for $B\in\dot{H}^{-1}\cap L^{2}(\mathbb{R}^{3})$. Examples of $\Gamma$. The assumptions are clearly satisfied by $\Gamma=\Lambda^{-a}$ (fractional powers of the negative Laplacian), with any $1\leq a\leq 2$. It is not difficult to check that as long as $1<a<2$, one can take less standard symbols $\Gamma=\Lambda^{-a}\log^{\alpha_{1}}(10+\Lambda)$, $\Lambda^{-a}\log^{\alpha_{1}}(10+\Lambda)\log^{\alpha_{2}}(10+\log(10+\Lambda))$, etc, for any $\alpha_{i}$. We now introduce the function space $\mathcal{Y}^{1}(\mathbb{R}^{n})$ defined by $\mathcal{Y}^{1}(\mathbb{R}^{n})=\left\\{f\in\mathcal{D}^{\prime}(\mathbb{R}^{n})\,|\,\int_{\mathbb{R}^{n}}{(1+|\xi|)}|\hat{f}(\xi)|\,\mathrm{d}\xi=:\|f\|_{\mathcal{Y}^{1}}<+\infty\right\\},$ which is a slight variant of the $\mathcal{X}^{1}$ space used in [32, 79]. We are now ready to state our main result. ###### Theorem A. We have the following local well-posedness and blow-up criterion for (1.1). * (i) (Local well-posedness) For any $s>\frac{5}{2}$, the system (1.1) is locally well-posed in $H^{s}(\mathbb{R}^{3})$: given a divergence-free initial data $B_{0}\in H^{s}$, there exist $T>0$ and a unique solution $B\in C([0,T];H^{s}(\mathbb{R}^{3}))$ with $B(t=0)=B_{0}$. * (ii) (Blow up criterion) Assume that $B$ is a solution to (1.1) belonging to $C([0,T];H^{s}(\mathbb{R}^{3}))$. Then, $\limsup_{t\to T}\|B\|_{H^{s}}=+\infty\quad\text{ if and only if }\quad\int_{0}^{T}{\|B\|_{\mathcal{Y}^{1}}}\,\mathrm{d}t=+\infty.$ Therefore, the solution can be continued past $T$ if and only if $\int_{0}^{T}{\|B\|_{\mathcal{Y}^{1}}}\,\mathrm{d}t<\infty$. When the multiplier $\Gamma$ is more singular, we can prove local well- posedness under the presence of a dissipative term. For concreteness, we focus on the case $\Gamma=\Lambda^{-a}$ with $0\leq a<1$, whose local well-posedness is not covered by Theorem A in the inviscid case. ###### Theorem B. Consider the fractionally dissipative system $\left\\{\begin{aligned} &\partial_{t}B+\nabla\times((\nabla\times\Lambda^{-a}[B])\times B)=-\Lambda^{b}B,\\\ &\nabla\cdot B=0.\end{aligned}\right.$ (1.6) Then, for $0\leq a<1$ and $1-a<b$, the Cauchy problem for (1.6) is locally well-posed in $C([0,T];H^{s}(\mathbb{R}^{3}))$ in the same sense as in Theorem A, as long as $s>\frac{7}{2}-a$. Sharpness of Theorems A and B. We emphasize that the local well-posedness statements in Theorems A and B are not simple consequences of standard energy estimates, even in the case of $\Gamma=\Lambda^{-a}$. To explain the sharpness of Theorem A, one can note that already in the case of 3D Euler where $\Gamma=\Lambda^{-2}$, $\nabla V$ in the right hand side of (1.2) is already as singular as $B$, so that as soon as $\Gamma=\Lambda^{-a}$ with $a<2$, $\nabla V$ is strictly more singular than $B$. Therefore, to close the a priori estimate in Sobolev spaces in the range $a\in[1,2)$, a cancellation structure and sharp commutator estimate must be used, which can be considered as an extension of the one proved in [11] for local well-posedness of gSQG equations in the “singular” regime. Furthermore, we believe that the threshold $a=1$ is sharp for local well-posedness; in a recent work [65], strong ill- posedness of the Cauchy problem for the electron–MHD system ((1.1) with $\Gamma=1$) was established in any sufficiently regular Sobolev, Hölder and even Gevrey spaces. From the paper [65], there are many reasons to believe that the loss of derivative in the electron–MHD system is precisely equal to 1, suggesting ill-posedness of (1.1) in the whole range of multipliers $\Gamma=\Lambda^{-a}$ with $0<a<1$. Formally, taking $\Gamma=\Lambda^{-a}$ and plugging in the ansatz $j\simeq\Lambda^{-1}b^{z}$ (which is reasonable since $b^{z}$ should be comparable with $\nabla j$), we have that $\begin{split}\partial_{t}b^{z}+\nabla^{\perp}\Gamma b^{z}\cdot\nabla b^{z}-\nabla^{\perp}\Lambda^{-1}b^{z}\cdot\nabla\Lambda^{1-a}b^{z}\simeq 0,\end{split}$ which is expected to be ill-posed, due to the singular term $\nabla\Lambda^{1-a}b^{z}$. The methods developed in [65, 66, 16] should be applicable to give ill-posedness, but the proof will be highly involved since (1.1) is a three-dimensional system involving a non-local operator. Similarly, we expect the restriction $1-a<b$ in Theorem B to be sharp; indeed, in the electron–MHD case $a=0$, while Theorem B requires $b>1$, strong ill-posedness was shown whenever $b\leq 1$ ([65]). Furthermore, it can be considered as an extension of the local well-posedness results of [15, 21] in the case $b=2$. Organization of the paper. Theorems A and B will be proved in Section 3 below, after discussing a few basic properties in Section 2. Possible extensions and open problems will be discussed in Section 4. ## 2 Basic properties of the active vector system ### 2.1 Conservation laws We can also write (1.1) in another velocity formulation. Let us consider $u$, which is the solution of $\nabla\times u=B$ satisfying the “gauge fixing condition” $\nabla\cdot u=0$. We also assume $u$ decays sufficiently fast at infinity. Then, since (1.1) is written as $\nabla\times(\partial_{t}u+(\nabla\times u)\times\Delta\Gamma u)=0,$ there exists a scalar function $Q=Q(x,t)$ such that $\left\\{\begin{aligned} &\partial_{t}u+(\nabla\times u)\times\Delta\Gamma u=-\nabla Q,\\\ &\nabla\cdot u=0.\end{aligned}\right.$ (2.1) We recall that in the case of Euler equations, $Q=p+\frac{|u|^{2}}{2}$, where $p$ is the pressure. Comparing (2.1) with (1.1), we find the following relation between the “two velocities”, $V=\Delta\Gamma[u].$ (2.2) We present the basic conservation laws for the active vector system. Among others please note that the helicity defined below is independent of the multiplier $\Gamma$ in (1.1). ###### Proposition 2.1 (Conservation laws). For a sufficiently smooth and decaying solutions of (1.1), we have the following conservations. * (i) (Energy) Let us define the energy: $E(t)=\frac{1}{2}\int_{\mathbb{R}^{3}}|\Gamma^{\frac{1}{2}}B|^{2}\,\mathrm{d}x=\frac{1}{2}\int_{\mathbb{R}^{3}}|(-\Delta)^{\frac{1}{2}}\Gamma^{\frac{1}{2}}u|^{2}\,\mathrm{d}x.$ Then, $E(t)=E(0).$ * (ii) (Helicity) We define the helicity: $H(t)=\int_{\mathbb{R}^{3}}B\cdot\nabla\times(-\Delta)^{-1}B\,\mathrm{d}x=\int_{\mathbb{R}^{3}}B\cdot u\,\,\mathrm{d}x.$ Then, $H(t)=H(0)$. ###### Proof. * (i) Taking the $L^{2}(\mathbb{R}^{3})$ inner product of (1.1) with $\Gamma B$ and integrating by part, we have that $\begin{split}\frac{1}{2}\frac{\mathrm{d}E(t)}{\mathrm{d}t}&=-\int_{\mathbb{R}^{3}}\nabla\times((\nabla\times\Gamma B)\times B)\cdot\Gamma B\,\mathrm{d}x\\\ &=-\int_{\mathbb{R}^{3}}((\nabla\times\Gamma B)\times B)\cdot(\nabla\times\Gamma B)\,\mathrm{d}x=0.\end{split}$ * (ii) Using (1.1) and (2.1), and integrating by part we compute $\displaystyle\frac{\mathrm{d}H(t)}{\mathrm{d}t}$ $\displaystyle=\int_{\mathbb{R}^{3}}B_{t}\cdot u\,\,\mathrm{d}x+\int_{\mathbb{R}^{3}}B\cdot u_{t}\,\,\mathrm{d}x$ $\displaystyle=-\int_{\mathbb{R}^{3}}\nabla\times\left((\nabla\times\Gamma[B])\times B\right)\cdot u\,\,\mathrm{d}x-\int_{\mathbb{R}^{3}}B\cdot\left((\nabla\times u)\times\Delta\Gamma[u]+\nabla Q\right)\,\,\mathrm{d}x$ $\displaystyle=-\int_{\mathbb{R}^{3}}\ (\nabla\times\Gamma[B])\times B\cdot B\,\,\mathrm{d}x-\int_{\mathbb{R}^{3}}B\cdot B\times\Delta\Gamma[u]\,\,\mathrm{d}x-\int_{\mathbb{R}^{3}}B\cdot\nabla Q\,\mathrm{d}x$ $\displaystyle=0.$ This finishes the proof. ∎ We shall observe that the formulation (1.2), viewing $V$ as a velocity field, is useful in the Lagrangian formulation of (1.1). Let us introduce the particle trajectory mapping $X(\alpha,t)$ generated by $V$ defined by $\frac{\partial X(t,\alpha)}{\partial t}=V(t,X(t,\alpha))\quad;\quad X(0,\alpha)=\alpha.$ As an immediate consequence of (1.2) the following Cauchy formula $B(t,X(t,\alpha))=B_{0}(\alpha)\cdot\nabla X(t,\alpha)$ (2.3) holds. For details of the proof see e.g. [82, Lemma 1.4], which is obviously applicable to our case. Let the initial closed curve $\mathcal{C}_{0}$ be an integral curve of $B_{0}$, namely there exists $\lambda(\cdot):[0,1)\to\mathbb{R}$ such that $\mathcal{C}_{0}=\\{\eta(s)\in\mathbb{R}^{3}\,|\,s\in[0,1],\eta(0)=\eta(s),\,\eta^{\prime}(s)=\lambda(s)B_{0}(\eta(s))\\}.$ (2.4) and define $\mathcal{C}_{t}=X(t,\mathcal{C}_{0})=\\{X(t,\eta(s))\in\mathbb{R}^{3}\,|\,s\in[0,1],\eta(0)=\eta(s),\,\eta^{\prime}(s)=\lambda(s)B_{0}(\eta(s))\\}.$ (2.5) Then, we claim $C_{t}$ is an integral curve of $B(\cdot,t)$, namely $\frac{\partial X(t,\eta(s))}{\partial s}=\lambda(s)B(t,\eta(s))\quad\forall s\in[0,1],\quad t>0.$ (2.6) Indeed, using (2.3), we have $\displaystyle\frac{\partial X(t,\eta(s))}{\partial s}$ $\displaystyle=\frac{d\eta(s)}{ds}\cdot\nabla X(t,\alpha)=\lambda(s)B_{0}(\eta(s))\cdot\nabla X(t,\alpha)$ $\displaystyle=\lambda(s)B(t,\eta(s)).$ ### 2.2 $2+\frac{1}{2}$-dimensional case The system (1.1) formally satisfies translational invariance, and therefore we may consider special solutions which are not dependent on the third coordinate. Such a reduction is sometimes referred to as the $2+\frac{1}{2}$–dimensional reduction ([82]). Both 3D Euler and Hall-MHD systems have been extensively studied under this simplifying ansatz ([2, 48, 67, 65, 49, 91, 22]). From the $z$-independence and divergence-free condition, we can write $\begin{split}B=(-\partial_{y}j,\partial_{x}j,b^{z}).\end{split}$ For simplicity we set $G=\nabla\times\Gamma B$ and write $G=(g^{x},g^{y},g^{z})$. From $\begin{split}\partial_{t}B+B\cdot\nabla G-G\cdot\nabla B=0,\end{split}$ taking the $z$-component gives $\begin{split}\partial_{t}b^{z}-G\cdot\nabla b^{z}+B\cdot\nabla g^{z}=0.\end{split}$ We now note that $\begin{split}B_{h}=\nabla^{\perp}j,\qquad G_{h}=-\nabla^{\perp}\Gamma b^{z},\qquad g^{z}=\Gamma\Delta j.\end{split}$ ($V_{h}$ stands for the vector consisting of the first two components of $V$. Moreover, $\nabla^{\perp}=(-\partial_{y},\partial_{x})$.) Then, the equation for $b^{z}$ can be written as $\begin{split}\partial_{t}b^{z}+\nabla^{\perp}\Gamma b^{z}\cdot\nabla b^{z}+\nabla^{\perp}j\cdot\nabla\Gamma\Delta j=0.\end{split}$ Next, from $\begin{split}\partial_{t}b^{x}-G\cdot\nabla b^{x}+B\cdot\nabla g^{x}=0,\end{split}$ noting that $\begin{split}-G\cdot\nabla b^{x}+B\cdot\nabla g^{x}&=-g^{x}\partial_{x}b^{x}-g^{y}\partial_{y}b^{x}+b^{x}\partial_{x}g^{x}+b^{y}\partial_{y}g^{x}\\\ &=\partial_{y}(g^{x}b^{y})-\partial_{y}(g^{y}b^{x})\end{split}$ (we have used the divergence-free conditions $\partial_{x}b^{x}+\partial_{y}b^{y}=0$, $\partial_{x}g^{x}+\partial_{y}g^{y}=0$), we have that $\begin{split}\partial_{t}j+\nabla^{\perp}\Gamma b^{z}\cdot\nabla j=0.\end{split}$ We have arrived at the closed system for $b^{z}$ and $j$: $\left\\{\begin{aligned} &\partial_{t}b^{z}+\nabla^{\perp}\Gamma b^{z}\cdot\nabla b^{z}+\nabla^{\perp}j\cdot\nabla\Gamma\Delta j=0,\\\ &\partial_{t}j+\nabla^{\perp}\Gamma b^{z}\cdot\nabla j=0.\end{aligned}\right.$ (2.7) In particular, note that when $j_{0}=0$, then the second equation of (2.7) implies that $j(\cdot,t)=0$ for $t>0$ as long as the smooth solution persists. Then, we have simply $\begin{split}\partial_{t}b^{z}+\nabla^{\perp}\Gamma b^{z}\cdot\nabla b^{z}=0,\end{split}$ which is (gSQG). In this sense the (gSQG) systems are embedded into (1.1). In the case of $\Gamma=\Delta^{-1}$, this is nothing but the 2D Euler equations. ## 3 Proof of the main theorem We first prove the the following key commutator estimate. ###### Lemma 3.1. Let $\Lambda=(-\Delta)^{\frac{1}{2}}$ and $\Gamma$ satisfy the assumptions (1.3)–(1.5). For $g\in L^{2}(\mathbb{R}^{n})$ and ${b}\in\mathcal{Y}^{1}$, we have $\begin{split}\|{\mathbf{P}_{\geq 1}\Lambda^{\frac{1}{2}}[\Gamma^{\frac{1}{2}},b]\nabla g}\|_{L^{2}}\leq C\|{b}\|_{\mathcal{Y}^{1}}\|{g}\|_{L^{2}}\end{split}$ (3.1) and $\begin{split}\|{\Lambda^{\frac{1}{2}}\mathbf{P}_{<1}\Lambda^{\frac{1}{2}}[\Gamma^{\frac{1}{2}},b]\nabla g}\|_{L^{2}}\leq C\|{b}\|_{\mathcal{Y}^{1}}\|{g}\|_{L^{2}}.\end{split}$ (3.2) Here, $\mathbf{P}_{\geq 1}$ is defined with Fourier multiplier $\mathbf{1}_{|\xi|\geq 1}$ and $\mathbf{P}_{<1}=1-\mathbf{P}_{\geq 1}$. ###### Proof. Taking the Fourier transform of $f:=\Lambda^{\frac{1}{2}}[\Gamma^{\frac{1}{2}},b]\nabla g$, we have $\begin{split}\widehat{f}(\xi)=|\xi|^{\frac{1}{2}}\int_{\mathbb{R}^{n}}\left(\gamma^{\frac{1}{2}}(|\xi|)-\gamma^{\frac{1}{2}}(|\eta|)\right)\widehat{b}(\xi-\eta)i\eta\widehat{g}(\eta)\,\mathrm{d}\eta.\end{split}$ We bound, using the assumptions for $\gamma$, $\begin{split}|\gamma^{\frac{1}{2}}(|\xi|)-\gamma^{\frac{1}{2}}(|\eta|)|&=\frac{|\gamma(|\xi|)-\gamma(|\eta|)|}{\gamma^{\frac{1}{2}}(|\xi|)+\gamma^{\frac{1}{2}}(|\eta|)}{=}\frac{1}{\gamma^{\frac{1}{2}}(|\xi|)+\gamma^{\frac{1}{2}}(|\eta|)}\int_{|\xi|}^{|\eta|}-\gamma^{\prime}({\rho})\,\mathrm{d}{\rho}\\\ &\leq\int_{|\xi|}^{|\eta|}C\rho^{-1}\frac{\gamma(\rho)}{{\gamma^{\frac{1}{2}}(|\xi|)+\gamma^{\frac{1}{2}}(|\eta|)}}\,\mathrm{d}{\rho}\leq C\int_{|\xi|}^{|\eta|}{\rho}^{-1}\gamma^{\frac{1}{2}}({\rho})\,\mathrm{d}{\rho}\\\ &\leq C|\xi-\eta|\left(\frac{1}{(|\xi||\eta|)^{\frac{1}{2}}(|\xi|^{\frac{1}{2}}+|\eta|^{\frac{1}{2}})}+\frac{1}{|\xi||\eta|}\right)\\\ &\leq\frac{C|\xi-\eta|}{|\eta|}\left(|\xi|^{-\frac{1}{2}}+|\xi|^{-1}\right).\end{split}$ Here, we have assumed $|\xi|\leq|\eta|$, and used both of (1.4) and (1.3), but the same inequality can be trivially deduced when $|\xi|>|\eta|$ as well. Inserting this bound into the integral expression for $\widehat{f}(\xi)$, we have that $\begin{split}\left|\widehat{f}(\xi)\right|\leq C{(1+|\xi|^{-\frac{1}{2}})}\int_{\mathbb{R}^{n}}|\xi-\eta||\widehat{b}(\xi-\eta)||\widehat{g}(\eta)|\,\mathrm{d}\eta.\end{split}$ By Young’s convolution inequality, we obtain $\|{\mathbf{P}_{\geq 1}f}\|_{L^{2}}=C\|{\mathbf{1}_{|\xi|\geq 1}\widehat{f}}\|_{L^{2}}\leq C\|b\|_{\mathcal{Y}^{1}}\|g\|_{L^{2}}$ and $\begin{split}\|{\Lambda^{\frac{1}{2}}\mathbf{P}_{<1}f}\|_{L^{2}}=C\|{|\xi|^{\frac{1}{2}}\mathbf{1}_{|\xi|<1}\widehat{f}}\|_{L^{2}}\leq C\|b\|_{\mathcal{Y}^{1}}\|g\|_{L^{2}}.\end{split}$ This finishes the proof. ∎ When the operator $\Gamma$ is given precisely by $\Lambda^{-a}$, we can obtain the following sharp inequality, with a straightforward modification of the above proof. ###### Corollary 3.2. For $g\in L^{2}(\mathbb{R}^{n})$ and ${b}\in\dot{\mathcal{Y}}^{1}(\mathbb{R}^{n})$, we have $\begin{split}\|{\Lambda^{\frac{a}{2}}[\Lambda^{-\frac{a}{2}},b]\nabla g}\|_{L^{2}}\leq C\|{b}\|_{\dot{\mathcal{Y}}^{1}}\|{g}\|_{L^{2}},\quad\|{b}\|_{\dot{\mathcal{Y}}^{1}}:=\int_{\mathbb{R}^{n}}|\xi||\widehat{b}(\xi)|\,\mathrm{d}\xi.\end{split}$ (3.3) ###### Lemma 3.3. The following embedding relations hold. $H^{s}(\mathbb{R}^{n})\hookrightarrow\mathcal{Y}^{1}(\mathbb{R}^{n})\hookrightarrow\dot{W}^{1,\infty}(\mathbb{R}^{n})$ (3.4) for $s>\frac{n}{2}+1.$ ###### Proof. For all $x\in\mathbb{R}^{n}$ the following estimates hold. $\begin{split}|\nabla f(x)|&\leq\int_{\mathbb{R}^{n}}|\xi||\hat{b}(\xi)|\mathrm{d}\xi{\leq}\|f\|_{\mathcal{Y}^{1}}\leq\int_{\mathbb{R}^{n}}(1+|\xi|^{2})^{\frac{s}{2}}|\widehat{f}(\xi)|(1+|\xi|^{2})^{-\frac{s-1}{2}}\mathrm{d}\xi\\\ &\leq\left(\int_{\mathbb{R}^{n}}(1+|\xi|^{2})^{s}|\widehat{f}(\xi)|^{2}\mathrm{d}\xi\right)^{\frac{1}{2}}\left(\int_{\mathbb{R}^{n}}(1+|\xi|^{2})^{-s+1}\mathrm{d}\xi\right)^{\frac{1}{2}}\leq C\|{f}\|_{H^{s}},\end{split}$ where we used the fact $\int_{\mathbb{R}^{n}}(1+|\xi|^{2})^{-s+1}\,\mathrm{d}\xi<+\infty$ for $s>\frac{n}{2}+1$. Hence, we have shown $\|\nabla f\|_{L^{\infty}}\leq\|f\|_{\mathcal{Y}^{1}}\leq C\|f\|_{H^{s}}$, and the lemma is proved. ∎ Lastly, we shall need a simple commutator estimate. ###### Lemma 3.4. For two functions $f,g$ belonging to $H^{s}$, we have $\begin{split}\|{[\Lambda^{s},f\cdot\nabla]g}\|_{L^{2}}\leq C(\|{f}\|_{\mathcal{Y}^{1}}\|{g}\|_{H^{s}}+\|{g}\|_{\mathcal{Y}^{1}}\|{f}\|_{H^{s}}).\end{split}$ ###### Proof. Note that the Fourier transform of $H:=[\Lambda^{s},f\cdot\nabla]g$ is given by $\begin{split}\widehat{H}(\xi)=\int(|\xi|^{s}-|\eta|^{s})\widehat{f}(\xi-\eta)i\eta\widehat{g}(\eta)\,\mathrm{d}\eta.\end{split}$ Proceeding similarly with the proof of Lemma 3.1, we obtain that $\begin{split}\left|\widehat{H}(\xi)\right|\leq C\int(|\xi-\eta|^{s-1}+|\eta|^{s-1})|\xi-\eta||\widehat{f}(\xi-\eta)||\eta\widehat{g}(\eta)|\,\mathrm{d}\eta\end{split}$ which gives with Young’s convolution inequality that $\begin{split}\|{H}\|_{L^{2}}\leq C(\|{f}\|_{\mathcal{Y}^{1}}\|{g}\|_{H^{s}}+\|{g}\|_{\mathcal{Y}^{1}}\|{f}\|_{H^{s}}).\end{split}$ This gives the estimate. ∎ ###### Proof of Theorem A. In the proof below we shall find the following vector calculus identity useful: $\begin{split}\nabla\times(B\times F)=(\nabla\cdot F+F\cdot\nabla)B-(\nabla\cdot B+B\cdot\nabla)F.\end{split}$ (3.5) We compute $\begin{split}\frac{1}{2}\frac{\mathrm{d}}{\mathrm{d}t}\|{\Lambda^{s}B}\|_{L^{2}}^{2}&=-\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\Lambda^{s}\nabla\times((\nabla\times\Gamma B)\times B)\,\mathrm{d}x\end{split}$ The main contribution on the right hand side comes from $\begin{split}I_{1}:=-\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\nabla\times((\nabla\times\Gamma\Lambda^{s}B)\times B)\,\mathrm{d}x\end{split}$ and $\begin{split}I_{2}:=-\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\nabla\times((\nabla\times\Gamma B)\times\Lambda^{s}B)\,\mathrm{d}x.\end{split}$ We first treat $I_{1}$: defining $F=\nabla\times\Lambda^{s}B$ for simplicity, we have that $\begin{split}I_{1}&=-\int_{\mathbb{R}^{3}}F\cdot(\Gamma F\times B)\,\mathrm{d}x=\int_{\mathbb{R}^{3}}\epsilon_{ijk}\Gamma^{\frac{1}{2}}(F^{i}B^{k})\,\Gamma^{\frac{1}{2}}F^{j}\,\mathrm{d}x\\\ &=\int_{\mathbb{R}^{3}}\epsilon_{ijk}\Lambda^{\frac{1}{2}}[\Gamma^{\frac{1}{2}},B^{k}]F^{i}\,\Lambda^{-\frac{1}{2}}\Gamma^{\frac{1}{2}}F^{j}\,\mathrm{d}x,\end{split}$ where $\epsilon_{ijk}$ is the totally skew-symmetric tensor with the normalization $\epsilon_{123}=1$, and we used $\begin{split}\int_{\mathbb{R}^{3}}\epsilon_{ijk}B^{k}\Gamma^{\frac{1}{2}}(F^{i})\,\Gamma^{\frac{1}{2}}(F^{j})\,\mathrm{d}x=0\end{split}$ (repeated indices are being summed). We write $\begin{split}\int_{\mathbb{R}^{3}}\Lambda^{\frac{1}{2}}[\Gamma^{\frac{1}{2}},B^{k}]F^{i}\,\Lambda^{-\frac{1}{2}}\Gamma^{\frac{1}{2}}F^{j}\,\mathrm{d}x&=\int_{\mathbb{R}^{3}}\mathbf{P}_{\geq 1}\left(\Lambda^{\frac{1}{2}}[\Gamma^{\frac{1}{2}},B^{k}]F^{i}\right)\,\Lambda^{-\frac{1}{2}}\Gamma^{\frac{1}{2}}F^{j}\,\mathrm{d}x\\\ &\qquad+\int_{\mathbb{R}^{3}}\Lambda^{\frac{1}{2}}\mathbf{P}_{<1}\left(\Lambda^{\frac{1}{2}}[\Gamma^{\frac{1}{2}},B^{k}]F^{i}\right)\,\Lambda^{-1}\Gamma^{\frac{1}{2}}F^{j}\,\mathrm{d}x\end{split}$ Applying Lemma 3.1, and using the assumption (1.3), we have that $\begin{split}&\left|\int_{\mathbb{R}^{3}}\epsilon_{ijk}\Lambda^{\frac{1}{2}}[\Gamma^{\frac{1}{2}},B^{k}]F^{i}\,\Lambda^{-\frac{1}{2}}\Gamma^{\frac{1}{2}}F^{j}\,\mathrm{d}x\right|\\\ &\qquad\leq{C\sum_{i,j,k}\left(\|{\mathbf{P}_{\geq 1}\Lambda^{\frac{1}{2}}[\Gamma^{\frac{1}{2}},B^{k}]F^{i}}\|_{L^{2}}\|{\Lambda^{-\frac{1}{2}}\Gamma^{\frac{1}{2}}F^{j}}\|_{L^{2}}+\|{\Lambda^{\frac{1}{2}}\mathbf{P}_{<1}\Lambda^{\frac{1}{2}}[\Gamma^{\frac{1}{2}},B^{k}]F^{i}}\|_{L^{2}}\|{\Lambda^{-1}\Gamma^{\frac{1}{2}}F^{j}}\|_{L^{2}}\right)}\\\ &\qquad\leq C{\|{B}\|_{\mathcal{Y}^{1}}}\|\Lambda^{s}B\|_{L^{2}}{\left(\|{\Lambda^{-\frac{1}{2}}\Gamma^{\frac{1}{2}}\nabla\times\Lambda^{s}B}\|_{L^{2}}+\|{\Lambda^{-1}\Gamma^{\frac{1}{2}}\nabla\times\Lambda^{s}B}\|_{L^{2}}\right)}\\\ &\qquad\leq C{\|{B}\|_{\mathcal{Y}^{1}}}\|{B}\|_{H^{s}}^{2}.\end{split}$ We now treat $I_{2}$: using (3.5), $\begin{split}I_{2}&=-\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot(\Lambda^{s}B\cdot\nabla(\nabla\times\Gamma B))\,\mathrm{d}x+\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot((\nabla\times\Gamma B)\cdot\nabla)\Lambda^{s}B\,\mathrm{d}x\\\ &=I_{2,a}+I_{2,b}.\end{split}$ The term $I_{2,a}$ can be directly bounded by $\begin{split}I_{2,a}\leq C\|{\nabla^{2}\Gamma B}\|_{L^{\infty}}\|{B}\|_{H^{s}}^{2}\leq C{\|{B}\|_{\mathcal{Y}^{1}}\|B\|_{{H^{s}}}^{2}},\end{split}$ using (1.3) and $s>\frac{5}{2}$. For the second term, we note that $\begin{split}I_{2,b}=\int_{\mathbb{R}^{3}}((\nabla\times\Gamma B)\cdot\nabla)\frac{|\Lambda^{s}B|^{2}}{2}\,\mathrm{d}x=-\int_{\mathbb{R}^{3}}(\nabla\cdot(\nabla\times\Gamma B))\frac{|\Lambda^{s}B|^{2}}{2}\,\mathrm{d}x=0.\end{split}$ Finally, we treat the remainder. Using the identity (3.5), we observe that $\begin{split}&I_{1}+I_{2}+\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\Lambda^{s}\nabla\times((\nabla\times\Gamma B)\times B)\,\mathrm{d}x\\\ &\quad=\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\left\\{\Lambda^{s}(\nabla\times(G\times B))-\nabla\times(\Lambda^{s}G\times B)-\nabla\times(G\times\Lambda^{s}B)\right\\}\,\mathrm{d}x\\\ &\quad=\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\left\\{\Lambda^{s}(B\cdot\nabla G-G\cdot\nabla B)-B\cdot\nabla\Lambda^{s}G+\Lambda^{s}G\cdot\nabla B-\Lambda^{s}B\cdot\nabla G+G\cdot\nabla\Lambda^{s}B\right\\}dx\\\ &\quad=\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\\{[\Lambda^{s},B\cdot\nabla]G\\}dx-\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\\{[\Lambda^{s},G\cdot\nabla]B\\}dx\\\ &\qquad\qquad-\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\\{(\Lambda^{s}B\cdot\nabla)G\\}dx+\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\\{\Lambda^{s}G\cdot\nabla B\\}dx\\\ &\quad:=K_{1}+K_{2}+K_{3}+K_{4},\end{split}$ (3.6) where we set $G:=\nabla\times\Gamma B$. We note that $\begin{split}\|{\Lambda^{s}G}\|_{L^{2}}\leq C\|{B}\|_{H^{s}},\qquad\|{\nabla G}\|_{L^{\infty}}\leq C\|{B}\|_{{\mathcal{Y}}^{1}}.\end{split}$ To see the first inequality, $\begin{split}\|{\Lambda^{s}G}\|_{L^{2}}^{2}&\leq C\int|\xi|^{2s}|\xi\times\gamma(|\xi|)\hat{B}(\xi)|^{2}\mathrm{d}\xi\\\ &\leq C\int|\xi|^{2s+2}(|\xi|^{-1}+|\xi|^{-2})^{2}|\hat{B}(\xi)|^{2}\mathrm{d}\xi\leq C\|{B}\|_{H^{s}}^{2},\end{split}$ using the assumption (1.3) for $\gamma$. The proof of the other inequality is similar. Then, we can bound $K_{4}$ by $\left|K_{4}\right|\leq C\|{\Lambda^{s}B}\|_{L^{2}}\|{\Lambda^{s}G}\|_{L^{2}}\|{\nabla B}\|_{L^{\infty}}\leq C\|B\|_{\dot{\mathcal{Y}}^{1}}\|{B}\|_{H^{s}}^{2}.$ Next, we estimate $K_{3}$ easily as follows: $\begin{split}\left|K_{3}\right|\leq\int_{\mathbb{R}^{3}}|\Lambda^{s}B\cdot\\{\Lambda^{s}B\cdot\nabla G\\}|\,\mathrm{d}x\leq C\|{\nabla G}\|_{L^{\infty}}\|{B}\|_{H^{s}}^{2}\leq C\|B\|_{\dot{\mathcal{Y}}^{1}}\|{B}\|_{H^{s}}^{2}.\end{split}$ Lastly, using Lemma 3.4, we can directly estimate the first two terms by $\begin{split}\left|K_{1}+K_{2}\right|&\leq\|\Lambda^{s}B\|_{L^{2}}\left(\|{[\Lambda^{s},B\cdot\nabla]G}\|_{L^{2}}+\|{[\Lambda^{s},G\cdot\nabla]B}\|_{L^{2}}\right)\\\ &\leq C{(\|B\|_{\mathcal{Y}^{1}}+\|G\|_{\mathcal{Y}^{1}})}(\|{B}\|_{H^{s}}+\|{G}\|_{H^{s}})\|\Lambda^{s}B\|_{L^{2}}\\\ &\leq C\|B\|_{\mathcal{Y}^{1}}\|B\|_{H^{s}}^{2}.\end{split}$ This finishes the proof of $\begin{split}\frac{\mathrm{d}}{\mathrm{d}t}\|{B}\|_{{\dot{H}^{s}}}\leq C\|B\|_{\mathcal{Y}^{1}}\|{B}\|_{H^{s}}.\end{split}$ On the other hand, from (1.2), it is straightforward to obtain $\begin{split}\frac{\mathrm{d}}{\mathrm{d}t}\|{B}\|_{L^{2}}\leq C\|{\nabla^{2}\Gamma[B]}\|_{L^{\infty}}\|{B}\|_{L^{2}}\leq C\|B\|_{\mathcal{Y}^{1}}\|{B}\|_{H^{s}}.\end{split}$ Combining the above estimates together with (3.4), we have $\begin{split}\frac{\mathrm{d}}{\mathrm{d}t}\|{B}\|_{H^{s}}\leq C\|{B}\|_{H^{s}}^{2}.\end{split}$ Hence, there exists $T>0$ depending only on $\|{B_{0}}\|_{H^{s}}$, such that $\begin{split}\sup_{t\in[0,T]}\|{B(t,\cdot)}\|_{H^{s}}\leq 2\|{B_{0}}\|_{H^{s}}.\end{split}$ (3.7) Given the a priori estimate, it is a simple matter to prove existence and uniqueness of a solution belonging to $L^{\infty}([0,T];H^{s})$. For uniqueness, one may assume that there are two solutions $B$ and $\tilde{B}$ to (1.1) with the same initial data $B_{0}$, and that $B,\tilde{B}\in L^{\infty}([0,T];H^{s})$ for some $T>0$. Proceeding similarly as in the proof of (3.7), using the equation satisfied by the difference $B-\tilde{B}$, one can prove $\begin{split}\frac{d}{dt}\|{B-\tilde{B}}\|_{L^{2}}\leq C(\|{B}\|_{L^{\infty}([0,T];H^{s})}+\|{\tilde{B}}\|_{L^{\infty}([0,T];H^{s})})\|{B-\tilde{B}}\|_{L^{2}}.\end{split}$ (See [17, 27] for details of this argument for closely related systems.) For existence, we consider the regularized systems $\left\\{\begin{aligned} &\partial_{t}B+\nabla\times((\nabla\times\Gamma[B])\times B)-\nu\Delta B=0,\\\ &\nabla\cdot B=0,\\\ &B(t=0)=B_{0},\end{aligned}\right.$ (3.8) for each $\nu>0$. Given $B_{0}\in H^{s}$, it is straightforward to show local existence of a solution $B^{(\nu)}$ to (3.8) satisfying $B^{(\nu)}\in L^{\infty}([0,T_{\nu}];H^{s})$ for some $T_{\nu}>0$, using the mild formulation; one can follow the arguments of [15]. Since the a priori estimate (3.7) applies to the viscous solutions $B^{(\nu)}$, the lifespan $T_{\nu}$ can be bounded from below by some $T>0$ independent of $\nu>0$. There exists a weak limit $B^{(\nu)}\to B$ in $L^{\infty}([0,T];H^{s})$ (by taking a sub- sequence if necessary), and the limit is a solution to (1.1). Lastly, we argue that the solution $B$ actually belongs to $C([0,T];H^{s})$. Using time-translation and time-reversal symmetries, it suffices to prove strong convergence $B(t)\to B_{0}$ for $t\to 0^{+}$. To begin with, from uniform boundedness, we obtain weak convergence; this gives that $\begin{split}\|{B_{0}}\|_{H^{s}}\leq\liminf_{t\to 0^{+}}\|{B(t)}\|_{H^{s}}.\end{split}$ On the other hand, from (3), we see that $\begin{split}\|{B_{0}}\|_{H^{s}}-Ct\leq\|{B(t)}\|_{{H^{s}}}\leq\|{B_{0}}\|_{H^{s}}+Ct,\quad\forall t\in[0,T_{0}]\end{split}$ for some $T_{0},C>0$ depending only on $\|{B_{0}}\|_{H^{s}}$. This gives norm convergence $\|{B(t)}\|_{H^{s}}\to\|{B_{0}}\|_{H^{s}}$, which together with weak convergence concludes strong convergence. The proof of (i) is now complete. The proof of (ii) immediately follows from the inequality $\|B(t)\|_{H^{s}}\leq\|B_{0}\|_{H^{s}}e^{C\int_{0}^{t}\|B(s)\|_{\mathcal{Y}^{1}}ds},$ which is obtained from (3). ∎ ###### Proof of Theorem B. We shall simply establish an $H^{s}$–a priori estimate for (1.6), and omit the proof of existence and uniqueness. Similarly as in the proof of Theorem A, we compute $\begin{split}\frac{1}{2}\frac{\mathrm{d}}{\mathrm{d}t}\|{\Lambda^{s}B}\|_{L^{2}}^{2}+\|{\Lambda^{s+\frac{b}{2}}B}\|_{L^{2}}^{2}&=-\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\Lambda^{s}\nabla\times((\nabla\times\Lambda^{-a}B)\times B)\,\mathrm{d}x.\end{split}$ The main terms on the right hand side are given by $\begin{split}I_{1}:=-\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\nabla\times((\nabla\times\Lambda^{-a}\Lambda^{s}B)\times B)\,\mathrm{d}x\end{split}$ and $\begin{split}I_{2}:=-\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\nabla\times((\nabla\times\Lambda^{-a}B)\times\Lambda^{s}B)\,\mathrm{d}x.\end{split}$ With $F:=\nabla\times\Lambda^{s}B$, recall that $\begin{split}I_{1}=\sum_{i,j,k}\int_{\mathbb{R}^{3}}\epsilon_{ijk}[\Lambda^{-\frac{a}{2}},B^{k}]F^{i}\,\Lambda^{-\frac{a}{2}}F^{j}\,\mathrm{d}x=\sum_{i,j,k}\int_{\mathbb{R}^{3}}\epsilon_{ijk}\Lambda^{\frac{1}{2}}[\Lambda^{-\frac{a}{2}},B^{k}]F^{i}\,\Lambda^{-\frac{a+1}{2}}F^{j}\,\mathrm{d}x,\end{split}$ which can be bounded by $\begin{split}\left|I_{1}\right|\leq C\|{\Lambda^{\frac{1}{2}}[\Lambda^{-\frac{a}{2}},B^{k}]\Lambda^{\frac{a+1}{2}}}\|_{L^{2}\to L^{2}}\|{\Lambda^{-\frac{a+1}{2}}F^{j}}\|^{2}_{L^{2}}.\end{split}$ To bound the operator norm of the commutator, observe that $\begin{split}\Lambda^{\frac{1}{2}}[\Lambda^{-\frac{a}{2}},B^{k}]\Lambda^{\frac{a+1}{2}}=\Lambda^{\frac{1-a}{2}}[B,\Lambda^{\frac{1+a}{2}}]-\Lambda^{\frac{1}{2}}[B,\Lambda^{\frac{1}{2}}]\end{split}$ and $\begin{split}\|{\Lambda^{\frac{1-a}{2}}[B,\Lambda^{\frac{1+a}{2}}]}\|_{L^{2}\to L^{2}}+\|{\Lambda^{\frac{1}{2}}[B,\Lambda^{\frac{1}{2}}]}\|_{L^{2}\to L^{2}}\leq C\|{B}\|_{\dot{\mathcal{Y}}^{1}}.\end{split}$ This estimate can be proved easily using the Fourier transform and following the arguments of Lemma 3.1 above. Therefore, $\begin{split}\left|I_{1}\right|\leq C\|{B}\|_{\dot{\mathcal{Y}}^{1}}\|{\Lambda^{s+\frac{1-a}{2}}B}\|_{L^{2}}^{2}\leq C\|{B}\|_{H^{s}}\|{\Lambda^{s+\frac{1-a}{2}}B}\|_{L^{2}}^{2}.\end{split}$ Next, recall that with a cancellation, we have simply $\begin{split}I_{2}&=-\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot(\Lambda^{s}B\cdot\nabla(\nabla\times\Lambda^{-a}B))\,\mathrm{d}x.\end{split}$ Then, it is not difficult to obtain the bound $\begin{split}\left|I_{2}\right|\leq C\|{\nabla^{2}\Lambda^{-a}B}\|_{L^{\infty}}\|{B}\|_{\dot{H}^{s}}^{2}\leq C\|{B}\|_{H^{s}}^{3},\end{split}$ where we have used that $s>\frac{3}{2}+2-a$. Finally, proceeding in the same way as (3.6), it is not difficult to prove the following estimate for the remainder: $\begin{split}\left|I_{1}+I_{2}+\int_{\mathbb{R}^{3}}\Lambda^{s}B\cdot\Lambda^{s}\nabla\times((\nabla\times\Lambda^{-a}B)\times B)\,\mathrm{d}x\right|\leq C\|{B}\|_{H^{s}}^{3}.\end{split}$ Therefore, we conclude that $\begin{split}\frac{1}{2}\frac{\mathrm{d}}{\mathrm{d}t}\|{\Lambda^{s}B}\|_{L^{2}}^{2}+\|{\Lambda^{s+\frac{b}{2}}B}\|_{L^{2}}^{2}&\leq C\|{B}\|_{{H}^{s}}\|{\Lambda^{s+\frac{1-a}{2}}B}\|_{L^{2}}^{2}+C\|{B}\|_{H^{s}}^{3}\leq C\|{B}\|_{{H}^{s}}^{1+2\alpha}\|{\Lambda^{s+\frac{b}{2}}B}\|_{L^{2}}^{2(1-\alpha)}+C\|{B}\|_{H^{s}}^{3}\end{split}$ for some $0<\alpha<1$ since $1-a<b$. Using Young’s inequality, we can absorb the $\|{\Lambda^{s+\frac{b}{2}}B}\|_{L^{2}}$–term to the left hand side. Together with an $L^{2}$ estimate for $B$, this finishes the proof of the a priori estimate in $H^{s}$. ∎ ## 4 Discussion and open problems Extensions. The proof of local well-posedness can be adapted to three- dimensional domains without boundary $\mathbb{R}^{k}\times\mathbb{T}^{3-k}$ for any $0\leq k<3$. Moreover, one may consider the systems with a (fractional) dissipation term. It could be interesting to consider the critically dissipative case, as an analogy with critically dissipative SQG equations studied for instance in [5, 38, 37, 71]. Propagation of $\dot{H}^{-1}$. Given the formulation of the system in terms of $u$ and scaling of the conservation law, it is natural to add the assumption that $B\in\dot{H}^{-1}$. Indeed, $\|{B}\|_{\dot{H}^{-1}}$ is equivalent with $\|{u}\|_{L^{2}}$, and it is easy to see using the equation for $u$ that as long as $B\in L^{\infty}_{t}H^{s}$ with $s>\frac{5}{2}$, $\begin{split}\frac{\mathrm{d}}{\mathrm{d}t}\|{B}\|_{\dot{H}^{-1}}\lesssim\|{B}\|_{\dot{H}^{-1}}.\end{split}$ That is, the assumption $B_{0}\in\dot{H}^{-1}$ propagates in time as long as the solution remains smooth. Blow-up criteria. We remark that the blow-up criterion can be replaced with $\begin{split}\lim_{t\to T}\int_{0}^{t}\|{B(\tau,\cdot)}\|_{H^{\frac{5}{2}}}\,\mathrm{d}\tau=+\infty.\end{split}$ (4.1) To see this, we observe the logarithmic Sobolev estimate for $s>\frac{5}{2}$ $\begin{split}\|{B(\tau,\cdot)}\|_{\mathcal{Y}^{1}}\leq C\|{B(\tau,\cdot)}\|_{H^{\frac{5}{2}}}\log(10+\|{B(\tau,\cdot)}\|_{H^{s}}).\end{split}$ (4.2) Then, one may obtain the a priori estimate $\begin{split}\frac{\mathrm{d}}{\mathrm{d}t}\|{B}\|_{H^{s}}\leq C\|{B}\|_{H^{\frac{5}{2}}}\|{B}\|_{H^{s}}\log(10+\|{B}\|_{H^{s}}),\end{split}$ from which the blow-up criterion (4.1) follows by applying the Gronwall lemma to the differential inequality $y^{\prime}\leq ay\log y,\quad y=10+\|B(t)\|_{H^{s}},\quad a=C\|B(t)\|_{H^{\frac{5}{2}}}.$ The inequality (4.2), in turn, can be verified as follows. We first estimate $\displaystyle\int_{\mathbb{R}^{3}}|\xi||\hat{f}|\mathrm{d}\xi$ $\displaystyle\leq\int_{\\{|\xi|<R\\}}(1+|\xi|^{2})^{\frac{5}{4}}|\hat{f}|(1+|\xi|^{2})^{-\frac{3}{4}}\mathrm{d}\xi+\int_{\\{|\xi|\geq R\\}}(1+|\xi|^{2})^{\frac{s}{2}}|\hat{f}|(1+|\xi|^{2})^{\frac{1-s}{2}}\mathrm{d}\xi$ $\displaystyle\leq\left(\int_{\\{|\xi|<R\\}}(1+|\xi|^{2})^{\frac{5}{2}}|\hat{f}|^{2}\mathrm{d}\xi\right)^{\frac{1}{2}}\left(\int_{\\{|\xi|<R\\}}(1+|\xi|^{2})^{-\frac{3}{2}}\mathrm{d}\xi\right)^{\frac{1}{2}}$ $\displaystyle\qquad+\left(\int_{\\{|\xi|\geq R\\}}(1+|\xi|^{2})^{s}|\hat{f}|^{2}\mathrm{d}\xi\right)^{\frac{1}{2}}\left(\int_{\\{|\xi|\geq R\\}}(1+|\xi|^{2})^{1-s}\mathrm{d}\xi\right)^{\frac{1}{2}}$ $\displaystyle\leq C\|f\|_{H^{\frac{5}{2}}}\log(1+R)+C\|f\|_{H^{s}}R^{\frac{5}{2}-s}\quad\forall R>0.$ Choosing $R=(\|f\|_{H^{s}}/\|f\|_{\frac{5}{2}})^{H^{\frac{2}{2s-5}}}$, we obtain (4.2). Issue of global well-posedness. Global well-posedness of smooth solutions in the case of 3D Euler is a notoriously difficult open problem. In the electron–MHD case, finite time singularity formation is available for smooth and axisymmetric data ([21]) but unfortunately local well-posedness (more precisely, uniqueness) is not known for the data used in the proof. It should be interesting to investigate the possibility of singularity formation for the interpolating models (1.1). Of course, when the multiplier $\Gamma$ becomes “very regular”, e.g. $\Gamma=(1-\Delta)^{-N}$ with $N$ large, then global well-posedness for smooth and decaying solutions can be proved using the conservation of energy. #### Acknowledgments D. Chae was supported partially by NRF grant 2021R1A2C1003234. I.-J. Jeong was supported by the NRF grant 2022R1C1C1011051. We sincerely thank the anonymous referees for providing several helpful suggestions and pointing out errors as well as relevant references. #### Data availability statement Data sharing not applicable to this article as no datasets were generated or analyzed during the current study. ## References * [1] Marion Acheritogaray, Pierre Degond, Amic Frouvelle, and Jian-Guo Liu, _Kinetic formulation and global existence for the Hall-Magneto-hydrodynamics system_ , Kinetic & Related Models, 2011, 4 (4) : 901–918. doi: 10.3934/krm.2011.4.901. * [2] Claude Bardos and Edriss S. Titi, _Loss of smoothness and energy conserving rough weak solutions for the $3d$ Euler equations_, Discrete Contin. Dyn. Syst. Ser. S 3 (2010), no. 2, 185–197. MR 2610558 * [3] Tristan Buckmaster, Steve Shkoller, and Vlad Vicol, _Nonuniqueness of weak solutions to the SQG equation_ , Comm. Pure Appl. Math. 72 (2019), no. 9, 1809–1874. MR 3987721 * [4] D. Biskamp, E. Schwarz, and A. Zeiler, _Electron magnetohydrodynamic turbulence_ , Physics of Plasmas 6, 751 (1999); https://doi.org/10.1063/1.873312 * [5] Luis A. Caffarelli and Alexis Vasseur, _Drift diffusion equations with fractional diffusion and the quasi-geostrophic equation_ , Ann. of Math. (2) 171 (2010), no. 3, 1903–1930. MR 2680400 * [6] Daomin Cao, Guolin Qin, Weicheng Zhan, and Changjun Zou, _On the global classical solutions for the generalized SQG equation_ , J. Funct. Anal. 283 (2022), no. 2, Paper No. 109503, 37 pp. MR 4406410 * [7] , _Existence and stability of smooth traveling circular pairs for the generalized surface quasi-geostrophic equation_ , arXiv preprint. * [8] Angel Castro, Diego Córdoba, and Javier Gómez-Serrano, _Existence and regularity of rotating global solutions for the generalized surface quasi-geostrophic equations_ , Duke Math. J. 165 (2016), no. 5, 935–984. MR 3482335 * [9] Angel Castro, Diego Córdoba, and Javier Gómez-Serrano, _Uniformly rotating analytic global patch solutions for active scalars_ , Ann. PDE 2 (2016), no. 1, Art. 1, 34. MR 3462104 * [10] Angel Castro, Diego Córdoba, Javier Gómez-Serrano, and Alberto Martín Zamora, _Remarks on geometric properties of SQG sharp fronts and $\alpha$-patches_, Discrete Contin. Dyn. Syst. 34 (2014), no. 12, 5045–5059. MR 3223861 * [11] Dongho Chae, Peter Constantin, Diego Córdoba, Francisco Gancedo, and Jiahong Wu, _Generalized surface quasi-geostrophic equations with singular velocities_ , Comm. Pure Appl. Math. 65 (2012), no. 8, 1037–1066. MR 2928091 * [12] Dongho Chae, Peter Constantin, and Jiahong Wu, _Inviscid models generalizing the two-dimensional Euler and the surface quasi-geostrophic equations_ , Arch. Ration. Mech. Anal. 202 (2011), no. 1, 35–62. MR 2835862 * [13] , _Deformation and symmetry in the inviscid SQG and the 3D Euler equations_ , J. Nonlinear Sci. 22 (2012), no. 5, 665–688. MR 2982049 * [14] , _An incompressible 2D didactic model with singularity and explicit solutions of the 2D Boussinesq equations_ , J. Math. Fluid Mech. 16 (2014), no. 3, 473–480. MR 3247363 * [15] Dongho Chae, Pierre Degond, and Jian-Guo Liu, _Well-posedness for Hall-magnetohydrodynamics_ , Ann. Inst. H. Poincaré Anal. Non Linéaire 31 (2014), no. 3, 555–565. MR 3208454 * [16] Dongho Chae, In-Jee Jeong, and Sung-Jin Oh, _Illposedness via degenerate dispersion for generalized surface quasi-geostrophic equations with singular velocities_ , in preparation. * [17] , _Well-posedness for logarithmically singular surface quasi-geostrophic equations_ , in preparation. * [18] D. Chae and J. Lee. _On the blow-up criterion and small data global existence for the Hall-magnetohydrodynamics_ , J. Differential Equations 256 (2014), no. 11, 3835–3858. * [19] D. Chae and M. Schonbek. _On the temporal decay for the Hall-magnetohydrodynamic equations_ , J. Differential Equations 255 (2013) 3971–3982. * [20] D. Chae, R. Wan, J. Wu. _Local well-posedness for the Hall-MHD equations with fractional magnetic diffusion_ , J. Math. Fluid Mech. 17 (2015), no. 4, 627–638. * [21] Dongho Chae and Shangkun Weng, _Singularity formation for the incompressible Hall-MHD equations without resistivity_ , Ann. Inst. H. Poincaré Anal. Non Linéaire 33 (2016), no. 4, 1009–1022. MR 3519529 * [22] Dongho Chae and Jörg Wolf, _On partial regularity for the 3D nonstationary Hall magnetohydrodynamics equations on the plane_ , SIAM J. Math. Anal. 48 (2016), no. 1, 443–469. MR 3455137 * [23] Jiajie Chen and Thomas Y. Hou. _Finite time blowup of 2D Boussinesq and 3D Euler equations with $C^{1,\alpha}$ velocity and boundary._ Comm. Math. Phys., 383(3):1559–1667, 2021. * [24] X. Cheng, H. Kwon, and D. Li, _Non-uniqueness of steady-state weak solutions to the surface quasi-geostrophic equations_ , arXiv:2007.09591 (2021). * [25] Kyudong Choi and In-Jee Jeong. _Filamentation near Hill’s vortex_ , Communications in Partial Differential Equations (2022), DOI: 10.1080/03605302.2022.2139721. * [26] Kyudong Choi and In-Jee Jeong. _On vortex stretching for anti-parallel axisymmetric flows._ arXiv:2110.09079. * [27] Young-Pil Choi and In-Jee Jeong, _Classical solutions for fractional porous medium flow_ , Nonlinear Analysis 210 (2021), 112393. * [28] Peter Constantin, Analysis of Hydrodynamic Models, CBMS-NSF Regional Conference Series in Applied Mathematics, 90, (2017). * [29] Peter Constantin, _An Eulerian-Lagrangian approach for incompressible fluids: local theory_ , J. Amer. Math. Soc. 14 (2001), no. 2, 263–278. * [30] Peter Constantin, Geometric statistics in turbulence, SIAM Rev. 36 (1994), 73–98. * [31] Peter Constantin, On the Euler equations of incompressible fluids, Bull. Amer. Math. Soc., 44 (2007), no. 4, 603–621. * [32] Peter Constantin, Diego Córdoba, Francisco Gancedo, and Robert M. Strain, _On the global existence for the Muskat problem_ , J. Eur. Math. Soc. (JEMS) 15 (2013), no. 1, 201–227. MR 2998834 * [33] Peter Constantin, Charles Fefferman and Andrew J. Majda, Geometric constraints on potential singularity formulation in the 3-D Euler equations, Comm. PDE, 21, (1996) no. 3-4, 559–571. * [34] Peter Constantin, Andrew J. Majda, and Esteban Tabak, _Formation of strong fronts in the $2$-D quasigeostrophic thermal active scalar_, Nonlinearity 7 (1994), no. 6, 1495–1533. MR 1304437 * [35] Peter Constantin, Andrew J. Majda, and Esteban G. Tabak, _Singular front formation in a model for quasigeostrophic flow_ , Phys. Fluids 6 (1994), no. 1, 9–11. MR 1252829 * [36] Peter Constantin and Huy Quang Nguyen, _Global weak solutions for SQG in bounded domains_ , Comm. Pure Appl. Math. 71 (2018), no. 11, 2323–2333. MR 3862092 * [37] Peter Constantin, Andrei Tarfulea, and Vlad Vicol, _Long time dynamics of forced critical SQG_ , Comm. Math. Phys. 335 (2015), no. 1, 93–141. MR 3314501 * [38] Peter Constantin and Vlad Vicol, _Nonlinear maximum principles for dissipative linear nonlocal operators and applications_ , Geom. Funct. Anal. 22 (2012), no. 5, 1289–1321. MR 2989434 * [39] Antonio Córdoba, Diego Córdoba, and Francisco Gancedo, _Uniqueness for SQG patch solutions_ , Trans. Amer. Math. Soc. Ser. B 5 (2018), 1–31. MR 3748149 * [40] Diego Córdoba, Javier Gómez-Serrano, and Alexandru D. Ionescu, _Global solutions for the generalized SQG patch equation_ , Arch. Ration. Mech. Anal. 233 (2019), no. 3, 1211–1251. MR 3961297 * [41] Francisco de la Hoz, Zineb Hassainia, and Taoufik Hmidi, _Doubly connected V-states for the generalized surface quasi-geostrophic equations_ , Arch. Ration. Mech. Anal. 220 (2016), no. 3, 1209–1281. MR 3466846 * [42] M. Dai. _Local well-posedness of the Hall-MHD system in $H^{s}(\mathbb{R}^{n})$ with $s>n/2$_, Math. Nachr. 293 (2020), no. 1, 67–78. * [43] M. Dai. _Local well-posedness for the Hall-MHD system in optimal Sobolev spaces_ , J. Differential Equations, 289 (2021), 159–181. * [44] R. Danchin, J. Tan. _On the well-posedness of the Hall-magnetohydrodynamics system in critical spaces_ , Comm. Partial Differential Equations 46 (2021), no. 1, 31–65. * [45] R. Danchin, J. Tan. _The Global Solvability Of The Hall-magnetohydrodynamics System In Critical Sobolev Spaces._ https://doi.org/10.48550/arXiv.1912.09194 * [46] Sergey A. Denisov, _Infinite superlinear growth of the gradient for the two-dimensional Euler equation_ , Discrete Contin. Dyn. Syst. 23 (2009), no. 3, 755–764. MR 2461825 * [47] , _The centrally symmetric $V$-states for active scalar equations. Two-dimensional Euler with cut-off_, Comm. Math. Phys. 337 (2015), no. 2, 955–1009. MR 3339167 * [48] Ronald J. DiPerna and Andrew J. Majda, _Oscillations and concentrations in weak solutions of the incompressible fluid equations_ , Comm. Math. Phys. 108 (1987), no. 4, 667–689. MR 877643 * [49] Baoying Du, _Global regularity for the $2\frac{1}{2}$D incompressible Hall-MHD system with partial dissipation_, J. Math. Anal. Appl. 484 (2020), no. 1, 123701, 27. MR 4042622 * [50] Tarek Mohamed Elgindi, _Osgood’s lemma and some results for the slightly supercritical 2D Euler equations for incompressible flow_ , Arch. Ration. Mech. Anal. 211 (2014), no. 3, 965–990. MR 3158812 * [51] Tarek M. Elgindi. _Finite-time singularity formation for ${C}^{1,\alpha}$ solutions to the incompressible Euler equations on $\mathbb{R}^{3}$_, Ann. of Math. (2) 194 (2021), no. 3, 647–727. MR 4334974 * [52] Tarek M. Elgindi and In-Jee Jeong. _Finite-time singularity formation for strong solutions to the axi-symmetric 3D Euler equations._ Ann. PDE, 5(2):Paper No. 16, 51, 2019. * [53] Tarek M. Elgindi and In-Jee Jeong. _The incompressible Euler equations under octahedral symmetry: singularity formation in a fundamental domain._ Adv. Math., 393:Paper No. 108091, 63, 2021. * [54] Tarek Mohamed Elgindi and In-Jee Jeong, _Symmetries and Critical Phenomena in Fluids_ , Comm. Pure Appl. Math. 73 (2020), no. 2, 257–316. * [55] , _On Singular Vortex Patches, I: Well-posedness Issues_ , Mem. Amer. Math. Soc., to appear. * [56] , _On Singular Vortex Patches, II: Long-time dynamics_ , Trans. Amer. Math. Soc. 373 (2020), no. 9, 6757–6775. * [57] Susan Friedlander and Vlad Vicol, _On the ill/well-posedness and nonlinear instability of the magneto-geostrophic equations_ , Nonlinearity 24 (2011), no. 11, 3019–3042. MR 2844825 * [58] Francisco Gancedo, _Existence for the $\alpha$-patch model and the QG sharp front in Sobolev spaces_, Adv. Math. 217 (2008), no. 6, 2569–2598. MR 2397460 * [59] Javier Gómez-Serrano, _On the existence of stationary patches_ , Adv. Math. 343 (2019), 110–140. MR 3880855 * [60] Siming He and Alexander Kiselev, _Small scale creation for solutions of the SQG equation_ , Duke Math. J. 170 (2021), no. 5, 1027–1041. MR 4255049 * [61] Taoufik Hmidi and Joan Mateu, _Existence of corotating and counter-rotating vortex pairs for active scalar equations_ , Comm. Math. Phys. 350 (2017), no. 2, 699–747. MR 3607460 * [62] John K. Hunter, Jingyang Shu, and Qingtian Zhang, _Contour dynamics for surface quasi-geostrophic fronts_ , Nonlinearity 33 (2020), no. 9, 4699–4714. MR 4135092 * [63] , _Two-front solutions of the SQG equation and its generalizations_ , Commun. Math. Sci. 18 (2020), no. 6, 1685–1741. MR 4176355 * [64] Philip Isett and Vlad Vicol, _Hölder continuous solutions of active scalar equations_ , Ann. PDE 1 (2015), no. 1, Art. 2, 77. MR 3479065 * [65] In-Jee Jeong and Sung-Jin Oh, _On the Cauchy problem for the Hall and electron magnetohydrodynamic equations without resistivity I: illposedness near degenerate stationary solutions_ , Ann. PDE 8 (2022), 15. MR 4456288 * [66] , _Strong ill-posedness for degenerate dispersive equations_ , in preparation. * [67] In-Jee Jeong and Tsuyoshi Yoneda, _Vortex stretching and a modified zeroth law for the incompressible 3D Navier–Stokes equations_ , Math. Ann., (2021), 380(3), 2041–2072. * [68] T. Kato and G. Ponce, _Commutator estimates and the Euler and Navier-Stokes equations,_ Commun. Pure Appl. Math. Vol. XLI, (1988), 891–907. MR 0951744 * [69] Calvin Khor and José L. Rodrigo, _Local existence of analytic sharp fronts for singular SQG_ , Nonlinear Anal. 202 (2021), 112116, 22. MR 4156974 * [70] , _On sharp fronts and almost-sharp fronts for singular SQG_ , J. Differential Equations 278 (2021), 111–145. MR 4199394 * [71] A. Kiselev, F. Nazarov, and A. Volberg, _Global well-posedness for the critical 2D dissipative quasi-geostrophic equation_ , Invent. Math. 167 (2007), no. 3, 445–453. MR 2276260 * [72] Alexander Kiselev, _Small scales and singularity formation in fluid dynamics_ , Proceedings of the International Congress of Mathematicians—Rio de Janeiro 2018. Vol. III. Invited lectures, World Sci. Publ., Hackensack, NJ, 2018, pp. 2363–2390. MR 3966854 * [73] Alexander Kiselev and Fedor Nazarov, _A simple energy pump for the surface quasi-geostrophic equation_ , Nonlinear partial differential equations, Abel Symp., vol. 7, Springer, Heidelberg, 2012, pp. 175–179. MR 3289364 * [74] Alexander Kiselev, Lenya Ryzhik, Yao Yao, and Andrej Zlatoš, _Finite time singularity for the modified SQG patch equation_ , Ann. of Math. (2) 184 (2016), no. 3, 909–948. MR 3549626 * [75] Alexander Kiselev and Vladimir Šverák, _Small scale creation for solutions of the incompressible two-dimensional Euler equation_ , Ann. of Math. (2) 180 (2014), no. 3, 1205–1220. MR 3245016 * [76] Alexander Kiselev, Yao Yao, and Andrej Zlatoš, _Local regularity for the modified SQG patch equation_ , Comm. Pure Appl. Math. 70 (2017), no. 7, 1253–1315. MR 3666567 * [77] Alexander A. Kiselev, _Small scale creation in active scalars_ , Progress in mathematical fluid dynamics, Lecture Notes in Math., vol. 2272, Springer, Cham, [2020] ©2020, pp. 125–161. MR 4176667 * [78] Hyunju Kwon, _Strong ill-posedness of logarithmically regularized 2D Euler equations in the borderline Sobolev space_ , J. Funct. Anal. 280 (2021), no. 7, 108822. MR 4211020 * [79] Zhen Lei and Fanghua Lin, _Global mild solutions of Navier-Stokes equations,_ Comm. Pure Appl. Math. 64, (2011), no. 9, 1297–1304. MR 2839302 * [80] Dong Li, _Existence theorems for the 2D quasi-geostrophic equation with plane wave initial conditions_ , Nonlinearity 22 (2009), no. 7, 1639–1651. MR 2519682 * [81] M. J. Lighthill, _Studies on magneto-hydrodynamic waves and other anisotropic wave motions_ , Philos. Trans. Roy. Soc. London Ser. A 252 (1960), 397–430. MR 0148337 * [82] Andrew J. Majda and Andrea L. Bertozzi, _Vorticity and incompressible flow_ , Cambridge Texts in Applied Mathematics, vol. 27, Cambridge University Press, Cambridge, 2002. MR 1867882 * [83] Fabien Marchand, _Existence and regularity of weak solutions to the quasi-geostrophic equations in the spaces $L^{p}$ or $\dot{H}^{-1/2}$_, Comm. Math. Phys. 277 (2008), no. 1, 45–67. MR 2357424 * [84] Huy Quang Nguyen, _Global weak solutions for generalized SQG in bounded domains_ , Anal. PDE 11 (2018), no. 4, 1029–1047. MR 3749375 * [85] Koji Ohkitani, _Growth rate analysis of scalar gradients in generalized surface quasigeostrophic equations of ideal fluids_ , Phys. Rev. E (3) 83 (2011), no. 3, 036317, 8. MR 2788267 * [86] , _Asymptotics and numerics of a family of two-dimensional generalized surface quasi-geostrophic equations_ , Physics of Fluids 24 (2012), no. 095101. * [87] Hans Pécseli, _Waves and oscillations in plasmas_ , Series in Plasma Physics (2012). * [88] R. Shvydkoy, _Convex integration for a class of active scalar equations_ , J. Amer. Math. Soc. 24 (2011), no. 4, 1159–1174. MR 2813340 * [89] Henggeng Wang and Houyu Jia, _Local well-posedness for the 2D non-dissipative quasi-geostrophic equation in Besov spaces_ , Nonlinear Anal. 70 (2009), no. 11, 3791–3798. MR 2515299 * [90] Xiaoqian Xu, _Fast growth of the vorticity gradient in symmetric smooth domains for 2D incompressible ideal flow_ , J. Math. Anal. Appl. 439 (2016), no. 2, 594–607. MR 3475939 * [91] Kazuo Yamazaki, _Irreducibility of the three, and two and a half dimensional Hall-magnetohydrodynamics system_ , Phys. D 401 (2020), 132199, 21. MR 4034672 * [92] Andrej Zlatoš, _Exponential growth of the vorticity gradient for the Euler equation on the torus_ , Adv. Math. 268 (2015), 396–403. MR 3276599
11institutetext: Aix Marseille Univ, CNRS, CNES, LAM, Marseille, France 22institutetext: Institut Universitaire de France, Paris, France 33institutetext: Aix Marseille Univ, CNRS, LIS, Marseille, France 44institutetext: INAF-IAPS, via del Fosso del Cavaliere 100, I-00133 Roma, Italy 55institutetext: INAF – Astronomical Observatory of Capodimonte, via Moiariello 16, I-80131, Napoli, Italy 66institutetext: Division of Science, National Astronomical Observatory of Japan, 2-21-1 Osawa, Mitaka, Tokyo 181-8588, Japan # Supervised machine learning on Galactic filaments Revealing the filamentary structure of the Galactic interstellar medium A. Zavagno 1122 F.-X. Dupé 33 S. Bensaid 1133 E. Schisano 44 G. Li Causi 44 M. Gray 11 S. Molinari 44 D. Elia 44 J.-C. Lambert 11 M. Brescia 55 D. Arzoumanian 1166 D. Russeil 11 G. Riccio 55 S. Cavuoti , F.-X. Dupé and S. Bensaid contributed equally to the work presented in this article.55 (Received May 25 2022; accepted November 20 2022) ###### Abstract Context. Filaments are ubiquitous in the Galaxy, and they host star formation. Detecting them in a reliable way is therefore key towards our understanding of the star formation process. Aims. We explore whether supervised machine learning can identify filamentary structures on the whole Galactic plane. Methods. We used two versions of UNet-based networks for image segmentation. We used H2 column density images of the Galactic plane obtained with Herschel Hi-GAL data as input data. We trained the UNet-based networks with skeletons (spine plus branches) of filaments that were extracted from these images, together with background and missing data masks that we produced. We tested eight training scenarios to determine the best scenario for our astrophysical purpose of classifying pixels as filaments. Results. The training of the UNets allows us to create a new image of the Galactic plane by segmentation in which pixels belonging to filamentary structures are identified. With this new method, we classify more pixels (more by a factor of 2 to 7, depending on the classification threshold used) as belonging to filaments than the spine plus branches structures we used as input. New structures are revealed, which are mainly low-contrast filaments that were not detected before. We use standard metrics to evaluate the performances of the different training scenarios. This allows us to demonstrate the robustness of the method and to determine an optimal threshold value that maximizes the recovery of the input labelled pixel classification. Conclusions. This proof-of-concept study shows that supervised machine learning can reveal filamentary structures that are present throughout the Galactic plane. The detection of these structures, including low-density and low-contrast structures that have never been seen before, offers important perspectives for the study of these filaments. ###### Key Words.: Methods: statistics – Stars: formation – ISM: general ## 1 Introduction The Herschel infrared Galactic Plane Survey, Hi-GAL (Molinari et al. 2010), revealed that the cold and warm interstellar medium (ISM) is organized in a network of filaments in which star formation is generally observed above a density threshold corresponding to $A_{\rm{V}}$=7 mag (André et al. 2014; Könyves et al. 2020). The most massive stars are formed at the junction of the densest filaments, called hubs (Kumar et al. 2020). Because filaments host star formation and link the organization of the interstellar matter to the future star formation, studying them is central to our understanding of all properties related to star formation, such as the initial mass function, the star formation rate, and the star formation efficiency. Filaments are therefore extensively studied with observations at all wavelengths and numerical simulations (André et al. 2010; Molinari et al. 2010; Arzoumanian et al. 2011; Hacar et al. 2018; Arzoumanian et al. 2019; Shimajiri et al. 2019; Clarke et al. 2020; Priestley & Whitworth 2022; Hacar et al. 2022, and references therein). All these data reveal the complex structure of filaments and show a changing morphology, depending on the way (resolution or tracers) in which they are observed (Leurini et al. 2019). For example, high-resolution molecular line observations of Galactic filaments with the Atacama Large Millimeter Array (ALMA) show that they are made of fibers on a spatial scale ¡ 0.1 pc (Shimajiri et al. 2019; Hacar et al. 2018). Their complex morphology and dynamics are also revealed with 3D spectroscopic information (Mattern et al. 2018; Hacar et al. 2020) and show their key role in the accretion process from large (¿ 10 pc) to subparsec scales, funnelling material down to the star-forming cores. However, the way filaments form and evolve in the ISM is still debated (Hoemann et al. 2021; Hsieh et al. 2021). Recent results suggest that compression from neutral (H i) and ionized (H ii) shells could play an important role in forming and impacting the evolution of Galactic filaments (Zavagno et al. 2020; Bracco et al. 2020). Their detection in nearby galaxies, where they are also clearly linked to the star formation process, makes studying filaments even more important and universal (Fukui et al. 2019). Different algorithms are used to extract filaments from 2D images (Sousbie 2011; Schisano et al. 2014; Koch & Rosolowsky 2015; Zucker & Chen 2018; Schisano et al. 2020; Men’shchikov 2021) and from 3D spectral data cubes (Sousbie 2011; Chen et al. 2020). These algorithms often rely on a threshold definition (for intensity or column density). Nonetheless, a close visual inspection of 2D images and 3D cubes show that some filaments are missed by all these detection algorithms, especially when the filaments have low column density contrasts. This means that even large surveys of the Galactic plane cannot deliver a complete (unbiased) view of the filaments present there. Another limitation of these algorithms comes from their computation time, which can make some of them too expensive to envision a complete run on large- scale surveys data for multiple defined threshold and extraction parameters. Because the multi-wavelength information available on our Galaxy on all spatial scales is so rich, proposing another way of extracting filaments might allow a leap forward for an unbiased census of these data. In this paper, we explore the potential of supervised machine learning as a new way to reveal filaments from 2D images. Using Hi-GAL data, Schisano et al. (2020) extracted filaments from column density (N${}_{\rm{H_{2}}}$) images of the Galactic plane. Based on these data, we study the possibility for convolutional UNet- based networks (Fu & Mui 1981) to identify pixels as belonging to the filament class, based on the input information given as previously identified filament masks from Schisano et al. (2020). Except for the catalog of filament candidates published in Schisano et al. (2020) where faint filaments present in the Galactic plane are known to be missed, no complete extraction of filamentary structures in the Galactic plane exists so far. This fact motivates our work, in which we propose an alternative method that could allow us to go beyond the current possibilities. However, this fact also indicates that we worked with an incomplete ground truth (see Section 2.2) that renders an absolute evaluation of the method performances proposed here impossible. Nonetheless, we show that new filaments revealed by the UNet-based algorithm and not detected before are confirmed through imaging at other wavelengths, which gives us confidence that this new method can progress toward an unbiased detection of filaments. The paper is organized as follows: in Section 2 we describe the images and the information on the filament locations we used in the supervised learning. The supervised learning itself is described in Section 3. Results are presented in Section 4 and are discussed in Section 5. Conclusions are given in Section 6. ## 2 Data ### 2.1 Hi-GAL catalog Figure 1: Hi-GAL H2 column density image of the $l$=160-171° (top) and $l$=349-356° (bottom) regions produced using Hi-GAL images as described in Elia et al. (2013). The $l$=349-356° zone contains the bright star-forming regions NGC 6334 and NGC 6357. These two regions are used in this paper to illustrate the results. The dark pixels are saturated. The _Herschel_ infrared Galactic Plane Survey, Hi-GAL (Molinari et al. 2010), is a complete survey of the Galactic plane performed in five infrared photometric bands centered at 70, 160, 250, 350, and 500 $\mu$m. H2 column density (N${}_{\rm{H_{2}}}$) images were created for the whole Galactic plane following the method described in Elia et al. (2013) and Schisano et al. (2020). N${}_{\rm{H_{2}}}$ and dust temperature maps were computed from photometrically calibrated images. The _Herschel_ data were convolved to the 500 $\mu$m resolution (36″), and a pixel-by-pixel fitting by a single- temperature graybody was performed. An example of the column density image covering the $l$=349-356° region is presented in Figure 1. This region contains the bright and well-studied Galactic star-forming regions NGC 6334 and NGC 6357. Schisano et al. (2020) analyzed the whole Galactic plane by extracting filamentary structures from the H2 column density (N${}_{\rm{H_{2}}}$) maps. In their work, a filament is defined as a two-dimensional, cylindric-like structure that is elongated and shows a higher brightness contrast with respect to its surroundings. The extraction algorithm is based on the Hessian matrix $H(x,y)$ of the intensity map N${}_{\rm{H_{2}}}(x,y)$ to enhance elongated regions with respect to any other emission. The algorithm performs a spatial filtering and amplifies the contrast of small-scale structures in which the emission changes rapidly. Further filtering allows identifying the filamentary structures. Figure 2 shows an example of this filament extraction, reproducing the figure 3 of Schisano et al. (2020). We chose this figure because it shows the input we use in this work: the spine (blue line) and the branches (red lines, both shown in the bottom left panel) associated with a given filament. Schisano et al. (2020) defined a filament as traced by its associated region of interest (RoI; bottom right), which covers a larger area than the region that is defined with the spine plus branches. In this work we use this spine plus branches structure to define a filament because the early tests we made to train the networks with the input RoIs returned filamentary structures that were too large compared to the structure that is observed in the column density mosaics. This point is illustrated in Figure 21 and is discussed in Sect. 4.3. The analysis of the extracted structures from Schisano et al. (2020) resulted in the publication of a first catalog of 32 059 filaments that were identified over the entire Galactic plane. We used this published catalog of filaments and their associated spine plus branches as ground truth of the filament class for the training process (see Sect. 2.2). The method is described in Section 3. Figure 2: Illustration of the filament extraction method from Schisano et al. (2020, their Figure 3). The spine (blue line) and branches (red lines) associated with a filament used for the supervised training process are shown in the bottom left corner. The RoI (bottom right corner) is the zone used by Schisano et al. (2020) to define a filament. ### 2.2 Data preprocessing As methods based on deep learning strongly depend on the nature and on the representativity of the input data, we took particular attention to the construction of the data set. We used four input maps (see Fig. 3): 1) the N${}_{\rm{H_{2}}}$ mosaics obtained as part of the Hi-GAL survey products (Molinari et al. 2016; Schisano et al. 2020, e.g., Fig. 1), 2) the spines plus branches of the detected filaments from Schisano et al. (2020, e.g., Fig. 2 bottom left corner), 3) a background map (localization of nonfilament pixels), and 4) a missing-data map (see Appendix A). The origin and the ways in which these maps were obtained are presented in Appendix A. An example of these maps is given on Figure 3, illustrated for the two portions of the Galactic plane that are located at 160-171° and 349-356°. We obtained results for the whole Galactic plane, which we illustrate in two regions that we selected because they represent the diversity of column density and filaments content observed in the Galactic plane well. The 160-171° region samples a low column density medium (up to 8$\times$1021 cm-2) in which only a few filaments are detected, while the 349-356° region shows a rich content in filaments that are detected in a high column density medium (up to 9$\times$1022 cm-2, see Figure 3). For the training step, we merged all the individual mosaics (10°-long in longitude direction) into one global map using the reproject module for astropy (Robitaille et al. 2020, and Appendix A). Figure 3: Illustration of the Galactic regions located at 160-171° (left) and 349-356° (right) of the four input maps used for the supervised learning. From top to bottom, we show the N${}_{\rm{H_{2}}}$ column density map, the input filament masks, a background localization map, and the missing-data map (0 in purple, 1 in yellow). All these maps were obtained as explained in Appendix A. The filament and background mask maps are multiplied by the missing-data map before they were used in the training process. The red rectangle shown in each column density map represents the region we extracted to compute the performance of the training (see Sect. 4.1). Figure 4: Construction of patches of size $p\times p$ using a sliding window. As the four input maps are very large ($150000$ $\times$ $2000$ pixels), we split them into many patches that constitute the original data set. As shown in Figure 4, we split the maps into $p\times p$ patches. The size of $p=32$ pixels was chosen to preserve the information on the small filamentary structures. This size is also the minimum size accepted by the UNet architecture. The patches were generated by applying a sliding window of size $p$ (the patch size) to the global mosaic of H2 column density (N${}_{\rm{H_{2}}}$). To ensure the coherence between the four input maps, the four patches (N${}_{\rm{H_{2}}}$, spine+branches, background, and missing data) were taken using the same coordinates (see Figure 5). In order to avoid any common information between patches, the construction was made without any overlap between the patches. Figure 5: Building the data set using the four input maps (on the left) into a set of patches (on the right). On the left, the maps are the column density (top left), filament spine+branches (top right), missing data (bottom left), and background pixels (bottom right). ### 2.3 Data augmentation Deep neural networks are greedy algorithms. In spite of the huge size of the Galactic map, the final data set had merely $52.000$ patches after empty patches were removed. Here, we refer as “empty patches” to patches that only contain missed values (patches located on edges) or to patches that contain “0 pixels labeled data”. Fully unlabeled patches were removed from the patch data set. As unlabeled pixels represent more than $80\%$ of the data set, this resulted in the loss of many patches. The data augmentation is thus necessary in order to increase the number of training and validation patches and to thereby enable a sufficient convergence of the neural network during the training step (Goodfellow et al. 2016). Two types of rotation were used: rotation around the central pixel of the squared patch (0°(original), 90°, 180° , and 270°), and flipping the patch with respect to the $x$ \- and $y$ -axes. We also allowed a composition of both rotations. All possible transformations are equally probable, that is, we selected the applied transformation following a uniform distribution. To attenuate redundancy issues, the augmentation was done on the fly, meaning that at each batch, we produced a new set of patches using the augmentation process. With our setting, we virtually increased the number of patches by a factor equal to 64. Figure 6 shows some examples of the data augmentation process. Figure 6: Example of data augmentation results. ## 3 Method ### 3.1 Segmentation pipeline Our segmentation method relies on three components: a data preparation procedure, a neural network with an architecture dedicated to the recognition of filamentary structures, and a training procedure adapted to the N${}_{\rm{H_{2}}}$ data. After the neural network was trained, we used it to segment the N${}_{\rm{H_{2}}}$ map. The result of the segmentation process is a map in which the pixels are classified into two classes: either a filament pixel (identified as class 1), or a background pixel (identified as class 0). With these two classes, we produced an intensity map with values of 0 and 1. The values of the classification indicate whether a pixel belongs to the filament class (the reverse map shows the classification value according to which a pixel belongs to the background class). #### 3.1.1 Segmentation with UNets Automatic segmentation is a well-known issue in the artificial intelligence community. Its origins lie in computer vision. It is a well-studied problem today, especially where segmentation is mandatory for decision or prediction, typically for medical or biology images (Fu & Mui 1981; Alzahrani & Boufama 2021). It has also been used for a long time in astrophysics, with recent applications on galaxies (Zhu et al. 2019; Hausen & Robertson 2020; Bianco et al. 2021; Bekki 2021). Most previous methods are based on classical machine- learning methods, such as Support-Vector Machine (SVM) or Random Forest (Hastie et al. 2001). These methods must extract an adapted set of features in order to be sufficiently efficient: we have to be sure that the extracted features represent the subject we wish to study well. These features are usually given by expert knowledge of the problem. Most successful machine-learning methods for image processing tasks today are based on deep neural network methods (Goodfellow et al. 2016). These methods have the particularity of learning both the task and a representation of the data dedicated to the task itself. Thus, they are more powerful than methods based on hand-tuned features. While the first methods were dedicated to classification (e.g., AlexNet (Krizhevsky et al. 2012), LeNet (LeCun et al. 1989), or ResNets (He et al. 2016)), there are now many different architectures depending on the targeted task. For segmentation, one of the most promising neural networks is the UNet, which was introduced for medical segmentation (Ronneberger et al. 2015). Many extensions exist, for instance, UNet++ (Zhou et al. 2019) with layers to encode the concatenations, VNet (Milletari et al. 2016), which is dedicated to 3D data, WNet (Xia & Kulis 2017), which has a double UNet architecture, and Attention-UNet (Oktay et al. 2018), which combines UNet with attention layers (Goodfellow et al. 2016). Still, the UNet based architecture remains one of the most effective methods for automatic segmentation. In the context of astrophysical study, these neural networks have been successfully used in different contexts. For example, Bekki (2021) used the UNet to segment the spiral arms of galaxies. Bianco et al. (2021) used a UNet based neural network called SegUNet to identify H ii regions during reionization in 21 cm. Another variant based on UNet and inception neural networks was used to predict localized primordial star formation (Wells & Norman 2021). UNet was also used to segment cosmological filaments (Aragon- Calvo 2019). We recommend Hausen & Robertson (2020) for a good introduction to deep learning applied to astrophysical data. Figure 7: Illustration of the UNet5 from Ronneberger et al. (2015). UNet is a multiscale neural network based on convolutional and pooling layers, as presented in Figure 7. In addition to its simple structure, the strength of this network is an encoder-decoder-based architecture with skip connections. First, the encoder extracts features from the input image down to a coarse scale by using filters and max-pooling. Then, the decoder takes the coefficients at the coarse level and combines them with those from each layer of the encoder via the skip connections, in order to re-inject the details that were lost in the down-sampling (max-pool) step and thereby build a better semantic segmentation map. The final activation function of the network is done by the sigmoid function (Goodfellow et al. 2016), as we wish to have values in order to resolve a segmentation issue, $s(x)=\frac{\exp(x)}{\exp(x)+1}\leavevmode\nobreak\ .$ (1) This function guarantees an output between 0 and 1. Thus, the output of the network can be read as a probability map for the class 1 filament (see Goodfellow et al. 2016, Sect.6.2.2.2). However, in our case, both the nonequilibrium between the two classes (filament and background) and the incomplete ground-truth prevent the direct interpretation of the segmented map values as probabilities (see Kull et al. 2017, about sigmoid output and probabilities). In the following, we name the intensity value of the segmented maps ”classification value”. In this study, the quality of the results is assessed by comparing these classification thresholds with a given threshold (see Sect. 3.1.5). This multiscale mirror-like structure makes the UNet very suitable for image processing such as denoising (Batson & Royer 2019) or segmentation (Ronneberger et al. 2015). Moreover, UNet belongs to the family of fully convolutional networks. These networks are almost independent of the size of the input images (Long et al. 2015). In UNet, the size of the output image will be the same as that of the input if the input is large enough (the minimum size is $32\times 32$ pixels). Figure 8: Illustration of UNet++ from Zhou et al. (2019). The $X^{i,j}$ are the same convolutional layers as for UNet. The difference between UNet and UNet++ can be depicted in three main points: 1) convolution layers on skip pathways (in green), which reduces the semantic gap between encoder and decoder feature maps; 2) dense skip connections on skip pathways (in blue), which improves the gradient flow; and 3) deep supervision (in red), which enables model pruning (Lee et al. 2015). A recent and more powerful extension of the UNet model, UNet++, was proposed in (Zhou et al. 2019). This network belongs also to the fully convolutional networks. As illustrated in Fig. 8, the plain skip connections of UNet are replaced with a series of nested dense skip pathways in the UNet++ neural network. The new design aims at reducing the semantic gap between the feature maps of the encoder and decoder sub-networks that makes the learning task easier to solve for the optimizer. In fact, the model captures more efficiently fine-grained details when high-resolution feature maps from the encoder are gradually enriched before fusion with the corresponding semantically rich feature maps from the decoder. Note that these ”inner” layers have also a mirror-like structure allowing a larger multi-scale representation. However, it is worthy to note that UNet++ requires more data than UNet as the latter has less parameters to tune (see Table 3 in Zhou et al. 2019). #### 3.1.2 Local normalization Neural networks such as UNet are highly sensitive to the contrast inside the input images (or patches). This sensitivity comes from the filters that belong to the different convolution layers. In order to avoid this issue, input data are usually normalized, generally by performing a global min-max normalization (Goodfellow et al. 2016). This normalization allows us to temper the dynamic of the contrast while keeping useful physical information about the structures (morphology and gradient). However, in our case, the intensity of the N${}_{\rm{H_{2}}}$ map presents a very high dynamical range, and a global normalization would artificially weaken many filamentary structures. To avoid this issue, we performed a local min-max normalization on each patch. As shown in Figure 9, this normalization helps to deal with high-contrast variation in nearby regions of the image. However, while this approach solves this issue, the contrast still has high local variations in some cases, so that two nearby patches may show different normalization. Figure 9: The local min-max normalization of the patches helps to avoid contrast issue allowing a better definition of the filaments. #### 3.1.3 Training with UNet and UNet++ Training a neural network requires a loss function that computes the errors between the model and the ground truth. For segmentation, a recommended function is the binary cross-entropy (BCE), which casts the problem as a classification problem (Jadon 2020). For the sake of clarity, we introduce some notations before we give the expression of the loss function. Let $\\{x_{i}\\}_{i}$ be the set of normalized N${}_{\rm{H_{2}}}$ patches. Let $\\{y_{i}\\}_{i}$ be the set of segmentation target, that is, the set of binary patches with 1 for filaments pixels and 0 for background pixels. Let $\\{m_{i}\\}_{i}$ be the set of missing data patches, that is, the set of binary patches with 0 for missing pixels and 1 elsewhere. For a given value of $i$, $x_{i},y_{i}$, and $m_{i}$ share the same Galactic coordinates. The cross-entropy loss for a set of $n$ patches is given by $\begin{split}\mathcal{L}(\\{x_{i},y_{i},m_{i}\\}_{i};\theta)=\frac{1}{np^{2}}&\sum_{i=1}^{n}\sum_{k,l=1}^{p}m_{i}[k,l]\Big{(}y_{i}[k,l]\log(f_{\theta}(x_{i})[k,l])+\\\ &(1-y_{i}[k,l])\log(1-f_{\theta}(x_{i})[k,l])\Big{)}\end{split},$ (2) where $f_{\theta}$ is the function that applies the forward propagation, and $\theta$ are the weights of the neural network. By using $\\{m_{i}\\}_{i\in[i\ldots n]}$, we ensure that only labeled data are used. Figure 10: Five steps of an epoch during the training. For illustration purposes, we reduced the batch to a set of one patch. $\theta_{t}$ represents the weights of the neural network at epoch $t,$ and $\mu_{t}$ is the learning rate at epoch $t$. As we have many unlabeled pixels in the patches (see section 2.2), we have to adapt the training step to avoid inconsistencies. We summarize the different steps in Figure 10. First, in step (1), we take a set of patches (a batch) and then apply the augmentation process (step (2)) on these patches. During this step (2), we ensure that for a given $i$, the same transformation is applied on $x_{i},y_{i}$, and $m_{i}$. The following steps are about computing the prediction errors of the model on the patches and making the back-propagation of the gradient of these errors in order to update the weights of the neural networks (Goodfellow et al. 2016). Therefore, in step (3), we begin to apply the network $f_{\theta_{t}}$ on the patches (forward propagation). Since this step implies using the convolution layers in the network, we use both unlabeled and labeled pixels. This is important as the neural network needs the neighboring pixels to compute the value for one pixel. After we restrict the result (step (4)) to labeled and nonmissing pixels using the mask $m_{i}$, we can compute (step (5)) the errors on the restricted results compared to the ground truth. Finally, in step (6), we update the weight of the network using the back-propagation of the gradients of the errors. This is done by using a stochastic gradient descent scheme (Goodfellow et al. 2016) with a learning rate $\mu_{t}$ that changes during the training step following the epochs. #### 3.1.4 Building the segmentation map When the neural network has been trained, we can apply the model to segment an image. As we described before, during the creation of the data set (Section 2.2), we must deal with the high dynamic contrast in images. Again we propose to solve the issue by taking small patches and apply a local min-max normalization. Moreover, since two closed patches may have a different contrast, the normalization can lead to variation in the results when applying the neural network. Therefore, in order to resolve this issue, we propose to use an overlapping sliding window to obtain the patches: the segmentation result is then the average between the output of the neural network applied on the patches. These patches are distinct from those used for learning the network (see Section 2.2). As the variance of contrast between patches introduced variance inside the output results of the neural network, the overlap and averaging operation (see Fig.11) allows us to decrease the artifact that may appear (Pielawski & Wählby 2020). Figure 11: Segmentation process. It takes patches from an observation (a), then normalizes the patch and applies the segmentation model (b), the segmented patch is positioned at the same coordinates (c), and is finally weighted by coefficients (d) representing the number of patches in which each pixel appears. Because a sliding window with overlap is used, a given pixel is segmented several times (as long as it falls in the sliding window). Then, we obtain several segmentation values for the same pixel. The final segmentation value assigned to the pixel corresponds to the average of all the segmentation values computed from the contributing sliding windows. Thus, we apply the following segmentation procedure (illustrated in Figure 11). We browse the image using an overlapping sliding window that gives patches (step (a)). Each patch is then normalized using a min-max normalization (step (b)); here we avoid the missing data (around borders and saturated areas). Then, we apply the trained neural network on the patch to obtain the density map output and add on the output image at the coordinate of the patches (step (c)). Since we use an overlapping sliding window, the results are added to the output, and then we divide each pixel by a weight representing the number of patches in which the pixel appears (this is done using the weight map built in step (d)). #### 3.1.5 Metrics In supervised classification problems, the confusion matrix, also called error matrix, is computed in order to assess the performance of the algorithm. We refer to the filament and background classes as the positive (P) and the negative (N) classes, respectively. We also refer to the correctly and misclassified pixels as true (T) and false (F), respectively. In a binary classification problem, the confusion matrix is thus expressed as in Table 1. Table 1: Confusion matrix | Predicted ---|--- | | filament | background Actual | filament | TP | FN background | FP | TN The confusion matrix is evaluated on the estimated filament masks that are deduced from the segmented map at a classification threshold. It is important to recall, however, that the evaluation set is restricted to labeled data. The classification scores are thereafter derived from the confusion matrix. In this work, the recall, precision, and dice index defined in 3, 4 and 5, respectively, are used to evaluate the classifier performance. Maximizing recall and precision amounts to minimizing false-negative and false-positive errors, respectively, whereas maximizing the dice index amounts to finding the optimal tradeoff between the two errors. Therefore, the closer to 1 these scores, the better. $\displaystyle\mathrm{Recall}$ $\displaystyle=\frac{TP}{TP+FN}$ (3) $\displaystyle\mathrm{Precision}$ $\displaystyle=\frac{TP}{TP+FP}$ (4) $\displaystyle\mathrm{Dice}$ $\displaystyle=\frac{2\ TP}{2\ TP+FP+FN}$ (5) In addition to the recovery scores, we also calculate the rate of missed structures (MS) in segmentation for the same set of thresholds. The MS score can be defined as the ratio of the missed filament structures over all the input filament structures, $\mathrm{MS}=\frac{\text{number of missed structures}}{\text{total number of filament structures}}$ (6) Figure 12: Morphological reconstruction is a method that computes shapes from marked pixels called seeds. (a) We first compute the seeds using the intersection between the segmentation results and the ground truth. (b) We use the intersection pixels as seeds (see the red seeds in the bottom left corner). (c) We apply the reconstruction to obtain the filaments with at least one seed. The MS metric is based on morphological reconstruction (Vincent 1993; Soille 2003; Robinson & Whelan 2004). Figure 12 shows how we apply this method to assess which known filaments are recovered. First, we compute the intersection between the known filaments (ground truth) and the segmentation results (Figure 12(a)). Then we use this intersection as the seed for the reconstruction method: in Figure 12(b), the yellow elements are the seeds, and the red elements are the part of filaments that is missed. The morphological reconstruction takes the seeds to recover the known filaments by using a shape-constrained growing process (Figure 12(c)). Only the filaments for which at least one pixel is used as seed will be recovered (Robinson & Whelan 2004). By subtracting the morphological reconstruction result from the ground truth, we can identify the missed structures (see also Figure 17). Then we count the number of missed structures by using a direct labeling of the pixels where two neighboring pixels share the same label (Fiorio & Gustedt 1996). Then, we can compute the MS metric and qualitatively assess the recovery of filaments in terms of structures, rather than individual pixels. ### 3.2 Experimental setup Taking the analysis of the normalization in Section 2.2 into consideration, the patch size was fixed to the lowest value accepted by the UNet family, $p=32$. In order to have proper training, validation, and test steps, we randomly split our initial set of patches (Section 2.2) into three sets, namely, the training, validation, and test sets, with proportions of $80\%$, $10\%,$ and $10\%$, respectively. The random split ensures the presence of the two classes (filament and background) in the three sets. The patches in training and validation sets were then shuffled after each epoch to help avoiding unwanted bias (Goodfellow et al. 2016, see). The total number of epochs was set to $100$. UNet and UNet++ were trained using the Adam optimization scheme with a multistep learning rate (Kingma & Ba 2015; Ronneberger et al. 2015). During the first 30 epochs, the initial learning rate value was divided by $10$ every five epochs. Four initial values of the learning rate, $10^{-5},10^{-4},10^{-3},\text{and }10^{-2}$, were tested for both networks. We denote by UNet[$\mathrm{lr}$] (UNet++[$\mathrm{lr}$]) the UNet (UNet++) model learned with $\mathrm{lr}$ as the initial value of the learning rate, where $\mathrm{lr}\in\\{10^{-5},10^{-4},10^{-3},10^{-2}\\}$. A summary of the parameter values used in the training step is given in Table 2. In the segmentation step, an overlapping sliding window of size $32\times 32$ was applied, where an overlap of $30$ pixels was used in order to limit edge artifacts and to generate highly smooth segmentation. In order to compare the performance of UNet with UNet++, two zones of the global N${}_{\rm{H_{2}}}$ mosaic were excluded from the initial patch data set. Constrained by the limited number of patches in the data set, small zones were removed, namely, the zones 166.1-168.3° and 350.3-353.5°. The choice of the removed regions was motivated by assessing the network performance in regions with a highly diverse column density and filaments content. While 350.3-353.5°, removed from mosaic 349-356°, is dense and rich in filaments ($38541$ filament pixels), the region 166.1-168.3°, removed from the mosaic 160-171°, is sparser and contains fewer filaments ($1459$ filament pixels). The filament density is always inferred based only on the incomplete ground truth (labeled part of the data set). The two removed zones were then segmented by the learned models, and the segmentation quality was assessed using the evaluation scores described in Section 3.1.5. The evaluation scores were also computed on the fully segmented Galactic plane in order to have a global performance evaluation of models. Table 2: Experimental setup Parameter | Value ---|--- patch size ($p$) | 32 pixels dataset split | {80%, 10%, 10%} batch size | 64 patches epochs | 100 initial learning rate | {$10^{-5},10^{-4},10^{-3},10^{-2}$} 111Parameters used in the experimental setup for UNet and UNet++ training ## 4 Results ### 4.1 Scores and segmented mosaic analysis In order to evaluate the training performances, we discuss below the different scores we obtained for the different tested scenarios. Two neural networks were tested (UNet and UNet++) with four different initial learning rates for each. For an input column density image, the segmentation process (see Section 3.1.4) returns a classification value mask from which it is possible to identify pixels that likely belong to a filamentary structure. Nevertheless, we did not attempt to extract the filaments like Schisano et al. (2020) did. We postpone this physical analysis on the newly identified filaments to a follow-up work. Here we present the method as a proof-of-concept and analyze its performances and returned results (segmented map of the whole Galactic plane). We illustrate these results on two portions of the Galactic plane that were selected for their characteristics in terms of column density and filament content (as known from the input data set of filament mask based on the spine+branches). To evaluate the performances of the different scenarios, Figure 14 presents the BCE curves for the training set and the validation set during the first $30$ epochs. For all models, these loss curves reach a plateau around the tenth epoch, confirming the rapid convergence of UNet-based networks. The performance of UNet++[$10^{-2}$] was removed from the displayed results due to convergence issues. As shown in Figure 14 and confirmed by this analysis, the models have similar performances and converge to a training error around $0.01$, except for schemes with a starting learning-rate value of $10^{-5}$ , where higher errors are reported (more than $0.016$ and $0.012$ for UNet[$10^{-5}$] and UNet++[$10^{-5}$], respectively). The similar performances obtained for the two architectures and the different learning rates indicate that the method is robust. Validation errors are slightly higher than training errors, with a difference up to $0.0011$, except for UNet[$10^{-5}$] (around $0.002$) and UNet++[$10^{-5}$] (around $0.0018$). The reported values indicate that the learned models show low bias and low variance. For both UNet and UNet++, the best performance is with an initial learning- rate value of $10^{-3}$. Overall, the best model in terms of loss function is the UNet++[$10^{-3}$]. For each scenario, the best-performing model in validation was used to segment the test set and compute the underlying BCE (see Table 3). BCEs calculated on the test set corroborate the previous analysis: scenarios with initial learning rates in [$10^{-2}$, $10^{-3}$, $10^{-4}$] result in similar test errors that do not exceed $0.009$, whereas scenarios with initial learning rates of $10^{-5}$ show a higher error. The lowest test error corresponds to the scenario UNet++[$10^{-3}$]. Figure 13: Dice curve evolution of schemes reported in Figure 14 are displayed over the first $30$ epochs in training (continued lines) and validation (dashed lines) steps, at classification threshold values $0.8$, $0.6$, $0.4,$ and $0.2$. The displayed curves are aligned with the results deduced from BCE curves, where close performances were obtained with initial learning-rate values of $10^{-4},10^{-3},\text{and }10^{-2},$ and a poorer performance is obtained for schemes with a learning-rate value of $10^{-5}$. The highest dice score is reported for UNet++[$10^{-3}$] (purple), and the lowest performance corresponds to the UNet[$10^{-5}$] scheme, (red) especially at thresholds 0.2 and 0.4. Similarly to the BCE curves and for all displayed schemes, a plateau regime is reached within the first ten epochs. In Figure 13 the dice index curves for the training and validation sets is plotted for different classification thresholds. The dice index was also computed on a test set using the best-performing models in validation (the models used in Table 3), and the corresponding results are given in Figure 7. In the same way, the dice index results are in line with the BCE. These models were used afterwards to segment the removed zones 166.1-168.3° and 350.3-353.5°. Figure 14: BCE evolution over the first $30$ epochs in training (continued lines) and validation (dashed lines) steps. UNet++[$10^{-2}$] schemes is removed as its corresponding BCE diverged. The displayed schemes show similar performances. Models with a learning-rate value of $10^{-5}$ resulted in higher BCEs. The lowest error is reported for UNet++[$10^{-3}$] (purple), and the highest error corresponds to the UNet[$10^{-5}$] scheme (red). A plateau regime is reached within the first ten epochs for all models, which confirms the rapid convergence of the UNet-based networks. Figure 16 presents the results of the segmentation process for the different scenarios tested and presented in Table 2. All segmented maps are presented within the range $[0,1]$. Overall, the obtained images are in line with the analyzed BCEs. The scenarios with initial learning rates of $10^{-2}$, $10^{-3}$, $\text{and }10^{-4}$ return very similar segmented maps. A noticeable difference is seen in maps that were segmented by both UNet and UNet++ with an initial learning rate of 10-5. In these cases, the filamentary structures are broader than those obtained for the other scenarios, especially for UNet. Moreover, the intensity of the segmentation maps also varies for these last two scenarios, where the low-value structures are better revealed (intensity variation up to a factor of 10 in those zones). The ability of these two last scenarios (and of the UNet in particular) to better reveal structures with a lower classification threshold might be used to detect structures that are not well seen on the original map, either due to their low contrast and/or their low column density. A close visual inspection of the column density images confirms that features revealed by UNet[$10^{-5}$] and UNet++[$10^{-5}$] are low-contrast filaments that were present in the original images, but absent from our input catalog of filaments that was used as ground truth. (a) PR curves for 350.3-353.5° (b) PR curves for 166.1-168.3° (c) PR curves for the fully segmented Galactic plane Figure 15: P-R curves of the schemes reported in Figure 14, computed on the segmented removed zones (top) and the full Galactic plane (bottom). (a) P-R curves computed on the segmented 350.3-353.5° , which corresponds to the dense region that was removed from the patches data set. (b) P-R curves computed on the segmented 166.1-168.3° , which corresponds to the sparse region that we removed from the patches data set. (c) P-R curves computed on the full segmented Galactic plane. Unlike in Fig. 15(b), P-R curves obtained on the latter are close to those obtained in Figure 15(a). Figure 16: Segmented maps obtained for the models analyzed in Figure 14, zooming in on a part of the two regions that was removed from the training namely, the $l$=166.1-168.3° (top) and $l$=350.3-353.5° (bottom) (see the red zones identified in Figure 3). The segmented images are displayed in the range $[0,1]$ representing the classification value according to which a pixel belongs to the filament class. For each region, the first row shows results of the UNet segmentation, and the second row shows results of the UNet++ segmentation with initial learning-rate values of 103 (left), 10-4 to 10-5 (right). The UNet++ with a learning rate of 10-2 is not presented because of diverging results (see Section 4.1). The N${}_{\rm{H_{2}}}$ is shown at this position for each region. The regions are $0.45$°$\times 0.45$° wide. In Figure 15, precision-recall curves (P-R curves) are shown for 350.3-353.5° (top left) and 166.1-168.3° (top right). The P-R curve represents precision vs recall for different threshold values. It is used to estimate the optimal threshold that maximizes the dice index (a trade-off between precision and recall). The more the P-R curve tends to the $(1,1)$ corner, the better the model. In other terms, the larger the area under the curve, the better the model. The objective is to estimate the optimal threshold that returns a trade-off between precision and recall. For clarity sake, all curves are zoomed in from $0-1$ to $0.7-1$ for precision and recall. In all figures, black asterisks refer to the precision-recall values at the optimal threshold; the values of the latter are reported in Tables 4.2, 4.2, and B. Different approaches can be used to compute the optimal threshold, such as minimizing the difference between the precision and recall, or minimizing the Euclidean distance between the P-R curve and the optimal performance, corresponding to a precision-recall of ($1,1$). Here, we computed the optimal threshold as the one that maximized the dice index (trade-off between filament and background recovery). In Tables 4.2 and 4.2, we report four samples from these P-R curves corresponding to conservative ($0.8$), medium ($0.5$), relaxed ($0.2$) and optimal (giving the best Dice index) thresholds. When investigating the dense zone of 350.3-353.5°, we note that, for a given threshold, all the models give results with similar performances (Dice indices $>85\%$), except for UNet[$10^{-5}$] (74.79% at threshold 0.2). Close optimal threshold are also obtained for all models, where values are situated between $0.35$ and $0.48$. Note that at the conservative threshold 0.8, UNet[$10^{-5}$] and UNet++[$10^{-5}$] result in low recall values compared with the remaining scenarios (72.96% and 75.14%, respectively), confirming the results reported with the segmented map where salient filaments are detected with lower values in these two scenarios compared with the remaining ones. However, they tend to be more performing when decreasing classification threshold, especially at threshold 0.2 where they are performing better than the remaining scenarios (the best recall is of 98.1% with UNet[$10^{-5}$], followed by UNet++[$10^{-5}$] with a recall value of 97.26%). This can be explained by the low value structures that are better revealed in these two scenarios as noticed before in the segmented map. In Appendix B, precision, recall and Dice index are computed on the fully-segmented Galactic plane, to infer the global segmentation performance. The resulting global scores are inline with the ones obtained with the dense zone where, for a given threshold, all models show close performances, except UNet[$10^{-5}$] and UNet++[$10^{-5}$] slightly less performing. Moreover, the optimal thresholds obtained with the global segmentation are close to thresholds obtained for the dense mosaic, where values range from $0.3$ to $0.44$. This result suggests that either the training step is more driven by high density regions and/or that these regions better represent the global properties observed on the Galactic plane. Table 3: Binary Cross Entropy Model | BCE Score ---|--- UNet$[10^{-2}]$ | $0.0085$ UNet$[10^{-3}]$ | $0.0084$ UNet$[10^{-4}]$ | $0.0088$ UNet$[10^{-5}]$ | $0.0161$ UNet++$[10^{-3}]$ | $0.0081$ UNet++$[10^{-4}]$ | $0.0088$ UNet++$[10^{-5}]$ | $0.0122$ 222Binary Cross Entropy (BCE) evaluation on the test set for the schemes reported in Figure 14. The lowest (highest) achieved BCE is given in blue (red). BCE values in test are inline with performances in training and validation steps with UNet++[$10^{-3}$] and UNet[$10^{-5}$] resulting in the best and less performing schemes, respectively. When examining the scores of the sparse zone of 166.1-168.3° in Table 4.2, all the models result in similar performances in precision. However, the recall performance per threshold has a higher contrast, where UNet++[$10^{-5}$] shows lower recall values for thresholds 0.8, 0.5, and 0.2. While the background is well recovered (the lowest precision is of $98.97\%$), lower recall values are obtained compared to the dense zone, where we had to relax the threshold to $0.2$ in order to improve filament recovery and obtain recall values higher than $70\%$. Similarly to the dense zone, close optimal dice indices were obtained for all models, where the difference between the best dice given by UNet++[$10^{-4}$] ($99.38\%$) and the more poorly performing UNet[$10^{-2}$] ($98.8\%$) is less than 0.6%. Although interesting scores are obtained at the optimal thresholds, it is very important to underline the very low values of these thresholds in all scenarios. Optimal thresholds range from $0.01$ to $<0.03$, except for UNet[$10^{-5}$] ($0.12$). The obtained values reflect the difficulty of the different networks to reveal structures in this mosaic, where almost $40\%$ to $60\%$ of the filament pixels are detected with classification values lower than $0.5$ (see the segmented maps in Figure 16). Moreover, we clearly note the discrepancy between scores computed on the sparse zone and the global scores reported in Appendices B, which might suggest that the trained models are more successful in revealing filaments in dense than in lower density zones. It is difficult, however, to conclude about the origin of this discrepancy because the number of labeled pixels (in both filament and background) is very different between the sparse and the dense zone. Limited labels in the sparse mosaic also imply that the computed scores with higher uncertainties need to be considered. Nevertheless, visual inspection of the segmented maps in Fig. 16 results in similar trends as observed for the dense zone, where close performances are noted for all models except for the models with an initial learning rate of $10^{-5}$. These models reveal more details at moderate to low classification values. ### 4.2 Analysis of missed structures In addition to pixel-level scores, the structure-level score was also computed in order to evaluate filament recovery in terms of structures. When some pixels of a given filament are missed, it does not automatically imply that the whole structure is missed. In Tables 4.2 and 4.2, the MS rate is computed at different classification thresholds for dense and sparse zones. As expected, the higher the threshold, the more structures are missed. Overall, we note that at conservative (0.8) and moderate (0.5) thresholds, low MS rates are obtained for dense mosaics compared with the sparse mosaic. In the latter, more strongly contrasting values are obtained across the classification thresholds where MS rates range from 0 (all structures are revealed) to almost 60% (more than half of the structures are missed; see the MS rates in Table 4.2). In a region with a low concentration of filaments, missing (or detecting) a structure would have more impact on the MS rate variation than in a dense region. Similarly to pixel-level scores, the global MS rates reported in Table B are closer to the values obtained with the dense zone than the sparse zone, and this for the same reasons as we invoked for pixel-level scores. In order to learn more about the structure of missed filaments, a missed- structure map at a classification threshold of $0.8$ was built. In this map, any structure that was missed by any model is represented. Here, the map intensity encodes the number of models that missed the structure, so that values range from 0 (detected structure) to 7 models (missed by all models). In Figure 17, representative portions from the sparse and dense regions are displayed. We note the prevalence of structures that were missed by all the models (yellow). In fact, $50\%$ of the missed structures in the whole Galactic plane are the same for all the models. After a close visual inspection of the missed structures, two categories are reported. The first category consists of small structures, which is the most prevalent category. These structures either correspond to small isolated filaments and/or to small parts that are missed in larger filaments. The second category corresponds to larger filaments that are misidentified as filaments in the ground truth. For example, structures reported in Figure 17 (bottom right) at positions (349°, 1°) and (350.5°, $-$1°) are excluded from the filament class. There are also isolated square-shaped structures that are mislabeled as filaments in the ground truth, corresponding to saturated pixels (see Figure 17 (bottom right) at position (350.8°, 1°). Unfortunately, trained models fail to reject these saturation bins when they are entangled within a true large filamentary structure. Table 4: Segmentation scores (350.3-353.5°) l—*7l Scores (%)Model UNet [10-2] UNet [10-3] UNet [10-4] UNet [10-5] UNet++ [10-3] UNet++ [10-4] UNet++ [10-5] Precision at 0.8 99.67 99.76 99.84 99.89 99.71 99.76 99.65 Precision at 0.5 97.45 98.06 98.24 97.11 97.87 97.87 95.13 Precision at 0.2 88.10 89.75 88.79 60.42 90.12 87.59 78.58 Precision at thropt 96.01 95.63 96.45 96.38 96.06 96.48 94.32 Recall at 0.8 79.15 79.45 79.10 72.96 78.34 79.71 75.14 Recall at 0.5 91.15 90.86 91.63 91.75 90.45 91.73 90.97 Recall at 0.2 96.55 96.27 96.83 98.10 96.09 96.82 97.24 Recall at thropt 92.86 93.91 93.86 92.49 93.17 93.37 91.82 Dice at 0.8 88.24 88.45 88.27 84.33 87.74 88.62 85.67 Dice at 0.5 94.19 94.33 94.82 94.35 94.02 94.70 93.00 Dice at 0.2 92.13 92.89 92.64 74.79 93.01 91.98 86.92 Diceopt 94.41 94.76 95.14 94.40 94.59 94.90 93.05 MS at 0.8 4.25 4.96 4.96 7.79 4.96 5.31 6.02 MS at 0.5 1.06 1.24 1.24 1.42 1.59 1.42 1.06 MS at 0.2 0.35 0.18 0.35 0.18 0.35 0.18 0.18 MS at thropt 0.53 0.35 0.53 1.24 0.53 0.53 0.88 thropt 0.41 0.35 0.38 0.48 0.37 0.42 0.47 333Segmentation scores evaluated on the the dense zone of 350.3-353.5° segmented by the models reported in Figure 14. Precision, recall, dice index, and MS rate are evaluated at classification threshold values of $0.8$, $0.5$, $0.2,$ and the optimal threshold. The latter refers to the threshold value optimizing the dice index and estimated using P-R curves (see Figure 15(a)). Blue (red) refers to the best (least performing) model in each row. The bold scores correspond to the absolute best (if in blue) or lowest (if in red) performance per score. Table 5: Segmentation scores (166.1-168.3°) l—*7l Scores (%)Model UNet [10-2] UNet [10-3] UNet [10-4] UNet [10-5] UNet++ [10-3] UNet++ [10-4] UNet++ [10-5] Precision at 0.8 100.00 100.00 100.00 100.00 100.00 100.00 100.00 Precision at 0.5 100.00 100.00 100.00 100.00 100.00 100.00 100.00 Precision at 0.2 100.00 100.00 100.00 100.00 100.00 100.00 100.00 Precision at thropt 99.04 98.97 99.65 99.66 99.25 99.32 99.24 Recall at 0.8 25.91 28.38 25.22 25.29 33.45 30.64 20.84 Recall at 0.5 49.01 58.40 49.01 54.90 63.33 60.59 41.74 Recall at 0.2 77.11 87.32 81.49 94.59 87.18 87.73 72.86 Recall at thropt 98.56 98.83 98.90 99.04 99.31 99.45 99.04 Dice at 0.8 41.15 44.21 40.28 40.37 50.13 46.90 34.49 Dice at 0.5 65.78 73.73 65.78 70.88 77.55 75.46 58.90 Dice at 0.2 87.07 93.23 89.80 97.22 93.15 93.46 84.30 Diceopt 98.80 98.90 99.28 99.35 99.28 99.38 99.14 MS at 0.8 61.02 45.76 57.63 50.85 32.20 37.29 44.07 MS at 0.5 16.95 6.78 10.17 15.25 6.78 6.78 11.86 MS at 0.2 1.69 0.00 1.69 0.00 0.00 1.69 1.69 MS at thropt 0.00 0.00 0.00 0.00 0.00 0.00 0.00 thropt 0.03 0.03 0.02 0.12 0.01 0.02 0.02 444Same as in Table 4.2, but for the sparse zone of 166.1-168.3° Figure 17: Input filaments missed by the segmentation process on the $l$=160-171° (top) and $l$=349-356° (bottom) regions of the Galactic plane, respectively. We show. the input filament mask (left) and the missed structures at a classification threshold of $0.8$ in a cumulative way (for all the scenarios) (right). The unit (color-coding) for the missed structures maps corresponds to the number of tested scenarios (from 1 to 7) that missed a given structure. ### 4.3 New possible filaments revealed by deep learning Based on the performance analysis in Section 4.1, two groups of models can be derived based on the initial learning rate: 1) models with an initial learning rate of $10^{-4},10^{-3},\text{and }10^{-2}$ , and 2) models with an initial learning rate of $10^{-5}$. In the following, we present the segmentation results that we illustrate for the best model, UNet++[$10^{-3}$], on the two selected submosaics. Figure 18 presents the evolution of the segmented map for the two selected regions as a function of the segmented map threshold. Figure 18: Zoom-in on the evolution of segmentation results as a function of the classification threshold showing the filamentary structures estimated by UNet++ 10-3 in 160-171° (six top images) and 349-356° (six bottom images) regions of the Galactic plane. The original H2 column density image (top left in each group), the ground-truth input filament mask, and the corresponding segmented image at different thresholds (0.8, 0.5, 0.2, and the optimal threshold) are shown from top left to bottom right. The regions are $2$°$\times 2$° wide. In both regions, more pixels are classified as filaments by the training and segmentation processes than in the input structures (input filament mask used as the ground truth in the supervised training). The ratio (new filament pixels to input filament pixels) varies between 2 to 7, depending on the threshold value (from 0.8 to 0.3, the optimal threshold). The same conclusions are drawn for the whole Galactic plane. In Figure 19, the distribution of candidate filament pixels across the entire Galactic plane, estimated in bins of 4.8° $\times$ 0.16°, is displayed for the ground truth and the segmentation results at different classification thresholds. Even at a conservative classification threshold of 0.8, more pixels are labeled as filaments than in the ground truth used in the learning step. As expected, the more we decrease the classification threshold, the more pixels are labeled as filament. A close visual inspection of the segmented images indicates that structures observed at thresholds lower than the optimal value are also seen on the original column density image, but were not previously detected due to their low contrast with respect to the surrounding background emission. (a) Ground truth (b) Segmentation at classification threshold of 0.8 (c) Segmentation at classification threshold of 0.5 (d) Segmentation at classification threshold of 0.2 (e) Ratio of 18(b) to 18(a) Figure 19: Number density distribution of candidate filament pixels across the entire Galactic plane, estimated in bins of 4.8° $\times$ 0.16° and comparing the ground truth of Fig. 18(a) with the segmentation results of the model UNet++ 10-3 at classification thresholds of 0.8 in Fig. 18(b), 0.5 in Fig. 18(c), and 0.2 in Fig. 18(d). In Fig. 18(e), we display the ratio of candidate filament pixels in the segmentation at a classification threshold of 0.8 to the ground truth. The squared structures shown in Figure 18 (bottom) are saturated pixels corresponding to bright sources located in filaments. These structures also appear in the ground truth, and we therefore retrieve them when applying our model. Because we lack information about the column density in these saturated pixels, we left them as squares in the segmented maps. Figure 20 presents the same result as Figure 18, but for the binarized version of the segmented map. Figure 20: Zoom-in on the evolution of segmentation results generated by UNet++ 10-3 model in mosaics 160-171° (six top images) and 349-356° (six bottom images). Here, the estimated binary filament masks are displayed at classification thresholds 0.8, 0.5, 0.2, and the optimal threshold. The original H2 column density image (image with the color bar) and the true input filament mask are also displayed for comparison. The regions are $2$°$\times 2$° wide. The filamentary structures identified at a given threshold are now represented as 1 when the associated pixel belongs to the filament class 0 instead. This representation allows a more direct comparison with the input filament mask (spine plus branches; see Fig. 2). However, the classification value itself (that indicates whether a pixel belongs to the filament class) no longer appears in this representation. As for results presented in Figure 18, the threshold decrease has two effects: i) more pixels are identified as belonging to the filament class (new structures are detected, in particular, those with a faint-to-low contrast that are barely visible in the original N${}_{\rm{H_{2}}}$ image), and ii) a given structure becomes thicker. This last effect can be identified as a lowering of the corresponding column density threshold, where the highest threshold identifies the densest part of the filament. It is interesting to note that the optimal threshold (as well as lower threshold values) in both regions identifies the filamentary structures as observed in the original column density map, down to their external envelope emission, before reaching the background emission. This result is important because it will allow a precise study of the filament-background relation. The widening of the filamentary structures can also be seen as the definition of the RoI given by Schisano et al. (2020) (see also Figure 2, bottom right). Schisano et al. (2020) defined the RoI as the objects that define filamentary candidates in their catalog. This point is important because it implies that the comparison of our segmented map results with the RoI would lower the factor we derived that represents the number of pixels classified as filament as the RoI are always thicker than the spine plus branches we used as input in this work to train the network to learn what a filament is. In this work, we infer the filament mask as a ground truth from the filament spine plus branches, as shown on Figure 2 (bottom left). However, the RoI defines the filament in order to delineate its spread on the column density map, and this corresponds to the observed widening of the structures with the lowering of the threshold. The comparison of the different structures is illustrated in Figure 21. Figure 21: Comparison of the segmented map result using the UNet++ 10-3 with the input spine plus branches (black contours) and the RoI (red contours). Because we used spine plus branches as input to define a filament here, we kept this input structure as a reference to compare with the result of the segmentation process. A key point in this work is to ascertain the nature of the filamentary structures we reveal. Filaments are made of gas and dust. The filaments detected by Schisano et al. (2020) are traced in dust emission maps. Dust grains emit over a wide range of wavelengths, and they act as absorbers in the optical, near-, and mid-infrared parts of the spectrum. Because of their dusty composition, filaments are well visible as absorbing features at shorter wavelengths (optical, near-, and mid-infrared) in the Galactic plane, and these data can be used to ascertain the nature of the structures returned by the training and segmentation processes. Only filaments that are visible in absorption on a strong emission background can be detected at short wavelengths. (Sub)millimeter emission of cold dust also reveals filaments (Mattern et al. 2018; Leurini et al. 2019) and can be used to ascertain the nature of the structures without encountering the extinction problem. This empirical (data-based) validation of the results is a first step in the analysis. In Figure 22, we illustrate the interest of this multiwavelength analysis to ascertain the nature of new detected filaments on the Galactic region G351.776-0.527. This region hosts a high-mass star-forming region analyzed by Leurini et al. (2019). Figure 22: Segmented map obtained for star-forming region G351.776-0.527 (bottom left) compared with the 2MASS $K$-band image (top left), where filaments are observed in absorption, and the column density map (top right), where filaments are observed in emission. The filamentary structures visible in the image segmented by UNet++[10${-3}$] (bottom left) and displayed in the [0,1] range are both visible in the 2MASS $K$ and N${}_{\rm{{H_{2}}}}$ column density images, ensuring their nature. The G351.776-0.527 source is located at the center of the images, connected to a filamentary network (hub). The corresponding input filament mask is shown (bottom right). Region G351.776-0.527 is located at the center of Figure 22 and appears as a hub with filaments converging toward the saturated central point located at $l$=351.77°, $b=-0.538$°. A filamentary structure observed in the segmented map right of the bright central source that is not visible in the input filament mask is seen in the 2MASS $K$-band image, confirming its nature. The segmented image also suggests that the region might be located at the edge of a bubble. A bright ionized region, G351.46-00.44, is located nearby and could explain the high level of turbulence and the high-mass star formation observed in this zone (Lee et al. 2012). The large-scale view of this region, revealed by the segmented map with suggested multiple filament connections of the central source with the surrounding medium, has to be confirmed with high- sensitivity observations of dust emission that could be complemented by spectroscopic data of dense gas molecular tracers, keeping in mind that, as pointed by Hacar et al. (2022), filaments identified with Herschel data (i.e., using dust continuum emission) might be a different family of objects than those detected in molecular line tracers. ## 5 Discussion and future prospects The purpose of this work was to study the potential of supervised deep learning as a new way to detect filaments in images of the Galactic interstellar medium. At this stage, the filamentary structures are revealed, but the filaments themselves are not extracted, with a measurement of their physical properties, from the segmented images. While the first task requires semantic segmentation, the second task consists of instance segmentation (Gu et al. 2022). In the first, the task is limited to attributing a class to each pixel, which is done in this paper, whereas in the latter, existing filaments are in addition enumerated to allow a global statistical study. In this paper, we used UNet-based networks which are the most modern methods in semantic segmentation. The analyzed performance in Section 4 proves the efficiency of these networks not only in revealing structures already existing in the initial catalog, but also in adding new structures that have not been detected before and that are confirmed through a detection at shorter and/or longer wavelengths, namely, at near- and mid-infrared and/or (sub)millimeter wavelengths, respectively. In astrophysics, several independent estimators can be used to ascertain the true nature of the detected filamentary structures, such as expert knowledge or a knowledge based on a large statistical definition, such as the one used in citizen projects, of particular interest for machine learning (Christy et al. 2022). Results of numerical simulations and/or data obtained at other wavelengths can also be used. On the multiwavelength data side, for example, filaments are clearly visible at other wavelengths in the Galactic plane because they are composed of dust, and these data can be used to ascertain the nature of the structures returned by the segmentation process, as shown in Figure 22. The UNet-based networks are supervised deep-learning algorithm. In spite of the incomplete ground truth, these networks produced a good estimate of filamentary structures. It is important to build a more enriched ground truth, however, to solve more complex tasks such as instance segmentation. This might be possible by combining several existing catalogs of filaments obtained on the Galactic plane, for example, by combining the Hi-GAL catalog Schisano et al. (2020) with getSF extractions made on several regions of the plane (Men’shchikov 2021). Another possibility is to use filament segmentation by UNets as a prestep and then consider the produced filaments mask as the ground truth for the instance segmentation. Another crucial step in filament segmentation using deep-learning algorithms is data normalization. Filament detection depends not only on the intrinsic column density of the structure, but also on the column density and the structure of the background. By using local (per patch) normalization, the filament contrast relative to the neighboring background is enhanced. As illustrated in Figure 9, a classical local normalization method was used in this work in order to enhance low-contrast filaments that are in turn well integrated in the training process, allowing them to be representative. Recently, more sophisticated multiscale normalization has been used in the Hi- GAL image processing to highlight the faintest structures observed on the Galactic plane (Li Causi et al. 2016). These normalized data are very interesting as input for deep-learning networks. Unfortunately, no ground truth exists for these normalized images so far, which does not allow their use for the moment. An important result of the segmentation for astrophysical purposes is to determine the classification threshold (intensity of the segmentation map) that allows for an optimal detection of filaments. While the optimal thresholds reported in Table B allowed a good recovery of existing and new filaments, filaments are still missed at these classification thresholds, some because they were misclassified in the ground truth (see Sect. 4.2). Another key point is to consider a region-specific optimal threshold rather than a unique global one. According to results reported in Tables 4.2, 4.2, and B, the optimal threshold of a given model is affected by the column density of the studied zone. In this work, the optimal thresholds in Tables 4.2 and 4.2 were inferred in narrow zones, with a low number of labeled pixels for the sparse zone. To obtain a robust estimate of a region-specific optimal threshold, a split of the Galactic plane divided into large homogeneous zones in terms of filament concentration is envisioned. The optimal thresholds can then be inferred on these slices. From the computational point of view, we trained the different networks on a NVIDIA RTX 2080Ti. Table 6 gives the training time for the different scenarios for 100 epochs. As UNet++ is a larger network than UNet, it takes slightly more time to train: while UNet requires about $2.2$h, UNet++ asks for about $2.8$h. As the patch data set is small (around 210 MB), these times are not impacted by the loading of the data. When the training step is completed, the neural networks are usually faster on CPU than on GPU Goodfellow et al. (2016) because the transfer from CPU memory to GPU memory takes time. The segmentation of the mosaics was therefore built on an Intel CPU machine (i7-10610U). It took about $4$h per mosaic with an overlap of 30 pixels and 32$\times$32 patches. The total training and segmentation time is therefore estimated to be $6.5$h per mosaic. Table 6: Training time Model | Training time [hour] ---|--- UNet$[10^{-2}]$ | $2.08$ UNet$[10^{-3}]$ | $2.23$ UNet$[10^{-4}]$ | $2.19$ UNet$[10^{-5}]$ | $2.17$ UNet++$[10^{-3}]$ | $3.22$ UNet++$[10^{-4}]$ | $2.79$ UNet++$[10^{-5}]$ | $2.73$ 555Training time in hours for the schemes reported in Figure 14. The shortest (longest) training time, in blue (red), is achieved by UNet[$10^{-2}$] (UNet++[$10^{-3}$]). From the method point of view, future works will include an improvement of the segmentation process by using dedicated windows to build the patches as in Pielawski & Wählby (2020). These tools may dramatically lower the computational burden. We also investigate alternative ways of building a larger set of patches while keeping good statistical properties. The method used in this work guarantees a good preservation of the statistical distribution, but leads to a small number of patches. Although this method has some limitations, that is, the limited quantity of patches for the training and the incomplete ground truth, and although it reveals filament structures instead of extracting them, the net increase (a factor between 2 and 7 on the whole segmented map) of the number of pixels that belongs to the filament class and the robust detection of intrinsically faint and/or low contrast ones offers important perspectives. We currently explore the implementation of an augmented ground-truth data set using results of numerical simulations on Galactic filaments. The extraction and separation of filament pixels observed in the 2D segmented map is also ongoing, and we add 3D spectroscopic data. ## 6 Conclusions We explored whether deep-learning networks, UNet-5 and UNet++, can be used to segment images of the whole Galactic plane in order to reveal filamentary structures. * • Using molecular hydrogen column density image of the Galactic plane obtained as part of the Hi-GAL survey and filaments previously extracted by Schisano et al. (2020), we trained two different UNet-5 based networks with six different scenarios based on a different initial learning rate. * • We showed the results and estimated the performances of the different scenarios that we presented for two representative mosaics of the Galactic plane selected for their low and high column density density and filament content. * • We determined the best models for these mosaics based on machine-learning metrics. We focused the training estimates on the recovery of input structures (filaments and background) and defined for each mosaic and for the whole plane an optimal classification threshold that ensured the best recovery of input structures. * • We show that depending on the model and the selected threshold, new pixels classified as filament candidates increase by a factor between 2 to 7 (compared to the input spine+branches structures used as ground truth). This suggests that this new method has the potential of revealong filamentary structures that may not be extracted by non-ML-based algorithms. We point out the high potential of the produced database for future studies of filaments (statistical analysis or follow-ups). We will use the results of the numerical simulations to enrich the ground truth and assess the uncertainties on segmented maps. The astrophysical analysis of the produced database is ongoing and will be published in a separate paper. ###### Acknowledgements. This project has received financial support from the CNRS through the MITI interdisciplinary programs (Astroinformatics 2018, 2019). This project is part of the ongoing project BigSF funded by the Aix Marseille Université A*Midex Foundation (SB post-doctoral funding). AZ thanks the support of the Institut Universitaire de France. We thank the anonymous referee for their constructive comments that helped to improve the quality of the paper. This research has made use of ”Aladin Sky Atlas” developed at CDS, Strasbourg Observatory, France. ## References * Alzahrani & Boufama (2021) Alzahrani, Y. & Boufama, B. 2021, SN Computer Science, 2, 1 * André et al. (2014) André, P., Di Francesco, J., Ward-Thompson, D., et al. 2014, in Protostars and Planets VI, ed. H. Beuther, R. S. Klessen, C. P. Dullemond, & T. Henning, 27 * André et al. (2010) André, P., Men’shchikov, A., Bontemps, S., et al. 2010, A&A, 518, L102 * Aragon-Calvo (2019) Aragon-Calvo, M. A. 2019, Monthly Notices of the Royal Astronomical Society, 484, 5771 * Arzoumanian et al. (2011) Arzoumanian, D., André, P., Didelon, P., et al. 2011, A&A, 529, L6 * Arzoumanian et al. (2019) Arzoumanian, D., André, P., Könyves, V., et al. 2019, A&A, 621, A42 * Batson & Royer (2019) Batson, J. & Royer, L. 2019, in International Conference on Machine Learning, PMLR, 524–533 * Bekki (2021) Bekki, K. 2021, A&A, 647, A120 * Bernard et al. (2010) Bernard, J. P., Paradis, D., Marshall, D. J., et al. 2010, A&A, 518, L88 * Bianco et al. (2021) Bianco, M., Giri, S. K., Iliev, I. T., & Mellema, G. 2021, Monthly Notices of the Royal Astronomical Society, 505, 3982 * Bracco et al. (2020) Bracco, A., Bresnahan, D., Palmeirim, P., et al. 2020, A&A, 644, A5 * Chen et al. (2020) Chen, M. C.-Y., Di Francesco, J., Rosolowsky, E., et al. 2020, ApJ, 891, 84 * Christy et al. (2022) Christy, C. T., Jayasinghe, T., Stanek, K. Z., et al. 2022, PASP, 134, 024201 * Clarke et al. (2020) Clarke, S. D., Williams, G. M., & Walch, S. 2020, MNRAS, 497, 4390 * Elia et al. (2013) Elia, D., Molinari, S., Fukui, Y., et al. 2013, ApJ, 772, 45 * Fiorio & Gustedt (1996) Fiorio, C. & Gustedt, J. 1996, Theoretical Computer Science, 154, 165 * Fu & Mui (1981) Fu, K.-S. & Mui, J. 1981, Pattern recognition, 13, 3 * Fukui et al. (2019) Fukui, Y., Tokuda, K., Saigo, K., et al. 2019, ApJ, 886, 14 * Goodfellow et al. (2016) Goodfellow, I., Bengio, Y., & Courville, A. 2016, Deep learning (MIT press) * Gu et al. (2022) Gu, W., Bai, S., & Kong, L. 2022, Image and Vision Computing, 120, 104401 * Hacar et al. (2020) Hacar, A., Bosman, A. D., & van Dishoeck, E. F. 2020, A&A, 635, A4 * Hacar et al. (2022) Hacar, A., Clark, S., Heitsch, F., et al. 2022, arXiv e-prints, arXiv:2203.09562 * Hacar et al. (2018) Hacar, A., Tafalla, M., Forbrich, J., et al. 2018, A&A, 610, A77 * Hastie et al. (2001) Hastie, T., Tibshirani, R., & Friedman, J. 2001, The Elements of Statistical Learning: Data Mining, Inference, and Prediction, Springer series in statistics (Springer) * Hausen & Robertson (2020) Hausen, R. & Robertson, B. E. 2020, ApJS, 248, 20 * He et al. (2016) He, K., Zhang, X., Ren, S., & Sun, J. 2016, in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 770–778 * Hoemann et al. (2021) Hoemann, E., Heigl, S., & Burkert, A. 2021, arXiv e-prints, arXiv:2104.04541 * Hsieh et al. (2021) Hsieh, C.-H., Arce, H. G., Mardones, D., Kong, S., & Plunkett, A. 2021, ApJ, 908, 92 * Jadon (2020) Jadon, S. 2020, 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), 1 * Kingma & Ba (2015) Kingma, D. P. & Ba, J. 2015, in ICLR * Koch & Rosolowsky (2015) Koch, E. W. & Rosolowsky, E. W. 2015, MNRAS, 452, 3435 * Könyves et al. (2020) Könyves, V., André, P., Arzoumanian, D., et al. 2020, A&A, 635, A34 * Krizhevsky et al. (2012) Krizhevsky, A., Sutskever, I., & Hinton, G. E. 2012, Communications of the ACM, 60, 84 * Kull et al. (2017) Kull, M., Silva Filho, T. M., & Flach, P. 2017, Electronic Journal of Statistics, 11, 5052 * Kumar et al. (2020) Kumar, M. S. N., Palmeirim, P., Arzoumanian, D., & Inutsuka, S. I. 2020, A&A, 642, A87 * LeCun et al. (1989) LeCun, Y., Boser, B. E., Denker, J. S., et al. 1989, Neural Computation, 1, 541 * Lee et al. (2015) Lee, C.-Y., Xie, S., Gallagher, P., Zhang, Z., & Tu, Z. 2015, in Proceedings of Machine Learning Research, Vol. 38, Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics, ed. G. Lebanon & S. V. N. Vishwanathan (San Diego, California, USA: PMLR), 562–570 * Lee et al. (2012) Lee, E. J., Murray, N., & Rahman, M. 2012, ApJ, 752, 146 * Leurini et al. (2019) Leurini, S., Schisano, E., Pillai, T., et al. 2019, A&A, 621, A130 * Li Causi et al. (2016) Li Causi, G., Schisano, E., Liu, S. J., Molinari, S., & Di Giorgio, A. 2016, in Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series, Vol. 9904, Space Telescopes and Instrumentation 2016: Optical, Infrared, and Millimeter Wave, ed. H. A. MacEwen, G. G. Fazio, M. Lystrup, N. Batalha, N. Siegler, & E. C. Tong, 99045V * Long et al. (2015) Long, J., Shelhamer, E., & Darrell, T. 2015, in Proceedings of the IEEE conference on computer vision and pattern recognition, 3431–3440 * Mattern et al. (2018) Mattern, M., Kauffmann, J., Csengeri, T., et al. 2018, A&A, 619, A166 * Men’shchikov (2021) Men’shchikov, A. 2021, A&A, 649, A89 * Milletari et al. (2016) Milletari, F., Navab, N., & Ahmadi, S.-A. 2016, in 2016 fourth international conference on 3D vision (3DV), IEEE, 565–571 * Molinari et al. (2016) Molinari, S., Schisano, E., Elia, D., et al. 2016, A&A, 591, A149 * Molinari et al. (2010) Molinari, S., Swinyard, B., Bally, J., et al. 2010, A&A, 518, L100 * Oktay et al. (2018) Oktay, O., Schlemper, J., Folgoc, L. L., et al. 2018, arXiv preprint arXiv:1804.03999 * Pielawski & Wählby (2020) Pielawski, N. & Wählby, C. 2020, PloS one, 15, e0229839 * Priestley & Whitworth (2022) Priestley, F. D. & Whitworth, A. P. 2022, arXiv e-prints, arXiv:2203.04367 * Robinson & Whelan (2004) Robinson, K. & Whelan, P. F. 2004, Pattern Recognition Letters, 25, 1759 * Robitaille et al. (2020) Robitaille, T., Deil, C., & Ginsburg, A. 2020, Astrophysics Source Code Library, ascl * Ronneberger et al. (2015) Ronneberger, O., Fischer, P., & Brox, T. 2015, in International Conference on Medical image computing and computer-assisted intervention, Springer, 234–241 * Schisano et al. (2020) Schisano, E., Molinari, S., Elia, D., et al. 2020, MNRAS, 492, 5420 * Schisano et al. (2014) Schisano, E., Rygl, K. L. J., Molinari, S., et al. 2014, ApJ, 791, 27 * Shimajiri et al. (2019) Shimajiri, Y., André, P., Ntormousi, E., et al. 2019, A&A, 632, A83 * Soille (2003) Soille, P. 2003, Morphological Image Analysis: Principles and Applications, 2nd edn. (Berlin, Heidelberg: Springer-Verlag) * Sousbie (2011) Sousbie, T. 2011, MNRAS, 414, 350 * Traficante et al. (2011) Traficante, A., Calzoletti, L., Veneziani, M., et al. 2011, MNRAS, 416, 2932 * Vincent (1993) Vincent, L. 1993, IEEE Transactions on Image Processing, 2, 176 * Wells & Norman (2021) Wells, A. I. & Norman, M. L. 2021, The Astrophysical Journal Supplement Series, 254, 41 * Xia & Kulis (2017) Xia, X. & Kulis, B. 2017, arXiv preprint arXiv:1711.08506 * Zavagno et al. (2020) Zavagno, A., André, P., Schuller, F., et al. 2020, A&A, 638, A7 * Zhou et al. (2019) Zhou, Z., Siddiquee, M. M. R., Tajbakhsh, N., & Liang, J. 2019, IEEE transactions on medical imaging, 39, 1856 * Zhu et al. (2019) Zhu, G., Lin, G., Wang, D., Liu, S., & Yang, X. 2019, Solar Physics, 294, 117 * Zucker & Chen (2018) Zucker, C. & Chen, H. H.-H. 2018, ApJ, 864, 152 ## Appendix A Different input data maps All the input maps used in the training process (the column density map, the filament mask, the background mask, and the missing data mask maps; see Section 2.2) were originally produced from the individual Hi-GAL mosaics extending over 10° of longitude (see Schisano et al. (2020)). To facilitate the use of the four input maps during the training process, we merged the individual maps using the Python module reproject666https://reproject.readthedocs.io/en/stable/index.html. The patches were then cut as shown in Figure 4. The four input maps used in the supervised training process are presented in Figure 3. These maps are described below. ### N${}_{\rm{H_{2}}}$ map The column density maps N${}_{\rm{H_{2}}}$ were obtained as part of the Herschel observations of the Galactic plane, Hi-GAL. The column density N${}_{\rm{H_{2}}}$ maps were computed from the photometrically calibrated Hi- GAL mosaics following the approach described in Elia et al. (2013). The Herschel data were convolved to the 500 $\mu$m resolution ( 36$\arcsec$) and rebinned on that map grid. Then a pixel-by-pixel fitting with a single- temperature greybody function was performed, as described in Elia et al. (2013); Schisano et al. (2020). We directly used the data derived and presented in Schisano et al. (2020). ### Missing-data map The N${}_{\rm{H_{2}}}$ map presents local degraded zones (noise, saturation, and overlap issues) that we wished to exclude from the training process. Examples of these degraded zones are shown in Figure 23. (a) (b) (c) (d) Figure 23: Four sources of missing data. (a) Example of a complex boundary due to satellite scanning. (b) Example of a noisy pattern inside the column density map (in log scale) due to the satellite scan. (c) Saturated pixels. (d) Example of structured artifacts built by the mapping process. In Figure 23 we show the four sources of missing data. First, the boundaries of the mosaics show a grid-like pattern, as shown in Figure 22(a). These features are due to the scanning pattern followed by the Herschel satellite while performing the observations of an Hi-GAL tile. The grid pattern corresponds to the slewing phase of the scanning pattern, when the satellite inverted the scan direction. Second, the Hi-GAL mosaics were built using the Unimap mapmaking (Traficante et al. 2011) following a strategy to avoid large- scale intensity gradients over the image. We refer to Appendix A of Schisano et al. (2020) for a detailed description of how the mosaics were built. In short, UNIMAP mapmaker was run to simultaneously process the data of two adjacent Hi-GAL tiles, creating what the authors call a texel, spanning 4°$\times$ 2°. The UNIMAP processing is able to deal with the slewing region along the side covered by observation of both Hi-GAL tiles. This produces an image of better quality with respect to the simple mosaicking of the two tiles. Multiple texels are combined together to produce the overall mosaic that is 10° wide in longitude. While it is possible to directly process a larger portion of the Galactic plane with UNIMAP, this approach introduces large gradients in intensity over the entire mosaic. The UNIMAP mapmaker produces maps with an average intensity level equal to zero. Therefore, the texels require to be calibrated in flux (Bernard et al. 2010). In some cases, there are discrepancies in the calibration level of adjacent texels that introduce sharp variations in intensity in the overlapping region between texels. An example is show in Figure 22(b) and Figure 22(d). These variations are not physical and were masked out from the learning process. The final full map of missing data was built by combining all the types of possible missing data described above. They were then removed from the training. ### Filament-mask map Schisano et al. (2020) published a catalog of filaments detected in the Galactic plane from N${}_{\rm{H_{2}}}$ images obtained with Hi-GAL photometric data. The position of spines and branches is available through binary masks in which pixels belonging to these structures are tagged for the 32059 filaments published in the catalog (see Figure 2 and their figure 3). We used this information as an input mask to define our ground truth for the filaments. The ground truth was defined over the outputs of the Hi-GAL filament catalog, and it depends on the completeness of that catalog. This implies that the ground truth is not absolutely fully defined because it will miss the information from any feature that may not have been detected by Schisano et al. (2020). ### Background map On all pixels, the column density mosaic contains emission from both filaments and background (Schisano et al. 2020). We examined each 10° mosaic to define a background level as the lowest level of emission observed on the mosaic that does not overlap any filament branches that were detected and used as ground truth. This means that our definition of the background emission sets the background class on a very low number of pixels, but ensures that these pixels do not contain filaments. We work on a more precise definition of the background to allow more pixels to be labeled in this class. We are primarily interested in detecting filamentary structures here. Moreover, the local normalization applied to all patches before the training process (see Figure 9) tends to limit the impact of the background on the detection of filaments. This reduces the need for a precise definition of the background further. ## Appendix B Supplementary quantitative scores Table 7: Dice index Model | Dice index [$0.2$] | Dice index [$0.4$] | Dice index [$0.6$] | Dice index [$0.8$] ---|---|---|---|--- UNet$[10^{-2}]$ | $93.19$ | $94.34$ | $94.07$ | $92.06$ UNet$[10^{-3}]$ | $93.13$ | $94.37$ | $94.16$ | $92.43$ UNet$[10^{-4}]$ | $92.75$ | $94.09$ | $93.79$ | $91.9$ UNet$[10^{-5}]$ | $82.4$ | $89.6$ | $91.3$ | $87.66$ UNet++$[10^{-3}]$ | $93.46$ | $94.61$ | $94.28$ | $92.36$ UNet++$[10^{-4}]$ | $93.04$ | $94.07$ | $93.6$ | $91.57$ UNet++$[10^{-5}]$ | $90.24$ | $91.88$ | $91.02$ | $87.67$ 777Dice index evaluation on the test set for the schemes reported in Figure 14 at classification threshold values of $0.8$, $0.6$, $0.4,$ and $0.2$. Blue (red) refers to the best (less performing) scheme in each column. The bold scores correspond to the absolute best (if in blue) or lowest (if in red) dice index. The closer the dice index to $1$, the better. Dice index values in test were aligned with performances in training and validation steps as close performances are obtained for schemes with initial learning-rate values in $[10^{-4},10^{-3},10^{-2}]$ and lower performance are noted for schemes with a learning-rate value of $10^{-5}$. Moreover, UNet++[$10^{-3}$] and UNet[$10^{-5}$] result in the best ($94.61\%$ at $0.4$) and least performing ($82.4\%$ at $0.2$) schemes, respectively. Table 8: Segmentation scores l—*7l Scores (%)Model UNet [10-2] UNet [10-3] UNet [10-4] UNet [10-5] UNet++ [10-3] UNet++ [10-4] UNet++ [10-5] Precision at 0.8 99.71 99.74 99.75 99.73 99.74 99.74 99.72 Precision at 0.5 98.29 98.45 98.54 97.07 98.47 98.42 97.35 Precision at 0.2 91.43 92.08 91.75 64.85 92.45 91.17 86.16 Precision at thropt 95.65 95.85 95.88 94.95 95.75 95.94 93.61 Recall at 0.8 72.38 71.34 70.50 65.14 70.87 71.74 65.54 Recall at 0.5 90.13 89.90 89.13 89.25 89.78 89.78 85.59 Recall at 0.2 97.49 97.64 97.29 98.72 97.61 97.50 95.81 Recall at thropt 95.26 95.50 94.96 92.24 95.67 94.97 91.56 Dice at 0.8 83.88 83.19 82.62 78.81 82.87 83.46 79.10 Dice at 0.5 94.05 94.00 93.62 93.01 93.95 93.92 91.11 Dice at 0.2 94.41 94.82 94.49 78.31 95.01 94.27 90.76 Diceopt 95.46 95.68 95.42 93.58 95.71 95.45 92.57 MS at 0.8 14.36 14.81 16.51 17.58 15.28 15.10 17.69 MS at 0.5 4.61 4.40 5.25 4.07 4.62 4.55 5.83 MS at 0.2 0.83 0.77 0.81 0.28 0.85 0.82 1.10 MS at thropt 1.95 1.69 1.88 2.72 1.68 1.97 2.85 thropt 0.32 0.31 0.31 0.44 0.30 0.33 0.34 888Segmentation scores evaluated on the Galactic plane segmented by the models reported in Figure 14. Precision, recall, dice index, and MS rate are evaluated at classification threshold values of $0.8$, $0.5$, $0.2,$ and the optimal threshold. The latter refers to the threshold value optimizing the dice index and estimated using P-R curves (see Figure 15(c)). Blue (red) refers to the best (least performing) scheme in each row. The bold scores correspond to the absolute best (if in blue) or lowest (if in red) performance per score. Overall, the segmentation scores obtained on the fully segmented Galactic plane consolidate the results obtained with the zone of 350.3-353.5° in Table 4.2
# FoPro: Few-Shot Guided Robust Webly-Supervised Prototypical Learning Yulei Qin,1 Xingyu Chen,1 Chao Chen,1 Yunhang Shen,1 Bo Ren,1 Yun Gu,2 Jie Yang,2 Chunhua Shen3 ###### Abstract Recently, webly supervised learning (WSL) has been studied to leverage numerous and accessible data from the Internet. Most existing methods focus on learning noise-robust models from web images while neglecting the performance drop caused by the differences between web domain and real-world domain. However, only by tackling the performance gap above can we fully exploit the practical value of web datasets. To this end, we propose a Few-shot guided Prototypical (FoPro) representation learning method, which only needs a few labeled examples from reality and can significantly improve the performance in the real-world domain. Specifically, we initialize each class center with few- shot real-world data as the “realistic” prototype. Then, the intra-class distance between web instances and “realistic” prototypes is narrowed by contrastive learning. Finally, we measure image-prototype distance with a learnable metric. Prototypes are polished by adjacent high-quality web images and involved in removing distant out-of-distribution samples. In experiments, FoPro is trained on web datasets with a few real-world examples guided and evaluated on real-world datasets. Our method achieves the state-of-the-art performance on three fine-grained datasets and two large-scale datasets. Compared with existing WSL methods under the same few-shot settings, FoPro still excels in real-world generalization. Code is available at https://github.com/yuleiqin/fopro. ## Introduction The past decade has witnessed a revolution in computer vision with the advent of large-scale labeled datasets (_e.g._ , ImageNet (Deng et al. 2009)). However, a large collection of data are sometimes inaccessible, let alone the time-consuming and expensive annotations. On the contrary, there are abundant weakly labeled images on the Internet. Therefore, webly supervised learning (WSL) has attracted growing attention from researchers (Krause et al. 2016; Kaur, Sikka, and Divakaran 2017; Kolesnikov et al. 2019; Zhang et al. 2020; Tu et al. 2020; Liu et al. 2021; Zhang et al. 2021). Figure 1: Differences between web and real-world images. (a) Web dataset noise. (b) Web dataset bias. Queries and tags are directly used as weak labels without verification, bringing about a considerable proportion of noises in web datasets (_e.g._ , 20% in JMT-300M (Sun et al. 2017), 34% in WebVision (Li et al. 2017), and 32% in WebFG496 (Sun et al. 2021)). As shown in Fig. 1(a), various noises include label flipping errors, semantic ambiguity of polysemy queries, and outliers of unknown categories. To alleviate their effect, prior knowledge such as neighbor density (Guo et al. 2018), reference clean sets (Jiang et al. 2018; Lee et al. 2018), and side information (Zhou et al. 2020; Cheng et al. 2020) is explored for label correction and sample selection. Recently, Li et al. (Li, Xiong, and Hoi 2020) develop a self-supervised method with representative class prototypes (MoPro) to achieve satisfying performance. Most existing WSL methods are merely concerned with noise reduction. They ignore the model degradation in real-world scenarios because the performance on web domain testing sets is emphasized in previous model assessments. Domain gaps exist between images crawled from the web (_e.g._ , advertising photos, artworks, and rendering) and those captured in reality (see Fig. 1(b)). In this case, the better fitting of web images counteractively leads to worse generalization on practical applications. Few studies try to tackle such performance gaps by domain adaptation methods. For example, Xu et al. (Xu et al. 2016) distill knowledge from the web domain to the real-world domain. Niu et al. (Niu, Li, and Xu 2015) fine-tune pretrained models on real-world datasets. However, both of them need plenty of labeled data in the target domain, which impedes practicability. Figure 2: (a) The t-SNE (Van der Maaten and Hinton 2008) of the low- dimensional embeddings of web images substantiates that with the increase of $K$, class prototypes ($\times$) are regularized to approach few shots ($\Delta$) with dense intra-class and isolated inter-class distribution. (b) The diminished performance gap between the testing results of web (WebVision1k) and real-world (ImageNet1k) images confirms that FoPro takes full advantage of few shots to improve its generalization beyond the web domain, making web data truly useful in learning representations for actuality. (c) FoPro estimates noise-robust prototypes to pull instances nearby closer. Noisy samples are filtered by assessing their relation with prototypes. Only clean ones update prototypes in return. FoPro achieves better interpretability, discriminability, and generalization. Best viewed magnified. Unlike the methods above, our objective is to cost-efficiently mine web data for real-world applications. We handle both the noise and domain gap by resorting to a few human-labeled samples for guidance on whom to learn from and what to learn. In our setting, clean labeled examples are too scarce to train or fine-tune a deep model, and therefore alternative methods need to be developed in response. To this end, we propose a robust prototypical representation learning method with noisy web data and a few clean examples (see Fig.2). Motivated by the anchoring role of class prototypes (Li et al. 2020; Li, Xiong, and Hoi 2020), we introduce Few-shot guided Prototypes, termed as FoPro, to effectively deal with noise and domain gap. Technically, we project features of the penultimate layer of a classification model to low-dimensional embeddings. The critical problem is how to formulate a class-representative and domain-generalized prototype in the embedding space without being deviated by the dominating noises. Due to noise memorization (Arpit et al. 2017), simply averaging over instances with high prediction confidence does not promise a noise-robust estimation. Consequently, we first initialize each class prototype with realistic few shots as the cluster center. Secondly, intra-class distance is shortened by contrastive learning between web instances and prototypes. Then, high-quality web examples are involved in polishing prototypes to improve discriminability. Simultaneously, high similarity between prototypes and few shots is regularized to maximize the interpretability and generalizability of prototypes. Finally, we quantify the compatibility between instances and prototypes by the proposed relation module for sample selection and correction, which benefits prototype update in the next iteration. Specifically, the relation module learns a flexible and transferable metric to assess if a web image corresponds to its label. Besides, we set siamese encoders (He et al. 2020) and prototypes are only updated by the momentum encoder in a smooth and progressive way. Our contributions can be summarized as follows: * • We propose a new few-shot learning setting in WSL with abundant noisy web images and a few real-world images, which aims to improve the performance of WSL for real-world applications in a cost-efficient way. * • We present a new method under the setting above called FoPro, which simultaneously solves noise and data bias in an end-to-end manner. Experimental results show that our method can significantly improve the performance in real-world benchmark datasets. * • We propose a new relation module for label noise correction. It outperforms existing methods that use fixed metrics (_e.g._ , cosine distance) by evaluating instance-prototype similarity with a learnable metric. * • Extensive experiments on the fine-grained WebFG496 and the large-scale WebVision1k datasets confirm the superiority of FoPro over the state-of-the- art (SOTA) methods. Performance under the increasing $K$-shot settings demonstrates that FoPro utilizes few shots wisely to bridge the gap towards real-world applications. ## Related Work ### Webly Supervised Learning WSL aims to leverage vast but weakly-annotated web resources. Previous works utilize web images for tasks including classification (Bergamo and Torresani 2010; Wu et al. 2021; Yao et al. 2017, 2020), detection (Divvala, Farhadi, and Guestrin 2014; Shen et al. 2020), and segmentation (Shen et al. 2018; Jin, Ortiz Segovia, and Susstrunk 2017). Recently, noise cleaning methods such as self-contained confidence (SCC) (Yang et al. 2020) and momentum prototype (MoPro) (Li, Xiong, and Hoi 2020) are proposed to improve representation learning in WSL. SCC balances two supervision sources from web labels and predicted labels by introducing instance-wise confidence. MoPro targets model pretraining for several down- streaming tasks by combining self-supervised and webly-supervised techniques. Specifically, MoPro is closely related to ours since the contrast between instances and prototypes is used to learn discriminative features. Different from MoPro, we formulate a brand-new setting where a few samples labeled by experts are available. To assure that class prototypes are not misled by noise, an implicit constraint on distribution is achieved by enforcing high similarity between prototypes and few shots. Furthermore, we estimate the relation score between instances and prototypes to correct labels and discard out-of-distribution (OOD) samples. ### Learning from Noisy Labels Labels in human-annotated datasets can still be noisy due to lack of expert domain knowledge (Song et al. 2022). To prevent deep models from overfitting noisy labels, several studies have been conducted and can be categorized as: 1) robust architecture (_e.g._ , noise transition layer (Chen and Gupta 2015) and probability model (Xiao et al. 2015)); 2) regularization techniques (_e.g._ , label smoothing (Pereyra et al. 2017) and mix-up (Zhang et al. 2018)); 3) robust losses (_e.g._ , MAE (Ghosh, Kumar, and Sastry 2017) and GCE (Zhang and Sabuncu 2018)); 4) loss refinement (_e.g._ , reweighting (Wang, Liu, and Tao 2017) and bootstrapping (Reed et al. 2015)); 5) sample selection (_e.g._ , multi-model collaboration (Malach and Shalev-Shwartz 2017) and iterative strategies (Li, Socher, and Hoi 2019)). Hybrid approaches are designed in practice. For example, PeerLearn (Sun et al. 2021) develops a two- stage framework with peer models. Each model chooses clean samples independently and feeds them to train the other model. Different from the existing methods, we do not assume that samples with small losses or high confidence are clean. Instead, we maintain class prototypes and filter out noise by comparing instances and prototypes in a non-linear metric. Moreover, PeerLearn presumes that the percentage of noise is consistent across categories, which contradicts our observation. ### Contrastive Representation Learning Contrastive learning methods can be roughly categorized as: 1) context- instance contrast, where the relationship of local parts with respect to global context is learned (Kim et al. 2018); 2) instance-wise contrast, where similar image pairs are pulled closer with dissimilar pairs pushed farther (He et al. 2020; Chen et al. 2020a). Prototypical contrastive learning (PCL) (Li et al. 2020) encourages each image embedding to be adjacent to its assigned cluster prototype. However, their method is under an unsupervised setting where k-means clustering is used to generate prototypes. Our model is supervised by both numerous-yet-noisy web labels and limited-yet-clean few- shot labels. Besides, in PCL, batch embeddings in the current epoch are contrasted with the “outdated” prototypes in the previous epoch. FoPro keeps modifying prototypes smoothly all the time so that clean samples can be pinpointed by the latest features. Figure 3: Overview of FoPro. The encoder, classifier, and projector are trained to produce discriminative embeddings. Class prototypes are first initialized by few shots and then polished with clean samples for contrastive learning to regularize cluster distribution. Instance-wise contrastive loss is optimized simultaneously. The relation module learns a distance metric between an instance and its assigned class prototype. Finally, we adjust web labels for confidence-weighted hybrid target learning. ## Method In this section, a formal description of our few-shot WSL setting is presented, followed by the detailed explanation of FoPro. Fig. 3 illustrates the model architecture. ### Problem Statement Existing WSL setting aims to train a deep model $\mathcal{F}(\theta_{e};\theta_{c})$ with the optimal parameters of encoder $\theta_{e}^{*}$ and classifier $\theta_{c}^{*}$ from the web dataset $\mathit{D}^{w}=\\{(\mathbf{x}_{i}^{w},y_{i}^{w})\\}_{i=1}^{N^{w}}$. Here, $\mathbf{x}_{i}^{w}$ denotes an image, $y_{i}^{w}\in\\{1,...,C\\}$ is its class label. The number of classes and images are $C$ and $N^{w}$, respectively. Due to noise issues, $y_{i}^{w}$ might not equal to the ground- truth $y_{i}^{*}$. If $y_{i}^{w}\neq y_{i}^{*}$ and $y_{i}^{*}\in\\{1,...,C\\}$, $(\mathbf{x}_{i}^{w},y_{i}^{w})$ is viewed as an in-distribution (IND) sample with label-flipping error. If $y_{i}^{w}\neq y_{i}^{*}$ and $y_{i}^{*}\notin\\{1,...,C\\}$, then $(\mathbf{x}_{i}^{w},y_{i}^{w})$ is an out-of-distribution (OOD) sample. We propose a new WSL setting that additional real-world images are available with verified labels: $\mathit{D}^{t}=\\{(\mathbf{x}_{i}^{t},y_{i}^{t})\\}_{i=1}^{N^{t}}$ and $y_{i}^{t}=y_{i}^{*}$. The number of real-world samples is $N^{t}=K\cdot C$, where $K$ denotes $K$-shot per class. Our FoPro aims to achieve two goals with $\mathit{D}^{w}$: 1) to learn generalizable representations from high-quality examples; 2) to correct IND samples and discard OOD samples. ### Model Architecture The main components of FoPro include siamese encoder backbones, a classifier, a projector, a reconstructor, an auxiliary classifier, and a relation module. Our siamese encoder networks share the same architecture. Enlighted by MoCo (He et al. 2020), we update parameters of the first encoder $\theta_{e}^{1}$ by back-propagation and employ momentum update for the second encoder $\theta_{e}^{2}$: $\theta_{e}^{2}=m_{e}\theta_{e}^{2}+(1-m_{e})\theta_{e}^{1},$ (1) where $m_{e}$ is the momentum parameter. The plain and momentum encoders respectively extract features $\mathbf{v}_{i}^{\\{w;t\\}}$ and $\mathbf{v}_{i}^{\prime\\{w;t\\}}\in{\rm I\\!R}^{d_{e}}$ from inputs $\mathbf{x}_{i}^{\\{w;t\\}}$ and their augmented counterparts $\mathbf{x}_{i}^{\prime\\{w;t\\}}$. Note that our encoder is structure- agnostic, and its choices are up to specific tasks. All layers except the last fully connected (FC) layer are used. A classifier is trained to map features $\mathbf{v}_{i}^{\\{w;t\\}}$ to the predicted probabilities $\mathbf{p}_{i}^{\\{w;t\\}}$ over $C$ classes. It consists of one FC layer with softmax activation. A projector distills discriminative contents from features $\mathbf{v}_{i}^{\\{w;t\\}}$ for low-dimensional embeddings $\mathbf{z}_{i}^{\\{w;t\\}}\in{\rm I\\!R}^{d_{p}}$. It is composed of two FC layers and one ReLU layer. We follow (Chen et al. 2020a, b) to perform contrastive learning in the embedding space after projection. $\ell_{2}$-normalization is involved for unit-sphere constraint on $\mathbf{z}_{i}^{\\{w;t\\}}$. A reconstructor recovers $\tilde{\mathbf{v}}_{i}^{\\{w;t\\}}$ from $\mathbf{z}_{i}^{\\{w;t\\}}$, where $\tilde{\mathbf{v}}_{i}^{\\{w;t\\}}$ should be as close as possible to $\mathbf{v}_{i}^{\\{w;t\\}}$. Symmetric structure is adopted for the projector and reconstructor. An auxiliary classifier with one FC layer generates probabilities $\mathbf{q}_{i}^{t}$ over $C$ classes based on embeddings $\mathbf{z}_{i}^{t}$. Our relation module compares each pair of one instance embedding $\mathbf{z}_{i}^{\\{w;t\\}}$ and one class prototype $\mathbf{c}_{k}\in{\rm I\\!R}^{d_{p}},k\in\\{1,...,C\\}$ for distance measurement. Given the concatenated embeddings $[\mathbf{z}_{i}^{\\{w;t\\}},\mathbf{c}_{k}]$, it learns to score their closeness $r_{ik}\in{\rm I\\!R}$ by two FC layers with one ReLU layer. ### Training Strategy FoPro employs a four-stage training strategy. #### Stage 1: Preparation In this early stage, we warm up the system by learning common, regular patterns for the first $T_{1}$ epochs. As discovered by (Arpit et al. 2017), easy examples are reliably learned with simple patterns before the model overfits noise. We achieve this via training the encoder and classifier with cross-entropy loss. $\mathcal{L}_{i}^{cls}=-\log(\mathbf{p}_{i(y_{i})}^{\\{w;t\\}}).$ (2) Since $\mathbf{v}_{i}^{\\{w;t\\}}$ might contain redundant features that make outliers indistinguishable, we set a projector to only keep principal components. The previous method PCL stems from the analogy of autoencoder to PCA, and learns projection by minimizing the reconstruction loss for the projector and reconstructor. In preliminary experiments, however, we find that such optimization cannot give a good starting point for prototype initialization because $\mathbf{z}_{i}^{\\{w;t\\}}$ is not necessarily class- indicative. Therefore, an auxiliary classifier is applied on $\mathbf{z}_{i}^{t}$ to bring back its representation capacity. Only few shots are used here due to purity concerns. $\mathcal{L}_{i}^{prj}=\|\tilde{\mathbf{v}}_{i}^{\\{w;t\\}}-\mathbf{v}_{i}^{\\{w;t\\}}\|^{2}_{2}-\log(\mathbf{q}_{i(y_{i})}^{t}).$ (3) #### Stage 2: Incubation Clean few shots play an anchoring role in “territory” enclosure in the embedding space. Given extracted embeddings from the momentum encoder, we initialize prototypes by averaging few shots in each class. $\hat{\mathbf{c}}_{k}=\frac{1}{K}\sum_{y_{i}=k}{\mathbf{z}_{i}^{t}},\mathbf{c}_{k}=\frac{\hat{\mathbf{c}}_{k}}{\|\hat{\mathbf{c}}_{k}\|_{2}}.$ (4) In this stage, we begin to pull instances within one class towards the prototype for $T_{2}$ epochs. Besides, instance-level discrimination is encouraged by contrastive losses (Chen et al. 2020a) to improve separation across classes. $\mathcal{L}_{i}^{pro}=-\log\frac{\exp((\mathbf{z}_{i}^{w;t}\cdot\mathbf{c}_{y_{i}}-\delta^{w;t})/\phi_{y_{i}})}{\sum_{k=1}^{C}\exp((\mathbf{z}_{i}^{w;t}\cdot\mathbf{c}_{k}-\delta^{w;t})/\phi_{k})},$ (5) $\mathcal{L}_{i}^{ins}=-\log\frac{\exp(\mathbf{z}_{i}^{w;t}\cdot\mathbf{z}_{i}^{\prime w;t}/\tau)}{\sum_{j=1}^{Q}\exp(\mathbf{z}_{i}^{w;t}\cdot\mathbf{z}_{j}^{\prime w;t}/\tau)},$ (6) where $\delta^{w;t}$ refers to the margin coefficient, and $Q$ is the length of the memory bank for storing embeddings of visited instances. Temperature coefficients can be fixed as $\tau$ or class-dependent as $\phi_{k}$. We put constraints on learning representations with a high margin so that clean few shots gather around prototypes tightly, ensuring better justification and interpretability. Furthermore, to regularize the distribution of each class cluster, adjustable temperature coefficients (Li et al. 2020) are estimated based on concentration. $\phi_{k}=\frac{\sum_{y_{i}=k}\|\mathbf{z}_{i}^{w;t}-\mathbf{c}_{k}\|_{2}}{N_{k}^{w;t}\log(N_{k}^{w;t}+\alpha)},$ (7) where $N_{k}^{w;t}$ denotes the number of web and few-shot instances of class $k$, and $\alpha$ is a smoother. Embeddings of large, loose clusters will be penalized more to approach their prototypes, while those of small, tight clusters will be relaxed. #### Stage 3: Illumination With parameters of the encoder and projector fixed, the relation module learns to score the compatibility between one instance and each prototype. It sheds light on whether the given label of a web image is correct. We select clean samples $\mathit{D}^{r}$ for training the relation module. $\mathit{D}^{r}=\mathit{D}^{t}\cup\\{(\mathbf{x}_{i}^{w},y_{i}^{w})|\sum_{j=1}^{C}|(\mathbf{z}_{i}^{w}-\mathbf{c}_{y_{i}})\cdot\mathbf{c}_{j}|\leq\sigma\\},$ (8) where $\sigma$ is a threshold between 0 and 1. Such criterion comprehensively considers both the cosine distance between instance and prototypes, and the distance among prototypes. Then, the relation module is trained for $T_{3}$ epochs by: $\mathcal{L}_{i}^{rel}=-\log\frac{\exp(r_{iy_{i}})}{\sum_{k=1}^{C}\exp(r_{ik})}.$ (9) #### Stage 4: Verification Armed with “pretrained” model, we start label correction, OOD removal, prototype update, and continue noise-robust learning for $T_{4}$ epochs. Three sources of prior knowledge are incorporated for cleaning: 1) self-prediction; 2) instance-prototype similarity; 3) relation score. Rules for adjusting labels are detailed below: $\mathbf{s}_{i}^{w}=\beta\mathbf{p}_{i}^{w}+(1-\beta)[\mathbf{c}_{1},...,\mathbf{c}_{C}]^{T}\cdot\mathbf{z}_{i}^{w}$ (10) $\hat{y}_{i}^{w}=\left\\{\begin{array}[]{lcl}y_{i}^{w}&&{\textrm{if}\ r_{iy_{i}}>\gamma,}\\\ \arg\max_{k}\mathbf{s}_{i(k)}^{w}&&{\textrm{else if}\max_{k}\mathbf{s}_{i(k)}^{w}>\gamma,}\\\ y_{i}^{w}&&{\textrm{else if}\ \mathbf{s}_{i(y_{i})}^{w}>1/C,}\\\ \textrm{Null}\ (OOD)&&{\textrm{otherwise,}}\end{array}\right.$ (11) where $\gamma$ is a threshold between 0 and 1. Since fine-grained categories share highly similar visual patterns, the relation module is only used for positive verification of the initial web label. Besides, we introduce an alternative confidence measure from self-prediction and similarity for label reassignment. When the first two conditions are not satisfied, an image will be kept as hard example if its confidence is above average. Otherwise, it is discarded as OOD. Note that the basic control flow above is inspired by MoPro. We further improve it with the proposed relation module (Eq.11 cond. 1) to better evaluate the compatibility between instances and class prototypes, and thereafter enable accurate label-flipping-error correction and OOD removal without ignoring hard examples by mistake (Eq.11 conds. 2–4). After label adjustment, we exploit the predicted probabilities as pseudo- labels for self-training (Tanaka et al. 2018; Han, Luo, and Wang 2019). Such soft targets can be viewed as a regularizer on the classifier like label smoothing (Müller, Kornblith, and Hinton 2019) and self-knowledge distillation (Hinton et al. 2015). Instead of using a fixed coefficient, we follow (Yang et al. 2020) to leverage confidence on the corrected label for weighting soft and hard targets. $\displaystyle\mathcal{L}_{i}^{cls}=$ $\displaystyle-\log(\mathbf{p}_{i(y_{i})}^{t})-\mathbf{s}_{i(\hat{y}_{i})}^{w}\log(\mathbf{p}_{i(\hat{y}_{i})}^{w})$ (12) $\displaystyle-(1-\mathbf{s}_{i(\hat{y}_{i})}^{w})\sum_{k=1}^{C}\mathbf{p}_{i(k)}^{w}\log\mathbf{p}_{i(k)}^{w}.$ With the label-flipping errors and OOD reduced, class prototypes are updated by embeddings of the remaining clean examples from the momentum encoder. Exponential moving average (Li, Xiong, and Hoi 2020) is used for two reasons: 1) initialization by few shots remains to exert a profound anchoring effect. 2) smoothed transition is achieved to stabilize contrastive learning. For class $k$, web images with $\hat{y}_{i}^{w}=k$ and few shot images with $y_{i}^{t}=k$ are involved: $\hat{\mathbf{c}}_{k}=m_{p}{\mathbf{c}}_{k}+(1-m_{p})\mathbf{z}_{i}^{w;t},\mathbf{c}_{k}=\frac{\hat{\mathbf{c}}_{k}}{\|\hat{\mathbf{c}}_{k}\|_{2}}.$ (13) Note that reliable samples, which are selected by Eq. 11 per batch, also participate in training the relation module. The criterion by Eq. 8 is only used in stage 3. ## Experiments We train FoPro on web datasets and evaluate it on real-world testing sets. FoPro boosts $K$-shot performance and reaches the SOTA. Ablation study validates the relation module. ### Datasets Web Dataset | # Img. | # Cls. | Real-World ---|---|---|--- Web- FG496 | Bird | 18k | 200 | CUB200-2011 Air | 13k | 100 | FGVC-Aircraft Car | 21k | 196 | Stanford Car Web- Vision1k | | Web- --- Vision1k 2.44M | 1k | ImageNet1k Google500 | 0.61M | 500 | ImageNet500 Table 1: Statistics of web datasets. ##### WebFG496 (Sun et al. 2021) contains three fine-grained datasets sourced from Bing. The testing sets of CUB200-2011 (Wah et al. 2011), FGVC-Aircraft (Maji et al. 2013), and Stanford Car (Krause et al. 2013) are used. ##### WebVision1k (Li et al. 2017) is collected from Google and Flickr. The validation set of ImageNet1k (Deng et al. 2009) is used. Besides, we also use Google500 (Yang et al. 2020) where 500 out of 1k categories are randomly sampled with images only from Google (see Table 1). We randomly sample $K$ shots per class from the training sets of real-world datasets. Classification accuracy (%) is adopted as the evaluation metric for all experiments. ### Implementation Details ##### WebFG496 The B-CNN (Lin, RoyChowdhury, and Maji 2015) (VGG-16 (Simonyan and Zisserman 2014)) is used as encoder. We refer to (Sun et al. 2021) for the training settings: optimizer is Adam with weight decay of $1\times 10^{-8}$; batch size is 64; the learning rate is $1\times 10^{-4}$ and decays to 0 by cosine schedule; a warm-up policy increases the learning rate linearly for 5 epochs with the frozen encoders. ##### WebVision1k The ResNet-50 (R50) (He et al. 2016) is used as encoder. We refer to (Yang et al. 2020) for the training settings: batch size is 256; optimizer is SGD with the momentum of 0.9 and weight decay of $1\times 10^{-4}$; the learning rate is 0.01 and decays to 0 by cosine schedule. We refer to MoPro to set $m_{e}=0.999$, $m_{p}=0.999$, $d_{p}=128$, and $Q=8192$. In view of the dataset scale, we set $T_{1}=20$, $T_{2}=5$, $T_{3}=20$, $T_{4}=175$ for WebFG496 and set $T_{1}=15$, $T_{2}=5$, $T_{3}=10$, $T_{4}=30$ for WebVision1k/Google500. Preliminary experiments on WebFG496 show that $\gamma=0.6$ and $\beta=0.5$ work better than $\gamma=0.2$ and $\beta=0,0.25,0.75,1$. A lower $\gamma$ means a more relaxed criterion on clean sample selection, which might bring in noise and cause performance drop. The balanced combination of self-prediction and similarity terms performs more robust to noise than the biased cases. Other hyper-parameters are empirically set as: $\delta^{w}=0$, $\delta^{t}=0.5$, $\tau=0.1$, $\alpha=10$, $\sigma=20$. Their optimal values require meticulous fine-tuning, which is beyond consideration of the present study. Data augmentation includes random cropping and horizontal flipping. Strong augmentation on the inputs to the momentum encoder (He et al. 2020) additionally uses color jittering and blurring. Since birds might only differ in color, random rotation in 45 degrees is used instead. Experiments are conducted on a CentOS 7 workstation with an Intel 8255C CPU, 377 GB Mem, and 8 NVIDIA V100 GPUs. ### Results Method | Back- | WebFG496 ---|---|--- bone | Bird | Air | Car | Avg. Vanilla | R50 | 64.43 | 60.79 | 60.64 | 61.95 MoPro† | R50 | 71.16 | 76.85 | 79.68 | 75.90 SCC† | R50-D | 61.10 | 74.92 | 83.49 | 73.17 Vanilla | B-CNN | 66.56 | 64.33 | 67.42 | 66.10 Decouple | B-CNN | 70.56 | 75.97 | 75.00 | 73.84 CoTeach | B-CNN | 73.85 | 72.76 | 73.10 | 73.24 PeerLearn | B-CNN | 76.48 | 74.38 | 78.52 | 76.46 PeerLearn† | B-CNN | 76.57 | 74.35 | 78.26 | 76.39 FoPro($K$=0) | B-CNN | 77.79 | 79.37 | 86.99 | 81.38 FoPro($K$=1) | B-CNN | 78.07 | 79.87 | 88.01 | 82.03 FoPro($K$=16) | B-CNN | 85.54 | 86.40 | 91.51 | 87.81 * $\dagger$ Results are reproduced by ourselves with the official codes. Table 2: The SOTA results on fine-grained datasets. ##### Baselines Our FoPro is compared with vanilla backbones and the SOTA methods including SCC, MoPro, Decouple (Malach and Shalev-Shwartz 2017), CoTeach (Han et al. 2018), PeerLearn, MentorNet (Jiang et al. 2018), CurriculumNet (Guo et al. 2018), and CleanNet (Lee et al. 2018). Results of the SOTA methods that are trained and evaluated on the same datasets are directly quoted here. We also reproduce three closely-related methods of SCC, MoPro, and PeerLearn under $K$-shot settings with the officially released codes. Their default hyper- parameters are employed if the same web datasets are engaged. Otherwise, they are set the same as ours. Additionally, we modify the proposed method only to exhibit its applicability for 0-shot without specific design. In that case, web images with predicted probability of the target class over $\gamma$ are used to train the auxiliary classifier. In view of the dataset scale, prototypes are initialized by randomly sampled 16 and 50 web images per class from WebFG496 and WebVision1k/Google500, respectively. Method† | Back- | ImageNet1k | ImageNet500 ---|---|---|--- bone | Top 1 | Top 5 | Top 1 | Top 5 MentorNet | Inception ResNetV2 | 64.20 | 84.80 | – | – Curriculum- Net | Inception V2 | 64.80 | 83.40 | – | – Vanilla | R50-D | 67.23 | 84.09 | – | – SCC | R50-D | 67.93 | 84.77 | 68.84 | 84.62 SCC† | R50-D | 67.57 | 85.74 | 64.40 | 81.56 Vanilla | R50 | 65.70 | 85.10 | 61.54 | 78.89 CoTeach | R50 | – | – | 62.18 | 80.98 CleanNet | R50 | 63.42 | 84.59 | – | – MoPro | R50 | 67.80 | 87.00 | – | – MoPro† | R50 | 66.05 | 85.66 | 58.68 | 78.39 PeerLearn† | R50 | 52.57 | 73.35 | 42.04 | 61.71 FoPro($K$=0) | R50 | 67.03 | 85.57 | 68.59 | 86.03 FoPro($K$=1) | R50 | 67.55 | 86.31 | 69.11 | 86.19 FoPro($K$=16) | R50 | 68.83 | 87.83 | 72.02 | 89.38 * $\dagger$ Results are reproduced by ourselves with the official codes. Table 3: The SOTA results on large-scale datasets. Table 2 confirms the superiority of the proposed method on fine-grained datasets even under 0-shot. FoPro boosts the accuracy of vanilla backbones more significantly than the SOTA methods with respect to their backbones. FoPro reaches the optimal performance on large-scale datasets with $K$=16 (see Table 3). The vanilla R50-D (He et al. 2019) performs better than R50. Although FoPro is preceded by SCC and MoPro at first (0-shot), it rises steadily after exploiting a few real-world examples efficiently. Figure 4: The SOTA results under $K$-shot settings. Figure 5: One-shot real- world examples and the web images sorted by their distance to class prototypes. Best viewed magnified. Under the degeneration circumstance ($K$=0), FoPro outperforms the SOTA methods on WebFG496. The reason why FoPro ($K$=0) degrades slightly on ImageNet1k/500 lies in the high percentage of noises in WebVision1k/Google500. In that case, prototypes ($K$=0) are initialized and polished solely by noisy web examples without intervention from clean shots, which may not be class- representative. With few real-world examples ($K>0$) involved, FoPro regains its advantage over the SOTA methods. $K$ | WebFG496 Avg. | ImageNet1k | ImageNet500 ---|---|---|--- Top 1 | Gap | Top 1 | Gap | Top 1 | Gap 0 | 81.38 | – | 67.03 | 5.57 | 68.59 | 3.85 1 | +0.65 | – | +0.52 | 5.22 | +0.52 | 3.63 2 | +0.85 | – | +0.67 | 5.20 | +1.35 | 3.29 4 | +2.17 | – | +0.32 | 4.60 | +1.50 | 2.91 8 | +4.10 | – | +0.85 | 4.64 | +2.06 | 2.90 16 | +6.43 | – | +1.80 | 3.91 | +3.43 | 2.19 16 | 87.81 | – | 68.83 | – | 72.02 | – Ref. | 87.16† | – | 76.15‡ | – | 76.22‡ | – * $\dagger$ Official results of the B-CNN trained on FGVC-Aircraft, CUB200-2011, and Stanford Car are averaged. * ‡ Official results of the R50 trained on ImageNet1k by PyTorch are quoted respectively for 500 and 1k classes. Table 4: FoPro gains of $K$-shot over 0-shot. Gap refers to the differences between web and real-world testing results. ##### Effect of Few-Shots We explore the potential of FoPro by varying the number of real-world examples per class from 1 to 16. As shown in Fig. 4, FoPro achieves consistent performance growth with $K$ on fine-grained datasets. It surpasses SCC and PeerLearn by a large margin. On ImageNet1k, the abnormal case of $K$=4 is mainly due to sampling jittering. Since ImageNet contains many unreal images, it could not eliminate the possibility of sampling atypical images of certain classes. However, as $K$ increases, FoPro starts to take the lead. We believe more few shots directly refine the estimated prototypes for better representation. Clean samples can be more appropriately selected to promote discriminative feature learning. With amendment on cluster formation, FoPro also enjoys a higher level of interpretability in class centers with competitive performance. Table 4 reports the performance gap between WebVision1k/ImageNet1k and Google500/ImageNet500 when the testing sets of web domain are available. In line with Fig. 2(b), the reduced gap reflects that we bridge the noisy web domain and real-world domain with limited $K$ shots. FoPro approaches reference benchmarks that are trained on real-world datasets, corroborating its practical value that much labor of data collection and annotation can be saved. ##### Effect of Relation Module In Table 5, we study the effect of relation module for clean example selection. We remove FC layers and directly compare an instance and each prototype using cosine similarity. Results confirm that the proposed relation module discovers clean examples more precisely than the pre-defined similarity metric. By using a non-linear metric, we do not assume that element-wise comparison could solely separate matching or mismatching pairs. Besides, such a learnable metric is not sensitive to input variation and behaves better on noisy samples. $K$=1 | WebFG496 Avg. | ImageNet1k | ImageNet500 ---|---|---|--- w/o RM | 81.59 | 65.22 | 64.69 w RM | 82.03 | 67.55 | 69.11 Table 5: Ablation Study on the Relation Module (RM). ##### Visualization In Fig. 2(a), we visualize the low-dimensional embeddings with t-SNE for the randomly chosen 10 categories in WebVision as a demonstration. For convenience, all 16 real-world examples in each class are averaged and displayed as one few-shot example. Differences in the cluster distribution (from $K$=1 to $K$=16) are highlighted to show that: 1) the distance between each prototype and the few-shot example becomes shorter; 2) the density of class clusters is improved. From Fig. 5, we conclude the following insights: 1) Web images close to the estimated prototypes are clean and similar to real- world photos with limited post-processing. Our FoPro learns to sort out noise in web data for robust representation learning. 2) The proposed method generalizes across various domains such as product close-up, computer graphics, and screenshots. 3) Intra-class diversity (_e.g._ , wing postures of the sooty albatross), uncaptured salient parts (_e.g._ , the yellowish patch on the back head of the bobolink), and editing of tone curve (_e.g._ , colored body of the painted bunting and yellow-breasted chat) are the possible reasons why hard examples of 1-shot and clean web images still locate away from prototypes. ## Conclusion This paper introduces a new setting for webly-supervised learning, which optimizes the learning system with a large quantity of noisy web images and a few real-world images. Under this setting, we propose a few-shot guided prototypical representation learning method called FoPro, which simultaneously tackles noise and dataset bias in a cost-efficient manner. It is characterized by the guidance from a few real-world domain images for learning noise-robust and generalizable representations from web data. Experimental results demonstrate that our method can effectively utilize few-shot images to improve the performance of WSL on real-world benchmarks. Future work includes investigation of side information from web datasets (_e.g._ , captions, website titles, and user comments) and application extension to weakly- supervised object detection and segmentation. ## References * Arpit et al. (2017) Arpit, D.; Jastrzebski, S.; Ballas, N.; Krueger, D.; Bengio, E.; Kanwal, M. S.; Maharaj, T.; Fischer, A.; Courville, A.; Bengio, Y.; et al. 2017. A closer look at memorization in deep networks. In _International Conference on Machine Learning_ , 233–242. PMLR. * Bergamo and Torresani (2010) Bergamo, A.; and Torresani, L. 2010. Exploiting weakly-labeled web images to improve object classification: a domain adaptation approach. _Advances in Neural Information Processing Systems_ , 23. * Chen et al. (2020a) Chen, T.; Kornblith, S.; Norouzi, M.; and Hinton, G. 2020a. A simple framework for contrastive learning of visual representations. In _International Conference on Machine Learning_ , 1597–1607. PMLR. * Chen et al. (2020b) Chen, X.; Fan, H.; Girshick, R.; and He, K. 2020b. Improved baselines with momentum contrastive learning. _arXiv preprint arXiv:2003.04297_. * Chen and Gupta (2015) Chen, X.; and Gupta, A. 2015. Webly supervised learning of convolutional networks. In _Proceedings of the IEEE international conference on computer vision_ , 1431–1439. * Cheng et al. (2020) Cheng, L.; Zhou, X.; Zhao, L.; Li, D.; Shang, H.; Zheng, Y.; Pan, P.; and Xu, Y. 2020. Weakly supervised learning with side information for noisy labeled images. In _European Conference on Computer Vision_ , 306–321. Springer. * Deng et al. (2009) Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; and Fei-Fei, L. 2009. ImageNet: A large-scale hierarchical image database. In _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , 248–255. IEEE. * Divvala, Farhadi, and Guestrin (2014) Divvala, S. K.; Farhadi, A.; and Guestrin, C. 2014. Learning everything about anything: Webly-supervised visual concept learning. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 3270–3277. * Ghosh, Kumar, and Sastry (2017) Ghosh, A.; Kumar, H.; and Sastry, P. S. 2017. Robust loss functions under label noise for deep neural networks. In _Proceedings of the AAAI Conference on Artificial Intelligence_ , volume 31. * Guo et al. (2018) Guo, S.; Huang, W.; Zhang, H.; Zhuang, C.; Dong, D.; Scott, M. R.; and Huang, D. 2018. CurriculumNet: Weakly supervised learning from large-scale web images. In _European Conference on Computer Vision_ , 135–150. * Han et al. (2018) Han, B.; Yao, Q.; Yu, X.; Niu, G.; Xu, M.; Hu, W.; Tsang, I.; and Sugiyama, M. 2018\. Co-teaching: Robust Training of Deep Neural Networks with Extremely Noisy Labels. _Advances in Neural Information Processing Systems_ , 31. * Han, Luo, and Wang (2019) Han, J.; Luo, P.; and Wang, X. 2019. Deep self-learning from noisy labels. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 5138–5147. * He et al. (2020) He, K.; Fan, H.; Wu, Y.; Xie, S.; and Girshick, R. 2020. Momentum contrast for unsupervised visual representation learning. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 9729–9738. * He et al. (2016) He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016. Deep residual learning for image recognition. In _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , 770–778. * He et al. (2019) He, T.; Zhang, Z.; Zhang, H.; Zhang, Z.; Xie, J.; and Li, M. 2019. Bag of tricks for image classification with convolutional neural networks. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 558–567. * Hinton et al. (2015) Hinton, G.; Vinyals, O.; Dean, J.; et al. 2015. Distilling the knowledge in a neural network. _arXiv preprint arXiv:1503.02531_ , 2(7). * Jiang et al. (2018) Jiang, L.; Zhou, Z.; Leung, T.; Li, L.-J.; and Fei-Fei, L. 2018. MentorNet: Learning data-driven curriculum for very deep neural networks on corrupted labels. In _International Conference on Machine Learning_ , 2304–2313. PMLR. * Jin, Ortiz Segovia, and Susstrunk (2017) Jin, B.; Ortiz Segovia, M. V.; and Susstrunk, S. 2017. Webly supervised semantic segmentation. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 3626–3635. * Kaur, Sikka, and Divakaran (2017) Kaur, P.; Sikka, K.; and Divakaran, A. 2017. Combining weakly and webly supervised learning for classifying food images. _arXiv preprint arXiv:1712.08730_. * Kim et al. (2018) Kim, D.; Cho, D.; Yoo, D.; and Kweon, I. S. 2018. Learning image representations by completing damaged jigsaw puzzles. In _IEEE Winter Conference on Applications of Computer Vision_ , 793–802. IEEE. * Kolesnikov et al. (2019) Kolesnikov, A.; Beyer, L.; Zhai, X.; Puigcerver, J.; Yung, J.; Gelly, S.; and Houlsby, N. 2019. Large scale learning of general visual representations for transfer. _arXiv preprint arXiv:1912.11370_ , 2(8). * Krause et al. (2016) Krause, J.; Sapp, B.; Howard, A.; Zhou, H.; Toshev, A.; Duerig, T.; Philbin, J.; and Fei-Fei, L. 2016. The unreasonable effectiveness of noisy data for fine-grained recognition. In _European Conference on Computer Vision_ , 301–320. Springer. * Krause et al. (2013) Krause, J.; Stark, M.; Deng, J.; and Fei-Fei, L. 2013. 3d object representations for fine-grained categorization. In _Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops_ , 554–561. * Lee et al. (2018) Lee, K.-H.; He, X.; Zhang, L.; and Yang, L. 2018. CleanNet: Transfer Learning for Scalable Image Classifier Training with Label Noise. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 5447–5456. * Li, Socher, and Hoi (2019) Li, J.; Socher, R.; and Hoi, S. C. 2019. DivideMix: Learning with Noisy Labels as Semi-supervised Learning. In _International Conference on Learning Representations_. * Li, Xiong, and Hoi (2020) Li, J.; Xiong, C.; and Hoi, S. 2020. MoPro: Webly Supervised Learning with Momentum Prototypes. In _International Conference on Learning Representations_. * Li et al. (2020) Li, J.; Zhou, P.; Xiong, C.; and Hoi, S. 2020. Prototypical Contrastive Learning of Unsupervised Representations. In _International Conference on Learning Representations_. * Li et al. (2017) Li, W.; Wang, L.; Li, W.; Agustsson, E.; and Van Gool, L. 2017. WebVision database: Visual learning and understanding from web data. _arXiv preprint arXiv:1708.02862_. * Lin, RoyChowdhury, and Maji (2015) Lin, T.-Y.; RoyChowdhury, A.; and Maji, S. 2015. Bilinear CNN models for fine-grained visual recognition. In _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , 1449–1457. * Liu et al. (2021) Liu, H.; Zhang, H.; Lu, J.; and Tang, Z. 2021. Exploiting Web Images for Fine-Grained Visual Recognition via Dynamic Loss Correction and Global Sample Selection. _IEEE Transactions on Multimedia_ , 24: 1105–1115. * Maji et al. (2013) Maji, S.; Rahtu, E.; Kannala, J.; Blaschko, M.; and Vedaldi, A. 2013. Fine-grained visual classification of aircraft. _arXiv preprint arXiv:1306.5151_. * Malach and Shalev-Shwartz (2017) Malach, E.; and Shalev-Shwartz, S. 2017. Decoupling” when to update” from” how to update”. _Advances in Neural Information Processing Systems_ , 30. * Müller, Kornblith, and Hinton (2019) Müller, R.; Kornblith, S.; and Hinton, G. E. 2019. When does label smoothing help? _Advances in Neural Information Processing Systems_ , 32. * Niu, Li, and Xu (2015) Niu, L.; Li, W.; and Xu, D. 2015. Visual recognition by learning from web data: A weakly supervised domain generalization approach. In _Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition_ , 2774–2783. * Pereyra et al. (2017) Pereyra, G.; Tucker, G.; Chorowski, J.; Kaiser, Ł.; and Hinton, G. 2017. Regularizing neural networks by penalizing confident output distributions. _arXiv preprint arXiv:1701.06548_. * Reed et al. (2015) Reed, S. E.; Lee, H.; Anguelov, D.; Szegedy, C.; Erhan, D.; and Rabinovich, A. 2015\. Training Deep Neural Networks on Noisy Labels with Bootstrapping. In _International Conference on Learning Representations (Workshop)_. * Shen et al. (2018) Shen, T.; Lin, G.; Shen, C.; and Reid, I. 2018. Bootstrapping the performance of webly supervised semantic segmentation. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 1363–1371. * Shen et al. (2020) Shen, Y.; Ji, R.; Chen, Z.; Hong, X.; Zheng, F.; Liu, J.; Xu, M.; and Tian, Q. 2020\. Noise-aware fully webly supervised object detection. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 11326–11335. * Simonyan and Zisserman (2014) Simonyan, K.; and Zisserman, A. 2014. Very deep convolutional networks for large-scale image recognition. _arXiv preprint arXiv:1409.1556_. * Song et al. (2022) Song, H.; Kim, M.; Park, D.; Shin, Y.; and Lee, J.-G. 2022. Learning from noisy labels with deep neural networks: A survey. _IEEE Transactions on Neural Networks and Learning Systems_. * Sun et al. (2017) Sun, C.; Shrivastava, A.; Singh, S.; and Gupta, A. 2017. Revisiting unreasonable effectiveness of data in deep learning era. In _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , 843–852. * Sun et al. (2021) Sun, Z.; Yao, Y.; Wei, X.-S.; Zhang, Y.; Shen, F.; Wu, J.; Zhang, J.; and Shen, H. T. 2021. Webly supervised fine-grained recognition: Benchmark datasets and an approach. In _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , 10602–10611. * Tanaka et al. (2018) Tanaka, D.; Ikami, D.; Yamasaki, T.; and Aizawa, K. 2018. Joint optimization framework for learning with noisy labels. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 5552–5560. * Tu et al. (2020) Tu, Y.; Niu, L.; Chen, J.; Cheng, D.; and Zhang, L. 2020. Learning from web data with self-organizing memory module. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 12846–12855. * Van der Maaten and Hinton (2008) Van der Maaten, L.; and Hinton, G. 2008. Visualizing data using t-SNE. _Journal of Machine Learning Research_ , 9(11). * Wah et al. (2011) Wah, C.; Branson, S.; Welinder, P.; Perona, P.; and Belongie, S. 2011. The caltech-ucsd birds-200-2011 dataset. * Wang, Liu, and Tao (2017) Wang, R.; Liu, T.; and Tao, D. 2017. Multiclass learning with partially corrupted labels. _IEEE Transactions on Neural Networks and Learning Systems_ , 29(6): 2568–2580. * Wu et al. (2021) Wu, Z.-F.; Wei, T.; Jiang, J.; Mao, C.; Tang, M.; and Li, Y.-F. 2021. NGC: a unified framework for learning with open-world noisy data. In _Proceedings of the IEEE/CVF International Conference on Computer Vision_ , 62–71. * Xiao et al. (2015) Xiao, T.; Xia, T.; Yang, Y.; Huang, C.; and Wang, X. 2015. Learning from massive noisy labeled data for image classification. In _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 2691–2699. * Xu et al. (2016) Xu, Z.; Huang, S.; Zhang, Y.; and Tao, D. 2016. Webly-supervised fine-grained visual categorization via deep domain adaptation. _IEEE Transactions on Pattern Analysis and Machine Intelligence_ , 40(5): 1100–1113. * Yang et al. (2020) Yang, J.; Feng, L.; Chen, W.; Yan, X.; Zheng, H.; Luo, P.; and Zhang, W. 2020. Webly supervised image classification with self-contained confidence. In _European Conference on Computer Vision_ , 779–795. Springer. * Yao et al. (2020) Yao, Y.; Hua, X.; Gao, G.; Sun, Z.; Li, Z.; and Zhang, J. 2020. Bridging the web data and fine-grained visual recognition via alleviating label noise and domain mismatch. In _Proceedings of the ACM International Conference on Multimedia_ , 1735–1744. * Yao et al. (2017) Yao, Y.; Zhang, J.; Shen, F.; Hua, X.; Xu, J.; and Tang, Z. 2017. Exploiting web images for dataset construction: A domain robust approach. _IEEE Transactions on Multimedia_ , 19(8): 1771–1784. * Zhang et al. (2021) Zhang, C.; Bengio, S.; Hardt, M.; Recht, B.; and Vinyals, O. 2021. Understanding deep learning (still) requires rethinking generalization. _Communications of the ACM_ , 64(3): 107–115. * Zhang et al. (2020) Zhang, C.; Yao, Y.; Liu, H.; Xie, G.-S.; Shu, X.; Zhou, T.; Zhang, Z.; Shen, F.; and Tang, Z. 2020. Web-supervised network with softly update-drop training for fine-grained visual classification. In _Proceedings of the AAAI Conference on Artificial Intelligence_ , volume 34, 12781–12788. * Zhang et al. (2018) Zhang, H.; Cisse, M.; Dauphin, Y. N.; and Lopez-Paz, D. 2018. MixUp: Beyond Empirical Risk Minimization. In _International Conference on Learning Representations_. * Zhang and Sabuncu (2018) Zhang, Z.; and Sabuncu, M. 2018. Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels. _Advances in Neural Information Processing Systems_ , 31. * Zhou et al. (2020) Zhou, X.; Pan, P.; Zheng, Y.; Xu, Y.; and Jin, R. 2020. Large scale long-tailed product recognition system at alibaba. In _Proceedings of ACM International Conference on Information & Knowledge Management_, 3353–3356.
# Beyond Incompatibility: Trade-offs between Mutually Exclusive Fairness Criteria in Machine Learning and Law Meike Zehlike Zalando ResearchBerlinGermany<EMAIL_ADDRESS>, Alex Loosley Zalando SEBerlinGermany<EMAIL_ADDRESS>, Håkan Jonsson Zalando ResearchBerlinGermany<EMAIL_ADDRESS>, Emil Wiedemann Institute of Applied Analysis, Ulm UniversityHelmholtzstr. 1889081 UlmGermany <EMAIL_ADDRESS>and Philipp Hacker European New School of Digital Studies, European University ViadrinaFrankfurt (Oder)Germany hacker@europa- uni.de ###### Abstract. Trustworthy AI is becoming ever more important in both machine learning and legal domains. One important consequence is that decision makers must seek to guarantee a ‘fair’, i.e., non-discriminatory, algorithmic decision procedure. However, there are several competing notions of algorithmic fairness that have been shown to be mutually incompatible under realistic factual assumptions. This concerns, for example, the widely used fairness measures of ‘calibration within groups’ and ‘balance for the positive/negative class’. In this paper, we present a novel algorithm (FAir Interpolation Method: FAIM) for continuously interpolating between these three fairness criteria. Thus, an initially unfair prediction can be remedied to, at least partially, meet a desired, weighted combination of the respective fairness conditions. We demonstrate the effectiveness of our algorithm when applied to synthetic data, the COMPAS data set, and a new, real-world data set from the e-commerce sector. Finally, we discuss to what extent FAIM can be harnessed to comply with conflicting legal obligations. The analysis suggests that it may operationalize duties in traditional legal fields, such as credit scoring and criminal justice proceedings, but also for the latest AI regulations put forth in the EU, like the recently enacted Digital Markets Act. ## 1\. Introduction As machine learning (ML) progressively permeates all sectors of society, responsible and trustworthy AI development is more crucial than ever ((Dignum, 2019; Grodzinsky et al., 2011; Thiebes et al., 2021)). Increasingly, this debate spills over from the ethics and machine learning communities to the legal domain ((Laux et al., 2022)), where different regulatory frameworks for ensuring trustworthy AI are emerging. In this context, discrimination against protected groups in decisions facilitated by ML models remains a key challenge ((Wachter, 2022)). Non-discrimination legislation applying to biased AI systems does exist in the US and the EU, but loopholes and enforcement problems remain ((Barocas and Selbst, 2016; Hacker, 2018; Zuiderveen Borgesius, 2020; Wachter et al., 2021b; Wachter, 2022)). Hence, the AI Act, proposed by the EU Commission in 2021 and set to be enacted soon in the EU, adds further duties for ML developers such as the duty to evaluate of training data for potential biases (Article 10(2) AI Act), the duty to use representative training data (Article 10(3) AI Act) and the duty to prevent biased feedback loops (Article 15(3) AI Act) ((Veale and Borgesius, 2021; Ebers et al., 2021; Hacker, 2021)). This is backed up by a framework for AI liability proposed in September 2022 by the EU Commission ((Hacker, 2022a)). Moreover, the recently enacted EU Digital Markets Act (DMA111Regulation (EU) 2022/1925 of the European Parliament and of the Council of 14 September 2022 on contestable and fair markets in the digital sector, OJ 2022, L265/1.) now compels large online platforms to engage in ‘fair and non-discriminatory’ ranking (Article 6(5) DMA). The laudable idea behind such approaches is to force developers to consider, and potentially eliminate, biases during the design of ML models. Such non- discriminatory models are called ‘fair’ in the ML community ((Mitchell et al., 2021; Dwork et al., 2012)). However, foundational papers at the intersection of machine learning and the law have shown that, under realistic assumptions, several desirable fairness criteria are mutually incompatibility ((Kleinberg et al., 2016)). This poses a significant challenge as the law, in general, prohibits discrimination without explicitly favoring one metric over another. Rather, a core tenet of legal research and jurisprudence is that conflicting rights or interests need to be balanced, i.e., be partially realized to satisfy each of the mutually incompatible demands to the extent that conditions warrant ((De Vries, 2013; Morijn, 2006)). Lawmakers, courts and regulators therefore need fairness frameworks that move beyond the mere statement of incompatibility. Simultaneously, ML research may benefit from the legal tradition of balancing by inducing decision makers and developers to refine algorithms implementing a balance between the normative trade-offs inherent in the mutually incompatible fairness criteria ((Hertweck and Räz, 2022)). Our paper develops one such algorithm allowing decision makers to interpolate between three different incompatible fairness criteria singled out for their importance in previous research. These three fairness criteria are described below. There are various ways in which unfairness, i.e., discrimination against certain groups ((Friedman and Nissenbaum, 1996)), may enter an ML model ((Calders and Žliobaitė, 2013)). For example, the model may perpetuate historical biases present in the training data; use non-representative training data; or optimize for a target variable which unwittingly encodes differential access of individuals or groups ((Barocas and Selbst, 2016)). The result of such biased modeling may be that false positive or false negative rates differ between groups of individuals. Specific fairness metrics such as equalized odds, quantifying the difference of false positive and false negative rates between groups ((Hardt et al., 2016)), are used to quantify how much a classifier systematically rewards or punishes one group more than another with respect to ground truth. Even if a classifier is fair in terms of equalized odds, its prediction scores can lead to unfair outcomes if they are uncalibrated, as decision makers are not able to reliably interpret its output ((Zadrozny and Elkan, 2001; Jiang et al., 2012)).222A classifier is calibrated when, given a score $s$, $s$-fraction of all individuals receiving score $s$ are ground truth positive ((Dawid, 1982)). What is worse, if a classifier emits scores that are differently calibrated between groups, identical scores have different meanings for different protected groups, again leading to unfair outcomes ((Kleinberg et al., 2016)). Given that equality of false positive rates between groups, equality of false negative rates between groups, and calibration between groups are all important fairness criteria ((Corbett-Davies and Goel, 2018)), it would be ideal if a classifier could satisfy all three. Unfortunately, previous work shows that simultaneously fulfilling all three of these fairness criteria is impossible, except under highly constrained circumstances such as when a classifier achieves 100% accuracy with each instance receiving a prediction score of exactly 0 or 1 ((Chouldechova, 2017; Kleinberg et al., 2016; Pleiss et al., 2017)). Pleiss et al. (2017) attempt to find a relaxed notion of error rate difference that can be simultaneously achieved alongside group calibration. However, they demonstrate that one may only simultaneously achieve group calibration with a relaxed version of one of either equalized false positive or false negative rates. Furthermore, they show that classifiers simultaneously satisfying two of these three fairness criteria must engage in the equivalent of randomizing a certain percentage of predictions. Thus, the authors suggest that practitioners should only attempt to satisfy one of the three fairness criteria based on importance, instead of trying to simultaneously fulfill a relaxed notion of two of them. Often, however, all three fairness notions embody valid concerns of different groups ((Corbett-Davies and Goel, 2018)). Merely satisfying one of them may not do justice to the various interests and rights of the groups, and, as discussed above, might even amount to unjustified discrimination if the law demands a balance rather than an absolute priority of one criterion. Building on the assumption that a decision makers should determine, within the constraints of the law, which combination of the three fairness criteria is most crucial towards a fair result in practice, our work introduces a new post-processing algorithm called FAir Interpolation Method (FAIM). It allows decision makers to consciously and continuously interpolate between fairness criteria. The algorithm is based on the notion that, given a classifier, a representative evaluation data set, and the corresponding prediction scores for each group, any one of the three fairness criteria could be satisfied by applying a certain fairness criteria specific score mapping function mapping scores from the distribution outputted by the model to a fair score distribution. Given three such fairness criteria specific score distributions, FAIM uses optimal transport to continuously interpolate between them. This leads to a mapping function that can be applied to new classifier predictions, post hoc, to achieve a desired, weighted combination of the respective fairness criteria. FAIM cannot be used to fully satisfy all three fairness criteria simultaneously, but it does give the practitioner a powerful tool to make a continuous compromise along the fairness criteria simplex, and thus to account for different concerns each fairness criterion encompasses. The remainder of the article is organized as follows. First, related work is outlined in detail. Next, the mathematical foundations regarding fairness criteria incompatibility and optimal transport are addressed. Based on this, FAIM is introduced, and tested empirically in three experiments. In the first experiment, FAIM is applied to a synthetic classifier that produces two distinct score distributions for two respective groups. Here the properties of FAIM are elucidated under controlled conditions. In the second experiment, FAIM is tested on the COMPAS data set, on which the COMPAS algorithm famously scores individuals for recidivism risk. In the last experiment, FAIM is applied to product rankings from the European e-commerce platform Zalando.333www.zalando.de Here, we evaluate our model in terms of typical fairness problems arising in e-commerce rankings. The discussion takes up the results of the experiments and adds a specific legal perspective. We show how FAIM may be harnessed to trade off conflicting legal requirements across a wide variety of domains: credit scoring, criminal justice decisions, and fair rankings according to the DMA. Overall, the paper brings together technical and legal perspectives on fairness criteria to take both discourses beyond incompatibility statements. In this way, decision makers may flexibly adapt ML models to different use cases in which varying interests of protected groups may necessitate different trade-offs between the involved fairness criteria. Simultaneously, gradually balancing the different fairness criteria, rather than prioritizing only one of them, will often be conducive to fulfilling legal requirements, both in established legal domains and in novel regulatory frameworks, such as the DMA. ## 2\. Related Work The incompatibility of various different fairness criteria in algorithmic decision making has been the subject of intense research in recent years. Friedler et al. (2016) discuss the contrast between individual and group fairness. While these results hint at an incompatibility between individual and group fairness and relate these notions to two opposing worldviews, there is no rigorous proof of such incompatibility, and no discussion of possible intermediate worldviews. Some of the results in Feldman et al. (2015) and Zehlike et al. (2020) can be viewed as proposals on how to continuously navigate between the two extremes. The recidivism prediction problem was treated in Kleinberg et al. (2016); Chouldechova (2017), with the result that a prediction algorithm must necessarily be unfair with regard to certain fairness criteria under realistic factual assumptions. In particular, Kleinberg et al. (2016) show the mutual incompatibility between calibration within groups and balance for the positive/negative classes in the (typical) case of unequal base rates. These fairness criteria also underlie the present contribution. The framework of Kleinberg et al. (2016) is thoroughly reviewed below; let us remark here only that the mentioned criteria of Kleinberg et al. (2016) acknowledge and accept the possibility of empirically observed disparities between different groups, but require the equal treatment of individuals in different groups conditional on their similarity according to ‘ground truth’. The competing criterion of calibration within groups concerns the accuracy of the algorithm and is shown to conflict with the other two criteria. The use of optimal transport methods in algorithmic fairness has become increasingly popular ((Dwork et al., 2012; Feldman et al., 2015; Gordaliza et al., 2019; Zehlike et al., 2020; Chiappa et al., 2020; Chiappa and Pacchiano, 2021)), applying optimal transport to trade-offs between individual vs group fairness, and accuracy vs fairness. Optimal transport is a mathematical tool that provides a very natural notion of distance of two probability distributions (the _Wasserstein_ or earth mover distance) and, moreover, gives an optimal (in a specific sense) way to translate between these distributions by means of an _optimal transport map_ or at least an _optimal transport plan_ , the latter requiring some form of randomization. In (Zehlike et al., 2020), particularly, an important role is played by the _Wasserstein-2 barycenter_ of several probability distributions and the _displacement interpolation_ that allows to continuously interpolate between various distributions. We make use of these techniques in the present article. The literature at the intersection of law and computer science has, in the past few years, increasingly discussed how legal non-discrimination requirements can be integrated into code, and vice versa ((Zehlike et al., 2020; Hacker, 2018; Wachter et al., 2021a, b; Zuiderveen Borgesius, 2020; Gerards and Borgesius, 2022; Wachter, 2020, 2022; Barocas and Selbst, 2016; Raghavan et al., 2020; Hellman, 2020; Nachbar, 2020)). This strand of research has focused particularly on remedying discrimination in criminal justice proceedings ((Berk et al., 2021; Starr, 2014; Mayson, 2019; Selbst, 2017; Katyal, 2019; Eaglin, 2017; Huq, 2019; Pruss, 2021)) and in other high-stakes AI decisions ((Veale et al., 2018)), such as employment contexts ((Raghavan et al., 2020)), credit scoring ((Bono et al., 2021)), or university admissions ((Zehlike et al., 2020)). More specifically, (Wachter et al. (2021a)) rightly point out that the reliance on fairness metrics based on ground truth may perpetuate biases if that ground truth itself is skewed against protected groups. This is a concern, for example, in criminal justice settings ((Berk et al., 2021; Bao et al., 2021)). We incorporate this constraint into the application perimeter of our algorithm (see Part 6.3). Conversely, in the e-commerce sector, numerous publications deal with the emerging competition law framework for digital markets, for example the DMA ((Eifert et al., 2021; Hacker, 2022b; Brouwer, 2021; Cabral et al., 2021; Laux et al., 2021; Podszun and Bongartz, 2021)). However, they do not, to our knowledge, specifically develop algorithms to solve legal questions arising in this field. Existing publications often use US law as the relevant yardstick ((Barocas and Selbst, 2016; Raghavan et al., 2020; Wachter, 2022; Selbst and Barocas, 2022; Hellman, 2020; Gillis and Spiess, 2019; Kim, 2017, 2022)), but have lately increasingly incorporated explicit discussions of EU law as well ((Zehlike et al., 2020; Hacker, 2018; Wachter et al., 2021a, b; Zuiderveen Borgesius, 2020; Gerards and Borgesius, 2022; Wachter, 2020, 2022; Veale and Binns, 2017)). Our paper builds on and expands this strand of research by offering a novel method to trade off and implement competing technical and legal fairness constraints. In three case studies, it demonstrates the potential of FAIM to integrate varying legal dimensions of fairness in ML settings, particularly in the criminal justice and credit scoring domain. The third case study tackles an, to our knowledge, entirely new field of law from an algorithmic fairness perspective: the EU Digital Markets Act (DMA). ## 3\. Mathematical Theory and Algorithmic Implementation This section introduces FAIM, our interpolation framework that allows a continuous shift between three mutually exclusive fairness notions, as presented in Kleinberg et al. (2016). We first provide a brief summary of the findings from Kleinberg et al. (2016) to make the reader familiar with the three fairness notions and establish necessary terminology. We then briefly present mathematical preliminaries from optimal transport theory. Finally we describe the mathematical theory of our algorithm as well as its implementation in parallel, to help the reader translate between the formulas and the code. We present the method in pseudocode in Algorithm 1. For the simplicity of presentation, we consider a population of individuals partitioned into two groups, indexed by $t=1,2$. Note first, that individuals need not be human, but could also be products, companies, etc., and second, that our method applies just as well to a setting with more than two groups. Each individual, regardless of group membership, is either truly ‘positive’ or ‘negative’ with respect to some trait of interest; i.e., in criminal justice, an individual would be ‘positive’ if they were to commit a certain type of crime within a given period in the future, and negative if not. The attributes ‘positive’ and ‘negative’ do not carry any normative meaning and are thus interchangeable, but we will refer to such a classification as ‘ground truth’: It is assumed that the assignments of the positive/negative label reflect reality (we will shortly discuss the limitations of this assumption). Furthermore, each individual is assigned a score by some algorithm (in the broadest sense). The score, which can take as its value any real number in the interval $[0,1]$, is supposed to reflect the probability that the given individual is positive. Therefore, if the ground truth were fully known to the algorithm, it would assign value 1 to positive individuals and 0 to negative ones. This unlikely scenario is known as _perfect prediction_. A note on terminology: When we say that an individual is truly negative/positive, we mean that the individual is negative/positive according to ground truth, regardless of their predicted score. This should not be confused with the common terminology in which a ‘true negative’ is an individual who is negative in ground truth and who, in addition, receives a negative prediction. In our setting, anyway, the predictions are not binary, but in terms of a continuous score between zero and one. ### 3.1. Incompatibility of Fairness Criteria The main result from Kleinberg et al. (2016) states that, in the setting described above, three natural fairness criteria are mutually incompatible except under trivial circumstances. We shortly summarize this incompatibility theorem here through a simple proof. Eight quantities are of importance in this context: * • $N_{t}$ is the total number of individuals in group $t$; * • $n_{t}$ is the number of (truly) positive individuals in group $t$; * • $x_{t}$ is the average score of the individuals in group $t$ that are (truly) negative; * • $y_{t}$ is the average score of the individuals in group $t$ that are (truly) positive. For the purpose of deriving the incompatibility theorem, it need not be assumed that the ground-truth dependent quantities $n_{t},x_{t},y_{t}$ be known; rather, _whatever their values might be_ , the fairness criteria proposed in the next paragraph are never met except in case of perfect prediction or equal base rates. Here, by _base rate_ for group $t$ we mean the ratio $n_{t}/N_{t}$, and _perfect prediction_ means that the score of each individual is 1 when she is positive and 0 when she is negative. The fairness criteria are as follows: * A) Calibration within groups: This measures the _accuracy of prediction_ of the score. The criterion requires that _the average score in each group should equal the ratio of (truly) positive individuals in that group_ , i.e., $\frac{x_{t}(N_{t}-n_{t})+y_{t}n_{t}}{N_{t}}=\frac{n_{t}}{N_{t}},\quad t=1,2.$ * B) Balance for the negative class: _The average score of (truly) negative individuals in group 1 should equal the average score of (truly) negative individuals in group 2_ , i.e., $x_{1}=x_{2}=:x.$ * C) Balance for the positive class: _The average score of (truly) positive individuals in group 1 should equal the average score of (truly) positive individuals in group 2_ , i.e., $y_{1}=y_{2}=:y.$ Let us illustrate the three requirements for the case of recidivism prediction. The members of each group (say, black and white offenders) are given a score that is intended to reflect the probability of the respective individual to recidivate within a given time period. Criterion A) then requires the following: The _average_ score within group $t$ should equal the ratio of actual recidivists in that group; so if, say, half of the members of the white group would turn out recidivist, then the average score given to white individuals should be $0.5$. Concerning B), given a non-recidivist individual, it is desirable that their score be independent of their group membership. It would have to be considered highly unfair if a black person received a higher score, and thus faced a higher likelihood of detention, than a white person, given that both would actually not commit an offense if released. This, in fact, is at the heart of the controversy around the COMPAS algorithm. Note that B) only considers scores on an average level: The _average_ score among all truly non-recidivist individuals should be equal between groups. Of course, this tells nothing about higher order statistical moments, such as the variance of the respective score distributions. The discussion for criterion C) is analogous. Suppose all the three requirements are satisfied, then this implies (1) $\displaystyle x(N_{1}-n_{1})+yn_{1}$ $\displaystyle=n_{1},$ $\displaystyle x(N_{2}-n_{2})+yn_{2}$ $\displaystyle=n_{2}.$ Regarding $N_{t},n_{t}$ as given parameters, this is a linear system of two equations for the two unknowns $x,y$. It has a unique solution if and only if the determinant of the coefficient matrix is nonzero. This determinant is computed as (2) $(N_{1}-n_{1})n_{2}-(N_{2}-n_{2})n_{1},$ which is zero if and only if either (1) $n_{1}\neq 0\neq n_{2}$and $\frac{N_{1}-n_{1}}{n_{1}}=\frac{N_{2}-n_{2}}{n_{2}}$, which is the case of equal base rates; or (2) one of $n_{1}$ or $n_{2}$ is zero, in which case (assuming $N_{1},N_{2}>0$) also the other one is zero, i.e., there are no positive individuals at all. Here, the choice $x=0$ and $y$ arbitrary yields a fair assignment according to Eq. 1. If there are no positive individuals, then everyone should get score zero, which is thus an instance of _perfect prediction_. For the first case, there are infinitely many solutions, one of which is $x=y=\frac{n_{1}}{N_{1}}=\frac{n_{2}}{N_{2}}$ (as mentioned in Kleinberg et al. (2016)). This is realizable by assigning the same score to everyone. (A mathematically simpler solution would be $x=0,y=1$, which however requires the algorithm to know who is positive and who isn’t and thus again requires perfect prediction.) Finally, let’s turn to the case when the determinant in Eq. 2 is nonzero, so there exists a unique solution to Eq. 1. It thus suffices to give one solution to rule out any other ones, and we can simply take $x=0,y=1$, i.e., perfect prediction, which therefore is the only possibility. In summary, we have thus verified Theorem 1.1, the main result, of Kleinberg et al. (2016), which we paraphrase as follows: ###### Theorem 1. For $t=1,2$, let $N_{t},n_{t}>0$. If a score function satisfies A), B), and C), then $\frac{n_{1}}{N_{1}}=\frac{n_{2}}{N_{2}}$ (equal base rates), or the positive individuals receive score 1 and the negative individuals receive score 0 (perfect prediction). ###### Remark 2. The formulation given here of criterion A) is weaker than the one stated in (Kleinberg et al., 2016) Page 4, where the criterion is required individually within each ‘bin’. As the computation here shows, the weaker requirement is already sufficient for incompatibility. ### 3.2. Mathematical Preliminaries on Optimal Transport Our goal is to design an algorithm that interpolates smoothly between the three fairness criteria discussed, as they are typically not attainable at the same time. The interpolation relies on the mathematical theory of optimal transport, which we now wish to recall. A probability measure $\nu$ on $[0,1]$ is called _absolutely continuous_ if it is representable by an integrable probability density function $f:[0,1]\to\mathbb{R}$, which means that $\nu((a,b))=\int_{a}^{b}f(x)dx$ for any $a,b\in(0,1)$ with $a\leq b$. In words, the probability that a random variable distributed according to $\nu$ has its value in the interval $(a,b)$ is given by the integral of the density function $f$ over said interval. For this entire section, we will assume that all occurring probability measures are absolutely continuous and have finite variance. Given two probability measures $\nu$ and $\mu$ on $[0,1]$, a _transport map_ $T:[0,1]\to[0,1]$ from $\nu$ to $\mu$ is a map such that, for every interval $(a,b)\subset[0,1]$, $\mu((a,b))=\nu(T^{-1}(a,b)),$ where $T^{-1}(a,b)$ denotes the preimage of $(a,b)$ under $T$, i.e., the set of all numbers that $T$ maps to the interval $(a,b)$. One often uses the notation $\mu=\nu\circ T^{-1}$ in this situation. This means the following: If a random variable $X$ with values in $[0,1]$ is distributed according to $\nu$, then $T$ is a transport map from $\nu$ to $\mu$ if and only if the random variable $T(X)$ is distributed according to $\mu$. As an example on $\mathbb{R}$, consider the two normal distributions $\mathcal{N}(0,1)$ and $\mathcal{N}(1,2)$. Then $T:\mathbb{R}\to\mathbb{R}$, $T(x)=2x+1$, would be a transport map from $\mathcal{N}(0,1)$ to $\mathcal{N}(1,2)$. However, there may be many transport maps: In this example, also $x\mapsto-2x+1$ would be a transport map. Among the many transport maps, one (or potentially several) may be _optimal_ in the following sense: ###### Definition 3 (Optimal transport map). Let $\nu,\mu$ be two absolutely probability measures on $[0,1]$ with finite variance. An _optimal transport map_ is a transport map between $\nu$ and $\mu$ that minimises the cost functional $C(\nu,\mu,T):=\int_{\mathbb{R}^{n}}|x-T(x)|^{2}d\nu(x)$ among all transport maps from $\nu$ to $\mu$. Under the stated assumptions on $\nu$ and $\mu$, Brenier (1987, 1991) ensures the existence and uniqueness444Uniqueness holds only $\nu$-almost everywhere; this means that the optimal transport map is not determined on any values outside the support of $\nu$. of an optimal transport map. This allows to define the so-called quadratic _Wasserstein distance_ between $\nu$ and $\mu$, given by $W_{2}(\nu,\mu):=C(\nu,\mu,T)^{1/2}.$ The Wasserstein distance forms a metric on the space of all absolutely continuous finite-variance probability measures on $[0,1]$. The notion of optimal transport map allows to define a kind of continuous interpolation between the measures $\mu$ and $\nu$: ###### Definition 4 (Displacement interpolation, cf. Remark 2.13 in (Ambrosio and Gigli, 2013)). Let $\nu,\mu$ be two probability measures on $[0,1]$ with unique optimal transport map $T$. The _displacement interpolation_ between $\nu$ and $\mu$ with interpolation parameter $\theta\in[0,1]$ is defined by $\nu^{\theta}((a,b))=\nu((T^{\theta})^{-1}(a,b)),$ where the map $T^{\theta}:[0,1]\to[0,1]$ is given by $T^{\theta}(x)=(1-\theta)x+\theta T(x)$. Clearly, $\nu^{0}=\nu$ and $\nu^{1}=\mu$. The measure $\nu^{\theta}$ thus is a measure ‘in between’ $\nu$ and $\mu$ that is closer to $\nu$ the smaller $\theta$ is chosen and closer to $\mu$ the larger it is. Finally, we mention the notion of _barycenter_ of a family of probability measures $\\{\nu_{k}\\}_{1,\ldots,N}$ with corresponding weights $\\{w_{k}\\}_{1,\ldots,N}$: ###### Theorem 5 (Barycenter in Wasserstein space (Agueh and Carlier, 2011)). Let $\\{\nu_{k}\\}_{1,\ldots,N}$ be a family of absolutely continuous probability with finite variance, and let $\\{w_{k}\\}_{k=1,\ldots,N}$ be positive weights with $\sum_{k=1}^{N}w_{k}=1$. Then there exists a unique probability measure $\nu$ on $[0,1]$ that minimizes the functional $\nu\mapsto\sum_{k=1}^{N}w_{k}W^{2}_{2}(\nu_{k},\nu).$ This measure is called the _barycenter_ of $\\{\nu_{k}\\}_{1,\ldots,N}$ with weights $\\{w_{k}\\}_{1,\ldots,N}$. With two groups and respective score distributions $\nu_{1},\nu_{2}$ and weights $w_{1},w_{2}$, the barycenter is given by the displacement interpolation $\nu=\nu_{1}\circ(T^{w_{2}})^{-1}$, where $T$ is the optimal transport map between $\nu_{1}$ and $\nu_{2}$ and $T^{w_{1}}$ is given in Def. 4. ### 3.3. The FAir Interpolation Method (FAIM) The goal of this paper is to transform a given model output into one which ‘balances’ between the three fairness criteria to an extent that the user is free to choose according to her specific situation. For instance, a model (like COMPAS) that has been used for a while and has been evaluated to give quite accurate predictions, but discriminate unfairly between groups, can be transformed into an improved algorithm that respects group fairness to a higher degree. Before we delve into the mathematics, let us describe more clearly the situation. Again we have two groups of individuals represented as (disjoint) sets $X_{1},X_{2}$, and for each group a given scoring algorithm, which for group $t$ ($t=1,2$) is simply a map $S_{t}:X_{t}\to[0,1]$. We assume this algorithm has been used for a sufficient amount of time such that reasonably reliable data is available on its performance; more precisely, we assume for each group a map $[0,1]\to[0,1]$, $s\mapsto\lambda^{+}_{t}(s)$ to be given, where $\lambda^{+}_{t}(s)$ reflects the proportion of truly positive555Recall that ‘truly positive’ means positive according to ground truth, regardless of the score received. individuals from group $t$ that were assigned score $s$. Alternatively, one may interpret $\lambda^{+}_{t}(s)$ as the probability that a randomly chosen individual from group $t$ whose score equals $s$ will be truly positive. We also define $\lambda^{-}_{t}(s):=1-\lambda^{+}_{t}(s)$ for all $s\in[0,1]$ as the proportion of truly negative instances among individuals of group $t$ and score $s$. From historical data, we know the score distributions in each group assigned by $S_{t}$, that is, we have two probability measures $\nu_{1},\nu_{2}$ such that for any (measurable) set $Q\subset[0,1]$, $\nu_{t}(Q)$ is the probability that an individual $x$ randomly chosen from group $t$ has a score in $Q$. ###### Remark 6. In practice, the question arises how to collect the data encoded in $\lambda_{t}^{+}$ and $\nu_{t}$. For the latter, it suffices to collect the previous outcomes of the scoring algorithm $S_{t}$, since $\nu_{t}(s)$ is the proportion of individuals from group $t$ that were assigned score $s$ in the past. The integral of $\nu_{t}(s)$ over all possible scores will then equal one, because the probability that a given individual was assigned _some_ score is one. For $\lambda_{t}^{+}$, we need information on the _ground truth_ , which is always a difficult and controversial matter. In addition to the data on past scores, we need data on the actual positivity or negativity of individuals. Though such data may be difficult to collect reliably, we wish to remark that such information is not just required for our method, but for _any_ reasonable quality control of the algorithm in question. Without any (at least approximate) determination of ground truth, there is generally no way to evaluate the performance of a given algorithm. In the first steps of Algorithm 1, we give for each of the criteria A), B), C) a procedure to transform a given algorithm that does not satisfy the respective criterion into one that does. #### 3.3.1. Criterion A) The original algorithm might not yet be correctly calibrated, i.e., the scores produced by $S_{t}$ do not yet give an accurate representation of the actual probability of being positive. To fix it, we compute a new map $S_{t}^{A}$ defined by (3) $S_{t}^{A}(x):=\lambda^{+}_{t}(S_{t}(x)),$ which means that an individual from group $t$ who initially received score $s$ will now receive score $\lambda^{+}_{t}(s)$, which is precisely its probability of being positive. In Algorithm 1, Lines 1–1 compute map $S_{t}^{A}$, and Lines 1–1 yield the probability distributions $\mu^{A}_{t}$. This probability, as discussed, is computed from historical data, which approximately reflects ground truth. Thus, although the new score will not in general be able to give a deterministic value (zero or one) of the particular individual’s true positivity, the new algorithm does satisfy requirement A) (even bin-wise, i.e., for each $s$). #### 3.3.2. Criterion B) To express the balance criterion, we need to determine the expected score of an individual $x$ from group $t$ on the condition that $x$ is negative. These conditional expectations for both groups should then be equal. Indeed, we can find a new score map such that the entire probability distributions of the score, conditional on negativity, coincide, thus giving statistical parity in score between the negative instances in each group. In particular, the risk of a false positive outcome will be equal for both groups. We do this as follows (see again Algorithm 1): First (Lines 1–1), we compute said conditional probabilities $\lambda_{t}^{-}$ and $\nu_{t}$ from the given data. By Bayes’ Theorem, the probability distribution for the score of an individual $x$ from group $t$ conditional on $x$ being negative is given as (4) $\sigma^{-}_{t}:=\frac{\lambda_{t}^{-}(s)\nu_{t}(ds)}{\int_{0}^{1}\lambda_{t}^{-}(s)d\nu_{t}(s)}.$ Secondly, we consider the Wasserstein-2 barycenter of $\sigma^{-}_{1},\sigma^{-}_{2}$ on $[0,1]$ with weights $\frac{N_{1}}{N_{1}+N_{2}}$ and $\frac{N_{2}}{N_{1}+N_{2}}$, respectively, which we denote by $\bar{\sigma}^{-}$ (Line 1). Then there are two (optimal) transport maps $T^{-}_{t}:[0,1]\to[0,1]$ such that (5) $\bar{\sigma}^{-}=\sigma^{-}_{1}\circ(T^{-}_{1})^{-1}=\sigma^{-}_{2}\circ(T^{-}_{2})^{-1}.$ Finally, set $S_{t}^{B}(x):=T^{-}_{t}(S_{t}(x))$ for $t=1,2$. Replacing $\nu_{t}$ with $\nu_{t}^{B}:=\nu_{t}\circ(T^{-}_{t})^{-1}$, which is the score distribution under the modified algorithm $S_{t}^{B}$, and replacing also $\lambda^{-}_{t}$ with $(\lambda^{-}_{t})^{B}:=\lambda_{t}^{-}\circ(T^{-}_{t})^{-1}$, which is the new probability that an individual from group $t$ with modified score is negative, we obtain via Eq. 4 (Line 3.3.2) (6) $\mu_{t}^{B}:=\sigma^{-}_{t}\circ(T_{t}^{-})^{-1}=\bar{\sigma}^{-}$ as the probability distribution of the modified score conditional on being negative. Since they agree for both groups, the new $S_{t}^{B}$ satisfies (a much stronger version of) B). input : $\texttt{rawScores[]}:$ an array with a score for each individual; $\texttt{stepsize}:$ a float that specifies the bin width of the truncated scores; $\texttt{groundTruthLabels[]}:$ an array with a binary ground truth label for each individual; $\texttt{groups[]}:$ an array with a group membership label for each individual; $\texttt{thetas[[]]}:$ a matrix with rows $\left[\theta^{A},\theta^{B},\theta^{C}\right]$ for each group $t$. output : $\texttt{fairScores[]}:$ a new score array with a fair score for each individual such that the fair score distributions map the fairness criteria given by thetas. 1 // Normalize scores to $[0,1]$ and discretize 2 $\texttt{rawScores[]}=\texttt{truncateScores(0, 1, stepsize)}$ // Step 1: Compute score map $S^{A}_{t}$, yielding $\mu^{A}_{t}$ (Eq.3) 3 criterionAScores[] = zeros(rawScores[].length) 4 for _$t$ in groups[].uniqueValues()_ do 5 for _$s$ in rawScores[].uniqueValues()_ do 6 groupMask = groups[] == $t$; groundTruthPositivesMask = groundTruthLabels[] == 1; scoreMask = rawScores[] == $s$; 7 $s^{A}_{t}$ = $\frac{\texttt{rawScores[groupMask \&\& scoreMask \&\& groundTruthPositivesMask].length}}{\texttt{rawScores[groupMask \&\& scoreMask].length}}$ // add key value pair to score map for criterion $A$. 8 $S^{A}_{t}$.add(($s,s^{A}_{t}$)) 9 end for // translate raw scores into fair scores w.r.t. criterion $A$: everybody with raw score $s$ gets fair score $s^{A}_{t}$ assigned. 10 criterionAScores[groupMask] = rawScores[groupMask].translate($S^{A}_{t}$) 11 $\mu^{A}_{t}$ = criterionAScores[groupMask].histogram() 12 13 end for // Step 2: compute $\mu^{B}_{t}$ 14 groupSigmas = [[]]; groupSizesInPercent = []; 15 groundTruthNegativesMask = groundTruthLabels[] == 0 16 for _$t$ in groups[].uniqueValues()_ do 17 groupMask = groups[] == $t$ 18 groupSizesInPercent.add(groupMask.length() / groups[].length()) // compute Eq. 4 19 $\sigma^{-}_{t}$ = zeros(rawScores.histogram().length()) 20 for _$s$ in rawScores[].uniqueValues()_ do 21 scoreMask = rawScores[] == $s$ 22 temp = $\frac{\texttt{rawScores[groupMask \&\& scoreMask \&\& groundTruthNegativesMask].length}}{\texttt{rawScores[groupMask \&\& groundTruthNegativesMask].length}}$ 23 $\sigma^{-}_{t}$.add(temp) 24 end for 25 groupSigmas.add($\sigma^{-}_{t}$) 26 end for // compute barycenter $\bar{\sigma}^{-}$ between groupSigmas 27 $\bar{\sigma}^{-}$ = wasserstein2barycenter(groupSigmas, groupSizesInPercent) 28 Algorithm 1 Algorithm FAIM. First, criteria $A,B$ and $C$ are individually fully satisfied, yielding distributions $\mu^{A}_{t},\mu^{B}_{t}$, and $\mu^{C}_{t}$. Then, the barycenter $\bar{\mu}_{t}$ of $\mu^{A}_{t},\mu^{B}_{t}$, and $\mu^{C}_{t}$ is computed, which yields optimal transport map $T_{t}$. $T_{t}$ is finally used to map from the original score distribution $\nu_{t}$ to the fair distribution $\bar{\mu}_{t}$. “emd” stands for earth mover distance. 26 for _$t$ in groups[].uniqueValues()_ do 27 groupMask = groups[] == $t$ // use $\bar{\sigma}^{-}$ to get optimal transport maps for ground truth negative individuals (Eq. 5) 28 $T^{-}_{t}$ = computeOptimalTransportMap(rawScores[groupMask && groundTruthNegativesMask].histogram(), $\bar{\sigma}^{-}$, kind=emd) // For ground truth negatives, translate rawScores into $S^{B}_{t}$. The histogram of $S^{B}_{t}$ forms $\mu_{t}^{B}$. Note, that $\mu^{B}_{t}$ is the score distribution for group $t$ that contains all individuals (negatives and positive), even though the translation is done for true negatives only. 29 $\mu_{t}^{B}$ = rawScores[groupMask && groundTruthNegativesMask].translate($T^{-}_{t}$).histogram() 30 end for // Step 3: compute $\mu^{C}_{t}$. We omit this part of the algorithm for brevity, since this is done the same way as $\mu^{B}_{t}$ (Lines 1\--3.3.2). The only difference is that we would use a groundTruthPositivesMask = groundTruthLabels[] == 1, instead of a groundTruthNegativesMask. // Step 4: compute final barycenter $\bar{\mu}_{t}$ which incorporates thetas[[]] and optimal transport maps $T_{t}$ for each group $t$. Then translate rawScores of group into fair scores. 31 for _$t$ in groups[].uniqueValues()_ do // compute $\bar{\mu}_{t}$ (Eq. 7) 32 $\bar{\mu}_{t}$ = wasserstein2barycenter($\mu^{A}_{t},\mu^{B}_{t},\mu^{C}_{t}$, thetas[$t$]) 33 groupMask = groups[] == $t$ 34 $T_{t}$ = computeOptimalTransportMap(rawScores[groupMask].histogram(), $\bar{\mu}_{t}$, kind=emd) 35 fairScores[groupMask] = rawScores[groupMask].translate($T_{t}$) 36 end for return _fairScores[] _ Let us, however, not ignore a small issue here: The map $T^{-}_{t}$ is, as mentioned above, only well-defined on the support of $\sigma_{t}^{-}$. For our intended application, there might well be scores that have never been given to a truly negative (or positive) individual: For example, if the original algorithm was not completely misguided, then it would most probably never have given score 1 to a truly negative individual, so that $\sigma_{t}^{-}$ would vanish near 1. For such scores we are essentially free to define the map $T$. The most natural and simple choice is to leave these scores unchanged, i.e., to set $T_{t}^{-}(x)=x$ for any $x$ outside the support of $\sigma_{t}^{-}$. #### 3.3.3. Criterion C) Replacing $-$ by $+$ everywhere in B), we obtain a new algorithm $S_{t}^{C}$ that satisfies C) and in particular assimilates the probabilities of a false negative outcome of the evaluation for both groups. #### 3.3.4. Combining the three procedures Each of the modified score maps $S_{t}^{A},S_{t}^{B},S_{t}^{C}$ gives rise to a corresponding score distribution $\mu_{t}^{A},\mu_{t}^{B},\mu_{t}^{C}$. For each group $t=1,2$, we therefore obtain a triangle in Wasserstein space. Let now $\theta_{t}^{A},\theta_{t}^{B},\theta_{t}^{C}\in[0,1]$ be three real parameters that add up to $1$, which are given as input to the algorithm. We then consider the weighted barycenter $\bar{\mu}_{t}$ in Wasserstein-2 space, i.e., the unique probability measure $\bar{\mu}$ that minimizes the functional (7) $\mu\mapsto\theta_{t}^{A}W_{2}^{2}(\mu,\mu_{t}^{A})+\theta_{t}^{B}W_{2}^{2}(\mu,\mu_{t}^{B})+\theta_{t}^{C}W_{2}^{2}(\mu,\mu_{t}^{C}).$ This barycenter embodies a ‘compromise’ between our three incompatible fairness criteria, where the parameters $\theta$ allow to continuously adjust the weights that the decision-maker wishes to put on the respective criterion (Line 3.3.2). The fair score of an individual in group $t$, based on the original (‘unfair’) score function $S_{t}$, is then determined as $T(S(x))$, where $T:[0,1]\to[0,1]$ is the optimal transport map from the original score distribution $\nu_{t}$ to the fair distribution $\bar{\mu}_{t}$ (Lines 3.3.2–3.3.2). ## 4\. Experiments We conduct extensive experiments on three use cases (one with synthetic data and two with real-world data) to show the effectiveness of FAIM to interpolate between the three different fairness criteria described in Section 3.1. First, we apply FAIM to synthetic data with two demographic groups and show how much of each type of fairness can be achieved by tuning each fairness preference $\theta^{A},\theta^{B}$ and $\theta^{C}$. As a motivating example, we consider the synthetic data to represent credit applicants from two different gender groups. Second, we apply the same analysis to the COMPAS data set, which contains protected group variables for gender, race, and age. We are aware of critical voices, such as (Bao et al., 2021), on using COMPAS to benchmark new fairness methods, and discuss whether recidivism risk is an appropriate use case for FAIM in Section 6.3. However, since Kleinberg et al. (2016), whose terminology we adopt, build their argumentation on the COMPAS case, we want to show how our method and findings compare to those of Kleinberg et al. (2016). Last, we study how the application of FAIM affects rankings on the European e-commerce platform Zalando, and whether it can be used to overcome the problem of popularity bias ((Bellogín et al., 2017)).666Data and code available under https://github.com/MilkaLichtblau/faim.git For each use case, FAIM is tested with the following fairness objectives: 1. (1) Fairness criterion A: Calibration within groups ($\theta^{A}=1$) 2. (2) Fairness criterion B: Balance for the negative class ($\theta^{B}=1$) 3. (3) Fairness criterion C: Balance for the positive class ($\theta^{C}=1$) 4. (4) An equally weighted combination of fairness criteria A), B), and C) ($\theta^{A}=\theta^{B}=\theta^{C}=\frac{1}{3}$) Note that any other weighted combination is possible, e.g., by setting $\theta^{A}=0.5$ and $\theta^{B}=\theta^{C}=0.25$ if one cares about criteria B) and C), but prefers a result that thoroughly calibrates scores. Note also, that the $\theta$-array can be configured on a per-group-level, but for simplicity of presentation we set the same $\theta$s for each group. For each fairness objective, we calculate performance metrics (accuracy, precision, and recall) and error rates including false negative rate (FNR) and false positive rate (FPR). Results from using FAIM are compared to those produced by the base model. ### 4.1. Experiments on Synthetic Data Figure 1. Synthetic true (solid lines) and predicted (dashed lines) score distributions for demographic two groups (orange and blue). The decision boundary at $\text{score}=0$ determines ground truth positive and negative labels based on the true scores, and predicted positive and negative labels based on the predicted scores. This simulates a situation in which the scoring algorithm overestimates the qualification of the advantaged group, and further disadvantages the disadvantaged group. As a motivating example for the synthetic data experiment, let us consider credit scoring. Increasingly, scoring agencies are turning to ML to predict creditworthiness ((Langenbucher, 2020; Langenbucher and Corcoran, 2021)). Now imagine that a company predicts ML-based credit scores for candidates stemming from two protected groups only, for example for men and women (no non-binary persons applied). Candidates receiving a score equal or above 0 are labeled offered a credit contract, while those below are rejected. #### 4.1.1. Data set. To model such a credit scoring situation, we generate a large synthetic data sets with 100,000 individuals and two demographic groups (one advantaged, e.g., male; and one disadvantaged, e.g., female) of approximately the same size. Each individual is assigned two scores, a _true score_ , and a _predicted score_. The true score is to be understood as an individual’s true probability of being part of a desirable positive class (e.g., creditworthy person), while the predicted score is meant to represent a score that has been assigned to an individual by a (synthetic) scoring algorithm. They are sampled from a multivariate normal distribution with covariance matrix $[[1,0.8],[0.8,1]]$ and true score means $(1,-1)$ for the advantaged and disadvantaged groups, respectively. The corresponding predicted score means are $(2,-3)$. Additionally, we define a synthetic decision threshold to yield ground truth and predicted labels: those with a true (resp. predicted) score above zero are considered ground truth (resp. predicted) positive (i.e., creditworthy, and receive a credit offer). Conversely, those with a true (resp. predicted) score below zero are considered ground truth (predicted) negative (i.e., not creditworthy, and are rejected). Figure 1 depicts this data set. Blue lines correspond to the advantaged group, with both predicted scores (dashed blue line) and true scores (solid blue line) normally distributed with peaks greater than zero. Orange lines correspond to the disadvantaged group, with both predicted scores (dashed orange line) and true scores (solid orange line) normally distributed with peaks less than zero. This data set reflects a scenario of different base rates between the groups, as the advantaged group has a higher true probability of being positive. In our example, this is the male group. However, the scoring algorithm has overestimated the capabilities of the blue, advantaged group (the predicted score distribution is centered to the right of the true score distribution), and underestimated the capabilities of the orange, disadvantaged group (the predicted score distribution is centered to the left of the true score distribution). Additionally, the algorithm does a better job in predicting the true score distribution for the advantaged group, than it does for the disadvantaged group (the blue distributions overlap more). To emphasize the disparity between the two demographic groups, the advantaged group is 5.3 times more likely to be labeled positive based on the true scores, compared to 6250 times more likely to be labeled positive based on the predicted scores from the (unfair) scoring algorithm. These two particular properties of disparities in model quality across groups are well studied in the algorithmic fairness literature ((Angwin et al., 2016; Zafar et al., 2017; Hardt et al., 2016)), which is why we choose this synthetic setting to present the functioning of FAIM. Note that we have plotted the true score distributions only for demonstration purposes to give a clear picture of the discriminatory scenario. They are usually not available, which is why FAIM relies on _observable_ ground truth labels, as explained in Section 3.3. Data set | Parameters | Performance | Error Rates ---|---|---|--- | | Accur. ($\Delta$) | Precision ($\Delta$) | Recall ($\Delta$) | FPR ($\Delta$) | FNR ($\Delta$) Synthetic | before FAIM | 0.852 | 0.853 | 0.852 | 0.138 | 0.157 blue | | 0.860 | 0.868 | 0.860 | 0.869 | 0.002 orange | | 0.844 | 0.869 | 0.844 | 0.000 | 0.990 Synthetic | $\theta^{A}=1$ | 0.885 (0.033) | 0.885 (0.033) | 0.884 (0.032) | 0.116 (-0.022) | 0.114 (-0.043) blue | | 0.884 (0.024) | 0.873 (0.005) | 0.884 (0.024) | 0.560 (-0.309) | 0.032 (0.032) orange | | 0.885 (0.041) | 0.874 (0.005) | 0.885 (0.041) | 0.032 (0.032) | 0.557 (-0.433) Synthetic | $\theta^{B}=1$ | 0.879 (0.027) | 0.882 (0.029) | 0.879 (0.027) | 0.076 (-0.062) | 0.166 (0.009) blue | | 0.877 (0.017) | 0.876 (0.008) | 0.877 (0.017) | 0.392 (-0.477) | 0.073 (0.073) orange | | 0.882 (0.038) | 0.873 (0.004) | 0.882 (0.038) | 0.016 (0.016) | 0.666 (-0.324) Synthetic | $\theta^{C}=1$ | 0.865 (0.013) | 0.873 (0.020) | 0.865 (0.013) | 0.208 (0.070) | 0.062 (-0.095) blue | | 0.854 (-0.006) | 0.868 (0.000) | 0.854 (-0.006) | 0.918 (0.049) | 0.001 (0.001) orange | | 0.877 (0.033) | 0.877 (0.008) | 0.877 (0.033) | 0.074 (0.074) | 0.389 (-0.601) Synthetic | $\theta^{A}=\theta^{B}=\theta^{C}$ | 0.883 (0.031) | 0.883 (0.030) | 0.883 (0.031) | 0.137 (-0.001) | 0.098 (-0.059) blue | | 0.884 (0.024) | 0.873 (0.005) | 0.884 (0.024) | 0.560 (-0.309) | 0.032 (0.032) orange | | 0.881 (0.037) | 0.875 (0.006) | 0.881 (0.037) | 0.057 (0.057) | 0.450 (-0.540) Table 1. We report the performance metrics and error rates of the synthetic experiment after FAIM has been applied, and the corresponding relative improvements or deteriorations (green or red values in parenthesis, respectively). The first line always refers to the whole data set, whereas “blue” and “orange” contain results disaggregated by group. The top row shows metrics before FAIM has been applied. #### 4.1.2. Results. Figure 2 shows the resulting score distributions of the synthetic data set after FAIM has been applied (left column), together with the corresponding transport maps that convert raw into fair scores (right column). The transport maps are to be read as follows: each raw score on the x-axis gets replaced by its corresponding value on the y-axis. Note that the transport maps necessarily differ by group, since the groups experience disparate treatment by our synthetic model which underestimated the capabilities of the disadvantaged group, and overestimated those of the advantaged. Each row of sub-figures corresponds to one of the scenarios described earlier. Consider Figures LABEL:fig:experiments:result:synthetic:condA and LABEL:fig:experiments:result:synthetic:TransportMapcondA as an example: they correspond to the experiment where we want to fully meet fairness criterion A). We see from the transport map (Fig. LABEL:fig:experiments:result:synthetic:TransportMapcondA), that the two score distributions are brought towards each other, as the disadvantaged group (orange) is assigned higher fair scores for the same raw scores. Figures LABEL:fig:experiments:result:synthetic:condB and LABEL:fig:experiments:result:synthetic:TransportMapcondB, as well as Figures LABEL:fig:experiments:result:synthetic:condC and LABEL:fig:experiments:result:synthetic:TransportMapcondC, correspond to the experiments that fully implement the balance criteria B) and C), respectively. Note that these two balance criteria are defined only on a subset of individuals, i.e., the true negatives and the true positives, respectively. Thus, FAIM uses only this subset of the true negative (resp. true positive) individuals to calculate the transport maps and, in turn, the fair scores. This is reflected in Figures LABEL:fig:experiments:result:synthetic:condB and LABEL:fig:experiments:result:synthetic:condC: the fair score distributions overlap _only for the truly negative (resp. positive) individuals_ , while the scores from the true positive (resp. negative) class are left untouched. Figures LABEL:fig:experiments:result:synthetic:compromise and LABEL:fig:experiments:result:synthetic:TransportMapcompromise show the results of using the fairness objective $\theta^{A}=\theta^{B}=\theta^{C}=\frac{1}{3}$. We see that this setting indeed yields a compromise between the three fairness criteria. Table 1 shows how FAIM affects classification performance and fairness. We understand fairness in terms of the three fairness objectives we seek to meet and therefore report metric and error rate differences in order to judge FAIM’s performance. The table reports the absolute values for each metric after FAIM has been applied, together with the changes relative to when FAIM is not applied (in parentheses). The top row shows the metrics before FAIM has been applied. Observe the results for our first experimental setting with $\theta^{A}=1$. We expect an overall performance increase, as the algorithm calibrates the predictions with respect to actual ground truth evaluation. We also expect the effect to be more pronounced for the orange group, because of the larger error between predicted and ground truth scores for that group. Both expectations are confirmed in the results. The chance of the orange group to be labeled positive is now 5.7% of the blue group, which marks an improvement of more than two orders of magnitude. Next, observe the results for the second and third fairness objectives where $\theta^{B}=1$, and $\theta^{C}=1$, respectively (balance criteria B) and C), respectively). When fulfilling criterion B), we expect the fair score distributions for ground truth negative individuals from the two groups to approximately match (recall Fig. LABEL:fig:experiments:result:synthetic:condB). Thus, the scores of ground truth negative blue and orange group individuals will decrease and increase, respectively. This corresponds to an FPR decrease for the blue group, which had a high base FPR including many false positive ground truth negative individuals scoring just slightly above the decision boundary whom become true negative after FAIM is applied. This also corresponds to either no or a slight FPR increase for the orange group, which had a base FPR of zero and ground truth negative individuals scoring far enough below the decision boundary that either none or only a few are mapped by FAIM above the decision boundary to become false positives. The results confirm our expectations. For the blue group, FAIM achieves an FPR of 39.2%, which marks an improvement w.r.t. the original FPR of 47.7%. We also see a slight increase of FPR for the orange group (1.6%). After FAIM is applied the FNR shows an improvement of 32.4%. Overall the probability of the orange group to receive a positive label is 5.3% of the blue group for $\theta^{B}=1$. When fulfilling criterion C), we expect the fair score distributions for true positive individuals to match (Fig. LABEL:fig:experiments:result:synthetic:condC), thus the false negative rates should improve, particularly for the orange group since most blue individuals were predicted positive by the original model. Again our expectations are confirmed by the results and declines in performance and error rates remain relatively small for both groups. Overall, the probability of the orange group for a positive label is 7.7% of the blue group. (a) (b) (c) (d) (e) (f) (g) (h) Figure 2. (Best seen in color). Left side: distributions of scores for each group after FAIM was applied under different fairness preferences (bold), together with the original distributions for comparison (transparent). Right side: transport maps that translate from the raw score (x-axis) to the respective fair score (y-axis). The gray dashed line marks the identity mapping. Note, that the two groups are assigned different fair scores for the same raw score. and that raw scores were normalized to $[0,1]$. Last, observe the last experimental setting $\theta^{A}=\theta^{B}=\theta^{C}$, which corresponds to a compromise between the three mutually exclusive fairness criteria. As is confirmed by Figures LABEL:fig:experiments:result:synthetic:compromise and LABEL:fig:experiments:result:synthetic:TransportMapcompromise, the results in Table 1 show that FAIM yields a compromise between calibration, balance for the true negatives, and balance for the true positives. It achieves similar performance improvements as aiming for calibration only ($\theta^{A}=1$), but better FPR and FNR improvements. Compared to the balance criteria, this setting achieves better performance improvements, but only slightly worse FPR and FNR. Overall, the probability of the orange group to receive a positive label is 9.7% of the blue group for $\theta^{A}=\theta^{B}=\theta^{C}$. (i) Accuracy rates disaggregated by decile score and gender. (j) Accuracy rates disaggregated by decile score and race. (k) Accuracy rates disaggregated by decile score and age categories. Figure 3. Accuracy rates disaggregated by COMPAS score. Across all demographics, accuracy rates show that the mid-range scores 4–7 do not provide very meaningful inside on recidivism risk. ### 4.2. Experiments on the COMPAS data set by (Angwin et al., 2016) #### 4.2.1. Data set COMPAS (Correctional Offender Management Profiling for Alternative Sanctions) is a commercial tool developed by Northpointe, Inc. to assess a criminal defendant’s likelihood of recidivating within a certain period of time. Based on the promise to enhance fairness in judicial decision making, COMPAS is used in several US states as a decision aid for judges, e.g., in parole cases. In 2016, Angwin et al. (2016) published an analysis of the tool based on a data set of criminal defendants from Broward County, Florida, in which they found the tool to be biased against certain groups (more on this below). From the data—as it was published by (Angwin et al., 2016)—we use decile_score (integers $\\{1,2,...,10\\}$) as predicted scores, and two_year_recid (boolean) as ground truth labels. Additionally, we construct groups based on sex, race, and age category using the features sex, race, and age_cat. To increase race group sizes, we merge races ‘Native American’ and ‘Asian’ into ‘Other’, leaving four race groups: ‘Caucasian,’ ‘African-American,’ ‘Hispanic,’ and ‘Other.’ These three features, i.e., the predicted score, the ground truth label, and the group, form the input for FAIM. To assess the impact of FAIM on the predictive performance and the group error rates, we perform an analysis similar to that carried out by Angwin et al. (2016). We also translate the predicted score into a binary label of high and low risk corresponding to predicted score $\geq 5$ and $<5$, respectively. This binarization is applied to both the predicted scores from the COMPAS data set and the resulting fair scores produced by FAIM. Additionally, we calculate the probability of the disadvantaged groups to be assigned a high risk label relative to the advantaged group while correcting for the seriousness of their crime, previous arrests, and future criminal behavior. #### 4.2.2. Performance Analysis of the Original Model To better interpret the results of applying FAIM to the COMPAS data set, we first provide a detailed performance analysis of the original COMPAS algorithm. Figure 3 shows the accuracy rates of the original model disaggregated by decile score and demographic group. Grey bars mark the accuracy rates measured for the aggregated data set. A striking insight is how poorly the model performs through the entire score range, but particularly for intermediate scores: within the range of 4–7, the accuracy rate for any demographic group hardly ever exceeds a level of 0.6. One could argue, of course, that a triangular pattern of accuracy rates (high accuracy for edge scores 0 and 10, lower accuracy for scores near the decision boundary of 5) would be expected from a well-calibrated model. After all, for a calibrated system, roughly 50% of individuals receiving a COMPAS score of 5 (corresponding to a normalized score of 0.5) would reoffend. Given that a COMPAS score $\geq 5$ corresponds to a prediction that an individual will reoffend, a calibrated system would achieve around 50% accuracy for the 5th score decile. It is important to understand however, that if the model was well-calibrated we would expect to see accuracy rates of _at least_ 0.9 for scores 1 and 10, _at least_ 0.8 for scores 2 and 8, and so forth, requiring an accuracy rate of _at least_ 0.5 for score 0.5. As shown, the model never achieves any of these minimum performance rates for any demographic group. Another interesting insight can be found by looking at the disparities of accuracy distributions across different demographic groups. Observe Figure 3: We see that women’s accuracy is higher than men’s in the lower scores, but vice versa for higher scores. This means that women are predominantly misclassified when they are assigned high scores, while men are predominantly misclassified, when they are assigned low scores. In other words, women are treated too harshly by the algorithm, while men are treated too gently (thus confirming the finding of Angwin et al. (2016)). The same is true for Hispanics in Figure 3, whose accuracy for the low scores is much higher than the one for the high score range. Surprisingly, Figure 3 reveals that young people, even though already being assigned relatively high scores, still seem to be treated too gently by the algorithm. #### 4.2.3. Experimental Results when Applying FAIM Data set | Parameters | Performance | Error Rates ---|---|---|--- | | Accur. ($\Delta$) | Precision ($\Delta$) | Recall ($\Delta$) | FPR ($\Delta$) | FNR ($\Delta$) COMPAS | before FAIM | 0.470 | 0.751 | 0.470 | 1.000 | 0.000 male | | 0.495 | 0.750 | 0.495 | 1.000 | 0.000 female | | 0.387 | 0.763 | 0.387 | 1.000 | 0.000 COMPAS | $\theta^{A}=1$ | 0.521 (0.051) | 0.751 (0.000) | 0.521 (0.051) | 0.739 (-0.261) | 0.185 (0.185) male | | 0.495 (0.000) | 0.750 (0.000) | 0.495 (0.000) | 1.000 (0.000) | 0.000 (0.000) female | | 0.613 (0.226) | 0.763 (0.000) | 0.613 (0.226) | 0.000 (-1.000) | 1.000 (1.000) COMPAS | $\theta^{B}=1$ | 0.530 (0.060) | 0.751 (0.000) | 0.530 (0.060) | 0.000 (-1.000) | 1.000 (1.000) male | | 0.505 (0.010) | 0.750 (0.000) | 0.505 (0.010) | 0.000 (-1.000) | 1.000 (1.000) female | | 0.613 (0.226) | 0.763 (0.000) | 0.613 (0.226) | 0.000 (-1.000) | 1.000 (1.000) COMPAS | $\theta^{C}=1$ | 0.530 (0.060) | 0.751 (0.000) | 0.530 (0.060) | 0.000 (-1.000) | 1.000 (1.000) male | | 0.505 (0.010) | 0.750 (0.000) | 0.505 (0.010) | 0.000 (-1.000) | 1.000 (1.000) female | | 0.613 (0.226) | 0.763 (0.000) | 0.613 (0.226) | 0.000 (-1.000) | 1.000 (1.000) COMPAS | $\theta^{A}=\theta^{B}=\theta^{C}$ | 0.530 (0.060) | 0.751 (0.000) | 0.530 (0.060) | 0.000 (-1.000) | 1.000 (1.000) male | | 0.505 (0.010) | 0.750 (0.000) | 0.505 (0.010) | 0.000 (-1.000) | 1.000 (1.000) female | | 0.613 (0.226) | 0.763 (0.000) | 0.613 (0.226) | 0.000 (-1.000) | 1.000 (1.000) Table 2. Evaluation of FAIM when applied to the COMPAS data set, disaggregated by gender. These are results only for an original COMPAS score of 5, because only individuals close to the decision boundary are reclassified after a score reassignment through FAIM. As such, for score ranges 1–4, and 6–10, neither error rates nor performance metrics change at all. Note also that, for one particular score, everybody is classified either positive or negative and the per-group error rates are thus either 0 or 1. Our experimental results using FAIM on the COMPAS data set mostly confirm an important hypothesis: our algorithm cannot fix an inherently flawed model (garbage in, garbage out). As described earlier, we assessed accuracy, precision and recall, as well as false positive and false negative rates, total and disaggregated by demographics (similarly to results shown in Table 1). Because the nominal changes were negligible, we performed this same analysis disaggregated by COMPAS score (as in Fig. 3). Together with the analysis of the transport maps, this revealed the following: to fulfill fairness criteria B) and C), FAIM does not change the scores drastically. This is understandable since FAIM does not correct for _accurate_ scores, but only for _equal distributions_ , as soon as $\theta^{A}=0$. Thus, only few individuals with original scores of 4 or 5 get reclassified from low to high risk, and vice versa. Those with original scores 1–4, or 6–10, do not cross the decision boundary between the low and high risk class and therefore, none of the metrics changes for these score ranges. Therefore, we exemplarily report results disaggregated by gender and _for score 5 only_ in Table 2, to showcase the behavior of FAIM on the COMPAS data set. We see that the algorithm behaves as expected and does, in fact, improve performance and error rates the way we expect it to. However, since the original model performance is so low (see first row section in Table 2), we would recommend to rather abandon this model entirely. It seems futile to change random guessing into fair random guessing. ### 4.3. E-commerce data set from Zalando #### 4.3.1. Data set For our last experiments, we use data from the e-commerce platform Zalando, one of Europe’s largest fashion and lifestyle retailers, operating in 25 European countries. We collected a data set of 81,048 articles on 27th October 2021, which belong to four different women clothes categories with a similar price range: skirts, jeans, trousers, and knitwear. Each data point contains the following information: a brand, a ranking score, and the number of impressions and clicks during the week after article collection. We look at the ratio of clicks over impressions and when it is above a certain threshold, assign a positive label to the article (as it got a positive weekly performance), or a negative label if it is below the threshold (as this article failed to gather enough clicks despite its visibility). Groups are assigned as follows: we compute a brand’s visibility as the average number of impressions per article for each brand and each category, and thus categorize brands having either low or high visibility. When this average is below the median, a brand is said to belong to the ‘low’ group, while if the number is above the third quartile, a brand is set to belong to the ‘high’ group. Articles of brands with visibility between the low and high group thresholds are discarded, leaving us with 62,461 articles. 70.5% of them belong to high visibility brands. Note, that the ranking scores are produced by a learning- to-rank model trained daily to optimize a surrogate normalized Discounted Cumulative Gain (nDCG) loss ((Järvelin and Kekäläinen, 2002)), but not to classify articles according to the ground truth labeling we defined above. However, because the nDCG relevance labels are based on customer interactions (i.e., implicit feedback), it is reasonable to assume that a good ranking model should be able to identify which article is likely to be clicked. Hence, the scores of an accurate click-through-rate classifier should provide a decent ranking. We observe in our data that, irrespective of whether those high articles are ground truth positive or negative, they receive higher scores than corresponding low articles. There are many reasons external to the ranking algorithm why customers would prefer well-known and highly visible brands: these brands have higher marketing budgets, they can afford a more aggressive pricing strategy thanks to economies of scale, and they benefit from the Matthew effect of accumulated advantage ((Perc, 2014)). Moreover, it is in the commercial interest of the platform to highlight such best selling items. At the same time, the platform may not want to reinforce the status quo but rather provide a level playing field for all brands to compete fairly. This makes it easier to attract new brands on the platform, offers a more diverse assortment to a potentially larger customer base, and can be seen as a good step to address the legal requirements discussed in 5.3. Data set | Parameters | Performance | Error Rates ---|---|---|--- | | Accur. ($\Delta$) | Precision ($\Delta$) | Recall ($\Delta$) | FPR ($\Delta$) | FNR ($\Delta$) Zalando | before FAIM | 0.624 | 0.623 | 0.192 | 0.079 | 0.808 high | | 0.586 | 0.622 | 0.238 | 0.122 | 0.762 low | | 0.717 | 0.747 | 0.012 | 0.002 | 0.988 Zalando | $\theta^{A}=1$ | 0.654 (0.030) | 0.585 (-0.039) | 0.518 (0.327) | 0.252 (0.173) | 0.482 (-0.327) high | | 0.623 (0.038) | 0.585 (-0.037) | 0.604 (0.366) | 0.361 (0.239) | 0.396 (-0.366) low | | 0.729 (0.012) | 0.576 (-0.171) | 0.186 (0.174) | 0.055 (0.053) | 0.814 (-0.174) Zalando | $\theta^{B}=1$ | 0.621 (-0.003) | 0.625 (0.001) | 0.169 (-0.022) | 0.070 (-0.010) | 0.831 (0.022) high | | 0.577 (-0.008) | 0.625 (0.003) | 0.189 (-0.050) | 0.095 (-0.027) | 0.811 (0.050) low | | 0.725 (0.008) | 0.623 (-0.124) | 0.095 (0.083) | 0.023 (0.021) | 0.905 (-0.083) Zalando | $\theta^{C}=1$ | 0.622 (-0.003) | 0.623 (-0.001) | 0.177 (-0.015) | 0.073 (-0.006) | 0.823 (0.015) high | | 0.579 (-0.007) | 0.625 (0.003) | 0.196 (-0.043) | 0.099 (-0.023) | 0.804 (0.043) low | | 0.725 (0.008) | 0.608 (-0.139) | 0.104 (0.092) | 0.027 (0.025) | 0.896 (-0.092) Zalando | $\theta^{A}=\theta^{B}=\theta^{C}$ | 0.633 (0.009) | 0.617 (-0.007) | 0.259 (0.067) | 0.110 (0.031) | 0.741 (-0.067) high | | 0.594 (0.009) | 0.618 (-0.004) | 0.294 (0.055) | 0.153 (0.031) | 0.706 (-0.055) low | | 0.727 (0.010) | 0.601 (-0.146) | 0.124 (0.112) | 0.033 (0.031) | 0.876 (-0.112) Table 3. We report the performance metrics and error rates of the experiment on Zalando data after FAIM has been applied, and the corresponding relative improvements or deteriorations (green or red values in parenthesis, respectively). The first line always refers to the whole data set, whereas “high” and “low” contain results disaggregated by group. The top row shows metrics before FAIM has been applied. #### 4.3.2. Results From Table 3 we see that low visibility brands do indeed have a relative disadvantage over high visibility brands. The low recall, and the high false negative rate indicate that many relevant items from the low group are not shown to the customer. This situation is improved by FAIM in all four scenarios, albeit to varying degrees. Our results also show that the algorithm yields expected improvements: When criterion A) is desired, accuracy and recall is improved for both groups. Pursuing criterion B) or C) levels the respective error rates. Finally, a compromise between all three criteria is found when setting $\theta^{A}=\theta^{B}=\theta^{C}$. (a) (b) (c) (d) (e) (f) Figure 4. (Best seen in color). Product rankings with low and high visibility brands under different settings. Colors indicate group membership, stripes indicate a positive label, i.e., these are the products customers find more relevant. Fig. LABEL:fig:experiments:result:zalando:synthetic exemplifies an optimal outcome, where all relevant products are in the top positions, and the visibility of brands A and B is distributed according to their share of relevant products. Fig. LABEL:fig:experiments:result:zalando:real shows the distribution of relevant (and non-relevant) products from low and high brands as shown by Zalando’s ranking algorithm by the time of data collection. We see that famous brands do indeed have an advantage over less-known brands, as there are almost no products from the low group in the first few bins. Fig. LABEL:fig:experiments:result:zalando:condA—LABEL:fig:experiments:result:zalando:compromise show how the ranking changes after FAIM is applied with different values for $\theta^{A}$, $\theta^{B}$, and $\theta^{C}$. Depending on our preference for calibration, balance for the positive class, or balance for the negative class, we see varying degrees of visibility improvement for the low group. For a better understanding on the effect of FAIM on the rankings, observe Figure 4. In Figure LABEL:fig:experiments:result:zalando:synthetic we show a fictive example for an ideal ranking: imagine a data set with two groups A and B. A has 1,000 items: 200 positive and 800 negative. B has 500 items: 300 positive and 200 negative. If we assume a perfect ranker, we can sort those items simply by their score. When we group them in bins of 50 items each, we would expect the first 10 bins to contain only positive items, and the next 20 bins to have only negative items. Furthermore, the positive bins should display a ratio of 300/200 for groups B and A respectively, while the negative bins should display a ratio of 200/800. Figure LABEL:fig:experiments:result:zalando:real shows the ranking as collected from the Zalando website on 27th October 2021, which by virtue of being real is not ideal. Since each bin contains more than 2,000 items, it is enough to focus on the first bins, as other items are shown very rarely, unless customers are actively looking for them. We see that the first five bins contain mostly high brands, meaning that relevant products from low brands are rarely shown to the customer, unless they start to use product filters or full text search. Again, there are various reasons why a customer would prefer well-known over less- known brands, but if we wanted to (or had to by legislation) mitigate the visibility disparities between high and low brands, FAIM provides a convenient approach (see Section 5.3 for a discussion on FAIM’s legal meaning for an e-commerce scenario such as ours). Figures LABEL:fig:experiments:result:zalando:condA– LABEL:fig:experiments:result:zalando:compromise show rankings after FAIM has been applied with different values for $\theta^{A}$, $\theta^{B}$, and $\theta^{C}$. In all cases, FAIM distributes visibility better across both brand groups, particularly for relevant products, which is the interesting case for an e-commerce platform such as Zalando. ## 5\. Application of the Algorithm in Legal Scenarios Given the _contextual agility_ ((Wachter et al., 2021b)) of EU regulations with respect to the definition of fairness, it is advantageous to introduce a similar flexibility in the operationalization of fairness concepts. FAIM has a wide range of applications in scenarios in which the law compels the decision maker to make decisions in both an accurate _and_ a non-discriminatory way. Technically speaking, this may result in a trade-off between competing fairness and accuracy measures embodying the contemplated criteria, i.e., calibration; balance for the negative class; balance for the positive class. Importantly, the weighting of the different measures via the $\theta$ interpolation parameters is a deeply normative choice which reflects competing visions of what is supposed to be considered fair, and legal, in a certain situation. Depending on the domain of application, the achievement of calibration, the balancing of false positives or of false negatives might be considered most important. FAIM affords the advantage of allowing the decision maker to flexibly adapt to these different scenarios by choosing different $\theta$ weights. The broad discussion of different fairness metrics both in computer science and the law ((Wachter et al., 2021a; Hellman, 2020; Binns, 2018; Friedler et al., 2021; Kim, 2017; Pessach and Shmueli, 2020; Mitchell et al., 2021; Ghoash et al., 2021)) suggests that no single metric such as statistical parity, error balance or equalized odds will fit all contexts. This also applies to the fairness measures under discussion here ((Corbett-Davies and Goel, 2018)). As pointed out before, an important prerequisite for those metrics is the reliability of ground truth ((Bao et al., 2021)). If these data points are collected in a way that systematically disfavors one protective group, conditioning on ground truth data risks perpetuating imbalances included in them ((Bao et al., 2021; Wachter et al., 2021a; Chouldechova, 2017)). Within these constraints, ground truth measures like the ones underlying FAIM can nevertheless be helpful tools in many situations to capture normative desiderata, particularly if used alongside strategies to improve the correctness and representativeness of ground truth. Generally speaking, choosing, for example, between a stronger balance for the negative or the positive class will depend on the respective consequences of misclassification (false negative and false positive predictions). As a result of prediction errors, individual and social costs arise, and legal norms may be violated. These costs and norms will differ widely depending on the deployment context. In the following, we review three examples in which different normative considerations may lead a decision maker to adopt varying weights for the respective accuracy and non-discrimination measures: recidivism prediction, credit scoring, and fair ranking according to the Digital Markets Acts (DMA). ### 5.1. Recidivism Prediction In this paper, we consider recidivism prediction for technical and argumentative reasons: the original papers by Kleinberg et al. (2016) and Chouldechova (2017)showing incompatibility between fairness metrics did use the COMPAS case, and we directly build on their work. Hence, we use COMPAS as an example to clarify how FAIM works. However, when considering recidivism prediction, the first questions that arise from a juridical perspective are whether one legally may, or policy-wise should, use algorithmic tools such as the COMPAS model to assess recidivism risk in Criminal Justice proceedings at all (see also 6.3). We do not wish to make any claim as to the legitimacy of the usage of these tools in this paper; there are, if anything, good reasons to be quite skeptical about it. As mentioned, COMPAS is a proprietary model used in pretrial settings in the US to determine whether potential offenders should be detained until their trial, based on a prediction of recidivism likelihood. Similar models are currently employed in Canada, the UK, and Spain ((Watch, 2019, p. 122)). While the use of such instruments has sparked significant controversy, and critique, in the legal literature ((Starr, 2014; Mayson, 2019; Selbst, 2017; Katyal, 2019; Eaglin, 2017; Huq, 2019; Pruss, 2021)), courts such as the Wisconsin Supreme Court have condoned their use under certain safeguards,777 State v. Loomis, 881 N.W.2d 749 (Wis. 2016). and their deployment is currently on the rise ((Hamilton, 2021); (Watch, 2019, p. 122)). Within the scope of this paper, we cannot offer an in-depth discussion of the promises and perils of algorithmic recidivism risk assessment (see, e.g., (Solow-Niederman et al., 2019; Kleinberg et al., 2018), and 6.3). Rather, we would like to point out that, to the extent that such tools are used at all, they must clearly fulfill minimum requirements seeking to safeguard normative and legal principles of the jurisdictions in which they are deployed. Importantly, the impossibility theorems concerning various fairness metrics mentioned above apply equally in the case of algorithmic and human risk assessments. While the influence of fairness considerations on human risk predictions, for example by a judge, remains difficult to elucidate, FAIM allows to make the relevant trade-offs transparent and to rank the involved fairness metrics in varying degrees of priority. In training the COMPAS model, the developers chose to prioritize calibration, which inevitably led to differing false positive and false negative rates given imperfect prediction and differing base rates between the involved ethnic groups ((Dieterich et al., 2016; Larson et al., 2016)). While accuracy remains significant in recidivism risk assessment—and is actually quite low with the COMPAS algorithm ((Dressel and Farid, 2018))—the core normative trade-off arguably occurs between the respective importance of equalizing false positive and false negative predictions. On the one hand, one could argue that the prevention of false positive outcomes should be prioritized because, under the rule of law, it is of utmost importance not to detain any person without sufficient reason. Under this reading, the greatest weight should be put on aligning false positive rates between groups, so that the burden of being unduly sent to prison is shared equally between the respective groups. On the other hand, one could claim that matching false negative predictions is crucial because they unduly spare individuals time in prison, affording a significant individual advantage to them. Under this reading, that undeserved benefit should not accrue to any protected group to a greater extent.888 Note that the fact false negative predictions constitute a particular risk to individuals and to society at large, with potentially dangerous offenders roaming free, is likely irrelevant here: it should not matter to victims, nor to society, by members of what protected group re- offences are committed. Criterion B) balances the negative classes, but does not (directly) reduce their size. FAIM allows for the establishment of an intermediary position such that both balances, for the positive and the negative class, are fulfilled to an equal degree, even though only partially. To the extent that such instruments are used at all, and that the data they are based on are considered adequate, FAIM may therefore operationalize a policy compromise between factually irreconcilable goals. From a different perspective, however, balancing false positives may be considered more important: losses loom larger than gains. Time in prison constitutes a highly significant restriction of personal freedom, interrupting private lives and careers, potentially endangering physical and mental health. Hence, lawmakers or judges may come to the conclusion that it is more important to share the burden of false positive predictions equally between groups than the unwarranted benefit of false negative predictions. FAIM then allows to prioritize the former criterion. ### 5.2. Credit Scoring Entirely different normative considerations are present in the case of credit scoring. Imagine a bank uses an ML-based scoring algorithm to assess the creditworthiness of loan applicants, as in our synthetic experiment. Increasingly, ML is indeed used for these purposes ((Langenbucher, 2020; Langenbucher and Corcoran, 2021)). Here, accuracy facilitates so-called responsible lending, i.e., loan decisions in which the credit institution intends to ensure that the borrower is not overburdened by the repayment obligations. After the financial crisis of 2008/09, responsible lending has become a cornerstone of financial law. In the EU, for example, the obligation to lend responsibly is enshrined in Article 8 of the Consumer Credit Directive 2008/48/EC, and other EU law instruments install a comprehensive compliance and supervision regime demanding regular audits to ensure the accuracy of credit scoring models (Art. 174 et seqq. of the Capital Requirements Regulation 575/2013, CRR). Hence, accuracy should certainly receive significant weight in the case of credit scoring. As Art. 174(1) CRR puts it, statistical models used by banks need to have ‘good predictive power’. However, with default base rates usually differing between protected groups, high degrees of calibration will lead to an imbalance in the false positive or false negative rates between the groups. A positive label, in credit scoring, means that the loan request is denied, a negative label that it is granted (because the risk of default is low enough). Clearly, false negative predictions may inflict financial damage on the lender if the credit cannot be repaid, but also potentially on the borrower, who may face financial penalties, eviction and future encumbrance due to a negative credit record. False positive predictions, on the other hand, give rise to opportunity costs: the lender does not earn interest payments, and the borrower does not obtain access to credit. Disparities concerning false negative or false positive rates affect access to credit or default rates between protected groups, and are therefore relevant for compliance with non-discrimination legislation. In the EU, rules on indirect discrimination forbid even seemingly neutral practices putting protected groups at a particular disadvantage, unless these differences can be justified.999 See, e.g., Art. 2(b) of Directive 2004/113/EC; Art. 2(2)(b) of Directive 2000/43/EC. Higher false positive rates clearly constitute a disadvantage for the affected group; even higher false negative rates can be viewed as a burden on the more often wrongly classified group, however, as they entail the concrete risk of default. Since accuracy and error rates cannot be balanced between protected groups at the same time under non-trivial conditions, the law will have to accept reasonable trade- offs between these competing and mutually exclusive obligations. It cannot and does not demand what is impossible (nemo ultra posse obligatur). From the perspective of legal doctrine, this will arguably take the route of the justification of a possible disadvantage. The damage inflicted must then be proportionate, given the reasons which can be advanced for prioritizing the other fairness criteria. Given these preconditions, equalizing false negative credit predictions might be considered more important in the case of high-stakes loans (e.g., large sums and little collateral). Here, a default would be particularly disruptive, and hence that burden should be shared equally between protected groups. As a consequence, the $\theta$ weight of the balance for the positive class (Criterion C) should be higher than that for the negative class (Criterion B). Conversely, in the case of low-stakes loans (e.g., low credit volume; sufficient collateral besides basic necessities of the borrower, e.g. beyond the house a family lives in; or limited personal liability in case of default), false positive predictions might be more deleterious than false negative ones because the damage is limited if the borrower defaults on the credit. Wrongfully denying access to credit may then have larger opportunity costs. Under such circumstances, the $\theta$ weight of the balance for the negative class (Criterion B) should be higher than that for the positive class (Criterion C). As an example, calibration could be set to 0.5, balance for the negative class to 0.35, and balance for the positive class to 0.15. Legally speaking, the fact that an unequal rate of false negative predictions persists will then arguably be justified by the overriding importance of aligning false positive predictions and calibration between groups. ### 5.3. Fairness in E-Commerce Rankings Finally, our method may be used to implement emerging legal notions of fairness in e-commerce rankings. Particularly in EU law, a growing number of provisions aim to safeguard the impartiality of rankings in online contexts. At the most general level, these rules have shifted from the prohibition of self-preferencing via a focus on transparency to, so far, under-researched concepts of fairness in the Digital Markets Act (DMA). To the best of our knowledge, our contribution constitutes the first attempt to operationalize the fair ranking provisions of the DMA on both a technical and a legal basis. #### 5.3.1. Competition Law The oldest and best-known rule of fairness in e-commerce rankings is derived from the prohibition, in general competition law, to abuse a dominant position (Art. 102 TFEU in EU law). It has long been argued that many dominant online undertakings, such as Google or Amazon, need to be scrutinized under this provision due to their dual role as providers of online marketplaces – on which third parties directly sell their goods to consumers – and as direct sellers of goods on their platforms ((Padilla et al., 2020; Graef, 2019)). Hence, such platforms may be considered competitors of the very business customers they serve on their marketplaces (vertical integration). As a consequence, dominant online platforms must not unduly preference their own offers vis-à-vis those of their business customers in rankings they display as a result of consumer search queries. This rule has led to some of the most spectacular fines in recent EU competition law. For example, the General Court of the European Union, in November 2021, affirmed the 2017 decision by the EU Commission to fine Google with €2.4 billion for engaging in self-preferencing in its online comparison shopping service.101010GCEU, Case T-612/17 (Google Shopping). A similar issue is at stake in the Amazon buy box case, in which the Italian Competition Authority, in December 2021, imposed a record fine of €1.1 billion on several Amazon companies for tying access to the buy box to the use of Amazon’s own logistics channel (Fulfillment by Amazon).111111Italian Competition Authority, Case A528, Amazon, Press release, 9 December 2021. The upshot of these rulings is that dominant platforms which serve a dual role of marketplace and seller must abstain from systematically tweaking their rankings to their own benefit. This rule against self-preferencing, deduced from Art. 102 TFEU and pertinent case law, may be considered a first substantive fairness element for the order of the ranking itself ((Podszun and Bongartz, 2021)). However, it only applies to undertakings which dominate a certain market. #### 5.3.2. Transparency Often, though, it is difficult for outsiders, even for the business customers, to determine how these rankings come about. Two new provisions therefore install explicit transparency provisions for rankings in EU law ((Hacker and Passoth, 2022; Eifert et al., 2021)). First, Art. 5 of the so-called P2B (Platform to Business) Regulation (EU) 2019/1150 obliges online intermediaries and search engines, independent of their market power, to disclose the main parameters of ranking and their relative importance. The provision has been in effect since July 2020. It is supposed to foster the predictability and understanding of the ranking for business users, and to foster competition between different intermediaries with respect to the ranking parameters (Recital 24 of the P2B Regulation). In a similar fashion, second, the new Art. 6a of the Consumer Rights Directive (CRD) has required online marketplaces from the end of May 2022 on to divulge to consumers the main parameters for rankings based on consumer search queries as well as their relative importance. According to the new Art. 2(1)(n) of the Unfair Commercial Practices Directive, ‘online marketplace’ denotes any software operated by or on behalf of a trader which allows consumers to conclude distance contracts with other traders or consumers. Therefore, the rule does not apply to companies only selling goods directly to consumers, but is again independent of market power. Essentially, scholars have argued, these provisions introduce an obligation for the global explanation of the ranking model ((Hacker and Passoth, 2022; Grochowski et al., 2021)). It may be fulfilled, for example, by using interpretable machine learning models (e.g., linear or logistic regression ((Rudin, 2019))) or by using global post-hoc explanations for black box models (e.g., DNNs) such as an average over local SHAP values ((Lundberg and Lee, 2017)) or SpRAy ((Lapuschkin et al., 2019)). However, the provisions do not add any specific, substantive fairness criteria concerning the order of the ranking itself. #### 5.3.3. The EU Digital Markets Act The DMA, which will come into effect in the EU in 2023, takes these existing provisions yet one step further by introducing a novel fairness condition for rankings. Its importance is hard to overstate: failure to comply with the DMA provisions may be sanctioned by fines of up to 10% of the global annual turnover of the offender (Art. 30(1) DMA), and up to 20% in the case of a repeated violation (Art. 30(2) DMA). The DMA includes a whole range of provisions aimed at strengthening contestable markets and fair competition in online environments by designing specific rules applicable to so-called gatekeepers, in essence large online platforms. ##### Fairness in the DMA. Art. 6(5) DMA specifically tackles rankings,121212 In the final version, it applies, beyond rankings, also to indexing and crawling. including when delivered via a virtual assistant (Recital 51 DMA; for example, a personal voice assistant). Art. 6(5) DMA first reiterates the prohibition of self- preferencing known from general antitrust law.131313 For differences to Art. 102 TFEU, see, e.g., (Eifert et al., 2021). The second sentence of Art. 6(5) DMA starts by specifying that rankings need to be transparent. As Recital 52 suggests, this could refer to the transparency provisions of the P2B Regulation. Importantly, Art. 6(5) DMA goes on to demand that gatekeepers must, in general, apply ‘fair and non- discriminatory conditions’ to their rankings. In the legal space, this has sparked a discussion on how these criteria may be interpreted ((Brouwer, 2021; Hacker, 2022b)).141414For a discussion of non-discriminatory rankings in general competition law, see Graef (2019). So far, however, no clear guidance or consensus has emerged on what these additional fairness requirements for rankings under the DMA could mean for gatekeepers. This raises the question of how the obligation to apply fair and non- discriminatory conditions to _rankings_ could be interpreted. We cannot explore this conceptual issue in depth in this paper. If, however, courts concluded in the future that fairness and non-discrimination under the DMA clause was somehow linked to accuracy, false positives and false negatives, we can smoothly operationalize the DMA fairness condition with FAIM. This is significant insofar as previous discussions of fairness in algorithmic contexts have invariably focused on non-discrimination law ((Zehlike et al., 2017; Asudeh et al., 2019; Singh and Joachims, 2018; Zehlike et al., 2022b)), and not on fairness conditions in e-commerce regulation, such as the DMA.151515The DMA applies to a range of digital services, so-called core platform services (Art. 2(2) DMA). These include social media and video- sharing services, for example; online intermediation, typical for e-commerce, is another core platform service according to Art. 2(2)(a) DMA. ##### DMA Ranking with FAIM. FAIM allows to specifically and flexibly re-order gatekeeper rankings such that certain conditions are prioritized, depending on the exact, and as of yet unknown, binding interpretation of Art. 6(5) DMA. In the case of e-commerce rankings, one hypothetical interpretation by courts could be that accuracy as well as an equal false positive rate are the most important parameters among the three conditions. Under such a reading, false negative predictions might be considered less important than false positive predictions as most consumers will not even see or engage with those items which are ranked at low positions. If a platform, preemptively or as a result of a court or agency order, wishes to avoid that criteria such as incumbent status play a role in the top-ranked products, $\theta$ values can be set accordingly. They could, for example, convey high priority to equalizing the false positive rate (Criterion B) between high-visibility and low-visibility brands.161616To be sure, there may be valid reasons for relegating certain brands to low- visibility status, for example reasons relating to product quality. Nonetheless, one may wish to equalize error rates between the groups. For instance, platforms may seek to limit false negative rates in the low group since even such brands may, eventually, produce some high-quality items; see also next para. A second hypothetical interpretation would focus on false negatives. False negative predictions unduly demote products in the ranking and thus deprive affected items of visibility and hence of transactions. If a regulatory agency or court found that the false-negative rate needs to be balanced between certain groups, FAIM can again be tuned to implement such a requirement by choosing corresponding $\theta$ values (high values for Criterion C). In future work, we will seek to identify further distinctions which may be problematic under the DMA fairness clause and which may be remedied by FAIM. Importantly, the fairness clause in the DMA provides an opportunity to reflect upon fairness criteria in e-commerce rankings beyond general non- discrimination law and the prohibition of self-preferencing in competition law. ## 6\. Discussion and Limitations ### 6.1. Normative Choices and Applicability To explicitly name the normative choices FAIM implies, we use the concepts by Friedler et al. (2021) of two different metric spaces— _construct space_ and _observable space_. Construct space denotes an idealized reality in which all feature values are correctly registered, and a decision is based on those features. However, decision makers usually do not have access to this realm of an objective truth. Rather, they need to base their model on observable space, which contains the feature values the decision maker has access to via measurement., and the decisions based on them. Importantly, ground truth may correctly reflect, or at least closely approximate, construct space—or it may not. This implies a first limitation for our method. FAIM operates in observable space, and does not undo any harm that might arise from a potentially biased mapping between construct and observable space, e.g., by systematic measurement errors. This is because we obtain our ground truth labels from observing behavior of scored individuals, such as: “did someone who was labeled creditworthy actually repay the loan?” Such observations naturally incorporate historic and on-going discrimination, as they usually do not ask: “If a person who was labeled creditworthy did not repay the loan, was that because she can genuinely not handle money, or because she belongs to a disadvantaged group?” To be sure, this question is crucial for the societal mitigation of discrimination in the long term. However, with the data we have at our disposition, we usually cannot answer it exhaustively. Hence, FAIM must generally content itself with correcting unequal calibration or disparate error rates, defined by a divergence from ground truth, that disproportionately affect a certain group of individuals. As such disparities often correlate with the membership of a disadvantaged group in society, it may also correct discrimination caused by skewed mappings between construct and observable space to a certain extend (as we have seen in our experimental results). However, this is only a side-effect and depends on the nature of discrimination that is present in the data. If ground truth is close to reality, i.e., to construct space, then we can cure biased mappings between construct and observable space; otherwise, not. For the same reasons, FAIM addresses _technical_ and _emergent_ , but not _pre-existing_ bias, as defined by Friedman and Nissenbaum (1996). We therefore advise FAIM to be used in settings where ground truth observations are sufficiently reliable (though this is a necessary condition for any data-driven decision making), and _it can be assumed that groups can be meaningfully compared by the respective measures of qualification or quality_. As an example for a non-comparable qualification measure across genders, consider the h-index ((Hirsch, 2005)), which suffers from strong gender bias. Research shows that women, for the same amount and quality of research, receive significantly lower h-index values because it takes them longer to publish, and their work is less cited than men’s ((Caplar et al., 2017)). In cases where groups cannot be meaningfully compared, a method that implements statistical parity or other affirmative action policies independent of ground truth (such as FA*IR by Zehlike et al. (2017, 2022a)) is an appropriate choice (see also (Wachter et al., 2021a)). ### 6.2. Limitations of FAIM ##### Continuous Mathematical Framework. As stated, we work with a _continuous_ scale of possible scores (in the real interval $[0,1]$) and accordingly with continuous probability distributions. In practice, of course, there will be only a discrete set of possible scores, and a finite set of individuals under scrutiny. The reason we work in a continuous setting is of a mathematical nature, in that it allows us to use certain tools from the theory of optimal transport, which in a discrete setting would complicate the theory considerably. It is common and reasonable practice to approximate discrete models by continuous ones, as long as the number of individuals considered is large and their scores are sufficiently spread out over a finite scale. Thus, our framework would not be suitable for a situation where there are only very few possible score values, e.g., only three or four. Further discussion in this direction, for a related problem, can be found in Zehlike et al. (2020). ##### Discontinuities in score distribution. In criterion B) (we use this case as an example to explain the problem, but the same applies to criterion C), we only balance the scores for the negative class and ignore the true positives. This means that the scores of the true positives remain unchanged by FAIM. This might become a problem for rare scores of true negative individuals for whom there was no example in the data when calculating the optimal transport (OT) matrix. Remember that our algorithm calculates an optimal transport matrix to translate raw scores into fair scores for _previously unseen_ individuals. This means that for new data points, we do not know the ground truth label. Instead, FAIM consults the OT matrix that corresponds to a certain group and setting of $\theta$ and checks if it finds a raw score entry in that matrix. In such a case, the following scenario might arise when trying to optimize for criterion B) only: Suppose a true negative person’s score lies at the edge of the overall true negative score distribution, i.e., only few others (in the world) have the same score while being true negative. In addition, critically, no true negative person had this score when the OT matrix was computed. In this case, the algorithm does not have an entry for translation in the OT matrix. As a result, this individual would be considered true positive, and their score would be left untouched. ### 6.3. Critique of using COMPAS as an experimental scenario Bao et al. (2021) raise the problem that Risk Assessment Instrument (RAI) data sets such as COMPAS are used for benchmarking fairness methods in a purely technical fashion without recognizing the domain- and data-specific concerns of Criminal Justice (CJ). As a result, technical fairness method contributions could be interpreted as substantive interventions in the CJ domain. However, the authors rightly argue, the latter presupposes careful considerations of the domain specifics that may invalidate such results. They raise several serious concerns about COMPAS: outcome label bias due to construct invalidity; measurement bias that cannot be calibrated due to lack of (reliable) ground truth; protected attribute measurement bias; racial bias in covariates; and distribution issues due to label selection bias. We recognize and acknowledge the concerns raised by Bao et al. However, since our contribution is an improvement on a previously published method presented on the COMPAS data set, we deem it important to enable a comparison of our results with the original. Thus, we publish the results from our experiments on the COMPAS data set only as a technical comparison with the results in Kleinberg et al. (2016), and under the caveat that reliable ground truth is not available for the COMPAS data set. Hence, the results on applying FAIM to COMPAS are not intended as a contribution to real-world outcomes in the criminal justice domain. Rather, we discuss how FAIM could be used in a pretrial setting _if_ the model and data were considered adequate. ### 6.4. Data Missingness Data missingness is a validity concern for the study of any fairness criterion, as noted by Goel et al. (2021), and is not unique for FAIM. As is the case for many fairness criteria, FAIM assumes the availability of ground truth data that is representative of the underlying distribution, and the absence of systematic censorship ((Kallus and Zhou, 2018)) in the data generation process. If systematic censorship occurs due to biased decision making, then this prevents us from sampling from the ground truth distribution needed for an unbiased fairness criterion, thus making evaluation problematic. Concerning the three data sets used in this paper: as mentioned, we use the COMPAS data set to allow a direct comparison with the metric proposed by Kleinberg et al. (2016), and we thus do not want to modify the data set, for example through recovery or imputation. As shown by Goel et al. (2021), it is possible to recover parts of the distribution on certain assumptions about the causal structure of the decision-making process, but making such assumptions would not allow comparison with the results of Kleinberg et al. (2016) or allow us to make any general conclusions about applying FAIM to COMPAS. The e-commerce data set is subject to systematic censorship in the sense that articles classified as ”low visibility” would receive less exposure and thus do not have opportunities to gain more exposure. Modifying the exploration- exploitation strategy for the learning to rank model would allow us to sample from the missing data to a larger degree, but performing such experiments would be prohibitively costly. Finally, to further investigate the effect of data missingness on FAIM, we could introduce synthetic missingness to the synthetic data set, but would have to do so under the assumption of different causal structures, which are in an undefined and infinite search space. Thus, it is unclear what the relevant causal structures would be that would allow us to make meaningful conclusions. ## 7\. Conclusion In this paper, we have presented FAIM, a re-scoring algorithm that interpolates between three important criteria from the literature on fairness in machine learning: calibration, balance for the negative class, and balance for the positive class. Previous research has shown that they are mutually exclusive under reasonable factual assumptions. However, in many practical and legal scenarios, it is unsatisfactory to fully prioritize one criterion while neglecting the other two. FAIM allows to find an optimal compromise between the three criteria through three per-group parameters $\theta^{A},\theta^{B}$, and $\theta^{C}$ using optimal transport theory. The parameter settings can be flexibly adapted to various scenarios. In this way, decision makers can balance competing interests and rights of protected groups; and they may adapt one model to different legal requirements. We presented mathematical theory together with a pseudocode implementation to help readers with different backgrounds to follow our approach. The model was tested in extensive experiments on real-world data sets from the criminal justice and the e-commerce domain. Our experiments have shown that FAIM is effective under all scenarios, but that an inherently flawed model or data set should not be “fixed” via fairness constraints. As such, we want to stress that our model is not a means of ethical reasoning, i.e., it can not determine whether a candidate should or should not receive bail or credit, or whether a product should be included in a top-$k$ ranking. Instead, it ensures an optimal compromise for mutually exclusive fairness constraints according to the desires of a decision maker, expressed through the choices of $\theta^{A},\theta^{B}$, and $\theta^{C}$. However, we also stressed that any choice of $\theta$ values must, of course, remain within the confines of the law. Hence, we discussed how FAIM may be harnessed to comply with competing legal non-discrimination obligations in three concrete settings corresponding to our experiments: credit scoring, bail decisions, and emerging notions of fairness in e-commerce rankings. The discussion shows that our model can handle both traditional legal fields and novel obligations under the recently enacted DMA. To our knowledge, our paper is the first to explore the DMA from both a technical and a legal perspective. ## 8\. Acknowledgments We want to thank Geraud Le Falher and Pak-Hang Wong for their support and insightful suggestions to improve the paper. The authors do not have any funding to disclose other than their affiliation. ## References * (1) * Agueh and Carlier (2011) Martial Agueh and Guillaume Carlier. 2011. Barycenters in the Wasserstein space. _SIAM Journal on Mathematical Analysis_ 43, 2 (2011), 904–924. * Ambrosio and Gigli (2013) Luigi Ambrosio and Nicola Gigli. 2013. A user’s guide to optimal transport. In _Modelling and optimisation of flows on networks_. Springer, 1–155. * Angwin et al. (2016) Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner. 2016. Machine bias. _ProPublica, May_ 23, 2016 (2016), 139–159. * Asudeh et al. (2019) Abolfazl Asudeh, HV Jagadish, Julia Stoyanovich, and Gautam Das. 2019. Designing fair ranking schemes. In _Proceedings of the 2019 International Conference on Management of Data_. 1259–1276. * Bao et al. (2021) Michelle Bao, Angela Zhou, Samantha Zottola, Brian Brubach, Sarah Desmarais, Aaron Horowitz, Kristian Lum, and Suresh Venkatasubramanian. 2021. It’s COMPASlicated: The Messy Relationship between RAI Datasets and Algorithmic Fairness Benchmarks. _arXiv preprint arXiv:2106.05498_ (2021). * Barocas and Selbst (2016) Solon Barocas and Andrew D Selbst. 2016. Big data’s disparate impact. _Calif. L. Rev._ 104 (2016), 671. * Bellogín et al. (2017) Alejandro Bellogín, Pablo Castells, and Iván Cantador. 2017\. Statistical biases in information retrieval metrics for recommender systems. _Information Retrieval Journal_ 20, 6 (2017), 606–634. * Berk et al. (2021) Richard Berk, Hoda Heidari, Shahin Jabbari, Michael Kearns, and Aaron Roth. 2021. Fairness in criminal justice risk assessments: The state of the art. _Sociological Methods & Research_ 50, 1 (2021), 3–44. * Binns (2018) Reuben Binns. 2018\. Fairness in Machine Learning: Lessons from Political Philosophy. In _Conference on Fairness, Accountability and Transparency_. 149–159. * Bono et al. (2021) Teresa Bono, Karen Croxson, and Adam Giles. 2021\. Algorithmic fairness in credit scoring. _Oxford Review of Economic Policy_ 37, 3 (2021), 585–617. * Brenier (1987) Yann Brenier. 1987\. Décomposition polaire et réarrangement monotone des champs de vecteurs. _CR Acad. Sci. Paris Sér. I Math._ 305 (1987), 805–808. * Brenier (1991) Yann Brenier. 1991\. Polar factorization and monotone rearrangement of vector-valued functions. _Communications on pure and applied mathematics_ 44, 4 (1991), 375–417. * Brouwer (2021) Dennis Jeroen Brouwer. 2021\. Towards a ban of discriminatory rankings by digital gatekeepers?: Reflections on the proposal for a Digital Markets Act. _Internet policy review_ (2021). * Cabral et al. (2021) Luis Cabral, Justus Haucap, Geoffrey Parker, Georgios Petropoulos, Tommaso M Valletti, and Marshall W Van Alstyne. 2021. The EU digital markets act: a report from a panel of economic experts. _Publications Office of the European Union_ (2021). * Calders and Žliobaitė (2013) Toon Calders and Indrė Žliobaitė. 2013\. Why unbiased computational processes can lead to discriminative decision procedures. In _Discrimination and privacy in the information society_. Springer, 43–57. * Caplar et al. (2017) Neven Caplar, Sandro Tacchella, and Simon Birrer. 2017\. Quantitative evaluation of gender bias in astronomical publications from citation counts. _Nature astronomy_ 1, 6 (2017), 1–5. * Chiappa et al. (2020) Silvia Chiappa, Ray Jiang, Tom Stepleton, Aldo Pacchiano, Heinrich Jiang, and John Aslanides. 2020\. A General Approach to Fairness with Optimal Transport.. In _AAAI_. 3633–3640. * Chiappa and Pacchiano (2021) Silvia Chiappa and Aldo Pacchiano. 2021. Fairness with Continuous Optimal Transport. _arXiv preprint arXiv:2101.02084_ (2021). * Chouldechova (2017) Alexandra Chouldechova. 2017\. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. _Big data_ 5, 2 (2017), 153–163. * Corbett-Davies and Goel (2018) Sam Corbett-Davies and Sharad Goel. 2018. The measure and mismeasure of fairness: A critical review of fair machine learning. _arXiv preprint arXiv:1808.00023_ (2018). * Dawid (1982) A Philip Dawid. 1982\. The well-calibrated Bayesian. _J. Amer. Statist. Assoc._ 77, 379 (1982), 605–610. * De Vries (2013) Sybe A De Vries. 2013\. Balancing fundamental rights with economic freedoms according to the European Court of Justice. _Utrecht Law Review_ 9, 1 (2013), 169–192. * Dieterich et al. (2016) William Dieterich, Christina Mendoza, and Tim Brennan. 2016\. COMPAS risk scales: Demonstrating accuracy equity and predictive parity. _Northpointe Inc._ (2016). * Dignum (2019) Virginia Dignum. 2019\. _Responsible artificial intelligence: how to develop and use AI in a responsible way_. Springer Nature. * Dressel and Farid (2018) Julia Dressel and Hany Farid. 2018. The accuracy, fairness, and limits of predicting recidivism. _Science advances_ 4, 1 (2018), eaao5580. * Dwork et al. (2012) Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. 2012. Fairness through awareness. _Proceedings of the 3rd innovations in theoretical computer science conference_ , 214–226. * Eaglin (2017) Jessica M. Eaglin. 2017\. Constructing Recidivism Risk. _Emory Law Journal_ 67, 1 (2017), 59–122. * Ebers et al. (2021) Martin Ebers, Veronica RS Hoch, Frank Rosenkranz, Hannah Ruschemeier, and Björn Steinrötter. 2021\. The European commission’s proposal for an artificial intelligence act—a critical assessment by members of the robotics and AI law society (RAILS). _J_ 4, 4 (2021), 589–603. * Eifert et al. (2021) Martin Eifert, Axel Metzger, Heike Schweitzer, and Gerhard Wagner. 2021. Taming the Giants: The DMA/DSA Package. _Common Market Law Review_ 58, 4 (2021), 987–1028. * Feldman et al. (2015) Michael Feldman, Sorelle A Friedler, John Moeller, Carlos Scheidegger, and Suresh Venkatasubramanian. 2015\. Certifying and removing disparate impact. In _proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining_. 259–268. * Friedler et al. (2016) Sorelle A. Friedler, Carlos Scheidegger, and Suresh Venkatasubramanian. 2016. On the (im)possibility of fairness. _arXiv preprint arXiv:1609.07236_ (2016). * Friedler et al. (2021) Sorelle A. Friedler, Carlos Scheidegger, and Suresh Venkatasubramanian. 2021. The (Im)possibility of Fairness: Different Value Systems Require Different Mechanisms For Fair Decision Making. _Commun. ACM_ 64, 4 (2021), 136–143. * Friedman and Nissenbaum (1996) Batya Friedman and Helen Nissenbaum. 1996. Bias in computer systems. _ACM Transactions on Information Systems (TOIS)_ 14, 3 (1996), 330–347. * Gerards and Borgesius (2022) Janneke Gerards and Frederik Zuiderveen Borgesius. 2022\. Protected Grounds and the System of Non-Discrimination Law in the Context of Algorithmic Decision-Making and Artificial Intelligence. _Colo. Tech. LJ_ 20 (2022), 1. * Ghoash et al. (2021) Bishwamittra Ghoash, Debabrota Basu, and Kuldeep S. Meel. 2021\. Algorithmic Fairness Verification with Graphical Models. _arXiv preprint arXiv:2109.09447_ (2021). * Gillis and Spiess (2019) Talia B Gillis and Jann L Spiess. 2019. Big data and discrimination. _The University of Chicago Law Review_ 86, 2 (2019), 459–488. * Goel et al. (2021) Naman Goel, Alfonso Amayuelas, Amit Deshpande, and Amit Sharma. 2021. The Importance of Modeling Data Missingness in Algorithmic Fairness: A Causal Perspective. In _Proceedings of the AAAI Conference on Artificial Intelligence_ , Vol. 35. 7564–7573. * Gordaliza et al. (2019) Paula Gordaliza, Eustasio Del Barrio, Gamboa Fabrice, and Jean-Michel Loubes. 2019. Obtaining fairness using optimal transport theory. In _International Conference on Machine Learning_. PMLR, 2357–2365. * Graef (2019) Inge Graef. 2019\. Differentiated treatment in platform-to-business relations: EU competition law and economic dependence. _Yearbook of European Law_ 38 (2019), 448–499. * Grochowski et al. (2021) Mateusz Grochowski, Agnieszka Jabłonowska, Francesca Lagioia, and Giovanni Sartor. 2021\. Algorithmic Transparency and Explainability for EU Consumer Protection: Unwrapping the Regulatory Premises. _Critical Analysis of Law_ 8 (2021), 43–63. * Grodzinsky et al. (2011) Frances S Grodzinsky, Keith W Miller, and Marty J Wolf. 2011\. Developing artificial agents worthy of trust:“Would you buy a used car from this artificial agent?”. _Ethics and information technology_ 13, 1 (2011), 17–27. * Hacker (2018) Philipp Hacker. 2018\. Teaching fairness to artificial intelligence: existing and novel strategies against algorithmic discrimination under EU law. _Common Market Law Review_ 55, 4 (2018). * Hacker (2021) Philipp Hacker. 2021\. A legal framework for AI training data—from first principles to the Artificial Intelligence Act. _Law, Innovation and Technology_ 13, 2 (2021), 257–301. * Hacker (2022a) Philipp Hacker. 2022a. The European AI Liability Directives – Critique of a Half-Hearted Approach and Lessons for the Future. _arXiv preprint arXiv:2211.13960_ (2022). * Hacker (2022b) Philipp Hacker. 2022b. KI und DMA – Zugang, Transparenz und Fairness für KI-Modelle in der digitalen Wirtschaft. _Gewerblicher Rechtsschutz und Urheberrecht_ 75 (2022), 1278–1285. * Hacker and Passoth (2022) Philipp Hacker and Jan-Hendrik Passoth. 2022. Varieties of AI Explanations Under the Law. From the GDPR to the AIA, and Beyond. In _International Workshop on Extending Explainable AI Beyond Deep Models and Classifiers_. Springer, 343–373. * Hamilton (2021) Melissa Hamilton. 2021\. Evaluating Algorithmic Risk Assessment. _New Criminal Law Review_ 24, 2 (2021), 156–211. * Hardt et al. (2016) Moritz Hardt, Eric Price, and Nati Srebro. 2016. Equality of opportunity in supervised learning. _Advances in neural information processing systems_ 29 (2016), 3315–3323. * Hellman (2020) Deborah Hellman. 2020\. Measuring Algorithmic Fairness. _Virginia Law Review_ 106 (2020), 811–866. * Hertweck and Räz (2022) Corinna Hertweck and Tim Räz. 2022. Gradual (In) Compatibility of Fairness Criteria. In _AAAI Conference on Artificial Intelligence_. * Hirsch (2005) Jorge E Hirsch. 2005\. An index to quantify an individual’s scientific research output. _Proceedings of the National academy of Sciences_ 102, 46 (2005), 16569–16572. * Huq (2019) Aziz Z. Huq. 2019\. Racial Equity in Algorithmic Criminal Justice. _Duke Law Journal_ 68, 6 (2019), 1043–1134. * Järvelin and Kekäläinen (2002) Kalervo Järvelin and Jaana Kekäläinen. 2002. Cumulated gain-based evaluation of IR techniques. _ACM Transactions on Information Systems (TOIS)_ 20, 4 (2002), 422–446. * Jiang et al. (2012) Xiaoqian Jiang, Melanie Osl, Jihoon Kim, and Lucila Ohno-Machado. 2012. Calibrating predictive model estimates to support personalized medicine. _Journal of the American Medical Informatics Association_ 19, 2 (2012), 263–274. * Kallus and Zhou (2018) Nathan Kallus and Angela Zhou. 2018. Residual unfairness in fair machine learning from prejudiced data. In _International Conference on Machine Learning_. PMLR, 2439–2448. * Katyal (2019) Sonia K. Katyal. 2019\. Private Accountability in the Age of Artificial Intelligence. _UCLA Law Review_ 66, 1 (2019), 54–141. * Kim (2022) Pauline Kim. 2022\. Race-aware algorithms: Fairness, nondiscrimination and affirmative action. _California Law Review, Forthcoming_ (2022). * Kim (2017) Pauline T. Kim. 2017\. Auditing Algorithms for Discrimination. _University of Pennsylvania Law Review Online_ 166 (2017), 189–203. * Kleinberg et al. (2018) Jon Kleinberg, Himabindu Lakkaraju, Jure Leskovec, Jens Ludwig, and Sendhil Mullainathan. 2018\. Human decisions and machine predictions. _The quarterly journal of economics_ 133, 1 (2018), 237–293. * Kleinberg et al. (2016) Jon Kleinberg, Sendhil Mullainathan, and Manish Raghavan. 2016\. Inherent trade-offs in the fair determination of risk scores. _arXiv preprint arXiv:1609.05807_ (2016). * Langenbucher (2020) Katja Langenbucher. 2020\. Responsible AI-based credit scoring–a legal framework. _European Business Law Review_ 31, 4 (2020). * Langenbucher and Corcoran (2021) Katja Langenbucher and Patrick Corcoran. 2021. Responsible AI Credit Scoring–A Lesson from Upstart. com. In _Digital Finance in Europe: Law, Regulation, and Governance_. De Gruyter, 141–180. * Lapuschkin et al. (2019) Sebastian Lapuschkin, Stephan Wäldchen, Alexander Binder, Grégoire Montavon, Wojciech Samek, and Klaus-Robert Müller. 2019\. Unmasking Clever Hans predictors and assessing what machines really learn. _Nature communications_ 10, 1 (2019), 1–8. * Larson et al. (2016) Jeff Larson, Surya Mattu, Lauren Kirchner, and Julia Angwin. 2016\. How we analyzed the COMPAS recidivism algorithm. _ProPublica, Technical Report_ (2016). * Laux et al. (2021) Johann Laux, Sandra Wachter, and Brent Mittelstadt. 2021\. Taming the few: Platform regulation, independent audits, and the risks of capture created by the DMA and DSA. _Computer Law & Security Review_ 43 (2021), 105613. * Laux et al. (2022) Johann Laux, Sandra Wachter, and Brent Mittelstadt. 2022\. Trustworthy Artificial Intelligence and the European Union AI Act: On the Conflation of Trustworthiness and the Acceptability of Risk. _Available at SSRN 4230294_ (2022). * Lundberg and Lee (2017) Scott M Lundberg and Su-In Lee. 2017. A unified approach to interpreting model predictions. _Advances in neural information processing systems_ 30 (2017). * Mayson (2019) Sandra G. Mayson. 2019\. Bias in, Bias out. _Yale Law Journal_ 128, 8 (2019), 2218–2301. * Mitchell et al. (2021) Shira Mitchell, Eric Potash, Solon Barocas, Alexander D’Amour, and Kristian Lum. 2021\. Algorithmic Fairness: Choices, Assumptions, and Definitions. _Annual Review of Statistics and Its Application_ 8 (2021), 141–163. * Morijn (2006) John Morijn. 2006\. Balancing fundamental rights and common market freedoms in Union law: Schmidberger and Omega in the light of the European Constitution. _European Law Journal_ 12, 1 (2006), 15–40. * Nachbar (2020) Thomas B Nachbar. 2020\. Algorithmic fairness, algorithmic discrimination. _Fla. St. UL Rev._ 48 (2020), 509. * Padilla et al. (2020) Jorge Padilla, Joe Perkins, and Salvatore Piccolo. 2020\. Self-preferencing in markets with vertically-integrated gatekeeper platforms. _Available at SSRN 3701250_ (2020). * Perc (2014) Matjaž Perc. 2014\. The Matthew effect in empirical data. _Journal of The Royal Society Interface_ 11, 98 (2014), 20140378\. * Pessach and Shmueli (2020) Dana Pessach and Erez Shmueli. 2020. Algorithmic Fairness. _arXiv preprint arXiv:2001.09784_ (2020). * Pleiss et al. (2017) Geoff Pleiss, Manish Raghavan, Felix Wu, Jon Kleinberg, and Kilian Q Weinberger. 2017. On Fairness and Calibration. In _Advances in Neural Information Processing Systems_ , I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Eds.), Vol. 30. Curran Associates, Inc. https://proceedings.neurips.cc/paper/2017/file/b8b9c74ac526fffbeb2d39ab038d1cd7-Paper.pdf * Podszun and Bongartz (2021) Rupprecht Podszun and Philipp Bongartz. 2021. The Digital Markets Act: Moving from Competition Law to Regulation for Large Gatekeepers. _Journal of European Consumer and Market Law_ 10, 2 (2021), 60–67. * Pruss (2021) Dasha Pruss. 2021\. Mechanical Jurisprudence and Domain Distortion: How Predictive Algorithms Warp the Law. _Philosophy of Science_ 88, 5 (2021), 1101–1112. * Raghavan et al. (2020) Manish Raghavan, Solon Barocas, Jon Kleinberg, and Karen Levy. 2020. Mitigating bias in algorithmic hiring: Evaluating claims and practices. In _Proceedings of the 2020 conference on fairness, accountability, and transparency_. 469–481. * Rudin (2019) Cynthia Rudin. 2019\. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. _Nature Machine Intelligence_ 1, 5 (2019), 206–215. * Selbst (2017) Andrew D. Selbst. 2017\. Disparate Impact in Big Data Policing. _Gerogia Law Review_ 52, 1 (2017), 109–196. * Selbst and Barocas (2022) Andrew D Selbst and Solon Barocas. 2022. Unfair Artificial Intelligence: How FTC Intervention Can Overcome the Limitations of Discrimination Law. _UCLA School of Law, Public Law Research Paper_ 24 (2022). * Singh and Joachims (2018) Ashudeep Singh and Thorsten Joachims. 2018. Fairness of exposure in rankings. In _Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining_. 2219–2228. * Solow-Niederman et al. (2019) Alicia Solow-Niederman, YooJung Choi, and Guy Van den Broeck. 2019\. The institutional life of algorithmic risk assessment. _Berkeley Tech. LJ_ 34 (2019), 705. * Starr (2014) Sonja B. Starr. 2014\. Evidence-Based Sentencing and the Scientific Rationalization of Discrimination. _Stanford Law Review_ 66 (2014), 803–872. * Thiebes et al. (2021) Scott Thiebes, Sebastian Lins, and Ali Sunyaev. 2021. Trustworthy artificial intelligence. _Electronic Markets_ 31, 2 (2021), 447–464. * Veale and Binns (2017) Michael Veale and Reuben Binns. 2017. Fairer machine learning in the real world: Mitigating discrimination without collecting sensitive data. _Big Data & Society_ 4, 2 (2017), 2053951717743530\. * Veale and Borgesius (2021) Michael Veale and Frederik Zuiderveen Borgesius. 2021\. Demystifying the Draft EU Artificial Intelligence Act—Analysing the good, the bad, and the unclear elements of the proposed approach. _Computer Law Review International_ 22, 4 (2021), 97–112. * Veale et al. (2018) Michael Veale, Max Van Kleek, and Reuben Binns. 2018\. Fairness and accountability design needs for algorithmic support in high-stakes public sector decision-making. In _Proceedings of the 2018 chi conference on human factors in computing systems_. 1–14. * Wachter (2020) Sandra Wachter. 2020\. Affinity profiling and discrimination by association in online behavioral advertising. _Berkeley Tech. LJ_ 35 (2020), 367. * Wachter (2022) Sandra Wachter. 2022\. The Theory of Artificial Immutability: Protecting Algorithmic Groups under Anti-Discrimination Law. _Tulane Law Review, Forthcoming_ (2022). * Wachter et al. (2021a) Sandra Wachter, Brent Mittelstadt, and Chris Russell. 2021a. Bias preservation in machine learning: the legality of fairness metrics under EU non-discrimination law. _W. Va. L. Rev._ 123 (2021), 735. * Wachter et al. (2021b) Sandra Wachter, Brent Mittelstadt, and Chris Russell. 2021b. Why fairness cannot be automated: Bridging the gap between EU non-discrimination law and AI. _Computer Law & Security Review_ 41 (2021), 105567. * Watch (2019) Algorithm Watch. 2019\. Automating Society Report 2019. * Zadrozny and Elkan (2001) Bianca Zadrozny and Charles Elkan. 2001. Obtaining calibrated probability estimates from decision trees and naive bayesian classifiers. In _Icml_ , Vol. 1. Citeseer, 609–616. * Zafar et al. (2017) Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P Gummadi. 2017. Fairness beyond disparate treatment & disparate impact: Learning classification without disparate mistreatment. In _Proceedings of the 26th international conference on world wide web_. 1171–1180. * Zehlike et al. (2017) Meike Zehlike, Francesco Bonchi, Carlos Castillo, Sara Hajian, Mohamed Megahed, and Ricardo Baeza-Yates. 2017. Fa* ir: A fair top-k ranking algorithm. In _Proceedings of the 2017 ACM on Conference on Information and Knowledge Management_. 1569–1578. * Zehlike et al. (2020) Meike Zehlike, Philipp Hacker, and Emil Wiedemann. 2020\. Matching code and law: achieving algorithmic fairness with optimal transport. _Data Mining and Knowledge Discovery_ 34, 1 (2020), 163–200. * Zehlike et al. (2022a) Meike Zehlike, Tom Sühr, Ricardo Baeza-Yates, Francesco Bonchi, Carlos Castillo, and Sara Hajian. 2022a. Fair Top-k Ranking with multiple protected groups. _Information Processing & Management_ 59, 1 (2022), 102707\. * Zehlike et al. (2022b) Meike Zehlike, Ke Yang, and Julia Stoyanovich. 2022b. Fairness in Ranking, Part I: Score-based Ranking. _ACM Computing Surveys (CSUR)_ (2022). * Zuiderveen Borgesius (2020) Frederik J Zuiderveen Borgesius. 2020\. Strengthening legal protection against discrimination by algorithms and artificial intelligence. _The International Journal of Human Rights_ 24, 10 (2020), 1572–1593.
# Life on Titan May Signal Early Life in the Universe Abraham Loeb Astronomy Department, Harvard University, 60 Garden St., Cambridge, MA 02138, USA ###### Abstract The temperature of the cosmic microwave background (CMB) was equal to the surface temperature of Saturn’s moon Titan, $94$K, at a redshift $z=33.5$, after the first galaxies formed. Titan-like objects would have maintained this surface temperature for tens of Myr irrespective of their distance from a star. Titan has the potential for the chemistry of familiar life in its subsurface water ocean, as well new forms of life in the rivers, lakes and seas of liquid methane and ethane on its surface. The potential future discovery of life on Titan would open the possibility that the earliest lifeforms emerged in metal-rich environments of the earliest galaxies in the universe, merely 100 Myr after the Big Bang. ## 1 Introduction The cosmic microwave background (CMB) provides a universal heating source of temperature (Fixsen et al., 1996), $T_{\rm cmb}=94{\rm K}\times[(1+z)/34.5]$ at a cosmological redshift $z$. Interestingly, this temperature matches the surface temperature of Saturn’s largest moon, Titan at $z\sim 34$, about 90 Myr after the Big Bang. Hence, a Titan-like object at that early time would have maintained this temperature for tens of Myr, sufficient for primitive life to form in its liquid reservoirs or atmosphere, irrespective of its distance from a star. ## 2 First Objects The standard cosmological model predicts that the first generation of stars and galaxies formed before $z\sim 34$ (Loeb & Furlanetto, 2013). Based on the measured cosmological parameters (Planck Collaboration et al., 2020), the first star-forming halos collapsed at $z\sim 71$ on our past light cone and at $z\sim 77$ within the entire Hubble volume (Loeb, 2014), including the delay from the streaming motion of baryons relative to dark matter (Fialkov et al., 2012). Hydrodynamical cosmological simulations predict that the first galaxies formed population III stars that were predominantly massive (Loeb & Furlanetto, 2013). For massive stars that are dominated by radiation pressure and shine near their Eddington luminosity $L_{\rm E}=1.3\times 10^{39}~{}{\rm erg~{}s^{-1}}(M_{\star}/10M_{\odot})$, the lifetime is independent of stellar mass $M_{\star}$ and is universally a few Myr, set by the nuclear efficiency of converting rest-mass to radiation, 0.7$\%$, namely $\sim(0.007M_{\star}c^{2})/L_{\rm E}=3~{}{\rm Myr}$ (Bromm et al., 2001). Consequently, the subsequent delay in dispersing heavy elements from the first stellar winds or pair-instability supernovae could have been as short as a few Myr, only a few percent of the age of the Universe at $z\sim 34$. The supernova ejecta could have produced high-metallicity islands that were not fully mixed with the surrounding primordial gas, leading to efficient formation of planets and moons within them. Altogether, this suggests that massive stars and supernovae were able to enrich the interstellar medium in the cores of the earliest galaxies with heavy elements before $z\sim 34$, leading to metal-rich pockets of gas inside of which the second generation of stars could have formed, accompanied by Titan-like objects. ## 3 Prospects of Life on Titan The temperature coincidence between Titan’s surface and the CMB at $z\sim 34$ raises the fascinating possibility of testing how early life could have arisen in the Universe by studying Titan. In other words, the question of whether Titan hosts life has cosmic implications. In the Solar system, Titan is the only object besides Earth that has rivers, lakes and seas on its surface, as well as a cycle of methane and ethane liquids raining from clouds, flowing across its surface and evaporating back into the atmosphere, similarly to Earth’s water cycle. Titan is also thought to have a subsurface ocean of water. Its atmosphere is primarily nitrogen like Earth’s, but with a $\sim 5\%$ contribution of methane. Titan’s landscape is covered with dark dunes of hydrocarbon grains, primarily around the equatorial regions. Gravity measurements by the Cassini spacecraft revealed that Titan has an underground ocean of liquid water, likely mixed with salts and ammonia (Lopes et al., 2019). Radio signals detected by the Huygens probe in 2005 strongly suggested the presence of an ocean 55-80 km below the icy surface, allowing for the chemistry of life-as-we-know-it. In addition, Titan’s bodies of liquid methane and ethane might serve as a foundation for the chemistry of life-as- we-do-not-know-it on the moon’s surface. Whether the physical conditions on Titan gave birth to these forms of life is unknown. The realization that Titan’s atmosphere is rich in organic compounds led to the proposal that it produced the chemical precursors of life. In particular, stable cryogenic cell membranes could arise from compounds observed in Titan’s atmosphere (Stevenson et al., 2015). The proposed chemical base for these membranes is acrylonitrile, which was detected in Titan’s atmosphere by Cassini and ALMA (Desai et al., 2017; Palmer et al., 2017). Analysis of data from the Cassini-Huygens mission reported anomalies in the atmosphere near the surface which could be consistent with the presence of lifeform of methane-consuming organisms, but may alternatively be due to abiotic chemical processes (Strobel, 2010; Clark et al., 2010). Laboratory experiments (Hörst et al., 2010) indicate that when discharge power is applied to a combination of gases like those in Titan’s atmosphere, they make prebiotic molecules such as the five nucleotide bases of DNA and RNA as well as amino-acids, among many other compounds. ## 4 Discussion The thermal gradients needed for life can be supplied by geological variations on the surface of early Titan-like objects. Examples for sources of free energy are geothermal energy powered by the object’s gravitational binding energy at formation and radioactive energy from unstable elements produced by the earliest supernova. If life persisted at $z\lesssim 34$, it could have also been transported to newly formed objects through panspermia (Ginsburg et al., 2018). Given the above considerations, the search for life on Titan could open the possibility that life may have started at a redshift of $z\sim 34$ in the standard cosmological model of our Universe. In addition to studying Titan, the feasibility of life in the early universe can be further tested by searching for planets with atmospheric bio-signatures around low-metallicity stars in the Milky Way galaxy or its dwarf galaxy satellites. Such stars represent the closest analogs to the first generation of stars at early cosmic times. ## Acknowledgements This work was supported in part by Harvard’s Black Hole Initiative, which is funded by grants from JFT and GBMF. ## References * Bromm et al. (2001) Bromm, V., Kudritzki, R. P., & Loeb, A. 2001, ApJ, 552, 464, doi: 10.1086/320549 * Clark et al. (2010) Clark, R. N., Curchin, J. M., Barnes, J. W., et al. 2010, Journal of Geophysical Research (Planets), 115, E10005, doi: 10.1029/2009JE003369 * Desai et al. (2017) Desai, R. T., Coates, A. J., Wellbrock, A., et al. 2017, ApJ, 844, L18, doi: 10.3847/2041-8213/aa7851 * Fialkov et al. (2012) Fialkov, A., Barkana, R., Tseliakhovich, D., & Hirata, C. M. 2012, MNRAS, 424, 1335, doi: 10.1111/j.1365-2966.2012.21318.x * Fixsen et al. (1996) Fixsen, D. J., Cheng, E. S., Gales, J. M., et al. 1996, ApJ, 473, 576, doi: 10.1086/178173 * Ginsburg et al. (2018) Ginsburg, I., Lingam, M., & Loeb, A. 2018, ApJ, 868, L12, doi: 10.3847/2041-8213/aaef2d * Hörst et al. (2010) Hörst, S., Yelle, R. V., Buch, A., et al. 2010, in AAS/Division for Planetary Sciences Meeting Abstracts, Vol. 42, AAS/Division for Planetary Sciences Meeting Abstracts #42, 36.20 * Loeb (2014) Loeb, A. 2014, International Journal of Astrobiology, 13, 337, doi: 10.1017/S1473550414000196 * Loeb & Furlanetto (2013) Loeb, A., & Furlanetto, S. R. 2013, The First Galaxies in the Universe * Lopes et al. (2019) Lopes, R. M. C., Wall, S. D., Elachi, C., et al. 2019, Space Sci. Rev., 215, 33, doi: 10.1007/s11214-019-0598-6 * Palmer et al. (2017) Palmer, M. Y., Cordiner, M. A., Nixon, C. A., et al. 2017, Science Advances, 3, e1700022, doi: 10.1126/sciadv.1700022 * Planck Collaboration et al. (2020) Planck Collaboration, Aghanim, N., Akrami, Y., et al. 2020, A&A, 641, A6, doi: 10.1051/0004-6361/201833910 * Stevenson et al. (2015) Stevenson, J., Lunine, J., & Clancy, P. 2015, Science Advances, 1, 1400067, doi: 10.1126/sciadv.1400067 * Strobel (2010) Strobel, D. F. 2010, Icarus, 208, 878, doi: 10.1016/j.icarus.2010.03.003
# Efficient Stereo Matching on Embedded GPUs with Zero-Means Cross Correlation Qiong Chang, Aolong Zha, Weimin Wang, Xin Liu, Masaki Onishi, Lei Lei, Meng Joo Er, Tsutomu Maruyama ###### Abstract Mobile stereo-matching systems have become an important part of many applications, such as automated-driving vehicles and autonomous robots. Accurate stereo-matching methods usually lead to high computational complexity; however, mobile platforms have only limited hardware resources to keep their power consumption low; this makes it difficult to maintain both an acceptable processing speed and accuracy on mobile platforms. To resolve this trade-off, we herein propose a novel acceleration approach for the well-known zero-means normalized cross correlation (ZNCC) matching cost calculation algorithm on a Jetson Tx2 embedded GPU. In our method for accelerating ZNCC, target images are scanned in a zigzag fashion to efficiently reuse one pixel’s computation for its neighboring pixels; this reduces the amount of data transmission and increases the utilization of on-chip registers, thus increasing the processing speed. As a result, our method is 2X faster than the traditional image scanning method, and 26% faster than the latest NCC method. By combining this technique with the domain transformation (DT) algorithm, our system show real-time processing speed of 32 fps, on a Jetson Tx2 GPU for 1,280x384 pixel images with a maximum disparity of 128. Additionally, the evaluation results on the KITTI 2015 benchmark show that our combined system is more accurate than the same algorithm combined with census by 7.26%, while maintaining almost the same processing speed. ## I INTRODUCTION Stereo matching is a key algorithm for depth detection in computer vision, but its usability is still limited because attaining high accuracy requires a very high computational complexity. By achieving both high accuracy and processing speed on mobile platforms, it can be used in many applications, including auto-driving, autonomous robots, and so on. Thus far, many researchers have focused upon accelerating stereo matching on mobile platforms. Most of them focus on accelerating the two most computationally intensive stages: cost calculation and cost aggregation. During cost calculation, each pixel in the reference image is first matched with several pixels in the target image one by one. Next, the similarity between any two pixels is quantified by a numerical value (cost) calculated by a matching method, such as the sum of absolute differences (SAD), census, or convolution neural network (CNN). Then, to further improve the accuracy of the matching system, the cost of each pixel within a certain region is expected to be aggregated together to represent the similarity between any two regions; this is called matching-cost aggregation. Many methods can be used to determine the range of the matching regions, such as semi-global matching (SGM) or domain transformation (DT). Various combinations of the above two stages not only result in different matching accuracies, but also different computational complexities; this leads to differing processing speeds. Many researches implement their stereo-matching systems on FPGAs, respectively. Wang [1] first combines a simplified SGM with the simple absolute differences and census matching algorithms, processing 1,024x768 pixel images with 96 disparities at 67 fps. Mohammad [2], Zhang [3] and Kuo [4] use the census algorithm to calculate their matching costs. Mohammad [2] combines census with a cross-aggregation method to achieve a good error rate of less than 9.22% and a high processing speed of faster than 200 fps on the KITTI 2015 benchmark [5]. Zhang [3] uses a box filter to aggregate matching costs and achieves a high processing speed of 60 fps for 1080p images. Kuo [4] uses a two-pass aggregation method and achieves the same processing speed as [3]. Oscar [6] uses SAD to calculate the matching cost and combines it with SGM. Due to SGM’s high accuracy for even the simple SAD matching algorithm, it can still achieve a lower error rate of 8.7%, except that its speed is reduced to 50 fps. Additionally, Zhang [7] develops a special ASIC to accelerate the implementation of SGM and achieves a processing speed of 30fps for 1080p images. Here, due to the limitations of floating-point decimal calculation, both the FPGA-based and the dedicated ASIC-based systems usually use methods such as SAD and census to obtain integer cost values. Although this is conducive to implementation on them, it also limits improvement in the matching accuracy. Furthermore, hardware-based systems typically need long development cycles and are also difficult to maintain. The recent advent of embedded GPUs has allowed the development of many systems [8] [9]. Compared to FPGAs, embedded GPU-based systems have short development cycles [10]. In addition, they are easy to maintain and port on other platforms. Wang [11], Smolyanskiy [12] and Tonioni [13] implement their systems on a Jetson Tx2 embedded GPU using the CNN; according to the evaluation results of KITTI 2015 [5] benchmark, their accuracies are high, with error rates between 3.2% and 6.2%. However, due to the significant calculations of CNN-based methods, their processing speeds are only a few fps, far below the requirements for practical applications. Daniel [14] constructs a fast stereo-matching system on a Jetson Tx2 GPU. It also combines census algorithm with SGM to achieve an error rate of 8.66% and a processing speed of 29 fps on KITTI 2015 benchmark. It is currently the best system for balancing the accuracy and processing speed on mobile GPUs; however, due to the census-matching method, this system is still not accurate enough, even if implemented on GPUs that are good at floating- point decimal calculations. According to [15], the matching accuracy by normalized cross correlation (NCC) is better than that by census because it has a higher ability to withstand changes in gain and bias. Furthermore, zero-means NCC (ZNCC)–an improved version of NCC–provides strong robustness because it also tolerates uniform brightness variations [16][17]. However, ZNCC has not been widely used on the mobile systems with limited hardware resources because of its higher computational complexity. In this paper, we accelerate ZNCC on a Jetson Tx2 embedded GPU and make it possible to achieve a comparable processing speed to that of census with a higher matching accuracy. The main contributions of this paper are as follows: * • We introduce a new calculation method, zigzag scanning based zero-means normalized cross correlation (Z2-ZNCC) to reuse the computational results of a pixel for the calculations of its neighbors. This makes it possible to reduce data transfer between the global memory of the GPU and increase the processing speed. * • We propose a strategy to make efficient use of registers during zigzag scanning to achieve higher parallelism of GPU threads driven by GPU cores. * • We design GPU-implementation algorithms for two parallel summation methods used in our Z2-ZNCC and comprehensively compare their performance. * • We create FastDT, an upgraded version of the GPU-based domain transformation method presented in [18] by removing the cost-value shifting step and increasing the flag code. Then, we combine it with Z2-ZNCC to construct a real-time stereo-matching system on an embedded GPU. The experimental results demonstrate that our method is 2X faster than the traditional image-scanning method and 26% faster than the latest NCC method [19]. Furthermore, our system achieves a processing speed of 32 fps and an error rate of 9.26% for 1,242x375 pixel images when the maximum disparity is 128 on a KITTI 2015 dataset. It is one of the few embedded GPU-based real-time systems, with an accuracy much higher than others. This paper extends our previous work (short paper) [20] from the following aspects. * • We design two parallel summation methods which maximize the processing speed of Z2-ZNCC depending on the template size. * • We introduce an efficient two-step implementation technique for domain transformation, which not only maintains a high processing speed, but also a high accuracy of Z2-ZNCC. * • We conduct comprehensive experiments to examine the impact of various conditions on the processing speed of Z2-ZNCC. The rest of the paper is organized as follows. Section II introduces the related works. Section III reviews ZNCC and DT calculation methods. Section IV discusses the GPU implementation of Z2-ZNCC and FastDT. Section V shows the evaluation results. Finally, Section VI presents the conclusions and our future work. ## II Related works Recently, many researchers have focused upon accelerating the performance of NCC-based methods and applying them into stereo-matching systems. Lin [21] proposes an optimization method for ZNCC calculation on a general platform. This method divides the standard equation into four independent parts and calculates the correlation coefficient efficiently using sliding windows. Hence, the computational complexity of ZNCC could be reduced from the original order. The computation time becomes constant for the window size; however, a large memory is required to store the calculation results for reuse. Although this method works nearly 10X faster than traditional ones, it is not applicable to embedded GPUs because of their limited memory space. Rui [22] implements a fast ZNCC-based stereo-matching system on GTX 970M GPU. This work focuses on the use of integral images to calculate the mean and standard deviation efficiently. Rui’s system runs approximately 9X faster than a single-threading CPU implementation and about 2X faster than eight- threading; however, according to our evaluation, this approach is inefficient for a small-size window (less than 9x9) because obtaining an integral image itself also requires calculation costs. These costs are mainly caused by data transfer of the integration results, which cannot be ignored for an embedded GPU with a high memory latency. Han [19] implements an NCC-based stereo-matching system on a Jetson Tx2 GPU. This method divides the equation into three parts, each with an identical control flow but different data locations. All intermediate results are stored on shared memory evenly so as to accelerate the calculation through reuse. However, as mentioned in Section I, the heavy use of shared memory reduces the parallelism of GPU blocks. ## III Algorithms and Optimizations ### III-A Zero-means Normalized Cross Correlation (ZNCC) ZNCC is used to calculate matching costs between a reference pixel $I_{R}(x,y)$ in the reference image and a series of target pixels $I_{T}(x-d,y)$ in the target image. $d$ is called disparity, and its range is $[0,D)$, where $D$ is a constant called maximum disparity. The function of ZNCC is given as follows: $C(x,y,d)=\frac{\displaystyle\sum_{(x,y)\in{W}}{\Delta{I_{R}}(x,y)\cdot{\Delta{I_{T}}(x-d,y)}}}{\sigma_{R}(x,y)\cdot{\sigma_{T}(x-d,y)}},$ (1) where $\begin{split}\sigma_{R}(x,y)&=\sqrt{\displaystyle\sum_{(x,y)\in{W}}{\Delta{I_{R}(x,y)^{2}}}},\\\ \sigma_{T}(x-d,y)&=\sqrt{\displaystyle\sum_{(x,y)\in{W}}{\Delta{I_{T}(x-d,y)^{2}}}},\\\ \end{split}$ and $\begin{split}\Delta{I_{R}(x,y)}&=I_{R}(x,y)-\overline{I_{R}(x,y)},\\\ \Delta{I_{T}(x-d,y)}&=I_{T}(x-d,y)-\overline{I_{T}(x-d,y)}.\\\ \end{split}$ Here, $\overline{I_{R}(x,y)}$ and $\overline{I_{T}(x-d,y)}$ are the averages of the pixel values in the matching windows $W$ surrounding $I_{R}(x,y)$ and $I_{T}(x-d,y)$, respectively. $C(x,y,d)$ in (1) is the correlation coefficient (i.e., the matching cost) between $I_{R}(x,y)$ and $I_{T}(x-d,y)$; its range is $[0,1]$ (the closer to one, the more similar the two windows). $\sigma_{R}(x,y)$ and $\sigma_{T}(x-d,y)$ are the standard deviations of the pixel values in the two windows and are used to normalize the correlation coefficient between them. Each reference pixel $I_{R}(x,y)$ needs to be matched with $D$ target pixels $I_{T}(x-d,y)$; here, $\Delta{I_{R}(x,y)}$ and $\Delta{I_{T}(x-d,y)}$ can be calculated in advance because the calculations of these terms are closed in each image. As such, the total number of calculations can be reduced. However, when the size of the matching window $W$ is $l^{2}$ (where $l=2r+1$ represents the side length of window $W$) $l^{2}$-times the memory space is needed for each image because each window has $l^{2}$ differences. To further reduce the ZNCC’s computation complexity, (1) can be rewritten as follows: $\begin{split}C(x,y,d)=\frac{\displaystyle\sum_{(x,y)\in{W}}{\Pi_{RT}(x,y,d)}-l^{2}\cdot{\overline{\Pi_{RT}(x,y,d)}}}{\sigma_{R}(x,y)\cdot{\sigma_{T}(x-d,y)}},\end{split}$ (2) where $\begin{split}\Pi_{RT}(x,y,d)&=I_{R}(x,y)\cdot{I_{T}(x-d,y)},\\\ \overline{\Pi_{RT}(x,y,d)}&=\overline{I_{R}(x,y)}\cdot{\overline{I_{T}(x-d,y)}},\\\ \end{split}$ and $\begin{split}\sigma_{R}(x,y)&=\sqrt{\displaystyle\sum_{(x,y)\in{W}}{I_{R}(x,y)^{2}}-l^{2}\cdot{\overline{I_{R}(x,y)}^{2}}},\\\ \sigma_{T}(x-d,y)&=\sqrt{\displaystyle\sum_{(x,y)\in{W}}{I_{T}(x-d,y)^{2}}-l^{2}\cdot{\overline{I_{T}(x-d,y)}^{2}}}.\\\ \end{split}$ In this calculation, $C(x,y,d)$, $\sigma_{R}(x,y)$, and $\sigma_{T}(x-d,y)$ are calculated from four values; $\overline{I_{R}(x,y)}^{2}$, $\sum{I_{R}(x,y)^{2}}$, $\overline{I_{T}(x-d,y)}^{2}$, and $\sum{I_{T}(x-d,y)^{2}}$, rather than from $\Delta{I_{R}(x,y)}$, and $\Delta{I_{T}(x-d,y)}$ as shown in (1). These four values are only related to their respective images and can all be calculated in advance. Thus, both $\sigma_{R}(x,y)$ and $\sigma_{T}(x-d,y)$ can be calculated efficiently without calculating $\Delta{I_{R}(x,y)}$ and $\Delta{I_{T}(x-d,y)}$ for each pixel $l^{2}$ times. This transformation not only helps to reduce the total calculation amount, but also reduce the required memory space. Of course, ZNCC-based matching is performed in a fixed size window, which is not accurate for irregular patterns in reality; therefore, the matching cost $C(x,y,d)$ is usually combined with various aggregation methods to improve the matching accuracy. ### III-B Domain Transformation (DT) In cost aggregation step, the matching costs of all similar pixels in the same area (e.g., the area within the pink-dashed line in Fig.1 (a)) are added together. Domain Transformation (DT) [23] is an effective algorithms for use at this stage. Unlike other algorithms, DT avoids over-fitting of cost propagation by using the gradients of adjacent pixels to weight their costs in different directions, rather than judging the boundary of each area in advance. The advantage of DT is that there is no need to segment each area for cost aggregation separately, and it is suitable for parallel processing to increase the aggregation speed. In DT, the matching cost of each pixel is aggregated from four different directions, while propagating its cost to four neighboring pixels is done according to the following equations: Figure 1: Domain transformation $\displaystyle C_{L}(x,y,d)$ $\displaystyle=C(x,y,d)\small{+}C_{L}(x\small{-}1,y,d)\cdot{W_{L}(x,y)},$ (3) $\displaystyle C_{R}(x,y,d)$ $\displaystyle=C_{L}(x,y,d)\small{+}C_{R}(x\small{+}1,y,d)\cdot{W_{R}(x,y)},$ (4) $\displaystyle C_{U}(x,y,d)$ $\displaystyle=C_{R}(x,y,d)\small{+}C_{U}(x,y\small{-}1,d)\cdot{W_{U}(x,y)},$ (5) $\displaystyle C_{D}(x,y,d)$ $\displaystyle=C_{U}(x,y,d)\small{+}C_{D}(x,y\small{+}1,d)\cdot{W_{D}(x,y)}.$ (6) Here, $C_{L}$, $C_{R}$, $C_{U}$, and $C_{D}$ represent the aggregated costs for each pixel from four different directions (left, right, up, and down). In this aggregation, for example, the boundary condition for (3) is given by $C_{L}(0,y,d)=C(0,y,d)$. $W_{L}$, $W_{R}$, $W_{U}$ and $W_{D}$ represent the corresponding weights calculated by the gradient between adjacent pixel values according to the following equations: $W_{L}(x,y)=a^{1+\frac{\sigma_{s}}{\sigma_{r}}\cdot{|I_{T}(x,y)-I_{T}(x-1,y)|}},$ (7) $W_{R}(x,y)=a^{1+\frac{\sigma_{s}}{\sigma_{r}}\cdot{|I_{T}(x,y)-I_{T}(x+1,y)|}},$ (8) $W_{U}(x,y)=a^{1+\frac{\sigma_{s}}{\sigma_{r}}\cdot{|I_{T}(x,y)-I_{T}(x,y-1)|}},$ (9) and $W_{D}(x,y)=a^{1+\frac{\sigma_{s}}{\sigma_{r}}\cdot{|I_{T}(x,y)-I_{T}(x,y+1)|}},$ (10) where $a=\exp{(-\frac{1}{\sigma_{s}})}.$ (11) In (7) to (11), $\sigma_{s}$ is a spatial parameter and $\sigma_{r}$ is an intensity range parameter. Both are used to adjust the weight caused by gradient changes in space and intensity. To simplify the calculation, the weight equations can be further simplified below (here, we only take (7) as an example): $\displaystyle\ln{W_{L}}$ $\displaystyle=\ln{(a^{1+\frac{\sigma_{s}}{\sigma_{r}}\cdot{|I_{T}(x,y)-I_{T}(x-1,y)|}})}$ $\displaystyle=(1+\frac{\sigma_{s}}{\sigma_{r}}\cdot{|I_{T}(x,y)-I_{T}(x-1,y)|})\cdot{\ln{a}}$ $\displaystyle=(1+\frac{\sigma_{s}}{\sigma_{r}}\cdot{|I_{T}(x,y)-I_{T}(x-1,y)|})\cdot{(-\frac{1}{\sigma_{s}})}$ $\displaystyle=-\frac{1}{\sigma_{s}}-\frac{|I_{T}(x,y)-I_{T}(x-1,y)|}{\sigma_{r}},$ then $\displaystyle W_{L}$ $\displaystyle=\exp(-\frac{1}{\sigma_{s}}-\frac{|I_{T}(x,y)-I_{T}(x-1,y)|}{\sigma_{r}})$ $\displaystyle=K\cdot{\exp(-\frac{|I_{T}(x,y)-I_{T}(x-1,y)|}{\sigma_{r}})},$ where $\displaystyle K=\exp(-\frac{1}{\sigma_{s}}).$ (14) $K$ is a constant coefficient used to ease the calculation of $W_{L}$. Figure.1 shows an example of cost aggregation for pixel $I_{T}(x,y)$. Figure.1 (a) shows part of the reference image centered on $I_{T}(x,y)$. Figures.1 (b) to 1 (e) show the cost propagation process from different directions. Three curves with different colors represent the changes in the gradient, weights, and propagated costs, respectively. According to (III-B), the weight is calculated by the gradient and then used to weight the propagated cost value. As shown in Figs.1 (b) to 1 (e), the weight changes in the opposite direction to the change in gradient value, thereby ensuring that cost propagation can be performed normally among non-edge pixels (Fig.1 (b)) and can also be interrupted at edge pixels (Fig.1 (c)). When the propagation from down to up is completed (i.e., when the final aggregation result $C_{D}(x,y,d)$ is obtained), $C(x,y,d)$ is replaced by $C_{D}(x,y,d)$, and used in the following stages. ### III-C Winner-Take-ALL (WTA) After calculating matching cost for maximum disparity $D$ times, the target pixel $I_{T}(x-d,y)$ that is most similar to reference pixel $I_{R}(x,y)$ is determined as: $D_{map}(x,y)=\mathop{\rm arg~{}min}\limits_{d}{(1-C(x,y,d))}.$ (15) As shown in this equation, the value of $d$ that minimizes $(1-C(x,y,d))$ is chosen as the disparity of the reference pixel $I_{R}(x,y)$. ## IV Implementation Implementing ZNCC and DT on an embedded GPU is a key challenge for realizing a fast and accurate mobile stereo-vision system. In this section, we first introduce the architecture of Jetson Tx2 GPU; then, we describe the acceleration approaches of ZNCC and DT, respectively. ### IV-A GPU Architecture and CUDA Programming Model The Jetson Tx2 has 2 streaming multi-processors (SMs); each SM runs in parallel using 128 cores (256 cores in total) and has two types of on-chip memory: register memory and shared memory. Their sizes are limited, but their access latencies are very low. This GPU also has a global memory (off-chip), which is usually used to hold all data for processing. Due to the high latency of access to the off-chip memory, the most important point for achieving high performance on the GPU is to minimize the amount of data transfer between on- chip and off-chip memory. In our implementation, we use the GPGPU programming model CUDA [24]. CUDA abstractly defines the GPU core, SM, and GPU itself as thread, block, and grid, respectively. A grid is composed of blocks and a block is composed of threads. Every 32 threads execute the same instruction, which is called a warp. The warps are scheduled serially by the SMs. Users can define the number of the abstract resources according to their requirements, which may exceed the physical GPU resources; then, the CUDA driver schedules abstract resources to work upon physical resources. Since the total number of registers and shared-memory space are fixed, the amount of these resources allocated to each thread and block determines how many of them can be active. The more allocated, the fewer threads and blocks can be activated, which resulting in reduced performance. By storing intermediate calculation results and reusing them afterwards, the total amount of calculation can be reduced; however, more hardware resources are needed to store these results, which limits the number of active threads. On the other hand, by recalculating them each time, the required amount of hardware resources can be reduced, and more threads can be active. Balancing the hardware-resource usage and total amount of calculation is a key point for achieving high performance, especially on embedded GPUs with limited hardware resources. ### IV-B Implementation of ZNCC Our ZNCC-acceleration approach includes two steps: (1) calculation of the means and the sums of squares in matching windows, and (2) calculation of the correlation coefficients of each pixel by zigzag scanning. #### IV-B1 Summation For the ZNCC-acceleration approach, the means and sums of squares in each matching window are calculated in advance (Section III-A). Taking the example of the pixel values in the reference image, we describe the two methods as follows (for simplicity, we only describe the sum in the reference image): Figure 2: Summation methods Method 1: As shown in Fig.2 (a), $S_{R}(x,y)$, the sum of the pixel values in each matching window $W$ is calculated as follows: $\begin{split}S_{R}(x,y)&=\displaystyle\sum_{(x,y)\in W}I_{R}(x,y).\\\ \end{split}$ (16) Here, $(2r+1)^{2}$ pixel values are simply added around the center pixel $I_{R}(x,y)$. The summation of each window is performed independently by different CUDA threads. Since the two adjacent windows for two adjacent pixels share $2r\times{(2r+1)}$ pixels, the data from $(2r+1)$ rows or columns are excepted to be cached in the same shared memory. The allocated memory space grows as the window size increases. The number of columns and rows processed by each CUDA block at the same time depends upon the size of the shared memory allocated; for smaller-sized windows, less hardware resources are required for each thread, and higher parallelism can be expected; however, as the window size increases, more hardware resources are required, and fewer threads can be active. Therefore, this method is not suitable for the summation of large windows. Method 2: This method performs the summation using the integral image $B_{R}(x,y)$: $\displaystyle B_{R}(x,y)=\sum_{u=0}^{x}\sum_{v=0}^{y}I_{R}(u,v).$ (17) As shown in Fig.2 (b), $S_{R}(x,y)$ is calculated using four points in the integral image, regardless of the window size: $\begin{split}S_{R}(x,y)&{\small=}B_{R}(x+r,y+r)+B_{R}(x{\small-}r{\small-}1,y{\small-}r{\small-}1)\\\ &{\small-}B_{R}(x+r,y{\small-}r{\small-}1){\small-}B_{R}(x{\small-}r{\small-}1,y+r).\end{split}$ (18) In this calculation method, To calculate the sum for the pixels on row $y$, only two rows ($y-r-1$ and $y+r$) of the integral image are needed. Thus, this method is suitable for the summation of large windows. Obtaining an integral image in parallel requires two steps: 1. 1. integrate each row of $I_{R}$, defined as $B_{H}$, 2. 2. integrate each column of $B_{H}$, defined as $B_{R}$. Figure 3: Integration along the $x$-axis To generate the integral image, all image data need to be loaded from global memory to the shared memory. However, due to the limitation of the shared memory, the pixels in one row are divided into several segments and integrated partly as shown in Fig.3 (a). Considering the limitations of data sharing between different GPU blocks, we only use one block (Block0 in Fig.3 (a)) for the integration of one row, which means that each segment is integrated by the same block one by one (Round1, Round2,…) rather than processing multiple blocks in parallel. In each block, a two-layer parallel-integration strategy based on the Kogge-Stone Adder algorithm is used as shown in Fig.3 (b) and (c), since the threads work in the units of warp. The Kogge-Stone Adder method is used because of its high computational efficiency and suitability for thread-level parallel operations on GPUs. It is not necessary to check the parity of the operand index for each stage. Figure.3 (b) shows the parallel integration on the Warp layer. In Fig.3 (b), “$\circ$” denotes the shared memory used in the current segment. Before integration, all threads in each warp are initialized with a variable $S1$, which represents the sum of the previous segment (step 1). For the first segment in one row, $S1$ is initialized to 0\. Then, each warp performs the integration independently and propagates its intermediate results by shared memory according to the Kogge-Stone Adder method (step 2). In this step, the result of each Warp $i$ is added to Warp $i+2^{t-1}\ (i+2^{t-1}<N)$ through the shared memory, with $t$ representing the number of repetitions increasing from 1 to $\lfloor{\log_{2}{N}\rfloor}$ and $N$ representing the number of threads in each warp. Then, the sum of the current segment is updated by the last thread (step 3); at the same time, the integrated result of each thread is transferred to the global memory for the integration along the $y$-axis. Figure 4: Stereo matching on the GPU Figure.3 (c) shows the parallel integration on the Thread layer. After the initialization shown in Fig.3 (b) step1, each thread loads the corresponding pixel value from the global memory to the registers represented by “$\bullet$”, and adds it to the initial value $S1$. Then, integration is performed through the register shift among the threads in the same warp using the method shown in Fig.3 (b) step 2, and the last result is stored in the shared memory. By repeating the above steps until the integration of the last segment ends, the integral image of each row can be calculated and used to obtain the entire integral image along the $y$-axis. Here, by transposing the matrix [25], integration can be transformed from vertical to horizontal. However, this may be less effective than a direct calculation when the vertical range is small, because the memory overhead required by matrix transposition itself reduces the parallelism of the GPU blocks. In this paper, we perform a sequential column-wise integration rather than using a parallel-computing method, because the height of the image set is less than 400 pixels. #### IV-B2 Z2-ZNCC on Stereo Matching After the summations above, the terms in (2) can be easily calculated with the exception of $\sum_{(x,y)\in W}{\Pi_{RT}(x,y,d)}$ is omitted in the following discussion to simplify the description. Here, we show that $\sum{\Pi_{RT}(x,y,d)}$ can be calculated efficiently by scanning the image in a zigzag fashion. Unlike the other summations, $\Pi_{RT}(x,y,d)$ represents a 3D-matching result between reference pixels and multiple target pixels under different disparities. Efficient calculation of $\sum{\Pi_{RT}(x,y,d)}$ is the most critical part of our implementation. ##### Task Assignment As shown in Fig.4 (a), $I_{R}(x,y)$ is matched with $D$ pixels $I_{T}(x-d,y)$. To calculate $\sum{\Pi_{RT}(x,y,d)}$ for each $I_{R}(x,y)$, $(2r+1)^{2}$ pixels around $I_{R}(x,y)$ and $(2r+1)\times{(D+2r)}$ pixels around $I_{T}(x-d,y)\ (d\in[0,D))$ are required. For this matching, one block is assigned because the pixel data loaded into the shared memory can be reused to match adjacent pixels. In each block, $D$ threads are assigned to perform the matching in parallel for each corresponding $d$, as shown in Fig.4 (b). Each thread $i$ calculates $\sum{\Pi_{RT}(x,y,i)}$ using the pixels in the windows $W$ (centered at $I_{R}(x,y)$) and windows $W^{\prime}$ (centered at $I_{T}(x-i,y)$). Here, $\Pi_{RT}(x,y,i)$ is calculated element by element, and their sum is calculated efficiently via our approach described below. After calculating $\sum{\Pi_{RT}(x,y,d)}$ (as shown in Fig.4 (c)), the matching cost $C(x,y,d)$ can be calculated according to (2) and then stored in the global memory for use in the cost-aggregation stage. With this task assignment, the data once loaded to the shared memory from the global memory can be efficiently reused for the calculation of adjacent pixels when $D$ is sufficiently large. ##### Zigzag Scanning Figure 5: Zigzag scanning In our approach, the image is scanned in a zigzag fashion, as shown in Fig.5 (a) along the $x$ and $y$ axis, while $\sum{\Pi_{RT}(x,y,d)}$ is calculated for each pixel in parallel along the $d$-axis. $V_{Z}$ pixels in a column are processed first from top to bottom; then, the same processing is repeated on the next column. This scanning method is repeated from left to right. In this zigzag scanning, the rows are segmented in the same way as in summation Method 2, and $(2r+V_{Z})\times{(2r+H_{Z})}$ pixels in the reference image and $(2r+V_{Z})\times{(2r+H_{Z}+D)}$ pixels in the target image are loaded into the shared memory respectively as shown in Fig.5 (b) (where $H_{z}$ is a constant decided by the shared-memory size). With this zigzag scanning, after $\sum{\Pi_{RT}(x,y,d)}$ was calculated, $\sum{\Pi_{RT}(x,y+1,d)}$ and $\sum{\Pi_{RT}(x+1,y,d)}$ can be easily calculated as follows: $\begin{split}\sum_{(x,y)\in W}{\Pi_{RT}(x,y{\small+}1,d)}&{\small=}\displaystyle\sum_{\Delta{x}{\mbox{\scriptsize=}}-r}^{r}{\sum_{\Delta{y}{\mbox{\scriptsize=}}-r}^{r}{\Pi_{RT}(x{\small+}\Delta{x},y{\small+}\Delta{y},d)}}\\\ &{\small+}\displaystyle\sum_{\Delta{x}=-r}^{r}{\Pi_{RT}(x{\small+}\Delta{x},y{\small+}r{\small+}1,d)}\\\ &{\small-}\displaystyle\sum_{\Delta{x}=-r}^{r}{\Pi_{RT}(x{\small+}\Delta{x},y{\small-}r,d)},\\\ \end{split}$ (19) $\begin{split}\sum_{(x,y)\in W}{\Pi_{RT}(x{\small+}1,y,d)}&{\small=}\displaystyle\sum_{\Delta{x}{\mbox{\scriptsize=}}-r}^{r}{\sum_{\Delta{y}{\mbox{\scriptsize=}}-r}^{r}{\Pi_{RT}(x{\small+}\Delta{x},y{\small+}\Delta{y},d)}}\\\ &{\small+}\displaystyle\sum_{\Delta{y}=-r}^{r}{\Pi_{RT}(x{\small+}r{\small+}1,y{\small+}\Delta{y},d)}\\\ &{\small-}\displaystyle\sum_{\Delta{y}=-r}^{r}{\Pi_{RT}(x{\small-}r,y{\small+}\Delta{y},d)}.\\\ \end{split}$ (20) As shown in these two equations, the advantage of using the zigzag scanning method is that as long as the sums of different rows and columns such as $\sum_{\Delta{x}\in[-r,r]}$ ${\Pi_{RT}(x{\small+}\Delta{x}}$,${y{\small-}r,d)}$ and $\sum_{\Delta{y}\in[-r,r]}{\Pi_{RT}(x{\small-}r,y{\small+}\Delta{y},d)}$ can be stored in the memory, they can be reused to efficiently calculate other sums along both directions. However, storing these intermediate results along the two directions requires a huge number of registers, which may reduce the total efficiency. ##### Z2-ZNCC To solve this problem, we propose a strategy for efficiently using registers. Figure.6 shows the processing flow of the summation of $(2r+1)^{2}$ pixel window. In this example, $V_{Z}=2$, meaning that two rows, $y$ and $y+1$, are processed during one zigzag scanning. The calculation process is as follows: * • Step 1: $\sum{\Pi_{RT}(x,y,d)}$ is first calculated in order and stored in the register $RS$; then, it can be used to calculate the matching cost $C(x,y,d)$. During this step, in order to calculate $\sum{\Pi_{RT}(x,y+1,d)}$ efficiently, the sum of $2r+1$ pixels on row $y-r$ is stored in the register $R0$. * • Step 2: The difference between $RS$ and $R0$ is calculated and stored in $RS$ to calculate $\sum{\Pi_{RT}(x,y+1,d)}$. * • Step 3: $\sum{\Pi_{RT}(x,y,d)}$ is still necessary for calculating $\sum{\Pi_{RT}(x+1,y,d)}$, but its value of $RS$ was discarded in Step2. On the other hand, the sum on row $y-r$ in $R0$ is no longer necessary. Thus, the difference stored in $RS$ is added back to $R0$ to recalculate $\sum{\Pi_{RT}(x,y,d)}$. This irregular procedure minimizes the number of registers used for this calculation and makes more threads active. * • Step 4: The sum of row $y+r+1$ is calculated and added to $RS$. Then, $\sum{\Pi_{RT}(x,y+1,d)}$ is obtained and used to calculate the matching cost $C(x,y+1,d)$. * • Step 5: $\sum{\Pi_{RT}(x,y+1,d)}$ is stored in register $R1$ to calculate $\sum{\Pi_{RT}(x+1,y+1,d)}$ in the same way. At this point, $\sum{\Pi_{RT}(x,y,d)}$ and $\sum{\Pi_{RT}(x,y+1,d)}$ are stored in $R0$ and $R1$ respectively, and these values are used to calculate $\sum{\Pi_{RT}(x+1,y,d)}$ and $\sum{\Pi_{RT}(x+1,y+1,d)}$. * • Step 6,7: To calculate $\sum{\Pi_{RT}(x+1,y,d)}$ from $\sum{\Pi_{RT}(x,y,d)}$ in the same way, the sums of $2r+1$ pixels in columns $x-r$ and $x+r+1$ are required. In our implementation, to make more threads active by reducing the memory usage as much as possible, these sums are not stored in the memory during the above calculations. Then, the difference between the pixels in columns $x-r$ and $x+r+1$ is calculated and summed. In Step6, the difference of the uppermost pixels is calculated and stored in $RS$, and in Step 7, the differences of the other pixels are added to $RS$. Finally, $RS$ becomes the difference between $\sum{\Pi_{RT}(x,y,d)}$ and $\sum{\Pi_{RT}(x+1,y,d)}$. * • Step 8: The accumulated difference is added to $R0$ and then $\sum{\Pi_{RT}(x+1,y,d)}$ can be obtained, and $C(x+1,y,d)$ is calculated. * • Step 9: The difference stored in $RS$ is updated to calculate $\sum{\Pi_{RT}(x+1,y+1,d)}$ by adding and subtracting $\Pi_{RT}$ on the four corners. * • Step 10: The accumulated difference $RS$ is added to $R1$ and then $\sum{\Pi_{RT}(x+1,y+1,d)}$ is obtained. Using this method, we only need $V_{Z}+1$ registers for each thread to perform the summation. In our implementation, only the intermediate results along the $x$ axis are held on registers, while those along the $y$ axis are recalculated. This strategy is chosen because $V_{Z}$ is smaller than $H_{Z}$. By repeating the above calculation continuously, the overall processing speed can be greatly improved by limiting the number of registers for each thread, and by making more threads active. Figure 6: Efficient zigzag scanning ### IV-C Implementation of DT Since DT is performed based on the ZNCC computation results, the task assignment is the same as that of ZNCC shown in Fig.4. In DT, because the cost values must be aggregated in four directions (left to right, right to left, up to down, and down to up), a large memory space is required. The size of on- chip shared memory is too small for this purpose, and we need to use the off- chip global memory. Here, because both the ZNCC ((1)) and weighting operations ($\eqref{equ:k1}$ to ${\eqref{equ:k4}}$) usually generate floating-point cost values (32 bits), there is not only causes a great burden on data transmission, but also a greater requirement for requires more shared-memory space, which reduces the parallelism of multi-thread processing. To solve these problems, it is usually good to use shorter integers (16 bits or even 8 bits) to represent the cost values instead of the floating-point data type. However, a serious problem needs to be addressed here. The magnitude of the aggregated cost values around the boundary of each area decreases due to the use of weight. Hence, the magnitude is sufficiently small for 16-bit integers but not for textureless regions. Figure.7 (a) shows an example of cost aggregation in a large white wall with less texture. Let $P(x^{\prime},y^{\prime})$ be a pixel that belongs to the wall in the image; its cost values $C(x^{\prime},y^{\prime},d)$ are accumulated from all pixels in this area according to (3) to (6). Due to the size of the wall (around 250x300 pixels) and the range of the ZNCC result ([0,1] as mentioned in Section III-A), the accumulated floating-point cost values $C(x^{\prime},y^{\prime},d)$ will largely exceed the upper limit on a 16-bit integer, causing overflow. As such, these cost values cannot be simply converted to integers by multiplying by a coefficient. In [18], we propose a solution to this problem by shifting the matching cost of census to quickly reduce the value range and compressing the 16-bit data into an 8-bit data with a 1-bit flag code. However, the shifting method is not suitable for ZNCC because of its decimal cost value; furthermore, a 1-bit flag code can only specify two positions on a 16-bit integer, which may reduce accuracy. Therefore, we upgrade the original solution by using a two-step strategy to reduce the aggregated costs’ data width and burden of transmission: * • Use a cost-value normalization with a nearly zero-mean to represent the original floating-point cost values with 16-bit short integers. * • Apply a data encoding & decoding method to further replace the normalized 16-bit short integers with 8-bit by using a 2-bit flag code. The details of this strategy are as follows: ##### Cost-value Normalization with Nearly Zero-Mean Figure.7 (b) shows the change in the cost values of $P(x^{\prime},y^{\prime})$ along the disparity (from 0 to $D-1$). Curve1 represents the change of $C(x^{\prime},y^{\prime},d)$ obtained by (6) (where $C_{D}(x^{\prime},y^{\prime},d)$ is used as final $C(x^{\prime},y^{\prime},d)$ as described above), and $C(x^{\prime},y^{\prime},d_{min})$ shows the minimum value along the curve. The corresponding disparity $d_{min}$ is the result obtained based on the magnitude relationship of $C(x^{\prime},y^{\prime},d)$ according to (15). Therefore, as long as the magnitude relationship remains unchanged, changing the values of $C(x^{\prime},y^{\prime},d)$ will not affect obtaining the correct $d_{min}$. Additionally, since the range of $C(x^{\prime},y^{\prime},d)$ is narrow (as shown in Fig.7, Cost_gap, the difference between the max and min of $C(x^{\prime},y^{\prime},d)$ is much smaller than the values of $C(x^{\prime},y^{\prime},d)$ themselves. Since the range of Cost_gap is narrow, each cost value can be nearly zero-mean normalized by subtracting $C(x^{\prime},y^{\prime},d_{arb})$, where $d_{arb}$ is an arbitrary disparity ($d_{arb}\neq{d_{min}}$). By subtracting the median value, the range of $C(x^{\prime},y^{\prime},d)$ can be minimized; in our implementation, however, an arbitrary value $C(x^{\prime},y^{\prime},d_{arb})$ is used to simplify the calculation. Then (6) can be changed to: $\displaystyle C^{\prime}_{D}(x,y,d)$ $\displaystyle=C^{\prime}_{U}(x,y,d)+C^{\prime}_{D}(x,y+1,d)\cdot{W_{D}(x,y)}$ $\displaystyle-C^{\prime}_{U}(x,y,d_{arb}),$ (21) where $C^{\prime}$ represents the normalized cost value such that the mean approaches zero. This effectively suppresses the increase in the aggregated cost values. To further ensure that these values will not cause an overflow, the cost-value normalization is extended in all directions. At the same time, each accumulated floating-point cost value is scaled up to a 16-bit signed integer via an integer coefficient $T$. The value of $T$ needs to be carefully determined according to the actual situation. The larger the value, the higher the accuracy but also the greater the risk of overflow. This approach not only effectively reduces the burden of data transmission, but facilitates the further reduction of data width using the following method. Figure 7: Cost-value normalization with nearly zero-mean ##### Data Encoding & Decoding After cost normalization, the value of $C^{\prime}(x^{\prime},y^{\prime},d_{arb})$ is reduced to 0 and some values, including $C^{\prime}(x^{\prime},y^{\prime},d_{min})$, become negative (when $d_{arb}\neq{d_{min}}$). Therefore, it is possible to further reduce the data width by focusing only upon the negative values while ignoring the positive ones. Our method includes two stages: encoding and decoding. Figure.8 shows the process of our method. In the encoding stage, the 16-bit cost value is first compressed into an 8-bit code containing a 6-bit value and a 2-bit flag. This 6-bit value represents the 6 significant bits of the 16-bit integer, and the 2-bit flag shows its position. In the decoding stage, each 8-bit code is decompressed into a 16-bit approximation by putting the 6-bit value in the 16-bit integer on the position specified by the 2-bit flag code. Between the stages, four adjacent 8-bit codes are packed into a 32-bit integer for more efficient transmission on a GPU. The details of the process in Fig.8 can be described as follows: * • Encoding 1. 1. As mentioned above, only negative values are used. Therefore, all positive values are set to 0 by checking the sign bit. 2. 2. Since two’s complement is used to represent the negative values, we first test the 4-bit $Data_{16}[14:11]$ to find whether it is ${}^{\prime}1111^{\prime}$. If not, ($Data_{16}[14:9]$) is chosen as the 6-bit code and ${}^{\prime}01^{\prime}$ is attached to it as the 2-bit flag to show the position of the 6-bit code in the original 16-bit integer. Then, an 8-bit dataset, $Data_{8}$, is constructed from the 16-bit integer. 3. 3. If $Data_{16}$[14:11] is ${}^{\prime}1111^{\prime}$, the next 4-bit code ($Data_{16}[10:7]$) is checked in the same way. If it is not ${}^{\prime}1111^{\prime}$, $Data_{16}[10:5]$ is chosen as the 6-bit code and the flag code ${}^{\prime}10^{\prime}$ is attached to show its position. 4. 4. Finally, if $Data_{16}[14:11]$ and $Data_{16}[10:7]$ are both ${}^{\prime}1111^{\prime}$, $Data_{16}[6:1]$ is chosen as the 6-bit code (no checking is necessary) and the 2-bit flag code ${}^{\prime}11^{\prime}$ is attached. * • Decoding 1. 1. We first check whether $Data_{8}$ is ${}^{\prime}0^{\prime}$ or not. If it is, $Data_{16}$ is also set to ${}^{\prime}0^{\prime}$; if not, its flag code $Data_{8}[1:0]$ is checked. 2. 2. If the 2-bit flag code is ${}^{\prime}11^{\prime}$, the 6-bit code at $Data_{8}[7:2]$ is copied to $Data_{16}[6:1]$; if the flag code is ${}^{\prime}10^{\prime}$, the 6-bit code is copied to $Data_{16}[10:5]$; if the flag code is ${}^{\prime}01^{\prime}$, the 6-bit code is copied to $Data_{16}[14:9]$. 3. 3. Then, the bits on the left-hand side of the copied 6-bit code in $Data_{16}$ are set to ${}^{\prime}1^{\prime}$ and the bits on the right side are set to ${}^{\prime}0^{\prime}$. Figure 8: Data encoding & decoding This method uses the 2-bit flag code, which specifies four position; and the position of the 6-bit code is not continuous on a 16-bit integer. While this is more accurate than the 1-bit flag code which specifies two positions, our approach still loses more information than general 16-bit to 6-bit data-width reduction. However, according to our experiments, high speed processing is possible without losing too much accuracy. ## V Evaluation $3$$5$$7$$9$$11$$13$$15$$0.5$$1$$1.5$$2$$2.5$$3$$3.5$$4$Side LengthTime (${ms}$)M1(BS=64x16)M1(BS=64x8)M1(BS=64x4)M2(BS=64)M2(BS=128)M2(BS=256) (a) Regcount = 32 $3$$5$$7$$9$$11$$13$$15$$0.5$$1$$1.5$$2$$2.5$$3$$3.5$$4$Side LengthM1(BS=64x16)M1(BS=64x8)M1(BS=64x4)M2(BS=64)M2(BS=128)M2(BS=256) (b) Regcount = 48 $3$$5$$7$$9$$11$$13$$15$$0.5$$1$$1.5$$2$$2.5$$3$$3.5$$4$Side LengthM1(BS=64x16)M1(BS=64x8)M1(BS=64x4)M2(BS=64)M2(BS=128)M2(BS=256) (c) Regcount = 60 Figure 9: Processing speed comparison for summation $3$$5$$7$$9$$11$$13$$15$$10$$17$$24$$31$$38$$45$$52$Side LengthTime (ms)$V_{Z}$=2$V_{Z}$=3$V_{Z}$=4$V_{Z}$=5$V_{Z}$=6 (a) Regcount=32, $H_{Z}$=32 $3$$5$$7$$9$$11$$13$$15$$10$$17$$24$$31$$38$$45$$52$Side Length$V_{Z}$=2$V_{Z}$=3$V_{Z}$=4$V_{Z}$=5$V_{Z}$=6 (b) Regcount=32, $H_{Z}$=64 $3$$5$$7$$9$$11$$13$$15$$10$$17$$24$$31$$38$$45$$52$$65$Side Length$V_{Z}$=2$V_{Z}$=3$V_{Z}$=4$V_{Z}$=5$V_{Z}$=6 (c) Regcount=32, $H_{Z}$=128 $3$$5$$7$$9$$11$$13$$15$$10$$17$$24$$31$$38$$45$$52$Side LengthTime (ms)$V_{Z}$=2$V_{Z}$=3$V_{Z}$=4$V_{Z}$=5$V_{Z}$=6 (d) Regcount=48, $H_{Z}$=32 $3$$5$$7$$9$$11$$13$$15$$10$$17$$24$$31$$38$$45$$52$Side Length$V_{Z}$=2$V_{Z}$=3$V_{Z}$=4$V_{Z}$=5$V_{Z}$=6 (e) Regcount=48, $H_{Z}$=64 $3$$5$$7$$9$$11$$13$$15$$10$$17$$24$$31$$38$$45$$52$Side Length$V_{Z}$=2$V_{Z}$=3$V_{Z}$=4$V_{Z}$=5$V_{Z}$=6 (f) Regcount=48, $H_{Z}$=128 $3$$5$$7$$9$$11$$13$$15$$10$$17$$24$$31$$38$$45$$52$Side LengthTime (ms)$V_{Z}$=2$V_{Z}$=3$V_{Z}$=4$V_{Z}$=5$V_{Z}$=6 (g) Regcount=60, $H_{Z}$=32 $3$$5$$7$$9$$11$$13$$15$$10$$17$$24$$31$$38$$45$$52$Side Length$V_{Z}$=2$V_{Z}$=3$V_{Z}$=4$V_{Z}$=5$V_{Z}$=6 (h) Regcount=60, $H_{Z}$=64 $3$$5$$7$$9$$11$$13$$15$$10$$17$$24$$31$$38$$45$$52$Side Length$V_{Z}$=2$V_{Z}$=3$V_{Z}$=4$V_{Z}$=5$V_{Z}$=6 (i) Regcount=60, $H_{Z}$=128 Figure 10: Processing speed evaluation for Z2-ZNCC $3$$5$$7$$9$$11$$13$$15$$10$$25$$40$$55$$70$$85$$100$Side LengthTime (ms)Original(D=96)Original(D=128)Z2-ZNCC(D=96)Z2-ZNCC(D=128)RTNCC(D=90) Figure 11: Processing speed comparison for Z2-ZNCC We implemented our acceleration approach on an embedded GPU Jetson Tx2 and evaluated 1. 1. summation, 2. 2. Z2-ZNCC, 3. 3. FastDT, and 4. 4. the processing speed and matching accuracy of the stereo-matching system based on the Z2-ZNCC, census, semi-global matching (SGM), and FastDT, algorithms using the KITTI 2015 [5] benchmark. ### V-A Evaluation of Summation The processing speeds of the two summation methods described in Section IV are compared using 1,280x384 pixel images. In our evaluation, the maximum number of registers (Regcount) for each thread is limited to 32, 48, and 60; then comparisons are performed for different window sizes and GPU blocks. Figure.9 compares the results of the two summation methods. In each graph, M1 and M2 represent the two methods and BS represent the GPU block size. The $x$-axis represents matching windows’ side length from 3 to 15 and the $y$-axis represents their corresponding processing times. In Method 1, each block processes 64 columns and three sets of rows: 16, 8, and 4\. As the side length increases, the processing time increases accordingly. This is occurs not only due to the increase in the amount of the calculation, but also to the increase in memory occupancy, which reduces the number of active threads. On the other hand, in Method 2, each block processes three sets of columns: 64, 128, and 256. Because the sum is calculated based on the integral image, the processing time does not change with side length. Here, we note that in all cases, the processing time does not change significantly with the GPU-block size because the parallelism of threads is not affected. For all three different block sizes in Method 2, the processing times for the first two steps of obtaining an integral image is about 355 $\mu{s}$ and 496 $\mu{s}$, respectively, and the total time including the averaging calculation is close to 1.4 ms. Due to the integration, the processing speed of Method 2 is not as fast as that of Method 1 when the window size is smaller than 9x9. Therefore, the methods can be chosen according to actual requirements. However, when BS = 64x16 and Regcount = 48 (or Regcount = 60), Method 1 becomes invalid when the side length is larger than 9 (or 7); this is because the large number of threads causes the number of allocated registers to exceed the allowable upper limit. Therefore, for Method 1, a small GPU block should be chosen so as to ensure that the summation can be performed correctly. ### V-B Evaluation of Z2-ZNCC Figure.10 shows the evaluation results for Z2-ZNCC using the same 1,280x384 size images with a maximum disparity of 128. To show the performance of Z2-ZNCC more clearly, Method 2 is used for the summation because of its consistent processing speed. In addition to Regcount, the evaluation is performed under various $V_{Z}$ and $H_{Z}$ conditions, as described in Section.IV. $V_{Z}$ is changed from 2 to 6 and $H_{Z}$ is set to 32, 64, and 128. A large $V_{Z}$ means that the rate of data reuse is high; however, it also requires a larger Regcount, which will affect the parallelism of threads. Similarly, a large $H_{Z}$ suggests that a large number of $\Pi_{RT}(x,y,d)$ need to be calculated at the same time, which also requires many registers. According to this figure, we note that in most cases, when $V_{Z}$ = 2, Z2-ZNCC performs the worst because of its low data-reuse rate. Furthermore, the larger the side length, the worse the performance, because step 7 in Fig.6 does not work effectively. For $V_{Z}$ = 5 or 6, the performance is still poor when Regcount = 32 (because of the limited number of registers) or when $H_{Z}$ = 128 (because of too much use of registers). Additionally, for the case shown in Fig.10 (c), the larger the $V_{Z}$ value, the worse the performance of Z2-ZNCC, with the exception of $V_{Z}$ = 2. Comparing Figs.10 (f) and (i) with (c), we can see that, by increasing the Regcount to 48 and 60 respectively, the performance can be improved accordingly. However, the difference between (f) and (i) is not large, meaning that increasing Regcount does not improve performance proportionally. For other cases, the processing time increases with the side length without any obvious outliers. The results show that increasing Regcount improves performance more than changing the $H_{Z}$, as shown in Fig.10 (a), (d), and (g). Finally, when $V_{Z}$ = 4, Z2-ZNCC always performs consistently. This means that it achieves a good balance between hardware resources and calculations according to our evaluations. In addition, we also compared Z2-ZNCC to other methods under various maximum disparity values. Figure.11 shows the results of five methods: two Z2-ZNCC methods, two original progressive-scan methods, and RTNCC [19]. $D$ represents the maximum disparity value and is set to 90, 96 and 128. Based on the summation-speed comparison, we use Method 1 when the window sizes are smaller than 9x9 and Method 2 when the window sizes are larger than 7x7. In the Z2-ZNCC methods, $V_{Z}$ = 4, $H_{Z}$ = 32 and Regcount = 60. The proposed Z2-ZNCC methods work faster than other methods where the disparity is 96 and 128. When the side length is 3 and $D$ = 128, the processing time is 18.05 ms for Original and 11.31 ms for Z2-ZNCC: a 38% increase in speed. When the side length is changed to 15, the processing time is 72.09 ms for Original and 34.35 ms for Z2-ZNCC, meaning that the processing time is reduced by more than half. This is mainly because the Z2-ZNCC methods can reuse data in the vertical direction but the original methods cannot. Furthermore, compared with the latest RTNCC (its $D$ is only 90), our method for $D$ = 96 requires only 16.04 ms, which is 26% faster despite its computational complexity. Figure 12: Matching Result(KITTI 2015): (a) Left image & ground truth; (b) S1: Z2-ZNCC+SGM; (c) S2: Census+SGM; (d) S3: Census+FastDT; (e) S4: Z2-ZNCC+FastDT $2$$4$$6$$8$$10$$12$$14$$16$$18$$20$$22$$24$203350901703306601200error rate ($\%$)Time (ms)AnyNet (10 fps)RTS2NET (6.3 fps)MADNet (3.8 fps)StereoDNN (0.1 fps)RTNCC (46 fps)S3:Census+FastDT(35 fps)S2:Census+SGM(29 fps)S1:Z2-ZNCC+SGM(28 fps)S4:Z2-ZNCC+FastDT(32 fps) Figure 13: Comparison with other systems TABLE I: Comparison of Kernel Performance In Z2ZNCC. Method | GT(G/s) | GE(%) | IPW | IPC ---|---|---|---|--- Original ZNCC | 0.92 | 67.09 | 1.06e+05 | 3.05 Z2ZNCC | 1.16 | 76.95 | 3.06e+05 | 3.15 * • GT: gld_throughput GE: gld_efficiency IPW: inst_per_warp IPC:inst_per_cycle Table.I shows the comparison between Z2ZNCC and the original ZNCC in terms of kernel performance. We use gld_throughput and gld_efficiency to evaluate the performance of our kernel in memory access, and use inst_per_warp and inst_per_cycle for the computational efficiency. In particular, the number of our IPW is roughly three times the original, which shows that our method can effectively save on-chip resources to maintain a high degree of thread parallelism. ### V-C Evaluation of FastDT We evaluated FastDT in combination with Z2-ZNCC using the training set of the KITTI 2015 [5] benchmark because the accuracy must be evaluated precisely. 200 pairs of images are included in the training set, and their sizes are close to 1,250x375 with a maximum disparity value of 128. In our evaluation, Regcount = 48, T = 21 and the side length of ZNCC is set to 3. The two parameters in (III-B) are $\sigma_{s}$ = 5 and $\sigma_{r}$ = 52. $d_{arb}$ is set to 0 to effectively reduce the range of cost values because $d_{min}$ is rarely equal to zero in stereo matching. TABLE II: Comparison of DT in KITTI 2015. Direction | Float | Int32 | Int16 | Int8 (Ours) | Int16-NP ---|---|---|---|---|--- L2R∗(ms) | 13.10 | 13.76 | 13.63 | 13.68 | 13.59 R2L(ms) | 15.41 | 15.32 | 8.72 | 5.52 | 8.56 U2D(ms) | 15.12 | 15.05 | 8.74 | 7.22 | 8.65 D2U∗(ms) | 7.06 | 7.03 | 6.29 | 6.27 | 6.34 Frame Rate(fps) | 20 | 20 | 26 | 32 | 27 Error Rate(%) | 7.36 | 7.49 | 7.57 | 7.63 | 8.14 * • L2R∗:Z2-ZNCC & left to right. R2L: right to left. U2D: up to down. D2U∗: down to up & WTA. NP: non-normalization. Table II shows the processing time for each step of DT under four different methods, together with their accuracy. The four methods–Float, Int32, Int16, and Int8–use floating-point, integer, short integer, and character respectively during cost propagation. Because Z2-ZNCC calculates the cost value for each pixel serially, it is combined with the cost propagation from left to right (L2R). WTA is also combined with the cost propagation from down to up (D2U) to calculate the disparity map directly without transferring cost values back to the global memory. According to our observation, the cost values aggregated during the L2R step are not usually large. Therefore, in our implementation, cost-value normalization is performed only in the U2D steps of Float, Int32, and Int16, and in the R2L step of Int8. The encoding is performed at the end of the R2L step and the decoding is performed at the beginning of the U2D step. Additionally, to verify the necessity of cost-value normalization, we added the evaluation of Int16-NP, in which such normalization is not performed. In the L2R step, the processing time is roughly the same regardless of the data type because the transmission latency is hidden by the calculation time of Z2-ZNCC. On the other hand, the processing time of Int16 is only 8.72 ms in the R2L step and 8.74 ms in the U2D step, which are roughly half of the values for Float and Int32. For Int16-NP, since it is roughly the same as Int16 in terms of calculation and data transmission, there is no obvious difference in processing time. Our Int8 shows the fastest processing speed of all methods, even though data encoding & decoding is performed. This is because our method transfers data in 8-bit, which greatly reduces the latency of data transfer. Table.III shows the comparison between FastDT and the original DT in terms of global memory accessing. As the results of gld_throughput and gld_efficiency shown, our kernel significantly improves the efficiency of global memory access in both R2L and U2D directions. This is mainly benefits from our encoder compression method, which can improve transmission efficiency by bundling data. In our implementation, Int8 achieved the requirement for real-time processing with 32 fps, which is 60% faster than the 20 fps attained by Float. TABLE III: Comparison of Global Memory Efficiency In DT. Method | L2R | R2L | U2D | D2U ---|---|---|---|--- Original DT | GT(G/s) | 0.42 | 0.56 | 5.18 | 38.7 GE(%) | 50.3 | 20.29 | 59.39 | 94.16 FastDT | GT(G/s) | 0.41 | 11 | 17.08 | 22.6 GE(%) | 50.13 | 99.2 | 88.9 | 88.9 * • GT: gld_throughput GE: gld_efficiency. TABLE IV: Comparison of matching accuracies in KITTI 2015 Error Rate (%) | | | D1-bg | | | | D1-fg | | | | D1-all | ---|---|---|---|---|---|---|---|---|---|---|---|--- | S1 | S2 | S3 | S4 | S1 | S2 | S3 | S4 | S1 | S2 | S3 | S4 All / All | 6.60 | 6.98 | 15.12 | 7.77 | 13.58 | 17.04 | 23.56 | 16.71 | 7.76 | 8.66 | 16.52 | 9.26 All / Est | 6.54 | 6.82 | 12.56 | 7.61 | 13.54 | 16.92 | 22.78 | 16.52 | 7.71 | 8.51 | 14.24 | 9.10 Noc / All | 5.22 | 5.48 | 14.33 | 6.92 | 11.38 | 14.83 | 21.88 | 15.01 | 6.24 | 7.03 | 14.48 | 8.26 Noc / Est | 5.20 | 5.44 | 11.73 | 6.87 | 11.38 | 14.82 | 21.38 | 14.99 | 6.22 | 6.99 | 13.38 | 8.21 * • S1: Z2-ZNCC+SGM. S2: Census+SGM. S3: Census+FastDT. S4: Z2-ZNCC+FastDT. In the accuracy evaluation, Float has the lowest error rate of 8.16%. As the data width decreases, the error rate gradually increases. Compared with the error rate of 8.37% for Int16, that for Int16-NP is higher (8.81%) because of the overflow of the cost values. This shows that our cost normalization with nearly zero-mean works very effectively. Int8 has an error rate of 8.41%, only a 0.25% loss compared to Float. This means that our method can effectively increase the processing speed even as it maintains a high accuracy in stereo matching. ### V-D Evaluation of Stereo Matching Finally, to further clarify the effectiveness of our Z2-ZNCC and FastDT for stereo matching, we also combined them with the state-of-the-art algorithms SGM and census, respectively. Then, we compared their accuracies and processing speeds on a Jetson Tx2 GPU using the KITTI 2015 benchmark. SGM and census were chosen because the GPU system in [14] uses them and has a good performance (8.66% error rate, 29 fps) under the same conditions. Our implementation of the SGM is almost the same as [14]. The difference is that to clarify the role of Z2-ZNCC, we do not use the stream function. The two parameters in SGM–$P1$ and $P2$–are set to 18 and 185, respectively, because the results of Z2-ZNCC are multiplied by the coefficient $T$. Figure.12 shows the comparisons of three systems based on use our proposed methods, and one other system [14]. The four systems are Z2-ZNCC+SGM, Census+SGM [14], Census+FastDT and Z2-ZNCC+FastDT; they are expressed as S1, S2, S3, and S4, respectively. Four pairs of images are selected from the training set to clearly show the difference in the results of these systems. Figure.12 (a) shows the reference images and their ground truths. Figures.12 (b) to 12 (e) show the results of S1 to S4, respectively. The top half of each figure shows disparity map and the bottom half the corresponding error image compared with the ground truths. The color bar denotes the disparity range of [0,128), with blue representing the farthest objects and red the closest. As for the four sets of results No.000045, No.000076, No.000094, and No.000144, S1 in Fig.12 (b) shows a clear advantage in terms of accuracy. Its error rates of 3.61%, 2.01%, 2.85%, and 4.75% are obviously lower than those of other systems. Compared with S2 in Fig.12 (c), S1 works better in photometric distortions and weak-pattern areas, as shown by the red boxes in Figs.12 (c) and (d). This means that ZNCC has stronger robustness than census. S3 shows a disadvantage in accuracy. Its error rates are the highest among the four systems. This is because, in addition to the less accurate matching by census, DT also easily causes a fattening effect, making some details disappear, as shown in the yellow boxes in Figs.12 (d) and (e). By replacing census with ZNCC, the accuracy of S4 has been greatly improved, as shown in Fig.12 (e). This means that as long as a high precision is achieved in the cost matching stage, DT will not induce an excessive loss in accuracy. Table IV compares the matching accuracies of the four systems using the testing set of KITTI 2015 benchmark. The result is consistent with the above, and the order of accuracy is S1$>$S2$>$S4$>$S3. The use of ZNCC improves the accuracies of the census- based systems by 0.9% (S1 vs. S2) and 7.26% (S4 vs. S3). Figure.13 shows a comparison with other systems in terms of processing speed and accuracy. The $x$-axis represents the error rate and the $y$-axis represents the time required for processing in ms; thus, the closer the evaluation results are to the origin, the higher the accuracy and the faster the processing speed. All of the CNN-based systems (AnyNet [11], StereoDNN [12], MADNet [13], and RTS2Net [26]) achieved high accuracy, but had no speeds exceeding 10 fps. RTNCC [19] and our S3 (Census+FastDT) algorithm achieved real-time processing speeds of 46 fps and 35 fps, respectively. However, their error rates are larger than 16%, which also limits their usability. S2 (Census+SGM) [14] and our S1 (Z2-ZNCC+SGM) have almost the same processing speed, but our accuracy is 0.5% lower. This shows that our method plays a significant role in stereo matching. As mentioned above, our S4 (Z2-ZNCC+FastDT) improves the accuracy of S3 from 16.52% to 9.26%. Compared with the systems based on SGM (S1 and S2), S4 still has a 1% lower accuracy, but its processing speed is 17% faster, allowing it to truly achieve real-time processing. ## VI Conclusion In this paper, we proposed an acceleration method for the zero-means normalized cross correlation (ZNCC) template-matching algorithm for stereo vision on an embedded GPU. This method helps us reuse intermediate calculation results efficiently without frequently transferring them among the hierarchy of memories, leading to a higher processing speed. It also helps to improve the matching accuracy because of its stronger robustness. We evaluated our systems based on this method by using the KITTI 2015 benchmark and showed their efficiency. Since the proposed method does not rely upon the GPU experimental architecture, it can be used on different GPUs; furthermore, it is not limited to GPUs and can be used on other hardware platforms like FPGAs. To further improve out systems’ performance, we are planning to combine it with other cost-aggregation algorithms. This will be done in future work. ## References * [1] Wenqiang Wang, et al: ”Real-Time High-Quality Stereo Vision System in FPGA”. IEEE Transactions on Circuits and Systems for Video Technology. 25(10): 1696-1708 (2015) * [2] Mohammad Dehnavi, Mohammad Eshghi: ”FPGA based real-time on-road stereo vision system,” Journal of Systems Architecture: Embedded Software Design, 81: 32-43 (2017) * [3] Xuchong Zhang, Hongbin Sun, Shiqiang Chen, Lin Song, Nanning Zheng: NIPM-sWMF: ”Toward Efficient FPGA Design for High-Definition Large-Disparity Stereo Matching”. IEEE Transactions on Circuits and Systems for Video Technology. 29(5): 1530-1543 (2019) * [4] Pin-Chen Kuo, et al: ”Stereoview to Multiview Conversion Architecture for Auto-Stereoscopic 3D Displays”. IEEE Transactions on Circuits and Systems for Video Technology. 28(11): 3274-3287 (2018) * [5] Moritz Menze, Andreas Geiger: ”Object scene flow for autonomous vehiclesm” IEEE Conference on Computer Vision and Pattern Recognition(CVPR), 2015: 3061-3070 * [6] Oscar Rahnama, et al: ”Real-Time Highly Accurate Dense Depth on a Power Budget Using an FPGA-CPU Hybrid SoC,” IEEE Transactions on Circuits and Systems II: Express Briefs, 66-II(5): 773-777 (2019) * [7] Xuchong Zhang, He Dai, Hongbin Sun, Nan-Ning Zheng: ”Algorithm and VLSI Architecture Co-Design on Efficient Semi-Global Stereo Matching”. IEEE Transactions on Circuits and Systems for Video Technology. 30(11): 4390-4403 (2020) * [8] Christoph Hartmann, Ulrich Margull: GPUart - An application-based limited preemptive GPU real-time scheduler for embedded systems. Journal of Systems Architecture: Embedded Software Design. 97: 304-319 (2019) * [9] Sparsh Mittal: A Survey on optimized implementation of deep learning models on the NVIDIA Jetson platform. Journal of Systems Architecture: Embedded Software Design. 97: 428-442 (2019) * [10] Mustafa U. Torun, Onur Yilmaz, Ali N. Akansu: ”FPGA, GPU, and CPU implementations of Jacobi algorithm for eigenanalysis”. Journal of Parallel and Distributed Computing. 96: 172-180 (2016) * [11] Yan Wang, et al: ”Anytime Stereo Image Depth Estimation on Mobile Devices,” IEEE International Conference on Robotics and Automation (ICRA), 2019: 5893-5900 * [12] Nikolai Smolyanskiy, Alexey Kamenev, Stan Birchfield: ”On the Importance of Stereo for Accurate Depth Estimation: An Efficient Semi-Supervised Deep Neural Network Approach,” IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPR), 2018: 1007-1015 * [13] Alessio Tonioni, Fabio Tosi, Matteo Poggi, Stefano Mattoccia, Luigi di Stefano: ”Real-Time Self-Adaptive Deep Stereo,” IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019: 195-204 * [14] Daniel Hernández Juárez, et al: ”Embedded Real-time Stereo Estimation via Semi-Global Matching on the GPU,” International Conference on Computational Science (ICCS), 2016: 143-153 * [15] Zhuoyuan Chen, et al: ”A Deep Visual Correspondence Embedding Model for Stereo Matching Costs,” IEEE International Conference on Computer Vision (ICCV), 2015: 972-980 * [16] Qiong Chang, Tsutomu Maruyama: ”Real-Time Stereo Vision System: A Multi-Block Matching on GPU,” IEEE Access 6: 42030-42046 (2018) * [17] Martin Werner, Benno Stabernack, Christian Riechert: ”Hardware implementation of a full HD real-time disparity estimation algorithm,” IEEE Transactions on Consumer Electronics, 60(1): 66-73 (2014) * [18] Qiong Chang, Aolong Zha, Masaki Onishi, Tsutomu Maruyama: ”A GPU Accelerator for Domain Transformation-Based stereo Matching,” International Conference on Algorithms, Computing and Artificial Intelligence (ACAI), 2019 (pp. 370-376). * [19] Cui, Han, and Naim Dahnoun. ”Real-Time Stereo Vision Implementation on Nvidia Jetson TX2,” Mediterranean Conference on Embedded Computing (MECO), 2019. p. 1-5. * [20] Qiong Chang, et al: Z2-ZNCC: ZigZag Scanning based Zero-means Normalized Cross Correlation for Fast and Accurate Stereo Matching on Embedded GPU. IEEE International Conference on Computer Design (ICCD), 2020: 597-600 * [21] Chuan Lin, Ya Li, Guili Xu, Yijun Cao: ”Optimizing ZNCC calculation in binocular stereo matching. Signal Processing: Image Communication, 52: 64-73 (2017) * [22] Fan Rui, and Naim Dahnoun. ”Real-time implementation of stereo vision based on optimised normalised cross-correlation and propagated search range on a gpu,” IEEE International Workshop on Imaging Systems and Techniques (IST), 2017:pp.1-6. * [23] Cuong Cao Pham and JaeWook Jeon. ”Domain Transformation-Based Efficient Cost Aggregation for Local Stereo Matching,” IEEE Transactions on Circuits and Systems for Video Technology, 2013: pp.1119-1130. * [24] NVIDIA Corporation: ”NVIDIA CUDA C programming guide,” 2019.Version 10.1.243. * [25] Qiong Chang, Tsutomu Maruyama: ”Real-Time High-Quality Stereo Matching System on a GPU,” IEEE International Conference on Application-specific Systems, Architectures and Processors (ASAP), 2018: 1-8 * [26] Pier Luigi Dovesi, et al: ”Real-Time Semantic Stereo Matching,” IEEE International Conference on Robotics and Automation (ICRA), 2020: 10780-10787.
# CUNI Non-Autoregressive System for the WMT 22 Efficient Translation Shared Task Jindřich Helcl Charles University, Faculty of Mathematics and Physics Institute of Formal and Applied Linguistics Malostranské náměstí 25, 11800 Prague, Czech Republic <EMAIL_ADDRESS> ###### Abstract We present a non-autoregressive system submission to the WMT 22 Efficient Translation Shared Task. Our system was used by Helcl et al. (2022) in an attempt to provide fair comparison between non-autoregressive and autoregressive models. This submission is an effort to establish solid baselines along with sound evaluation methodology, particularly in terms of measuring the decoding speed. The model itself is a 12-layer Transformer model trained with connectionist temporal classification on knowledge-distilled dataset by a strong autoregressive teacher model. ## 1 Introduction In the past few years, non-autoregressive (NAR) models for neural machine translation (NMT) attracted interest from the research community (Gu et al., 2018; Lee et al., 2018). Given the conditional independence between the output states, the decoding process can be parallelized across time steps. In theory, this leads to higher decoding speeds. Since efficient decoding is claimed to be the main motivation of non- autoregressive models, the Efficient Translation Shared Task seems to be the appropriate venue to provide fair comparison between these models and their autoregressive counterparts. However, all submissions to this task were autoregressive so far (Birch et al., 2018; Hayashi et al., 2019; Heafield et al., 2020, 2021). Recently, Helcl et al. (2022) pointed out common flaws in the evaluation methodology of NAR models. We found that optimized autoregressive models still achieve superior performance over NAR models. The only scenario where NAR models showed some potential is GPU decoding with batch size of 1 (latency). Nevertheless, optimized autoregressive models were still both faster and better in terms of translation quality. The main purpose of this submission is to provide a reasonable baseline to future non-autoregressive submissions. ## 2 Model In our experiments, we use the non-autoregressive model proposed by Libovický and Helcl (2018) based on Connectionist Temporal Classification (CTC; Graves et al., 2006). We submit models that have been trained as a part of Helcl (2022). ### Architecture. The architecture is a 6-layer Transformer encoder, followed by a state- splitting layer and another stack of 6 Transformer layers. The state-splitting layer takes the encoder states, project them into $k$-times wider states using an affine transformation, and then split the states into $k$-times longer sequence while retaining the original model dimension. In the submitted model, we set $k=3$. The latter 6 layers cross-attend to the states immediately after state-splitting. We use Transformer model dimension of 1,024, 16 attention heads and a dimension of 4,096 in the feed-forward sublayer. The defining property of non-autoregressive models is that the decoding process treats output states as conditionally independent. In this architecture, we set the output sequence length to $k\times T_{x}$ where $T_{x}$ is the length of the source sentence. To allow for shorter output sequences, the any output state can produce an empty token. The training loss is then computed using a dynamic programming algorithm as a sum of losses of all possible empty token alignments which lead to the same output sentence. The schema of the architecture is shown in Figure 1. Input token embeddingsEncoder$\mathbf{h}$$W_{\text{spl}}\mathbf{h}$$\mathbf{s}$DecoderConnectionist Temporal Classification$w_{1}$$w_{2}$$w_{3}$$\varnothing$$w_{4}$$\varnothing$$w_{5}$$w_{6}$$\varnothing$$\varnothing$$\varnothing$$w_{7}$$w_{8}$$\varnothing$$w_{9}$$\varnothing$Output tokens / null symbols Figure 1: The CTC-based model architecture. We show the original image from Libovický and Helcl (2018). ### Training. We train our model on the knowledge-distilled data generated by the provided teacher (Chen et al., 2021). We use learning rate of 0.0001 in a inverse square-root decay scheme with 8,000 warm-up and decay steps. ### Implementation. We implement and train our model in the Marian toolkit (Junczys-Dowmunt et al., 2018). For the CTC implementation, we use the warp-ctc library111https://github.com/baidu-research/warp-ctc. We release our code at https://github.com/jindrahelcl/marian-dev. The trained model (and a number of different variants including models in opposite translation direction) can be downloaded at https://data.statmt.org/nar. ## 3 Results We refer the reader to the original paper for more details about the evaluation and its results. The model we submitted is denoted in the paper as “large”. A summary of the results follows. ### Translation Quality. To summarize the main findings, the model achieves a competitive BLEU score (Papineni et al., 2002) on the WMT 14 news test set (Bojar et al., 2014), which serves as a comparison to other non-autoregressive models that use this test set as the de facto standard benchmark. When evaluated on the WMT 19 news test set, our model obtains BLEU of 47.8, and a COMET score (Rei et al., 2020) of 0.1485. Compared to an similarly-sized autoregressive teacher model with 50.5 BLEU and COMET of 0.4110, we see a somewhat surprising gap between the COMET scores while BLEU scores are relatively close. We hypothesize that the errors that the non-autoregressive model makes are out of the training domain of the COMET models, which makes them more sensitive towards this kind of errors. ### Decoding Time. We evaluated our models on the one million sentences benchmark used in the previous editions of this task (Heafield et al., 2021), and we tried to reproduce the official hardware setup to large extent. For CPU decoding, we measured time to translate the test set on an Intel Xeon 6354 server from Oracle Cloud, with 36 cores. We run the evaluation only in the batch decoding mode, as the models were too slow to decode with a single sentence in batch. With the submitted model, the translation on CPU took 7,434 seconds (using batch of 16 sentences). We used a single Nvidia A100 GPU for GPU decoding. In the latency setup, the translation took 7,020 seconds, and the batched decoding ($b=128$) took 782 seconds. When compared with other submissions to this task, we find that the smallest difference is indeed found in the GPU decoding latency setting. However, the optimized models submitted to last year’s round still achieved significantly better decoding times. ## 4 Conclusions We submit a non-autoregressive system to the Efficient Translation Shared Task to the WMT 22. The model is trained with connectionist temporal classification, which allows the generation of empty tokens and thus making generation of sentences of various length possible while retaining the conditional independence among output tokens without explicit length estimation. The main motivation of this submission is to provide a reasonable baseline system for future research. We believe that the sub-field of non- autoregressive NMT cannot progress without controlled decoding speed evaluation, which is exactly what the shared task organizers provide. ## Acknowledgements This project received funding from the European Union’s Horizon Europe Innovation programme under grant agreement no. 101070350 (HPLT). Our work has been using data provided by the LINDAT/CLARIAH-CZ Research Infrastructure, supported by the Ministry of Education, Youth and Sports of the Czech Republic (Project No. LM2018101). ## References * Birch et al. (2018) Alexandra Birch, Andrew Finch, Minh-Thang Luong, Graham Neubig, and Yusuke Oda. 2018\. Findings of the second workshop on neural machine translation and generation. In _Proceedings of the 2nd Workshop on Neural Machine Translation and Generation_ , pages 1–10, Melbourne, Australia. Association for Computational Linguistics. * Bojar et al. (2014) Ondřej Bojar, Christian Buck, Christian Federmann, Barry Haddow, Philipp Koehn, Johannes Leveling, Christof Monz, Pavel Pecina, Matt Post, Herve Saint-Amand, Radu Soricut, Lucia Specia, and Aleš Tamchyna. 2014. Findings of the 2014 workshop on statistical machine translation. In _Proceedings of the Ninth Workshop on Statistical Machine Translation_ , pages 12–58, Baltimore, Maryland, USA. Association for Computational Linguistics. * Chen et al. (2021) Pinzhen Chen, Jindřich Helcl, Ulrich Germann, Laurie Burchell, Nikolay Bogoychev, Antonio Valerio Miceli Barone, Jonas Waldendorf, Alexandra Birch, and Kenneth Heafield. 2021. The University of Edinburgh’s English-German and English-Hausa submissions to the WMT21 news translation task. In _Proceedings of the Sixth Conference on Machine Translation_ , pages 104–109, Online. Association for Computational Linguistics. * Graves et al. (2006) Alex Graves, Santiago Fernández, Faustino Gomez, and Jürgen Schmidhuber. 2006. Connectionist temporal classification: Labelling unsegmented sequence data with recurrent neural networks. In _Proceedings of the 23rd International Conference on Machine Learning_ , pages 369–376, Pittsburgh, PA, USA. JMLR.org. * Gu et al. (2018) Jiatao Gu, James Bradbury, Caiming Xiong, Victor O. K. Li, and Richard Socher. 2018\. Non-autoregressive neural machine translation. In _6th International Conference on Learning Representations, ICLR 2018_ , Vancouver, BC, Canada. * Hayashi et al. (2019) Hiroaki Hayashi, Yusuke Oda, Alexandra Birch, Ioannis Konstas, Andrew Finch, Minh-Thang Luong, Graham Neubig, and Katsuhito Sudoh. 2019. Findings of the third workshop on neural generation and translation. In _Proceedings of the 3rd Workshop on Neural Generation and Translation_ , pages 1–14, Hong Kong. Association for Computational Linguistics. * Heafield et al. (2020) Kenneth Heafield, Hiroaki Hayashi, Yusuke Oda, Ioannis Konstas, Andrew Finch, Graham Neubig, Xian Li, and Alexandra Birch. 2020. Findings of the fourth workshop on neural generation and translation. In _Proceedings of the Fourth Workshop on Neural Generation and Translation_ , pages 1–9, Online. Association for Computational Linguistics. * Heafield et al. (2021) Kenneth Heafield, Qianqian Zhu, and Roman Grundkiewicz. 2021. Findings of the WMT 2021 shared task on efficient translation. In _Proceedings of the Sixth Conference on Machine Translation_ , pages 639–651, Online. Association for Computational Linguistics. * Helcl et al. (2022) Jindřich Helcl, Barry Haddow, and Alexandra Birch. 2022. Non-autoregressive machine translation: It’s not as fast as it seems. In _Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies_ , pages 1780–1790, Seattle, United States. Association for Computational Linguistics. * Helcl (2022) Jindřich Helcl. 2022. _Non-Autoregressive Neural Machine Translation_. Ph.D. thesis, Charles University. * Junczys-Dowmunt et al. (2018) Marcin Junczys-Dowmunt, Roman Grundkiewicz, Tomasz Dwojak, Hieu Hoang, Kenneth Heafield, Tom Neckermann, Frank Seide, Ulrich Germann, Alham Fikri Aji, Nikolay Bogoychev, André F. T. Martins, and Alexandra Birch. 2018. Marian: Fast neural machine translation in C++. In _Proceedings of ACL 2018, System Demonstrations_ , pages 116–121, Melbourne, Australia. Association for Computational Linguistics. * Lee et al. (2018) Jason Lee, Elman Mansimov, and Kyunghyun Cho. 2018. Deterministic non-autoregressive neural sequence modeling by iterative refinement. In _Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing_ , pages 1173–1182, Brussels, Belgium. Association for Computational Linguistics. * Libovický and Helcl (2018) Jindřich Libovický and Jindřich Helcl. 2018. End-to-end non-autoregressive neural machine translation with connectionist temporal classification. In _Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing_ , pages 3016–3021, Brussels, Belgium. Association for Computational Linguistics. * Papineni et al. (2002) Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. Bleu: a method for automatic evaluation of machine translation. In _Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics_ , pages 311–318, Philadelphia, Pennsylvania, USA. Association for Computational Linguistics. * Rei et al. (2020) Ricardo Rei, Craig Stewart, Ana C Farinha, and Alon Lavie. 2020. COMET: A neural framework for MT evaluation. In _Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)_ , pages 2685–2702, Online. Association for Computational Linguistics.
# Identifying Different Layers of Online Misogyny Wienke Strathern School of Social Sciences and Technology Technical University of Munich 80333 Munich, Germany <EMAIL_ADDRESS> Juergen Pfeffer School of Social Sciences and Technology Technical University of Munich 80333 Munich, Germany <EMAIL_ADDRESS> ###### Abstract Social media has become an everyday means of interaction and information sharing on the Internet. However, posts on social networks are often aggressive and toxic, especially when the topic is controversial or politically charged. Radicalization, extreme speech, and in particular online misogyny against women in the public eye have become alarmingly negative features of online discussions. The present study proposes a methodological approach to contribute to ongoing discussions about the multiple ways in which women, their experiences, and their choices are attacked in polarized social media responses. Based on a review of theories on and detection methods for misogyny, we present a classification scheme that incorporates eleven different explicit as well as implicit layers of online misogyny. We also apply our classes to a case study related to online aggression against Amber Heard in the context of her allegations of domestic violence against Johnny Depp. We finally evaluate the reliability of Google’s Perspective API—a standard for detecting toxic language—for determining gender discrimination as toxicity. We show that a large part of online misogyny, especially when verbalized without expletive terms but instead more implicitly is not captured automatically. _Keywords_ Online misogyny $\cdot$ Hate speech detection ## 1 Introduction In May 2016 actress, model, and activist Amber Heard went public and accused her then-husband, actor Johnny Depp, of intimate partner violence. She described a turbulent relationship and reported that "Johnny verbally and physically abused me throughout our relationship"111https://www.chicagotribune.com/entertainment/ct-johnny-depp- amber-heard-statement-20160531- story.html. She publicly posted a picture of injuries and filed for divorce. This sparked a firestorm on social media and online news sites, with commentators offering wildly differing opinions as to what happened and who was to blame. Of course, it is not possible for an outsider to know exactly what happened in this incident or what the dynamics were in the relationship. However, many were quick to make accusations and blame one or the other. In recent years more attention has been paid to the role of women in society, unfortunately also because of cases of real hatred against them.222https://onlineviolencewomen.eiu.com/ According to the Pew Research Center report on online harassment (Vogels, 2021), women and men are similarly often abused or threatened online. However, women are more likely than men to report being sexually harassed (16% vs. 5%) or stalked (13% vs. 9%) online. Young women are particularly often affected by sexual harassment on the Internet—33% of women under 35 say they have been sexually harassed online. With the constant growth of social media and microblogging platforms, hatred of women is becoming more prevalent, creating numerous examples of how misogyny can spread almost uncontrolled (Jane, 2017b; Ging and Siapera, 2018, 2019). Misogyny, defined as hatred or prejudice against women is expressed linguistically in a variety of ways, including social exclusion, discrimination, hostility, threats of violence, and sexual objectification. A study reveals the sheer scale and nature of online abuse faced by women and provides a resource to researchers and engineers interested in exploring the potential of machine learning in content moderation.333https://decoders.amnesty.org/projects/troll-patrol/findings In order to handle hateful content and protect people, automated systems are being used extensively to identify potentially problematic content. But a series of Failure-to-Act reports uncovers the dark side of social media platforms, more often experienced by women who are active on social media: "how harassment, violent threats, image-based sexual abuse can be sent by strangers, at any time and in large volumes, directly into your DMs without consent and platforms do nothing to stop it"444https://counterhate.com/research/hidden-hate/. Machine learning algorithms are deployed to scan content and flag it for human moderators. For instance, the Perspective API developed by Google Jigsaw was used to flag potentially toxic content for review on Wikipedia and in the New York Times comments section.555https://perspectiveapi.com/case-studies/ One challenge is to capture the linguistic specifics of hate speech, polarizing and offensive statements. Udupa (2020) observed that users of online social media platforms have managed to bypass automatic hate speech detection methods by using creative indirect forms of linguistic expression. According to Strathern et al. (2022) alternative methods to recognize moral slurs could be successfully implemented. Since hate is expressed in many different ways, automated methods can lack context sensitivity when determining implicit hate. To shed light on this discrepancy, we first examine which scientific theories and methods deal with the topic of misogyny. In the second step, we examine more closely how, based on theory and empirical work, classes of misogyny are built according to which content of hate speech can be assigned. In this, we assume that, in addition to a large amount of explicit hate speech, there is also a significant proportion of implicit misogynistic hate. Consequently, another goal of our study is to examine how well automated approaches to detect toxic language can identify misogyny. We collected 240,000 tweets from 2019–2021 containing the tweet handle @realamberheard and selected the top 5,000 most retweeted tweets to label and score them according to the classes identified in the literature. We then had these 5000 tweets analyzed by the Google Perspective API toxicity metric. A major outcome of this study is that online misogyny cannot be satisfactorily identified with this automated toxicity identification tool. ## 2 Review of Theories and Methods on Misogyny Our study is motivated by work dealing with a) misogyny, its modeling and detection, b) the classification of hate speech and c) the verification of hate speech detectors. ### 2.1 Misogyny According to Allen (2021) a definition of misogyny is disputed. Studies examining online anti-feminist rhetoric have used alternative terms, including "gender cyber hatred" (Jane, 2017a), "cyber harassment" (Citron, 2014), "technological violence" (Ostini and Hopkins, 2015), "gender trolling" (Mantilla, 2013), "e-bile" and "gender hate speech" (Jane, 2015). According to Code (2003), misogyny is defined as any of the following acts or feelings: sexual violence against women, physical violence against women, exclusion of women, promotion of patriarchy, belittlement of women, and marginalization of women. This framework is supplemented by specific forms of online misogyny by Zuckerberg (2018). Jane (2015) identifies technological determinism as one paradigm of ‘flaming’. The author argues that flaming is a product of the digital medium, favoured by specific features of online communication systems such as anonymity, invisibility, and disinhibition. Further research on flaming did not show that online abuse is highly gender-specific (Lee, 2016). In this analysis, reactions to ‘flaming’ or inflammatory messages refer to expletives targeting women under the general heading of ‘insult’ rather than being placed within the broader framework of structural misogyny. In contrast, Herring and Martinson (2004) examined gender differences in communication styles and feminist responses to ‘trolling’ and found that the ‘gender nature’ of online abuse messages and hate speech is paramount. Online misogyny has offline effects that requires investigation and further research. Citron and Norton (2011) hypothesize that the gendered nature of online harassment and digital abuse is an important facet of women’s overall ‘digital citizenship’. Megarry (2014) studies the psychological consequences of online misogyny, such as pseudonymous involvement and withdrawal, noting that online misogyny limits women’s voice and presence and reduces their digital engagement. The case of Amber Heard was the subject of a study by Whiting et al. (2019). They conducted their study from a psychological perspective on the subject of domestic violence. The authors examined the commenting behavior of users on various social media platforms. To better understand typical types of social media reactions to allegations of domestic violence, the authors performed a content analysis on Facebook. Five main categories were extracted, namely victim blaming, perpetrator blaming, couple blaming, withholding judgment, and mixed reactions to the process. The respective main topics also contain subtopics on reactions to the allegations. ### 2.2 Modeling Misogyny Determining and classifying misogyny in comments is a major challenge for humans and computers. There are various definitions and approaches to modeling this complex social and linguistic phenomenon. Fersini et al. (2018) developed a machine learning classification approach to model misogyny. The main categories are based on gender studies theory and contain classes that are used to determine comments. The classes are: stereotyping and objectification, dominance, derailment, sexual harassment, threats of violence, and discrediting. The categorization starts after an a priori distinction of whether a tweet is classified as misogynistic or not. In a study by Farrell et al. (2019) a misogyny model was developed to examine the flow of extreme language in online communities on Reddit. Based on feminist language criticism, the author created nine lexica that capture specific misogyny rhetoric (physical violence, sexual violence, hostility, patriarchy, stoicism, racism, homophobia, disparagement, and inverted narrative), and used these lexica to examine how language evolves within and between misogynist groups. Recent work by Guest et al. (2021) presents a hierarchical taxonomy for online misogyny and an expert-labeled data set that allows automatic classification of misogyny content. The taxonomy consists of misogynistic content, broken down into misogynistic pejoratives and treatment, misogynistic disparagement, and gendered personal attacks. ### 2.3 Detecting Online Misogyny In addition to modeling misogyny and detecting hate speech, we find studies examining how politically and socially active women are treated in current public debates. To gain insight into gender discrimination, various automated methods are used. In a study by Rheault et al. (2019), the authors applied machine learning models to predict rudeness directed at Canadian politicians and US senators on Twitter. In particular, they test whether women in politics are more affected by online abuse, as recent media reports suggest. Another article by Beltran et al. (2021) examined gender insults towards Spanish female politicians. In an analysis of tweets written by citizens, the authors found evidence of gender slurs and note that mentions of appearance and infantilizing words are disproportionately common in texts addressing female politicians in Spain. The results show how citizens treat politicians differently depending on their gender. Fuchs and Schäfer (2021) presented the results of an exploratory analysis of misogynistic and sexist hate speech and abuse against female politicians on Twitter, using computer-assisted corpus linguistic tools and methods, supplemented by a qualitative in-depth study of abuse by four prominent female politicians in Japan. Studies by Bauer (2015), Ditonto et al. (2014), Herrnson et al. (2003), Lawless (2015) suggest that voters evaluate candidates from the perspective of gender stereotypes and test how this affects attitudes and voting behavior. ### 2.4 Hate Speech Classification The annotation of hate speech is important for automated classification tasks. The classification scheme and its underlying assumptions are crucial for annotation. There are different approaches to this process such as predefined word lists or more complex models. One of the main difficulties is the definition of hate speech and its interpretation and therefore correct application. Recently, the Gab Hate Corpus was published (Kennedy et al., 2022), which uses a specially developed coding typology for annotating hateful comments. It was developed based on a synthesis of hate speech definitions drawn from legal precedents, previous hate speech coding typologies, and definitions from psychology and sociology. In addition, it contains a hierarchical clustering for dehumanizing and violent speech, as well as indicators for target groups and rhetorical peculiarities. A study by Ben- David and Fernández (2016) examined how overt hate speech and covert discriminatory practices circulate on Facebook. They argue that hate speech and discriminatory practices are not only explained by the motivations and actions of the users, but also emerge through a network of connections between the platform’s politics, its technological capabilities, and the communicative actions of its users. The difficult task of capturing implicit and explicit statements was addressed in a study by Gao et al. (2017). The authors proposed a weakly supervised two-path bootstrapping approach for an online hate speech detection model that uses big unlabeled data to address several limitations of supervised hate speech classification methods, such as corpus bias and the enormous cost of annotation. The implicitness of linguistic statements is also the subject of a work by Frenda et al. (2022). The authors proposed a number of statistical and computational analyses that support reflections on indirect propositions that focus on the creative and cognitive aspects of implicitness. In a more recent work by ElSherief et al. (2021), implicit statements were used for machine learning tasks to introduce a theoretically based taxonomy of hate speech. Wiegand et al. (2021) studies the detection of implicitly abusive language, that is, abusive language that is not conveyed by abusive words. In their position paper, they explain why existing datasets make learning implicit abuse difficult and what needs to be changed in the design of such datasets. ### 2.5 Bypassing Hate Speech Detection Tricking or recalibrating automated methods results from the observation that the underlying assumptions of common machine methods do not adequately define group-specific hatred. That is, there seems to be a discrepancy between methods for operationalization tasks and the complexity of social processes. Against this background there are ways to trick hate speech detection methods or to test them for their measurement accuracy and validation. Both, cultural and associated linguistic peculiarities are thus taken into account. There are studies that try to capture culture- and language-specific hatred, which machines have difficulty recognizing. Zannettou et al. (2020) focused on examining the spread of antisemitic content. The authors carried out a large- scale quantitative analysis to discover abnormalities in language use. The results show that there are several distinct facets of antisemitic language, ranging from slurs to conspiracy theories, drawing on biblical literature and narratives expressed differently in the language. Herein, antisemitism is considered as one form of hate speech and the authors developed a method to comply with this. Another study by Gröndahl et al. (2018), evaluates empirically previously proposed models and datasets for classifying hate speech. The results show that none of the pre-trained models performs well when tested with a different dataset. The authors assert that the characteristics indicative of hate speech are not consistent across different datasets. The results show that the definitions of hate speech do not seem to be consistent and that they need further differentiation and context sensitivity. Another study by Hiruncharoenvate et al. (2015) examined ways to circumvent the observation of the state in the Chinese language, which suppresses free speech. In China, political activists use homophones (two words that are written differently and have different meaning but sound the same, e.g., brake/break) of censored keywords to avoid detection by keyword- matching algorithms. The authors claim that it is possible to expand this idea in a way that makes them difficult to counteract. One result of this work is to mathematically (and almost optimally) change the content of a post by replacing censored keywords with homophones. So, by tricking the system with linguistic creativity, they bypass the derived rules for automatic speech recognition on Weibo. ## 3 Overview of Misogyny Classes from the Literature Based on the theories and methods discussed above, we have developed a classification scheme for online misogyny that covers most of the aspects discussed in the related literature. These classes include explicit and implicit misogynistic language and are presented in the following. Some of these classes are close to each other in their definitions and are not always easy to distinguish. The case study in the second part of this manuscript will show that they significantly overlap when used for coding real-world messages. The goal of identifying misogyny classes was not to identify unambiguous definitions, but to cover a wide variety of aspects of hate against women. ### 3.1 Explicit Misogyny In explicit misogynistic statements users openly attack, insult, or even threaten a woman (Waseem et al., 2017; Gao et al., 2017). Based on the literature presented above, we have identified the following four subcategories of explicit misogyny. Call for action/violence. This class implies verbal threads that intend to punish a target physically. Statements in which users call for deletion, prison, boycott, or sending the target to a psychiatric institution (Fersini et al., 2018). Personal insult, denigration. Personal insults and denigration intend to cause harm to a target verbally. Statements containing harmful wishes, demeaning, threatening, denigrating, inciting, defaming, use of slur words (Fersini et al., 2018; Guest et al., 2021; Farrell et al., 2019). Gendered personal attack. Gendered personal attacks refer to stereotypes of women. From these stereotypes, verbal (misogynistic) attacks derive. Statements that contain misogynistic speech and swearwords, revenge porn, or are sexually motivated because the target is a woman (Fersini et al., 2018; Guest et al., 2021; Farrell et al., 2019). Weakness of character, intellectual inferiority. Making negative judgments of a woman’s moral and intellectual worth using explicit slur words. Statements that call a woman controlling, psychotic, a liar, hypocritical, narcissistic, or manipulative (Fersini et al., 2018; Guest et al., 2021; Farrell et al., 2019). ### 3.2 Implicit Misogyny Implicit statements of misogyny include cynicism and sarcasm, skepticism and distrust, insinuation, accusations, speculation and questioning of credibility, a demonstration of power, and taking a position (Waseem et al., 2017; Gao et al., 2017; ElSherief et al., 2021; Frenda et al., 2022). Cynicism, sarcasm. Cynicism and sarcasm represent a very derogatory attitude of a person towards others. It is expressed in an indirect form and is spiteful and bitter. Contextual knowledge is needed. Statements in which in a subliminal way, a rejecting attitude is shown (Whiting et al., 2019). Skeptical attitude, distrust. That includes “facts” or other details to undermine a woman’s account. Doubtfulness toward a woman’s claims or accusations. Questions whether the target had lied before and therefore cannot be trusted (Whiting et al., 2019). Imputation. Imputation is considered as assumptions that the target behavior is motivated by flawed motivations. That includes statements that show a moral judgment, containing comments where a woman is described as revenge-seeking, vindictive, attention-seeking, monetarily driven (Whiting et al., 2019). Allegation. The category implies actions in which the evidence and allegations are challenged suggesting intentionally motivated actions. Statements of users that offer facts that refute a woman’s account, evidence to the contrary (Whiting et al., 2019). Speculation, denying credibility. This category includes an investigative- style attitude. Speculations and doubts about the target’s behavior. Commenting on the case, e.g., of domestic violence and its severity, we find claims about how this case might affect future reporting, users offering life stories to undermine the target’s account, personal expertise, intent to prove something, credibility from experience, claiming special predictive power (Whiting et al., 2019). Demonstration of power. The category implies a power relation between one gender and the other. Statements in which support for the man is demonstrated (Fersini et al., 2018). Taking position. Taking position or ‘flipping the narrative’ encapsulates terms and expressions that refer to the relationship between the target and the perpetrator. Statements on who is the ‘perpetrator’ and who is the ‘victim’ (Fersini et al., 2018; Guest et al., 2021; Farrell et al., 2019). ### 3.3 Examples for Misogyny Classes In order to study the prevalence of these misogynistic classes on social media, we have collected and analyzed messages addressing Amber Heard’s Twitter account @realamberheard in a case study in the next section. Here, in Table 1, we show sample tweets to exemplify these classes. Since the content contains explicit hate speech and profanity, we have redacted the texts. Class | Example Tweet ---|--- Call for action/violence | Oh @realamberheard …. You ignorant witch. We ALL already know you’re the guilty one here. | Johnny’s innocence has been proven. You’re just trying to buy time, before you (hopefully) | have you sit your scronny ass in a jail cell. You speak nothing but venomous lies. #JohnnyDepp Personal insult, denigration | Seriously, how fucking sick you have to be to pull a "prank" like this on someone ? What kind of | gross bitch would think pooping in people’s bed is funny ? Well, apparently @realamberheard does. | #JusticeForJohnnyDepp Gendered personal attack | Not a johnny Depp fan but @realamberheard claims have more holes than swiss cheese. I dont | understand females who can’t make their own money and want to pocket off someone elses. It’s | hard to find a victim that no one sides with in todays world but I think we all call bs on AH. Weakness of character, | Look what headline just poped up on sky news! @realamberheard you dirty little Lier! intellectual inferiority | #AmberHeardIsALiar #JusticeForJohnnyDepp Cynicism, sarcasm | @realamberheard Yes, the excitement around #JusticeLeague was huge … definitely nothing to do | with you though. Imagine being in a 4 hour movie for 5 minutes and being the most insufferable | part of it. Skeptical attitude, distrust | I just noticed the ’actor/ activist’ claims in your biog @realamberheard !! Well, you certainly | are an actress for real!! Only trouble is that the majority of your acting seems to be done OFF | stage!! And you have set ’activism’ back decades dear!! Ugh, you are some piece of work! Imputation | @realamberheard @realamberheard Put your hand down and stop exploiting Evan’s story to sway | the public perception back in your favor. Don’t act like you didn’t break bread and hang out with | Marilyn Manson for years after his relationship with ERW/ your o Refutation | Listen bitch, I just saw a video about you demanding Depp supporter info for some legal implica- | tions!!If you want any info about me just DM me and I’ll be MORE than happy to bring you upto | speed!! @realamberheard I am allowed my opinion and you are scum (&u better pay my airfare!) Speculation, | @realamberheard You do not represent women nor survivors. I stand with Johnny Depp, Kate denying credibility | James, Jennifer Howell, Lily-Rose Depp, Hilda Vargas, Samantha McMillen, Katherine Kendall, | Trinity Esparza and ALL THE OTHER women and men who knows your true color Demonstration of Power | Justice for Johnny Depp outside @wbpictures studio where @realamberheard is currently filming | @aquamanmovie #JohnnyDepp #JusticeForJohnnyDepp #JOHNNY #AmberHeard Taking a position | @realamberheard is not a victim, she is the perpetrator. Table 1: Misogynic classes and example tweets ## 4 Case Study To assess the importance of the misogyny classes presented in the manuscript, we conducted a case study using Twitter data related to the celebrity domestic violence abuse case between Amber Heard and Johnny Depp. In the following, we describe the data and the annotation process as well as present quantitative results showing the prevalence of our explicit and implicit misogyny classes in the data. Kennedy et al. (2022) documented that the annotation of hate speech has been shown to lead to a high level of disagreement between the annotators, see also Ross et al. (2016). According to Mostafazadeh Davani et al. (2020) this is due to a combination of factors, including differences in understanding of the definition of hate speech, interpretation of the annotated texts, or assessment of the harm done to certain groups, i.e. inconsistent application of the definition of hate speech to different social groups. Data. By utilizing the Twitter Academic API (Pfeffer et al., 2023) we collected 266,579 original tweets (excluding re-tweets) in January of 2022 that contained the account @realamberheard in the tweet texts. This resulted in 266,579 tweets (2019: 64,334 tweets, 2020: 117,231 tweets, 2021: 85,014 tweets). For the annotation process, we extracted 5,000 tweets that have been retweeted most often. ### 4.1 Annotation Process For our case study we employed two annotators, a graduate student who is also a co-author on this paper and was instrumental in developing the misogyny classes (annotator $1$), as well as an undergraduate student who was new to the topic (annotator $2$). The annotators were briefed with an introduction to the topic in general and then presented with the misogyny classes. All the information presented together with coding examples was also shared in a coding manual. The manual also includes detailed descriptions of the individual coding steps and further explanations of the definition of the classes and the coding method according to the literature. We analyzed the entire tweet at the sentence and word level, including the use of emoticons and content on the websites following URLs appearing in tweets. We looked at images, memes, or quotes, and watched linked videos. Each tweet was rated by the annotators based on all of its content. If the tweet contained statements supporting Amber Heard was neutral, or contained advertising, we annotated this tweet as _other_ and ignored the tweet in the subsequent analytical steps. We used the eleven misogyny classes for annotation. After the annotation process, we created the explicit/implicit annotation from the eleven classes following the categorization described above. A single tweet could be annotated with multiple misogyny classes. If a tweet contained multiple sentences where one was implicit and one was explicit, we chose the explicit class due to the fact that a Tweet with explicit misogynistic content will be perceived as being explicit in its entirety. Coding 11 classes with multiple overlapping definitions will lead to low levels of completely identical annotations. However, when comparing the explicit/implicit/other classes among the two annotators, the overall level of agreement between the annotators was acceptable. We can report the following values for Krippendorff’s alpha (Krippendorff, 2011): explicit 0.779, implicit 0.736, other 0.867. ### 4.2 Prevalence of the Misogyny Classes For further analysis of this article, the annotator $1$ manually compared the annotations from both annotators for all 5,000 tweets and harmonized the annotations into a single mapping of tweets to misogyny classes. The frequencies and proportions of the classes in the overall dataset as well as in the misogynistic tweets can be seen in Table 2. Shockingly, two-thirds of the most retweeted tweets addressing Amber Heard’s Twitter account have been classified into explicit (35.6%) or implicit (30.3%) classes of misogyny. While explicit and implicit classes can overlap within tweets, the meta- classes explicit/implicit are mutually exclusive (see above). | Misogyny Class | Frequency | All | Misogyny ---|---|---|---|--- Explicit | Call for Action | 681 | 13.6% | 20.4% (35.6%) | Personal Insult | 1,649 | 33.0% | 49.5% | Gendered Personal Attack | 730 | 14.6% | 21.9% | Intellectual Inferiority | 1,325 | 26.5% | 39.8% Implicit | Cynicism/Sarcasm | 367 | 7.3% | 11.0% (30.3%) | Scepticism/Distrust | 461 | 9.2% | 13.8% | Imputation | 556 | 11.1% | 16.7% | Allegation | 546 | 10.9% | 16.4% | Speculation | 305 | 6.1% | 9.2% | Demonstration of Power | 459 | 9.2% | 13.8% | Taking up a Position | 181 | 3.6% | 5.4% N | | | 5,000 | 3,331 Table 2: Frequencies and proportions of misogyny classes in all 5,000 annotated tweets as well as proportions in 3,331 misogynistic tweets. ## 5 Comparing Misogyny Classes with Google’s Perspective API Google’s Perspective API is one of the standards for identifying toxic language on online platforms and is described as "the product of a collaborative research effort by Jigsaw and Google’s Counter Abuse Technology team exploring machine learning as a tool for better discussions online."666https://www.perspectiveapi.com/research/. In this section, we will test how well the toxicity scores of this API are capable of identifying online misogyny as operationalized with our eleven classes to get an understanding of how useful these approaches can be in automatically identifying online misogyny. We worked out the different attributes and evaluation methods of the API as the first step for comparison. In the second step, we applied the API to the same dataset of 5,000 tweets. For each tweet, the API specifies a range of values for each of its categories. In the third step, we compared the values using statistical methods and applied network analysis to show the co- occurrence of classes and their average toxicity value reported by the Perspective API. ### 5.1 Attributes of Perspective API The Perspective API predicts the perceived impact of a comment on a conversation by evaluating that comment with a set of emotional concepts called attributes, namely toxicity, severe toxicity, identity attack, offense, threat, and profanity. The returned values are in the range [0.1] and are an indicator of the likelihood that something will be perceived as toxic. The higher the score, the more likely it is that the patterns in the text are similar to the patterns in comments that others have identified as toxic. The values are intended to allow developers/users to set a threshold and ignore values below that value. Values around $0.5$ indicate that the model does not know if it is similar to toxic comments. The Google recommended threshold setting is $0.7$. These thresholds are central to interpretation. | Misogyny Class | Average Toxicity ---|---|--- Explicit (35.6%) | Call for Action | 0.504 ø= 0.572 | Personal Insult | 0.589 | Gendered Personal Attack | 0.619 | Intellectual Inferiority | 0.577 Implicit (30.3%) | Cynicism/Sarcasm | 0.356 ø= 0.493 | Scepticism/Distrust | 0.527 | Imputation | 0.557 | Allegation | 0.423 | Speculation | 0.572 | Demonstration of Power | 0.436 | Taking up a Position | 0.581 Other (34.1%) | Marketing/PR | 0.193 Table 3: Categories and Average Toxicity for Explicity and Implicity. ### 5.2 Measuring Toxicity for Misogyny Classes To measure the average toxicity for the misogyny classes, we compare Google’s probability score to our manual coding by summing up the codes divided by the number of tweets in each meta-class. The results show that the average toxicity score by Google for our category of explicit misogyny is $0.572$. For our category of implicit misogyny, the average score by Google is $0.493$. These numbers already are a strong indicator that toxicity, as identified with the Perspective API, is a poor predictor of our variable of online misogyny, and in particular of implicit hate against women. Table 3 reveals the average toxicity scores for each class. In Figure 1 we can further see the density distribution of toxicity scores for each of the meta-classes of tweets with explicit or implicit misogyny as well as others. Figure 1: Distribution of toxicity scores from Google’s Perspective API for tweets with explicit or implicit misogyny as well as tweets without misogynistic content. In the _other_ sub-figure we can clearly see that there are almost no tweets that have been identified by the Perspective API as toxic that we have not also classified as misogynistic—consequently, the automated coding does not create false positives. The explicit language used for the classes that we have summarized with the meta-class _explicit_ can be identified by the Perspective API to a certain degree, and the peak of the score distribution is above the standard threshold of $0.7$. In other words, tweets coded with explicit misogyny contain text patterns that are similar to the patterns in comments that have been identified as toxic when the Perspective API models have been trained. Unfortunately, the picture looks different when looking at the distribution of scores for the implicit misogyny classes. Here, the resulting toxicity scores are almost evenly distributed, having more scores with very low values than with very high values. Consequently, the tweets coded with implicit misogynistic classes do not reflect text patterns that are similar to the patterns that have been identified as toxic in the Perspective API’s training data. ### 5.3 Co-Occurrence Network of Misogyny Classes In addition to statistical analysis, we built a co-occurrence network that maps manual coded classes and the average toxicity scores by the Perspective API (3). Nodes represent the eleven classes and the edge value is the number of co-occurrences, i.e., the co-occurrence of classes within a tweet. The edge color is the edge value, and the node size is the proportion of the number per code divided by the number of tweets. The node color is the average toxicity value from the Perspective API where blue means low and red means high toxicity values. In the centre, we can find the dominant four explicit classes which are identified to a certain degree as being toxic. The classes are well connected with each other. Explicit abusive statements come with similar forms of abusive language. For implicit statements, the picture looks different. In the periphery, we can find the seven classes of our meta-class implicit. Implicit misogynistic statements occur more with various forms of explicit abusive language and less among each other. In many cases, something is said implicitly, but it co-occurs with an explicit abusive statement. As mentioned above, we decided to code a tweet as explicit if both classes occurred. But the network analysis reveals the co-occurrence of explicit and implicit abusive language against women within one statement. It offers a more qualitative comparison of stereotypical hating: statements that contain a demonstration of power come with inferiority and insults. A skeptical attitude comes with abusive terms of inferiority, imputation, gendered personal attacks, and insults. Statements of speculation and doubt come with sarcastic and gender-attacking language. Despite the proximity of all classes, the network reveals a distinction between explicit and implicit misogyny. Figure 2: Co-Occurrences of Categories within a Tweet ### 5.4 Interpretation and Discussion We asked how well an automated approach like Google’s Perspective API performs in detecting misogyny. Based on our study, two things become apparent: Google’s text model does recognize explicit misogyny in the text patterns as toxic. However, the model does not recognize implicit misogyny in text patterns as toxic. The interpretation of the following tweets underlines the challenges of detecting and understanding implicit/indirect hate: "@realamberheard It’s the way you think that posing this is going to change public perception of you. We heard what you did in your own words. A failure in the system isn’t uncommon, so thank you for proving that male victims will never be taken seriously." A user recapitulates what has happened, draws conclusions for men, and thanks the target person for that in a very calm manner. But reading the tweet with contextual knowledge makes one understand that the thankful gesture is a cynical one. No keyword of hate can be found here; the words are all positive, but the underlying assumption is an accusation against Amber Heard and against her gender. None of the scores indicates harm in this tweet: Toxicity: 0.28, Severe Toxicity: 0.17, Identity Attack: 0.26, Offense: 0.07, Threat: 0.21 and Profanity: 0.14. In another tweet, a user comments on what has happened and concludes that this behavior is not acceptable. The tweet contains a link to a screenshot in which impressions of what happened are reflected. Again, there is no harmful word, it all sounds positive in isolation, but clearly implies that this user is rejecting the behavior of the woman and at the same time accusing her of what she has done: "@realamberheard I had to translate to really understand where you’re coming from. And no I wouldn’t encourage my daughter or sister to do what you did (URL redacted)". But here as well, the scoring is very low. Toxicity: 0.20, Severe Toxicity: 0.12, Identity Attack: 0.11, Offense: 0.07, Threat: 0.16 and Profanity: 0.14. The following example can exemplify how the toxicity score can be influenced by a single word that is interpreted as negative, even though the tweet could be interpreted as being funny: "@realamberheard @USNatArchives She will forever be known as the lady who pooped on Johnny Depp’s bed." Toxicity: 0.69, Severe Toxicity: 0.15, Identity Attack: 0.74, Offense: 0.65, Threat: 0.34, and Profanity: 0.74. There may be several reasons for this discrepancy to detect misogyny. One reason could be that there was no misogynistic content in the training texts for the human annotators. Or misogyny was never defined as an annotation class, hence, annotator could not label it. Annotators could not be informed / trained on the topic of misogyny and, therefore, could not recognize and annotate it in the texts. Although we do not know how the data sets were constructed and the model trained, we can summarize that Google’s Perspective API struggles with identifying text patterns containing implicit misogynistic statements. ## 6 Conclusion In this manuscript, we have presented a classification scheme that incorporates 11 classes of misogyny and have described a data set that contains misogynistic content labels from Twitter. We have also provided a detailed coding book and a data set with all of the labels. The data set benefits from a detailed classification scheme based on the existing literature on online misogyny. The involvement of trained annotators and an adjudication process also ensures the quality of the labels. We applied the classification scheme to a case related to online aggression against Amber Heard in the context of her allegations of domestic violence against Johnny Depp. For 5,000 tweets, we identified online misogyny operationalized with our eleven classes for two-thirds of the tweets, one- third as explicit misogyny, and one-third as implicit misogyny. Finally, we evaluated the reliability of Google’s Perspective API for determining implicit misogyny and found that this approach can identify explicit misogyny to a certain extent, but fails with identifying implicit misogyny. Ethical considerations and limitation. Ethical considerations must be taken into account with regard to the training and supervision of the annotator. The annotator was an undergraduate student who first read the typology and coding manual and secondly performed a test of approximately 50 messages that had been previously annotated and approved by one of the authors. Kennedy et al. (2022) pointed out the pressing concern that annotators may experience trauma or similar negative effects such as desensitization when annotating hate speech. On the basis of our own annotation experiences, we would like to highlight these thoughts. Although no research has examined the effects of constant daily exposure to hate speech on human moderators, there is evidence that exposure to linguistic and visual violence online can have negative mental health effects, as shown by Kwan et al. (2020). We also provided the annotator with Kennedy’s suggested written guide 777https://www.apa.org/ptsd- guideline/ptsd.pdf to help detect changes in cognition and avoid secondary trauma. It advises the user to take breaks and not imagine traumatic situations. The annotator was asked to stay in contact with the author of the study if she notices symptoms of PTSD, which are also listed in the guide. The purpose of this guide is to normalize feelings that are negatively impacted by work, provide trauma-specific education, recognize signs of traumatic stress and provide a mechanism for support as a preventative measure against secondary traumatic stress. A limitation of this study is the fact that we do not know whether the Perspective API’s text models contained misogynistic content and we do not know whether the data sets contained implicit/indirect forms of hate. Furthermore, we do not know whether the annotators were informed or trained on the topic of misogyny or implicit/indirect forms of hate. However, our results show that there may be a lack of information on misogyny according to existing definitions. Implications and Future Work. Given real-world online aggression against women, it is probable that Google’s toxicity model would not identify it. Thus, a huge fraction of implicit misogyny texts would stay visible and would not be deleted or otherwise acted upon. Misogynous behavior and target classification still remain a very challenging problem. One approach may be to create lexicons capturing specific misogynistic rhetoric and improve annotation scheme. Another challenge is to capture the peculiarities of implicity or indirect forms of hate in language. Language is very context- sensitive, and a negative tone can be expressed without a clear negative key word. Moreover, implicit sentences depend decisively on the non-linguistic accompanying signals. With our work, we would like to enhance existing research on investigating linguistic distinction between implicit and group- specific hate rhetoric. Furthermore, as we have seen from the network perspective, aside from the technical solution questions arise on how and why these different sub-classes are closely connected. From a gender perspective, we ask why are these stereotypes so consistent over time? Due to the still increasing number of users and posts in social media, automated annotation based on machine learning is inevitable. There is no other way to handle the sheer volume of text. At the same time, it becomes apparent that the proportion of aggressive misogynistic speech is increasing sharply. An assessment and, if necessary, the deletion of unacceptable statements is imperative for the protection of people. Especially with regard to women, their protection is of immense importance to enable participation in public discourse and avoid withdrawing because of fear of being attacked or marginalized. However, the key to better handling the problem is to better understand the phenomenon of misogyny. ## 7 Acknowledgments The authors gratefully acknowledge the financial support from the Technical University of Munich - Institute for Ethics in Artificial Intelligence (IEAI). Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the IEAI or its partners. The labeled dataset and the coding book will be made publicly available. ## References * (1) * Allen (2021) Amy Allen. 2021\. Feminist Perspectives on Power. In _The Stanford Encyclopedia of Philosophy_ , Edward N. Zalta (Ed.). Metaphysics Research Lab, Stanford University. * Bauer (2015) Nichole M. Bauer. 2015\. Emotional, Sensitive, and Unfit for Office? Gender Stereotype Activation and Support Female Candidates. _Political Psychology_ 36 (2015). * Beltran et al. (2021) Javier Beltran, Aina Gallego, Alba Huidobro, Enrique Romero, and Lluís Padró. 2021. Male and female politicians on Twitter: A machine learning approach. _European Journal of Political Research_ 60 (2021), 239–251. * Ben-David and Fernández (2016) Anat Ben-David and Ariadna Matamoros Fernández. 2016\. Hate Speech and Covert Discrimination on Social Media: Monitoring the Facebook Pages of Extreme-Right Political Parties in Spain. _International Journal of Communication_ 10 (2016). * Citron (2014) Danielle Keats Citron. 2014\. Hate Crimes in Cyberspace - Introduction. _Harvard University Press_ (2014). * Citron and Norton (2011) Danielle K. Citron and Helen Norton. 2011. Intermediaries and Hate Speech: Fostering Digital Citizenship for Our Information Age. _Boston University Law Review_ 91 (2011). * Code (2003) Lorraine Code. 2003\. _Encyclopedia of Feminist Theories_. Routledge. * Ditonto et al. (2014) Tessa M. Ditonto, Allison J. Hamilton, and David P. Redlawsk. 2014. Gender Stereotypes, Information Search, and Voting Behavior in Political Campaigns. _Political Behavior_ 36 (2014), 335–358. * ElSherief et al. (2021) Mai ElSherief, Caleb Ziems, David Muchlinski, Vaishnavi Anupindi, Jordyn Seybolt, Munmun De Choudhury, and Diyi Yang. 2021\. Latent Hatred: A Benchmark for Understanding Implicit Hate Speech. In _Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing_. Association for Computational Linguistics, 345–363. * Farrell et al. (2019) Tracie Farrell, Miriam Fernandez, Jakub Novotny, and Harith Alani. 2019. Exploring Misogyny across the Manosphere in Reddit. In _Proceedings of the 10th ACM Conference on Web Science_. * Fersini et al. (2018) E Fersini, P Rosso, and M Anzovino. 2018. Overview of the Task on Automatic Misogyny Identification at IberEval 2018. In _EVALITA@CLiC-it_. * Frenda et al. (2022) Simona Frenda, Viviana Patti, and Paolo Rosso. 2022\. Killing me softly: Creative and cognitive aspects of implicitness in abusive language online. _Natural Language Engineering_ (2022). Publisher: Cambridge University Press. * Fuchs and Schäfer (2021) Tamara Fuchs and Fabian Schäfer. 2021. Normalizing misogyny: hate speech and verbal abuse of female politicians on Japanese Twitter. _Japan Forum_ 33, 4 (2021). * Gao et al. (2017) Lei Gao, Alexis Kuppersmith, and Ruihong Huang. 2017\. Recognizing Explicit and Implicit Hate Speech Using a Weakly Supervised Two-path Bootstrapping Approach. In _Proceedings of the Eighth International Joint Conference on Natural Language Processing_. * Ging and Siapera (2018) Debbie Ging and Eugenia Siapera. 2018. Special issue on online misogyny. _Feminist Media Studies_ 18 (2018). * Ging and Siapera (2019) Debbie Ging and Eugenia Siapera (Eds.). 2019. _Gender Hate Online: Understanding the New Anti-Feminism_. Springer International Publishing. * Gröndahl et al. (2018) Tommi Gröndahl, Luca Pajola, Mika Juuti, Mauro Conti, and N. Asokan. 2018. All You Need is "Love": Evading Hate Speech Detection. In _Proceedings of the 11th ACM Workshop on Artificial Intelligence and Security_. 2–12. * Guest et al. (2021) Ella Guest, Bertie Vidgen, Alexandros Mittos, Nishanth Sastry, Gareth Tyson, and Helen Margetts. 2021. An Expert Annotated Dataset for the Detection of Online Misogyny. In _Association for Computational Linguistics_. 1336–1350. * Herring and Martinson (2004) Susan C. Herring and Anna Martinson. 2004. Assessing Gender Authenticity in Computer-Mediated Language Use: Evidence From an Identity Game. _Journal of Language and Social Psychology_ 23, 4 (2004). Publisher: SAGE Publications Inc. * Herrnson et al. (2003) Paul S. Herrnson, J. Celeste Lay, and Atiya Kai Stokes. 2003\. Women Running “as Women”: Candidate Gender, Campaign Issues, and Voter-Targeting Strategies. _The Journal of Politics_ 65, 1 (2003). * Hiruncharoenvate et al. (2015) Chaya Hiruncharoenvate, Zhiyuan Lin, and Eric Gilbert. 2015\. Algorithmically Bypassing Censorship on Sina Weibo with Nondeterministic Homophone Substitutions. _Proceedings of the International AAAI Conference on Web and Social Media_ 9 (2015), 150–158. Number: 1. * Jane (2017a) Emma Jane. 2017a. Gendered cyberhate: A new digital divide? In _Theorizing Digital Divides_. Routledge. * Jane (2017b) Emma Jane. 2017b. _Misogyny Online: A Short and (Brutish) History_. Sage Publications. * Jane (2015) Emma A. Jane. 2015\. Flaming? What flaming? The pitfalls and potentials of researching online hostility. _Ethics and Information Technology_ 17 (2015), 65–87. * Kennedy et al. (2022) Brendan Kennedy, Mohammad Atari, Aida Mostafazadeh Davani, Leigh Yeh, Ali Omrani, Yehsong Kim, Kris Coombs, Shreya Havaldar, Gwenyth Portillo-Wightman, Elaine Gonzalez, Joe Hoover, Aida Azatian, Alyzeh Hussain, Austin Lara, Gabriel Cardenas, Adam Omary, Christina Park, Xin Wang, Clarisa Wijaya, Yong Zhang, Beth Meyerowitz, and Morteza Dehghani. 2022\. Introducing the Gab Hate Corpus: defining and applying hate-based rhetoric to social media posts at scale. _Language Resources and Evaluation_ 56 (2022). * Krippendorff (2011) Klaus Krippendorff. 2011\. _Computing krippendorff’s alpha-reliability_. Retrieved from: https://repository.upenn.edu/ascpapers/43. * Kwan et al. (2020) Irene Kwan, Kelly Dickson, Michelle Richardson, Wendy MacDowall, Helen Burchett, Claire Stansfield, Ginny Brunton, Katy Sutcliffe, and James Thomas. 2020\. Cyberbullying and Children and Young People’s Mental Health: A Systematic Map of Systematic Reviews. _Cyberpsychology, Behavior and Social Networking_ 23, 2 (2020). * Lawless (2015) Jennifer L. Lawless. 2015\. Female Candidates and Legislators. _Annual Review of Political Science_ 18 (2015). * Lee (2016) Hangwoo Lee. 2016\. Behavioral Strategies for Dealing with Flaming in An Online Forum. _The Sociological Quarterly_ 46, 2 (2016), 385–403. * Mantilla (2013) Karla Mantilla. 2013\. Gendertrolling: Misogyny Adapts to New Media. _Feminist Studies_ 39, 2 (2013), 563–571. * Megarry (2014) Jessica Megarry. 2014\. Online incivility or sexual harassment? Conceptualising women’s experiences in the digital age. _Women’s Studies International Forum_ 47 (2014), 46–55. * Mostafazadeh Davani et al. (2020) Aida Mostafazadeh Davani, Mohammad Atari, Brendan Kennedy, Shreya Havaldar, and Morteza Dehghani. 2020\. Hatred is in the Eye of the Annotator: Hate Speech Classifiers Learn Human-Like Social Stereotypes. In _31st Annual Conference of the Cognitive Science Society (CogSci)_. * Ostini and Hopkins (2015) Jenny Ostini and Susan Hopkins. 2015. Online harassment is a form of violence. _The Conversation_ 8 (2015), 1–4. Publisher: The Conversation Media Trust. * Pfeffer et al. (2023) Juergen Pfeffer, Angelina Mooseder, Jana Lasser, Luca Hammer, Oliver Stritzel, and David Garcia. 2023\. This Sample seems to be good enough! Assessing Coverage and Temporal Reliability of Twitter’s Academic API. _Proceedings of the International AAAI Conference on Web and Social Media_ (2023). * Rheault et al. (2019) Ludovic Rheault, Erica Rayment, and Andreea Musulan. 2019\. Politicians in the line of fire: Incivility and the treatment of women on social media. _Research & Politics_ 6, 1 (2019). * Ross et al. (2016) Björn Ross, Michael Rist, Guillermo Carbonell, Ben Cabrera, Nils Kurowsky, and Michael Wojatzki. 2016\. Measuring the Reliability of Hate Speech Annotations: The Case of the European Refugee Crisis. In _Proceedings of NLP4CMC III: 3rd Workshop on Natural Language Processing for Computer-Mediated Communication_ , Michael Beißwenger, Michael Wojatzki, and Torsten Zesch (Eds.). 6–9. * Strathern et al. (2022) Wienke Strathern, Mirco Schoenfeld, Raji Ghawi, and Juergen Pfeffer. 2022. Identifying Lexical Change in Negative Word-of-Mouth on Social Media. _Social Network Analysis and Mining_ 59, 12 (2022). * Udupa (2020) Sahana Udupa. 2020\. Artificial Intelligence and the Cultural Problem of Extreme Speech. _Social Science Research Council, (20 December 2020), online_ (2020). * Vogels (2021) Emily A. Vogels. 2021\. The State of Online Harassment. _Pew Research Center_ (2021). * Waseem et al. (2017) Zeerak Waseem, Thomas Davidson, Dana Warmsley, and Ingmar Weber. 2017. Understanding Abuse: A Typology of Abusive Language Detection Subtasks. In _Proceedings of the First Workshop on Abusive Language Online_. Association for Computational Linguistics, Vancouver, BC, Canada, 78–84. * Whiting et al. (2019) Jason Whiting, Rachael Dansby Olufowote, Jaclyn Cravens-Pickens, and Alyssa Banford Witting. 2019\. Online Blaming and Intimate Partner Violence: A Content Analysis of Social Media Comments. _The Qualitative Report_ 24 (2019). * Wiegand et al. (2021) Michael Wiegand, Josef Ruppenhofer, and Elisabeth Eder. 2021\. Implicitly Abusive Language – What does it actually look like and why are we not getting there?. In _Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies_. Association for Computational Linguistics, 576–587. * Zannettou et al. (2020) Savvas Zannettou, Joel Finkelstein, Barry Bradlyn, and Jeremy Blackburn. 2020. A Quantitative Approach to Understanding Online Antisemitism. In _Proceedings of the International AAAI Conference on Web and Social Media_ , Vol. 14. 786–797. * Zuckerberg (2018) Donna Zuckerberg. 2018\. _Not All Dead White Men. Classics and Misogyny in the Digital Age_. Harvard University Press.
# A Survey of Mobile Edge Computing for the Metaverse: Architectures, Applications, and Challenges Yitong Wang <EMAIL_ADDRESS>School of Computer Science and Engineering Nanyang Technological University Singapore Jun Zhao <EMAIL_ADDRESS>School of Computer Science and Engineering Nanyang Technological University Singapore ###### Abstract Metaverse is an emerging virtual universe where humans can have real-time interactions and solid social links like in the physical world, and it opens up a new era of Internet and interactions. In Metaverse, an immersive and photorealistic environment promotes social activities, including education, meetings, and shopping of digital avatars based on critical technologies, including 3D rendering, extended reality, digital twins, artificial intelligence, and Blockchain. However, the limitations of computation, storage, and energy resources restrict the development of Metaverse, and a series of system issues (e.g., latency, security, and battery-life) continue to arise. As a result, how to find corresponding measurements to mitigate unsatisfactory influences becomes the focus. Mobile edge computing (MEC) as a distributed computing paradigm offloads computation-intensive tasks to the edge of the network. It brings the resources as close as possible to the end devices, addressing the shortcomings mentioned above. In this paper, we propose a comprehensive survey of the MEC-based Metaverse. Particular emphasis is given to the technologies convergence, architectures, and application scenarios, e.g., BoundlessXR and CloudXR. Significantly, we introduce the potential future directions for developing Metaverse systems. ###### Index Terms: Metaverse, Edge Computing, Extended Reality, Augmented Reality, Virtual Reality, Blockchain. ## I INTRODUCTION Metaverse, as a universe of ternary-world interactions (i.e., physical world, human world, and digital world), enables humans as digital avatars in sub- Metaverses to carry out various social activities in the physical world, such as business meetings, learning, shopping, and digital assets trading[1]. Metaverse significantly revolutionises human manners of interactions with machines, from conventional 2D web browsing to immersive real-time interaction in 3D virtual environments, supported by high fidelity image technology, 3D rendering, Blockchain, digital twins, artificial intelligence (AI) and sophisticated sensor devices. Recently, Metaverse has been investigated by lots of commercial and gaming companies. For example, Tencent and Facebook have already invested in chat scenarios of Metaverse; even Facebook renamed itself “Meta” in 2021. The increasing number of mobile devices and users has caused the fact that wireless communication networks constantly evolved to cope with the strict requirement of high data rate, latency, and machine-to-machine(M2M) connection density[2]. 5th generational communication network that consists of specific attributes is the best choice to cater for the emerging Metaverse system. The advent of 5G boosts the performance indicators of communication networks by a multiple of 10 to 100 times, which meets the stringent requirements of various services and applications. It acts as a bridge for the transmission of original pose data and compressed image data generated by VR devices and edge cloud servers, respectively, which perfectly enhances the rendering performance of the Metaverse. Extended reality (XR)[3] technology incorporates mixed reality (MR), augmented reality (AR), virtual reality (VR), and any other technology that encompasses the fusion of all reality and virtualization. XR devices collect data, including biological data like eye-, hand-, and head-tracking, and the accumulation of user data from other social media platforms[4]. XR applications also provide multi-sensor immersiveness and real-time interactions for users. Augmented reality(AR)[5], as a technology that replicates the sensory perception of the real world in terms of time and space domains, projects additional augmentations generated by computers upon natural objects, combining real and virtual worlds to enhance people’s sensory experience further. In the last few years, both KLM Royal Dutch Airlines and British budget airline EasyJet have allowed passengers to check the size of their suitcases for boarding through AR technology. AR is performed based on realistic environments, and application services enrich reality with augmented components[6]. On the opposite side, VR is a part of MR where its surrounding environment is virtual, and people can fully have immersive experiences in virtual worlds with appropriate wearable equipment (e.g., VR Helmets or Glasses)[7]. To sum up, VR provides immersive interactivity in a digital world. AR delivers authentic experiences to customers with digital holograms, images, and videos based on real-world objects. MR acts as a transition, providing a distinctive experience between VR and AR[8]. Further developments in micro-sensors and XR technology are making XR equipment, such as helmet- mounted displays (HMDs), promising to be the primary endpoint for experiencing the Metaverse [9]. The exponential growth of mobile internet traffic has driven a dramatic evolution in computing paradigms. Modern mobile applications have more stringent requirements which cloud computing cannot satisfy, such as ultra-low latency, ultra-high throughput, ultra-high stability, and high spectral efficiency. Meanwhile, big data-associated Internet of Things devices causes an immense growth in the total traffic flows. To break the limitations of cloud computing and ensure better customer service performance, the emerging computing paradigm of edge computing, especially mobile edge computing, has gained the spotlight in both commercial and academic fields nowadays. Following European Telecommunications Standard Institute (ETSI), Mobile Edge Computing (MEC) is formally defined as a new platform with significant computing resources, including IT and Cloud-computing and it gets closer to the subscribers’ side [10]. Intending to extend 3GPP access scenarios such as Wi-Fi, the original MEC concept has now transitioned to “Multi-access Edge Computing” by ETSI. The remarkable characteristic of MEC is offloading computation tasks from core data centres to distributed edge servers located on the base stations. Thus, this typically distributed computing paradigm significantly reduces the communication latency and provides a relatively large amount of computation resources compared with conventional cloud computing paradigms. Also, the deployment of MEC relies heavily on the virtualized platforms[11]. Based on the evolution and combination of the various technologies, services, and equipment mentioned above, Metaverse technology is rapidly gaining attention and enhancement. However, there is a highly distant path to explore to realise the Metaverse fully. Firstly, transmitting the surging data flows together to cloud servers from end devices is contrary to what is being sought. The reason is that it not only aggravates network latency on the system but also causes network congestion and data loss when reaching cloud services via the core network. Furthermore, data leakage and asset authentication of the Metaverse are both obstacles encountered. Data security leads to the loss of all personal information and the forgery of avatar identities. As multiple administrators provide the assets and virtual currencies in Metaverse, how to authenticate and trade among the diverse assets in sub-Metaverses raises a concern. The contributions of our survey are as follows: * • We first introduce the relevant background and basic concepts of Metaverse and present a clear overview of the types of architectures, evolutionary and eventual outcomes in Metaverse. Notably, we provide a nuanced introduction to indicators and essential technologies of Metaverse. With these concise introductions, readers can keep abreast of the ongoing technological developments in Metaverse. * • We discuss the technological convergence of MEC and Metaverse and illustrate the role that MEC plays in Metaverse. Unlike previous papers on MEC-based Metaverse, we not only describe the contributions of MEC to various performance indicators but also emphasize the application of collaborative architectures of MEC with other computing paradigms in Metaverse. Specially, we point out that MEC is crucial in solving the latency, privacy, and energy challenges in Metaverse. * • We outline the research directions and challenges to pave the path toward future research attempts to realize higher performance in MEC-based Metaverse architectures. Furthermore, our survey aims to assist researchers in gaining a thorough and in-depth understanding of the MEC-based Metaverse and grasping a holistic view of the research in this area. The remainder of this paper is organized as follows. Section II introduces the present advanced developments of technologies in Metaverse, and the scenarios of mobile edge computing applied to the XR domain. Section III interprets the concepts, architectures, and features of the Metaverse and identifies some of the current challenges. An introduction to mobile edge computing, including its strengths and weaknesses, is presented in Section IV. Section V summarises the enhancements and facilitations when mobile edge computing is applied to the Metaverse and demonstrates the architectures of the Metaverse based on the collaboration of mobile edge computing with cloud and fog computing. Future research directions are shown in Section VI. Section VII provides the conclusion of the paper. Figure 1: The Metaverse architecture performs the interactions among virtual area, human area, and physical area ## II RELATED WORK We have witnessed many papers in terms of different aspects of Metaverse technologies and application scenarios of mobile edge computing. In this section, we discuss recent Metaverse and MEC research and show the main distinction with our paper. Fernandez et al.[12] review an in-depth survey of Metaverse from three main motivation perspectives, including privacy, governance, and ethics design. Specifically, they analyze privacy from sensory level, behaviour and communications, as well as users and bystanders. Also, further challenges of these three factors and a novel modular-based framework are presented in this paper. Xu et al.[13] explore the significance of using specific spectrums in Metaverse and investigate the vision of Blockchain technology, respectively. Cheng et al.[14] focus on some existing social VR platforms that have the potential to evolve into the architectures of the Metaverse and also conduct a test by looking in depth at the network operation and capabilities of two typical platforms. Lee et al.[15] present several innovations in the computation arts of Metaverse, proposing some research agendas about democratizing computational arts, digital privacy and safety for Metaverse artists, the identification of proprietary rights in digital artworks, and the directions of technological developments. For MEC, Siriwardhana et al.[16] emphasize the survey of mobile augmented reality (MAR) based on MEC and discusses future vital application areas for MAR. Lim et al.[17] introduce a study focused on edge-intelligence Metaverse in term of virtual city development. A stochastic optimal resource allocation scheme (SORAS) is further proposed byNg et al. based on stochastic integer programming with the aim of optimizing the cost-effectiveness of virtual service suppliers [18]. However, none of these recent papers focuses on the convergence of the Metaverse and MEC paradigm in terms of unique perspectives such as quality of experience, application scenarios, future challenges, and so on. ## III METAVERSE ### III-A Metaverse Introduction The term ’Metaverse’ first emerged in the 20th century in Neal Stephenson’s science fiction novel named ’Snow Crash’[19], but did not receive much attention at that time. Two primary factors have driven the Metaverse into focus currently: * • Social factors: The Covid-19 pandemic has led to a dramatic shift in the manners of people living and working via the Internet from outdoors to indoors[20], which has caused an increased frequency of daily activities via the Internet. Additionally, commercial enterprises’ strong focus on VR resulted in the re-emergence of the Metaverse in the public arena; * • Technical factors: XR ecosystem developments (including AR, VR, MR) facilitate the further enhancement of virtual scenarios and interactions between physical and virtual worlds; Emerging Web 3.0 converts “server-centric” networks into “user-centric” networks based on decentralized networks; MEC servers perform data processing at the edge of the network (i.e., closer to the user) to mitigate system latency; 5G networks with downlink (DL) speeds of 200Mbps meet the constraints including low latency, high throughput, and reliability of the Metaverse. Through the ternary nature of the Metaverse definition, Fig. 1 shows the architecture of the Metaverse, and the relationships between these three worlds can be explained in detail as follows: 1) Physical Area: The physical area provides the essential infrastructures to support the other two worlds, including computation, cache, transmission, and sensors. Data from around the environment and the human body are collected precisely by sensor infrastructures. Transmission infrastructures (e.g. artificial satellites strictly) ensure the stability and continuity of the entire network connection. Computationally intensive tasks are processed by servers located at different positions via robust computation infrastructures, depending on the type of these tasks. Also, cache infrastructures at various locations (i.e., local cache, edge cache, and cloud cache) can store major tasks and reduce the total latency of the entire task processing. 2) Virtual Area: In the Virtual world, the Metaverse engine facilitates and manages large volumes of digital information from both the physical and human worlds to enable large-scale Metaverse services[8]. * • Engine Layer: Metaverse engine layer consists of various technologies to provide a photorealistic environment, real-time interactions, non-fungible token (NFT)[21] and translation, respectively. 3D rendering technology provides a virtual environment with diverse spatiotemporal dimensions and attributes. Digital twins (DT) is crucial to the behaviours of avatars in Metaverse. Especially various human social activities are identified and performed via DT technology. AI plays a vital role in guaranteeing the reliability of Metaverse. By combining with other technologies, many machine learning algorithms are used to solve challenging tasks, such as computational resource allocation[22], predictions of human actions, and increased spectrum utilisation. * • Virtual Layer: As illustrated in Fig. 1, the virtual layer is composed of the virtual environment, digital avatars, virtual services/goods and digital assets, which is supported by the critical technologies in the engine layer. High-fidelity images are provided with 3D rendering and AI support by collecting and processing real-world environmental data via XR sensor devices. Digital avatars are capable of behaving exactly like humans in the real world based on DT technology. * • Virtual Universe: In contrast to the concept of a unified Metaverse, the whole Metaverse comprises a series of dispersed sub-Metaverses where avatars can be served distinctly. Distinct sub-Metaverses can be freed from the constraints of time and space. Due to the existence of multiple operators, creating multiple sub-universes becomes a fact, but their interconnection holds the key to future developments. 3) Human Area: In reality, people’s intricate social relationships and diverse activities contribute to the functioning of society. Crucially, as in the real world, people as avatars are at the centre of the virtual world and are constantly creating virtual objects. Currently, by wearing devices (e.g. XR HMDs, Wristband sensors), social activities such as business meetings, office learning and entertainment concerts can be realised in the virtual world via human-computer interaction (HCI) and especially brain-computer interaction (BCI). Human activities can be accurately mapped onto the virtual world through various Metaverse engines. And services in the virtual world can also continuously influence human interactions and perceptions in the real world. ### III-B Metaverse Characteristics Immersive: The ultimate goal of Metaverse development is to approximate the user’s experience in the real world gradually. Supported by various infrastructures and Metaverse engines, users can interact in a rendered and high-fidelity virtual universe. Immersive means that the user can adapt to the digital world and move around in the Metaverse without discomforts, such as dizziness and nausea. Interoperability: Multiple operating systems, such as Ethereum, Solana, and Polygon, co-exist in the Metaverse. Accordingly, a variety of value tokens that are derived from each technology continuously emerge. For example, the Bitcoin Blockchain supports a token called Bitcoin, while the Ethernet-backed tokens include SAND, MANA, AXS, and GALA. The interoperability of the Metaverse is guaranteed by the accessibility of connections and transactions between different sub-universes or currencies. Multi-technology: A few technological advancements have contributed to the emergence of the Metaverse. For example, 5G technology has improved peak transmission rate, time delay, and reliability to meet the requirement that transmissions must be completed in milliseconds or less, relieving users of dizziness caused by high latency (>20ms). Also, because downlink speed is proportional to the rendering resolution, higher downlink speeds can make virtual environments more realistic. In summary, multi-technology provides an immersive experience and high-fidelity virtual worlds based on augmented reality technology, generates a city twin of the physical society with digital twin technology [23], and creates an economic system based on Blockchain technology [24]. ### III-C Metaverse Technologies XR: XR technology covers AR, VR, and MR. Wearable VR devices provide users with a completely virtual scenario in which they can be fully immersive[25]. For example, Second life provides a virtual world in which players can survive. In this world, players can control their avatars to perform activities and enjoy personalized services. For AR/MR technology, augmentations are overlaid in the real world[26], e.g., users can play Pokémon GO on their mobile phones and catch virtual Pokemon. XR devices significantly enhance the user’s immersion in the Metaverse. Digital Twin: Digital twin (DT) projects humans, objects, and environments of the real world into the virtual world, creating digital clones that are visually indistinguishable from the real world in real-time[27]. By processing the input data, DT is able to manage and optimize physical objects periodically [28][29]. Significantly, the data flows between virtual objects and the physical world are bidirectional. The physical object transmits the collected data in multiple formats to the virtual twins, and the virtual twins convey the processed feedback to the physical object[30], further accelerating the intersection of the human and virtual worlds. Blockchain: Blockchain is considered to be a decentralized technology [31]. Unlike traditional centralized systems, Blockchain technology can help safeguard users’ data, identity information, and virtual assets with certificates (i.e., NFTs) from leakage and theft in the event of a threat. At the same time, Blockchain, as a distributed ledger, ensures that all data is protected from tampering and modification. Hence, Blockchain not only ensures data security and quality but also enables seamless data sharing and data interoperability and integrity[32]. Furthermore, the convergence of Blockchain and other key technologies can significantly empower the performence of the Metaverse system [33]. AI: AI plays an indispensable role in creating and rendering large-scale Metaverse. Conventional AI technologies are composed of supervised learning, semi-supervised learning, unsupervised learning, as well as reinforcement learning (RL)[34]. Nowadays, many advanced machine learning (ML) algorithms of supervised and reinforcement learning have been used for different challenging tasks, including automatic resource allocation, attack prevention, and network fault detection[35][36]. Based on massive Metaverse multimodal input data, high-quality Metaverse scenes can be created and rendered by AI via considerable data interference[37]. Also, deep learning (DL) and ML algorithms facilitate the provision of personalized services by making good decisions. ## IV MOBILE EDGE COMPUTING ### IV-A MEC Introduction The motivation for investigating MEC computing paradigms is tackling the tradeoff between computation and communication. As shown in Fig. 2, edge servers are deployed on the edge of networks, which provides computational resources. As a result, computational tasks are processed at the network’s edge as proximate to the user as possible. Different from fog computing and cloud computing, a MEC server is a node device which means that the decentralization of mobile communication networks is achieved[38]. The security and privacy of networks are enhanced greatly, and congestion during peak transmission could also be tackled. Nowadays, an increasing number of papers on mobile edge computing are investigated, especially in the fields of system architecture design and task offloading. Paper[39] proposes a task offloading strategy based on an improved auction algorithm, divided into two main phases: task offloading and task scheduling, considering the limited resources of smart devices and the stronger computing and storage capabilities of edge servers. Furthermore, this strategy provides the basis for offloading decisions in the decision phase by considering the time cost of task execution locally or at the edge as well as energy consumption to determine whether the computation task that needs to be processed will be offloaded to the edge of the network. Once the task is offloaded to the edge server, reducing computational latency and reducing transmission energy consumption to achieve the global optimum are the main objectives of the task scheduling phase. Paper [40] considers mobility at the user end, classifies cellular networks in more detail and gives more accurate state transfer probabilities. And this research models the VM migration process as a Markov decision process and uses a policy iteration algorithm to find the optimal solution, effectively reducing energy consumption and latency. ### IV-B MEC Application MEC is used as a key technology that guarantees the efficient realisation of diverse services, especially in the practical application of task offloading and resource allocation. In vehicular edge computing (VEC) networks[41], vehicles can act as an edge server at the edge of the Internet of Things (IoT) system to provide computing resources[42] and various services. Meanwhile, the collaboration between vehicle edge servers (VES) and fixed edge servers (FES) can provide a variety of offloads and computing options for network computing, offering various edge computing resources depending on the tasks. Ren et al. [43] demonstrate that edge servers can reduce the core network use by elaborating on the system framework of web-based services with MEC technology and further showing that MEC can address real-world Web AR deployments. Figure 2: MEC computing paradigm features the low-latency and energy- efficiency procession. ## V METAVERSE WITH MEC It is theoretically feasible to perform all computing tasks on the local XR devices or cloud servers, but this contradicts the desire to prolong battery life, lighten device weight and maintain ultra-low latency. The convergence of mobile edge computing and Metaverse results in a new generation of MEC technology[44][45], which is being used to break the dependency of devices on centralized cloud servers and to improve the high real-time performance and security of the entire system. Primarily, splitting rendering is generated after this convergence. As graphics are heavily rendered, on-device processes are augmented by being partitioned to IoT devices and edge clouds. The graphical rendering of the Metaverse system is enhanced by using photonic processing for head tracking and motion tracking of latency-sensitive devices. Furthermore, based on the design of stream processing technology, Metaverse systems can also effectively decrease the latency and have excellent performance as a high-performance distributed stream processing system (DSPS)[46]. More studies are explained as follows. Figure 3: Architecture of Mobile Edge Computing-Based Metaverse ### V-A MEC-Based Architecture Conventional cloud-based Metaverse architecture is not beneficial to implementing virtual functions and services via transmitting massive data such as movements of avatars and physical factors of surroundings to the centralized server. To mitigate negative influences caused by limitations, innovative MEC-based Metaverse architectures have resulted in optimization and a satisfactory performance: 1) Mobile Edge Computing-Based Metaverse: In the Metaverse of this architecture, multiple dynamic edge nodes are incorporated to perform users’ instructions, which significantly decrease the delay caused by users’ movements[47]. As shown in Fig. 3, the private data generated by the customers is transmitted to the specific edge server. Through the processing and delivery of the edge servers, customers enjoy a low-latency service and an immersive Metaverse experience. To cope with the data leakage when transmitting users’ data, federated learning(FL) is regarded as a critical technology. Consequently, parameters rather than private data from similar local models are uploaded to the FL layer for further training. 2) Fog-Edge-Cloud Metaverse: This distributed architecture is hierarchical and tiers different servers to eliminate the Metaverse fragmentation and computational bottlenecks.Kechadi et al.[48] proposed this hybrid computational architecture for Metaverse services. In cloud layers, Cloud servers exist to simulate the virtual worlds. The edge layer servers are dedicated to processing specific virtual buildings’ computation tasks and providing virtual environment animation. For fog servers located on the fog layer, servers are deployed to perform computation tasks of the virtual home environment and users’ movements. ### V-B Improvement User Side: In virtual worlds, accurate predictions and assessments of users’ behaviours, habitual preferences, and movement paths are crucial to providing users with an immersive and real-time experience. Based on the attributes of MEC, stable and reliable seamless services could be offered to users’ devices, and real-time services are also provided as a result of the location of MEC servers [49]. Furthermore, since edge servers have more computing capabilities than local devices(e.g. VR devices, HMDs, and AR glasses), they can efficiently process computation-intensive tasks and transfer feedback to mobile devices, which will reduce latency and thus alleviate user discomfort such as dizziness caused by latency. System Side: The two main merits of applying MEC to Metaverse systems today are enhanced resource utilisation and significantly reduced latency. Taking advantage of the fact that MEC servers come with their own computational resources[50], most Metaverse applications can be processed through these servers. At the same time, the wearability and small size of most devices lead to limitations on the battery life and computational resources, which is tackled by the MEC-based Metaverse. Simultaneously, the dynamic resource allocation[23] and edge intelligence-based technology[29] further improve the performance of all aspects of the Metaverse substantially. Figure 4: BoundlessXR consists of XR devices, 5G network, and edge-cloud. ### V-C Application Scenarios On-device processing is an essential part of performing Metaverse applications. In a stand-alone model, on-device processing is responsible for dealing with entire XR processing assignments. While augmented by the edge cloud, on-device processing provides rendering and tracking that are power- efficiency, high-performance and latency-sensitive. Nowadays, Boundless XR created by the Qualcomm uses split rendering to distribute the computation tasks between the edge servers and end devices to deliver a genuinely immersive XR experience over 5G. Specifically, the architecture is presented in Fig. 4. Firstly, VR devices send 6-degrees-of-freedom head pose data to the edge cloud over 5G. The edge cloud sends the compressed image to the devices after processing this data and rendering a new image. Finally, VR devices further perform on-device processing. Based on the 5G and MEC technologies, the motion-to-render-to-photon latency is less than 20ms, which meets Metaverse services’ requirements. NVIDIA’s CloudXR is also a novel solution for streaming XR data from any openVR/XR application on a distant server, which can also dynamically adjust network conditions and maximize image quality while minimizing effective latency. In addition to innovative applications by commercial companies, deep research in edge computing is also evolving in academia. A remarkable FACT algorithm was proposed in the paper[51] to balance three main factors, including network latency, computation latency, and system accuracy in an edge-based mobile augmented reality system. Applying this new paradigm to the MAR applications significantly addressed the optimization problems as a result of multiple mobile users and improved the performance of processing video frames in Metaverse. In order to optimize the computation resources and AR configurations on the edge servers, another optimization protocol called DARE was presented in the paper[52]. When considering multiple coherence times in the research of MEC-based MAR, the algorithm based on Lyapunov optimization in the paper[53] can effectively improve the reliability and the quality of augments (QoA) in Metaverse systems. Furthermore, when studied from the perspective of energy consumption and the interactions between MAR system parameters, the LEAF algorithm was studied by Wang et al.[54] in the specific energy-aware MEC-based MAR system. ## VI FUTURE RESEARCH CHALLENGES In this section, future challenges of MEC-based Metaverse are discussed from the following aspects. ### VI-A Computer Resource Allocation Currently, papers are primarily conducted on computation allocation. Tang et al. [55] proposed an optimization algorithm by considering the computation offloading strategies and resource allocation problems together and effectively solved the interuser interference due to load. Nath et al. [56] studied the cached content by four decision parameters were considered, including whether a given task needs to be cached, how much transmission power is appropriate to utilize during offloading, and how many MEC resources to assign to perform a task to further improve collaboration issues between MEC servers. However, these researches do not consider the dynamic nature of Metaverse systems. Specifically, the computational resources of each node are dynamic in real time rather than static. Therefore, idle computational resources will be unused inefficiently when the MEC is processing computationally intensive tasks after nodes are pre-determined. For further research, how to solve the problem of dynamic allocation of computational resources is a crucial challenge to be tackled. ### VI-B Mobility Management Metaverse applications rely heavily on users’ states and previous behaviours, meaning mobility is a key consideration when implementing applications sensitive to continuity and real-time [57]. On the other hand, user mobility and excellent mobility management ensure seamless service to enhance the user experience and QoS further. The alternative approaches are caching[58] and SDN controllers now. For caching, the main challenges are how the popularity of the content to be cached can be accurately predicted and how we can achieve collaborative caching across multiple cells to improve web caching performance. [59] For SDN controllers, Shah et al. [60] propose an SDN-based MEC-enabled 5G vehicular networks. They use four software modules developed by themselves in the SDN controller to develop the rules of QoS further. Furthermore, the SDN controller proposed can federate and coordinate the allocation of MEC resources to further provide continuous service and seamless coverage to mobile users. In future research directions, mobility management can be combined with artificial intelligence. Through this technologies convergence, the Metaverse system can learn the user’s behaviours on its own, predict how the user will move, and transmit information about the user’s state to the SDN controller in advance, thus providing a more reliable service. ### VI-C User Experience Digital avatars experience various activities in photorealistic sub-Metaverses and can obtain perceptions through sensor devices. The final evolution of the Metaverse is to integrate with the physical world, which means dark real-world forces also exist in virtual life. For example, robberies and car accidents in Metaverse badly influence humans in the real world and even cause physical pain. Moreover, the unfettered behaviour of avatars, like sexual assault, also causes psychological victimization of the user. Thus, establishing a code of ethics to constrain the behaviour of avatars is a crucial challenge to maintaining a safe community for users in virtual worlds[61]. Furthermore, user addiction[62] is another harmful influence on emotional health in Metaverse. ### VI-D Data For obstacles of Metaverse in the future, two main factors are supposed to be noticed: data leakage and data reliability. 1) data leakage: MEC-based metaverse architecture is not the same as the decentralized bitcoin architecture. Private and safety-sensitive data (such as avatar ID numbers and passwords, digital assets, and currencies) are transmitted to the edge cloud from end devices, which may suffer grievous attacks by potential adversaries. Lee et al.[63] identify three main ad fraud threats, including blind-spot tracking, gaze and controller cursor-jacking, and the misuse of assisted display in content sharing. 2) data quality:The data quality levels directly alter the metaverse applications’ QoS. High data quality significantly improves the QoE of users, while low level stands on the opposite side. 3) data reliability: Data sources and generation processes will cause the data to be unreliable if the sources are fake IDs and jitters occur during propagation. Hence, preventing leakage and improving the quality and reliability of data in Metaverse are also essential requirements during the development of realizing Metaverse. ### VI-E Delay The main goal of the Metaverse is to deliver the processed data to the XR device in real-time via the downlink and to render the virtual spatial environment. For example, the delay generated by VR devices should not exceed 20ms; otherwise, it will cause spatial vertigo and vomiting to the user in the Metaverse. Hence, the partitioning of computational tasks is a constant focus of researchers[64]. Furthermore, balancing the tradeoff between task splitting and delay is essential. For example, Lightweight and wearable XR devices can be worn easily but with fewer computational resources. ### VI-F Privacy Data privacy and user identity are the two significant aspects of Metaverse security that must be ensured. For data privacy, users’ virtual assets, facial and retinal data, and transaction records in the Metaverse are at risk of leakage, especially during data processing and transmission. Moreover, after one user’s behavioural data and interactions with other avatars have been collected and analyzed by hackers[19], a fake avatar will be made that can imitate the original avatar to carry out social activities in the Metaverse and even commit crimes. Bystander privacy in the Metaverse is also at risk of being acquired. Then, the protection of data and identity becomes the way forward. ## VII CONCLUSION In this paper, we have introduced an in-depth survey of Metaverse based on mobile edge computing. We have presented an overview of Metaverse and discussed its fundamentals, characteristics, core technologies, and existing issues. Afterwards, fundamental research and application scenarios like AR(Web-AR)/VR/MR of mobile edge computing are presented. Significantly, we analyze the convergence of these two prominent technologies from aspects including architectures, improved peculiarities, and application scenarios. Furthermore, we show the future research challenges of mobile edge-based Metaverse. We anticipate that this survey will shed light on the direction of the MEC and Metaverse and spark even more impressive research in the future. ## Acknowledgement This research is supported in part by Nanyang Technological University Startup Grant; in part by the Singapore Ministry of Education Academic Research Fund under Grant Tier 1 RG97/20, Grant Tier 1 RG24/20 and Grant Tier 2 MOE2019-T2-1-176. ## References * [1] J. Zhong and Y. Zheng, “Empowering future education: Learning in the Edu-Metaverse,” in _2022 International Symposium on Educational Technology (ISET)_. IEEE, 2022, pp. 292–295. * [2] P. Popovski, K. F. Trillingsgaard, O. Simeone, and G. Durisi, “5G Wireless Network Slicing for eMBB, URLLC, and mMTC: A Communication-Theoretic View,” _Ieee Access_ , vol. 6, pp. 55 765–55 779, 2018. * [3] J. Guan, J. Irizawa, and A. Morris, “Extended Reality and Internet of Things for Hyper-Connected Metaverse Environments,” in _2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)_. IEEE, 2022, pp. 163–168. * [4] J. Evans, “White Paper-The IEEE Global Initiative on Ethics of Extended Reality (XR) Report–Extended Reality (XR) Ethics in Medicine,” 2022. * [5] N. R. D. Torres, G. A. L. Morteo, and R. G. Cruz-Flores, “Conceptualization of Augmented Digital Ecosystem (ADE) in context of Metaverse and The Augmented Dimension,” in _2022 IEEE Mexican International Conference on Computer Science (ENC)_. IEEE, 2022, pp. 1–5. * [6] T. Braud, C. B. Fernández, and P. Hui, “Scaling-up AR: University Campus as a Physical-Digital Metaverse,” in _2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW)_. IEEE, 2022, pp. 169–175. * [7] M. Qu, Y. Sun, and Y. Feng, “Digital Media and VR Art Creation for Metaverse,” in _2022 2nd Asia Conference on Information Engineering (ACIE)_. IEEE, 2022, pp. 48–51. * [8] Y. Wang, Z. Su, N. Zhang, D. Liu, R. Xing, T. H. Luan, and X. Shen, “A Survey on Metaverse: Fundamentals, Security, and Privacy,” _arXiv preprint arXiv:2203.02662_ , 2022. * [9] M. Sugimoto, “Extended Reality (XR: VR/AR/MR), 3D Printing, Holography, AI, Radiomics, and Online VR Tele-Medicine for Precision Surgery,” in _Surgery and Operating Room Innovation_. Springer, 2021, pp. 65–70. * [10] E. M. E. Computing, I. Initiative _et al._ , “Mobile-Edge Computing: Introductory Technical White Paper,” _ETSI: Sophia Antipolis, France_ , pp. 1–36, 2014. * [11] Y. C. Hu, M. Patel, D. Sabella, N. Sprecher, and V. Young, “Mobile Edge Computing—A Key Technology towards 5G,” _ETSI white paper_ , vol. 11, no. 11, pp. 1–16, 2015. * [12] C. B. Fernandez and P. Hui, “Life, the Metaverse and Everything: An Overview of Privacy, Ethics, and Governance in Metaverse,” _arXiv preprint arXiv:2204.01480_ , 2022. * [13] H. Xu, Z. Li, Z. Li, X. Zhang, Y. Sun, and L. Zhang, “Metaverse Native Communication: A Blockchain and Spectrum Prospective,” _arXiv preprint arXiv:2203.08355_ , 2022. * [14] R. Cheng, N. Wu, S. Chen, and B. Han, “Will Metaverse be NextG Internet? Vision, Hype, and Reality,” _arXiv preprint arXiv:2201.12894_ , 2022. * [15] L.-H. Lee, Z. Lin, R. Hu, Z. Gong, A. Kumar, T. Li, S. Li, and P. Hui, “When Creators Meet the Metaverse: A Survey on Computational Arts,” _arXiv preprint arXiv:2111.13486_ , 2021. * [16] Y. Siriwardhana, P. Porambage, M. Liyanage, and M. Ylianttila, “A Survey on Mobile Augmented Reality with 5G Mobile Edge Computing: Architectures, Applications, and Technical Aspects,” _IEEE Communications Surveys & Tutorials_, vol. 23, no. 2, pp. 1160–1192, 2021. * [17] W. Y. B. Lim, Z. Xiong, D. Niyato, X. Cao, C. Miao, S. Sun, and Q. Yang, “Realizing the Metaverse with Edge Intelligence: A Match Made in Heaven,” _arXiv preprint arXiv:2201.01634_ , 2022. * [18] W. C. Ng, W. Y. B. Lim, J. S. Ng, Z. Xiong, D. Niyato, and C. Miao, “Unified Resource Allocation Framework for The Edge Intelligence-Enabled Metaverse,” _arXiv preprint arXiv:2110.14325_ , 2021. * [19] R. Di Pietro and S. Cresci, “Metaverse: Security and Privacy Issues,” in _2021 Third IEEE International Conference on Trust, Privacy and Security in Intelligent Systems and Applications (TPS-ISA)_. IEEE, 2021, pp. 281–288. * [20] A. E. Bates, R. B. Primack, B. S. Biggar, T. J. Bird, M. E. Clinton, R. J. Command, C. Richards, M. Shellard, N. R. Geraldi, V. Vergara _et al._ , “Global COVID-19 Lockdown Highlights Humans as Both Threats and Custodians of The Environment,” _Biological conservation_ , vol. 263, p. 109175, 2021\. * [21] R. García, A. Cediel, M. Teixidó, and R. Gil, “Semantics and Non-Fungible Tokens for Copyright Management on the Metaverse and Beyond,” _arXiv preprint arXiv:2208.14174_ , 2022. * [22] P. Si, J. Zhao, H. Han, K.-Y. Lam, and Y. Liu, “Resource Allocation and Resolution Control in the Metaverse with Mobile Augmented Reality,” _arXiv preprint arXiv:2209.13871_ , 2022. * [23] Y. Han, D. Niyato, C. Leung, C. Miao, and D. I. Kim, “A Dynamic Resource Allocation Framework for Synchronizing Metaverse with IoT Service and Data,” in _ICC 2022-IEEE International Conference on Communications_. IEEE, 2022, pp. 1196–1201. * [24] H. Ning, H. Wang, Y. Lin, W. Wang, S. Dhelim, F. Farha, J. Ding, and M. Daneshmand, “A Survey on Metaverse: the State-of-the-art, Technologies, Applications, and Challenges,” _arXiv preprint arXiv:2111.09673_ , 2021\. * [25] M. Xu, D. Niyato, J. Kang, Z. Xiong, C. Miao, and D. I. Kim, “Wireless Edge-Empowered Metaverse: A Learning-Based Incentive Mechanism for Virtual Reality,” in _ICC 2022-IEEE International Conference on Communications_. IEEE, 2022, pp. 5220–5225. * [26] M. McGill, “The IEEE Global Initiative on Ethics of Extended Reality (XR) Report–Extended Reality (XR) and the Erosion of Anonymity and Privacy,” _Extended Reality (XR) and the Erosion of Anonymity and Privacy-White Paper_ , pp. 1–24, 2021. * [27] C. A. Lynch, A. O. Adeyeye, and M. M. Tentzeris, “” Smart” Packaging of Self-Identifying and Localizable mmID for Digital Twinning and Metaverse Temperature Sensing Applications,” in _2022 IEEE 72nd Electronic Components and Technology Conference (ECTC)_. IEEE, 2022, pp. 87–92. * [28] M. Aloqaily, O. Bouachir, F. Karray, I. Al Ridhawi, and A. El Saddik, “Integrating Digital Twin and Advanced Intelligent Technologies to Realize the Metaverse,” _IEEE Consumer Electronics Magazine_ , 2022. * [29] D. Van Huynh, S. R. Khosravirad, A. Masaracchia, O. A. Dobre, and T. Q. Duong, “Edge Intelligence-based Ultra-Reliable and Low-Latency Communications for Digital Twin-enabled Metaverse,” _IEEE Wireless Communications Letters_ , 2022. * [30] Y. Wu, K. Zhang, and Y. Zhang, “Digital twin networks: a survey,” _IEEE Internet of Things Journal_ , vol. 8, no. 18, pp. 13 789–13 804, 2021\. * [31] Z. Wang, Q. Hu, M. Xu, and H. Jiang, “Blockchain-based Edge Resource Sharing for Metaverse.” * [32] T. R. Gadekallu, T. Huynh-The, W. Wang, G. Yenduri, P. Ranaweera, Q.-V. Pham, D. B. da Costa, and M. Liyanage, “Blockchain for the Metaverse: A Review,” _arXiv preprint arXiv:2203.09738_ , 2022. * [33] S. Badruddoja, R. Dantu, Y. He, M. Thompson, A. Salau, and K. Upadhyay, “Trusted AI with Blockchain to Empower Metaverse,” in _2022 Fourth International Conference on Blockchain Computing and Applications (BCCA)_. IEEE, 2022, pp. 237–244. * [34] F. Lin, W. Ning, and H. Zhai, “ICDVAE: An Effective Self-supervised Reinforcement Learning Method for behavior decisions of Non-player Characters in Metaverse Games,” in _2022 IEEE 10th Joint International Information Technology and Artificial Intelligence Conference (ITAIC)_ , vol. 10. IEEE, 2022, pp. 1727–1731. * [35] T. Huynh-The, Q.-V. Pham, X.-Q. Pham, T. T. Nguyen, Z. Han, and D.-S. Kim, “Artificial Intelligence for the Metaverse: A Survey,” _arXiv preprint arXiv:2202.10336_ , 2022. * [36] X. Zhou, C. Liu, and J. Zhao, “Resource allocation of federated learning for the metaverse with mobile augmented reality,” _arXiv preprint arXiv:2211.08705_ , 2022. * [37] O. Bouachir, M. Aloqaily, F. Karray, and A. Elsaddik, “AI-based Blockchain for the Metaverse: Approaches and Challenges,” in _2022 Fourth International Conference on Blockchain Computing and Applications (BCCA)_. IEEE, 2022, pp. 231–236. * [38] F. Vhora and J. Gandhi, “A Comprehensive Survey on Mobile Edge Computing: Challenges, Tools, Applications,” in _2020 Fourth International Conference on Computing Methodologies and Communication (ICCMC)_. IEEE, 2020, pp. 49–55. * [39] J. Sheng, J. Hu, X. Teng, B. Wang, and X. Pan, “Computation Offloading Strategy in Mobile Edge Computing,” _Information_ , vol. 10, no. 6, p. 191, 2019. * [40] Q. Wang, R. Xie, and T. Huang, “Research on Mobility Management of Mobile Edge Computing,” _ZTE TECHNOLOGY JOURNAL_ , 2018. * [41] Y. Liu, H. Yu, S. Xie, and Y. Zhang, “Deep Reinforcement Learning for Offloading and Resource Allocation in Vehicle Edge Computing and Networks,” _IEEE Transactions on Vehicular Technology_ , vol. 68, no. 11, pp. 11 158–11 168, 2019. * [42] K. Li, Y. Cui, W. Li, T. Lv, X. Yuan, S. Li, W. Ni, M. Simsek, and F. Dressler, “When Internet of Things Meets Metaverse: Convergence of Physical and Cyber worlds,” _arXiv preprint arXiv:2208.13501_ , 2022. * [43] P. Ren, X. Qiao, J. Chen, and S. Dustdar, “Mobile Edge Computing–A Booster for The Practical Provisioning Approach of Web-Based Augmented Reality,” in _2018 IEEE/ACM Symposium on Edge Computing (SEC)_. IEEE, 2018, pp. 349–350. * [44] W. Zhang, J. Chen, Y. Zhang, and D. Raychaudhuri, “Towards Efficient Edge Cloud Augmentation for Virtual Reality MMOGs,” in _Proceedings of the Second ACM/IEEE Symposium on Edge Computing_ , 2017, pp. 1–14. * [45] S. Dhelim, T. Kechadi, L. Chen, N. Aung, H. Ning, and L. Atzori, “Edge-enabled Metaverse: The Convergence of Metaverse and Mobile Edge Computing,” _arXiv preprint arXiv:2205.02764_ , 2022. * [46] J. Heo, K. Bhardwaj, and A. Gavrilovska, “Poster: Enabling Flexible Edge-Assisted XR,” in _2021 IEEE/ACM Symposium on Edge Computing (SEC)_. IEEE, 2021, pp. 465–467. * [47] L. Chang, Z. Zhang, P. Li, S. Xi, W. Guo, Y. Shen, Z. Xiong, J. Kang, D. Niyato, X. Qiao _et al._ , “6G-Enabled Edge AI for Metaverse: Challenges, Methods, and Future Research Directions,” _arXiv preprint arXiv:2204.06192_ , 2022. * [48] T. Kechadi, L. Chen, N. Aung, H. Ning, L. Atzori _et al._ , “Edge-enabled Metaverse: The Convergence of Metaverse and Mobile Edge Computing,” 2022. * [49] L.-H. Lee, T. Braud, P. Zhou, L. Wang, D. Xu, Z. Lin, A. Kumar, C. Bermejo, and P. Hui, “All One Needs to Know about Metaverse: A Complete Survey on Technological Singularity, Virtual Ecosystem, and Research Agenda,” _arXiv preprint arXiv:2110.05352_ , 2021. * [50] W. C. Ng, W. Y. B. Lim, J. S. Ng, Z. Xiong, D. Niyato, and C. Miao, “Unified Resource Allocation Framework for The Edge Intelligence-Enabled Metaverse,” in _ICC 2022-IEEE International Conference on Communications_. IEEE, 2022, pp. 5214–5219. * [51] Q. Liu, S. Huang, J. Opadere, and T. Han, “An Edge Network Orchestrator for Mobile Augmented Reality,” in _IEEE INFOCOM 2018-IEEE Conference on Computer Communications_. IEEE, 2018, pp. 756–764. * [52] Q. Liu and T. Han, “Dare: Dynamic Adaptive Mobile Augmented Reality with Edge Computing,” in _2018 IEEE 26th International Conference on Network Protocols (ICNP)_. IEEE, 2018, pp. 1–11. * [53] J. Liu and Q. Zhang, “Edge Computing Enabled Mobile Augmented Reality with Imperfect Channel Knowledge,” in _European Wireless 2019; 25th European Wireless Conference_. VDE, 2019, pp. 1–6. * [54] H. Wang and J. Xie, “User Preference Based Energy-Aware Mobile AR System with Edge Computing,” in _IEEE INFOCOM 2020-IEEE Conference on Computer Communications_. IEEE, 2020, pp. 1379–1388. * [55] L. Tang and H. Hu, “Computation Offloading and Resource Allocation for The Internet of Things in Energy-Constrained MEC-Enabled HetNets,” _IEEE Access_ , vol. 8, pp. 47 509–47 521, 2020. * [56] S. Nath and J. Wu, “Deep Reinforcement Learning for Dynamic Computation Offloading and Resource Allocation in Cache-Assisted Mobile Edge Computing Systems,” _Intelligent and Converged Networks_ , vol. 1, no. 2, pp. 181–198, 2020. * [57] Z. Huang and V. Friderikos, “Mobility Aware Optimization in the Metaverse,” _arXiv preprint arXiv:2209.08899_ , 2022. * [58] Y. Cai, J. Llorca, A. M. Tulino, and A. F. Molisch, “Joint Compute-Caching-Communication Control for Online Data-Intensive Service Delivery,” _arXiv preprint arXiv:2205.01944_ , 2022. * [59] Y. Liu, J. Liu, A. Argyriou, L. Wang, and Z. Xu, “Rendering-Aware VR Video Caching over Multi-Cell MEC Networks,” _IEEE Transactions on Vehicular Technology_ , vol. 70, no. 3, pp. 2728–2742, 2021. * [60] S. D. A. Shah, M. A. Gregory, S. Li, R. Fontes, and L. Hou, “SDN-Based Service Mobility Management in MEC-Enabled 5G and Beyond Vehicular Networks,” _IEEE Internet of Things Journal_ , 2022. * [61] J. Yu, A. Alhilal, P. Hui, and D. H. Tsang, “6G Mobile-Edge Empowered Metaverse: Requirements, Technologies, Challenges and Research Directions,” _arXiv preprint arXiv:2211.04854_ , 2022. * [62] J. Zhu, P. Ni, and G. Wang, “Activity Minimization of Misinformation Influence in Online Social Networks,” _IEEE Transactions on Computational Social Systems_ , vol. 7, no. 4, pp. 897–906, 2020. * [63] H. Lee, J. Lee, D. Kim, S. Jana, I. Shin, and S. Son, “AdCube: WebVR Ad Fraud and Practical Confinement of Third-Party Ads,” in _30th USENIX Security Symposium (USENIX Security 21)_ , 2021, pp. 2543–2560. * [64] Y. Cai, J. Llorca, A. M. Tulino, and A. F. Molisch, “Compute-and Data-Intensive Networks: The Key to The Metaverse,” _arXiv preprint arXiv:2204.02001_ , 2022.
# Enabling Fast Unit Commitment Constraint Screening via Learning Cost Model Xuan He1, Honglin Wen2, Yufan Zhang2 and Yize Chen1 1Hong Kong University of Science and Technology (Guangzhou) <EMAIL_ADDRESS><EMAIL_ADDRESS>2Shanghai Jiao Tong University {linlin00<EMAIL_ADDRESS> ###### Abstract Unit commitment (UC) are essential tools to transmission system operators for finding the most economical and feasible generation schedules and dispatch signals. Constraint screening has been receiving attention as it holds the promise for reducing a number of inactive or redundant constraints in the UC problem, so that the solution process of large scale UC problem can be accelerated by considering the reduced optimization problem. Standard constraint screening approach relies on optimizing over load and generations to find binding line flow constraints, yet the screening is conservative with a large percentage of constraints still reserved for the UC problem. In this paper, we propose a novel machine learning (ML) model to predict the most economical costs given load inputs. Such ML model bridges the cost perspectives of UC decisions to the optimization-based constraint screening model, and can screen out higher proportion of operational constraints. We verify the proposed method’s performance on both sample-aware and sample- agnostic setting, and illustrate the proposed scheme can further reduce the computation time on a variety of setup for UC problems. ## I Introduction Solving large-scale optimization problems is one of the cornerstones for many power system operation tasks, such as unit commitment (UC), optimal power flow (OPF), and capacity expansion planning. Many such applications require to solve the problem in a timely manner, as the solutions are essential to market clearing, power grid reliability, and power grid operations [1]. And in particular, as the DC power flow is often utilized in UC models, while the generators’ ON/OFF statutes are often described as integer variables, UC problems are usually formulated as mixed integer programming (MIP). There is a lot of computation burden to find the most economical UC solutions for such NP-hard problems on large-scale power networks. The computational burden of the UC problem can be significant, especially for large-scale transmission systems. This motivates the research on seeking surrogate models that are much smaller than the original UC problems yet ensure equivalence of constrains’ binding situation at optimal solution which should be the same as the original one. On the one hand, the active constraints can be kept with the constraint generation technique, where the violated constraints are gradually added into the surrogate model until the solution to the surrogate model is feasible to the original UC problem [2]. On the other hand, the redundant constraints can be screened out by solving a relaxed optimization for each constraint to identify whether the upper or lower bound of each constraint is redundant [3]. After securely screening constraints, the promise is to reduce the computation time significantly for the reduced UC problems which only involve a subset of original physical constraints in the UC problem. Although the total number of UC constraints is theoretically prohibitive with the above methods, empirical evidence shows that a vast majority of the constraints are still redundant and only a smaller subset of constraints could be binding (equality holds) given the region of load profiles [4]. Recent efforts have explored the potential of using cost-driven [5] and data-driven [6] ways to handle these issues. [5] adds an operational cost upper bound to the standard relaxed optimization so as to further narrow the subset of constraints; [6] proposes to use the k-nearest neighbor as a screening model to classify the constraint binding statuses using historical samples, yet both methods have tradeoffs in accuracy and efficiency. Indeed, the availability of historical power system operation records can provide rich information regarding operational decisions, constraint patterns and load profiles [7, 8]. Note that the system operating pattern corresponds to a unique region of load. The reexamination of the inactive or redundant constraints is needed, once the system operating pattern changes due to the change of the load. As such, [5] determines the cost upper bound for different aggregated net demand, i.e., load level. [9] considers the constraint screening for varying load ranges and results in a surrogate model that is applicable in a long operation period. In [10], the spatial correlation between nodal demands is taken into consideration for identifying the umbrella constraints. Figure 1: The schematic of our proposed method. We use the historical records involving load vector and UC cost to train a neural network model serving as cost model (a), and further use the trained model to get the upper bound of UC cost so as to screen constraints and obtain a surrogate model (b). However, the data-driven method used in [6, 7, 8] directly using machine learning (ML) predictions to classify if constraints are redundant, which may have simplicity but be hard to respond to the change of load region and without guarantee of equivalence. Meanwhile, the cost-driven method proposed in [5] can promise the solution accuracy, but it adds integer constraints to the constraint screening problem, making the screening procedure cumbersome to solve. Thus, it is required to come up with a method considering load region to achieve the accuracy-simplicity balance for the constraints screening problem. In this work, we investigate the potential of designing cost-driven paradigms to efficiently screen constraints in the unit commitment problem. We try to bridge the insights given by machine learning algorithms and the standard optimization-based constraint screening processes. Our approach also utilizes the historical data, but here ML prediction is used to help optimization-based method screen each constraint more efficiently. Specifically, we train a neural network to predict the optimal costs given load inputs [11]. Then we can conveniently upper-bound the search space of constraint screening problem by integrating the cost level predicted and optimized via the trained neural network model, where the formulation of the screening can be treated as a simple linear programming problem. Besides, our method can be flexibly integrated to screen constraints for either one specific load vector or for a given region of load, which we term as _sample-aware_ [3] and _sample- agnostic_ [9] constraint screening respectively. In sample-aware case, our proposed method can screen out about 90.03% of the redundant constraints. In sample-agnostic case, we can realize the UC cost prediction with relative error less than 1% and remove the redundant constraints without cost error along with saving the solution time. ## II Problem Setup ### II-A UC Problem Formulation In this paper, we assume the system operators need to decide both the ON/OFF statuses as well as dispatch level for all generators. As the realistic UC problem requires to take start-up and shut-down costs and logic constraints as well as ramp constraints into considerations, which make the analysis of multi-step constraints more complicated, we firstly consider the single-period UC problem as follows: $\displaystyle J(\bm{\ell})=\min_{\mathbf{u},\mathbf{x},\mathbf{f}}\quad$ $\displaystyle\sum_{i=1}^{n}c_{i}x_{i}$ (1a) s.t. $\displaystyle u_{i}\underline{x}_{i}\leq x_{i}\leq u_{i}\bar{x}_{i},\quad i=1,...,n$ (1b) $\displaystyle-\overline{\mathbf{f}}\leq\mathbf{Kf}\leq\overline{\mathbf{f}}$ (1c) $\displaystyle\mathbf{x}+\overline{\mathbf{A}}\mathbf{f}=\bm{\ell}$ (1d) $\displaystyle u_{i}\in\\{0,1\\},\quad i=1,...,n.$ (1e) In the UC problem, we optimize over the generator statuses $\mathbf{u}$, the generator dispatch $\mathbf{x}$ and the line power flow $\mathbf{f}$ to find the least-cost solutions with cost denoted as $J(\bm{\ell})$ in the objective function (1a). $c_{i}$ denotes the cost coefficient. Constraint (1b), (1c) and (1d) denotes the generation bound, the flow bound and the nodal power balance respectively. Note that the power flows are modeled as a DC approximation, while the phase angles are absorbed into the fundamental flows $\mathbf{f}\in\mathbb{R}^{n-1}$ [11, 12]; $K$ and $\bar{\mathbf{A}}$ map such fundamental flows to flow constraints and nodal power balance respectively. (1e) enforces the binary constraint of generator statuses, where $u_{i}=1$ indicates the generator is on. ### II-B Constraint Screening Since there are many redundant line flow constraints when seeking the optimal solution of UC problem with given load region or specific load vector, which brings unnecessary computation burden, constraint screening for the line flow constraints can be meaningful. Similar to [3], we relax the integer variables $\mathbf{u}$ in (1) as continuous variables in $[0,1]$, and the screening approach requires to iteratively solve the relaxed optimization problem to find the upper and lower flow values on each transmission line. If the upper and lower bound cannot be reached by the relaxed optimization problem, we can safely screen out that line flow constraint. For the case that the load region $\mathcal{L}$ is known, a _sample-agnostic constraint screening problem_ can be formulated for a group of operating scenarios, which can be given as follows, $\displaystyle\max_{\mathbf{u},\mathbf{x},\mathbf{f},\bm{\ell}}/\min_{\mathbf{u},\mathbf{x},\mathbf{f},\bm{\ell}}\quad$ $\displaystyle f_{j}$ (2a) s.t. $\displaystyle u_{i}\underline{x}_{i}\leq x_{i}\leq u_{i}\bar{x}_{i},\quad i=1,...,n$ (2b) $\displaystyle-\overline{\mathbf{f}}_{\mathcal{F}/j}\leq\mathbf{K}_{\mathcal{F}/j}\tilde{\mathbf{f}}\leq\overline{\mathbf{f}}_{\mathcal{F}/j}$ (2c) $\displaystyle\mathbf{x}+\overline{\mathbf{A}}\mathbf{f}=\bm{\ell}$ (2d) $\displaystyle 0\leq u_{i}\leq 1,\quad i=1,...,n$ (2e) $\displaystyle\bm{\ell}\in\mathcal{L};$ (2f) where $\mathcal{F}/j$ denotes all remaining entries of vectors or matrix which excludes those correspond to $f_{j}$. On the contrary, when the specific load vector is available, we can conduct the following _sample-aware constraint screening_ : $\displaystyle\max_{\mathbf{u},\mathbf{x},\mathbf{f}}/\min_{\mathbf{u},\mathbf{x},\mathbf{f}}\quad$ $\displaystyle f_{j}$ (3a) s.t. $\displaystyle u_{i}\underline{x}_{i}\leq x_{i}\leq u_{i}\bar{x}_{i},\quad i=1,...,n$ (3b) $\displaystyle-\overline{\mathbf{f}}_{\mathcal{F}/j}\leq\mathbf{K}_{\mathcal{F}/j}\tilde{\mathbf{f}}\leq\overline{\mathbf{f}}_{\mathcal{F}/j}$ (3c) $\displaystyle\mathbf{x}+\overline{\mathbf{A}}\mathbf{f}=\bm{\ell}$ (3d) $\displaystyle 0\leq u_{i}\leq 1,\quad i=1,...,n;$ (3e) where $\bm{\ell}$ is a known load vector for UC problem. The above formulations are both optimization-based approaches, which seek to find the limit of the flow while keeping all other flow and generation constraints satisfied. However, this approach still allows some line flow values causing unrealistic cost to reach the upper or lower bounds, and thus there are more redundant constraints reserved[5]. Therefore, it is interesting to consider the economical goal in the original UC problem, minimizing the system cost, to further safely screen out constraints. ## III Learning to Predict UC Costs ### III-A Learning Cost Predictors for Unit Commitment Problem As mentioned before, screening without cost objectives enlarge the possible value range of load variables $f_{j}$, which leads to conservative screening and keep more constraints as non-redundant. To close such gap, in this paper, we investigate if it is possible to tighten the search space of constraint screening by adding a cost constraint in the form of $\sum_{i=1}^{n}c_{i}x_{i}\leq\bar{C}$, where $\bar{C}$ is the upper bound whose value needs to be determined in the following sections. To achieve such goal, the adopted method should approximate the map between load input and system costs $J(\bm{\ell})$ well along with predicting system costs efficiently. Thus, in this paper, we use a neural network (NN) to find the upper bound. To train the NN model, we utilize the past record of UC solutions and the training loss between the output of NN model and the actual cost defined as follows, $\displaystyle L:=\left\|f_{NN}(\bm{\ell})-J(\bm{\ell})\right\|_{2}^{2};$ (4) where $f_{NN}(\bm{\ell})$ denotes the NN model given load inputs. In the next subsection, we detail how to connect NN’s predicted costs to the constraint screening problems. ### III-B Tightening the Search Space for Constraint Screening Note that the ML model is not directly applied for making operation or dispatch decisions, and alternatively, we are treating the ML prediction as a constraint to reduce the search space of optimization-based constraint screening problems (2) and (3). With such design, the resulting optimization problem can still find feasible decisions for the original UC problem. We can then add the neural network’s prediction as an additional constraint to the original constraint screening problem to further restrict the search space for each transmission’s flow bounds. In sample-agnostic case, to ensure feasibility of the constraint screening problem after adding the cost constraint for the whole load space along with restricting the searching space effectively, we need to find a predicted cost given by NN which can serve as the upper bound. Then projected gradient ascent (PGA) algorithm can be adopted to achieve this goal. PGA can find the upper bound iteratively by moving $\bm{\ell}$ in the gradient direction at each step along with projecting it onto $\mathcal{L}$, and the details are listed in Algorithm 1. Besides, in practice, the real load samples may be out of distribution, and incur costs which are over the upper bound and thus causing screening failure. Therefore, we use a relaxation parameter $\epsilon$ to adjust the obtained upper bound $\texttt{PGA}(f_{NN}(\bm{\ell}))$ and integrate it to (2). Then, we can get the following sample-agnostic screening problem considering the cost constraint, $\displaystyle\max_{\mathbf{u},\mathbf{x},\mathbf{f},\bm{\ell}}/\min_{\mathbf{u},\mathbf{x},\mathbf{f},\bm{\ell}}\quad$ $\displaystyle f_{j}$ (5a) s.t. $\displaystyle\eqref{Screening:gen}\eqref{Screening:flow}\eqref{Screening:balance}\eqref{Screening:u}\eqref{Screening:load}$ (5b) $\displaystyle\sum_{i=1}^{n}c_{i}x_{i}\leq\texttt{PGA}(f_{NN}(\bm{\ell}))(1+\epsilon).$ (5c) 0: Load distribution $\mathcal{L}$, trained NN model $f_{NN}(\bm{\ell})$, step size $\beta$. 0: Upper bound $\texttt{PGA}(f_{NN}(\bm{\ell}))$. 0: Random load vector $\bm{\ell}^{(0)}\in\mathcal{L}$, $k=0$. 1: while $\bm{\ell}^{(k)}$ doesn’t converge do 2: Update: $\texttt{PGA}(f_{NN}(\bm{\ell}))\leftarrow f_{NN}(\bm{\ell}^{(k)})$ 3: Calculate gradient $\nabla_{\bm{\ell}}f_{NN}(\bm{\ell})$ 4: Update: $\bm{\ell}^{(k+1)}\leftarrow\texttt{Proj}_{\mathcal{L}}(\bm{\ell}^{(k)}+\beta\nabla_{\bm{\ell}}f_{NN}(\bm{\ell}))$ 5: $k\leftarrow k+1$ 6: end while 7: Return $\texttt{PGA}(f_{NN}(\bm{\ell}))$ Algorithm 1 Projected Gradient Ascent Algorithm In sample-aware case, we predict and still relax the UC cost for each specific sample, as the predicted cost may be lower than the actual cost, which can result in an infeasible adjusted screening problem for the investigated sample. Then we add the relaxed cost to (3), and the adjusted sample-agnostic screening problem can be formulated as follows, $\displaystyle\max_{\mathbf{u},\mathbf{x},\mathbf{f}}/\min_{\mathbf{u},\mathbf{x},\mathbf{f}}\quad$ $\displaystyle f_{j}$ (6a) s.t. $\displaystyle\eqref{Screening2:gen}\eqref{Screening2:flow}\eqref{Screening2:balance}\eqref{Screening2:u}$ (6b) $\displaystyle\sum_{i=1}^{n}c_{i}x_{i}\leq f_{NN}(\bm{\ell})(1+\epsilon).$ (6c) Note that the upper bound given by the NN model will be a constant given load region or specific load vector, so the screening problems (5) and (6) can be treated as linear programming problems which are efficient to solve. ## IV Case Study Figure 2: The NN’s predicted costs of different load levels. Figure 3: Percentage of the reduced constraints (upper) and relative solution time (lower) of sample-agnostic screening on different load variation ranges. To evaluate the performance of the proposed constraint screening algorithm, we take the original optimization-based method as benchmark, and also compare our method with KNN method in this section. The predicting accuracy of the learning cost predictors, the computational efficiency and the solution accuracy of the reduced UC problem are examined over a wide range of problem settings. The details are given at https://github.com/Hexuan085/UC_SCREENING_ML. ### IV-A Simulation Setups We carry out the numerical experiments on IEEE 14-bus, IEEE 39-bus and IEEE 118-bus power systems. For each system, we consider the load with 0%, 25%, 50%, 75% and 100% variation which is defined as $r$ around the average nominal values $\overline{\bm{\ell}}$. When investigating the sample-aware constraint screening, the load level is known in our setting and is defined as $\overline{L}$. Then the load region $\mathcal{L}$ considered here can be represented as: $\displaystyle(1-r)\overline{\bm{\ell}}\leq$ $\displaystyle\bm{\ell}\leq(1+r)\overline{\bm{\ell}}$ (7a) $\displaystyle\sum_{i=1}^{n}$ $\displaystyle l_{i}=\overline{L}.$ (7b) TABLE I: Comparisons of the relative cost error and relative solution time for IEEE 118-bus system | Total cost error (%) | Total solution time (%) ---|---|--- Method Range | 25 | 50 | 75 | 100 | 25 | 50 | 75 | 100 KNN5 | 5.3 | 9.8 | 3.3 | 4.5 | 18.3 | 16.5 | 16.4 | 16.7 KNN10 | 0.8 | 0 | 0.2 | 1.7 | 18.5 | 19.1 | 19.7 | 20.3 Benchmark | 0 | 0 | 0 | 0 | 31.4 | 36.1 | 40.8 | 45.6 Ours | 0 | 0 | 0 | 0 | 21.7 | 33.4 | 39.5 | 45.7 TABLE II: Percentage of average reduced constraints of Sample-aware screening | Num.Gen. | Num.Lines | Benchmark | Ours | Actual | $\epsilon$ ---|---|---|---|---|---|--- 14-bus | 5 | 20 | 92.5% | 97.5% | 97.5% | 0.01 39-bus | 10 | 46 | 84.7% | 86.9% | 92.4% | 0.03 118-bus | 54 | 186 | 81.5% | 85.7% | 97.3% | 0.01 To generate samples for training and validating the neural network model and KNN model, we use uniform distribution to get different load vectors $\bm{\ell}\in\mathcal{L}$ for sample-agnostic case or random $\bm{\ell}$ for sample-aware case, and then solve (1) for all loads. The UC cost and the binding situation of each line flow constraint are recorded. Under each setting, we solve and collect 10,000 samples for each neural network with 20 percentage of generated data split as test samples, while for KNN we solve 2,000 samples only based on 118-bus system due to computation burden. Moreover, when evaluating the screening performance of the benchmark, the proposed method and KNN, we use the same validation data and consider 100 samples for each validation case. The used neural networks all have ReLU activation units and 4 layers, and corresponding neurons on each hidden layer are 50, 30, 30. We feed the load vector as input for the neural network and the output is the corresponding UC cost, then we further use the cost to solve (5) and (6). All simulations have been carried out on a laptop with a 2.50 GHz processor and 16G RAM. Specifically, all the optimization problems are modeled using Python and solved with CVXPY[13] powered by GPLK_MI solver [14]. ### IV-B Simulation results To ensure the effectiveness and scalability of the proposed cost predictors, we train the NNs for different load levels, and randomly select a specific load vector from each load level to predict the corresponding cost. The results are shown in Fig. 2, from which it can be seen that the predicted costs are almost equal to the actual costs obtained by solving (1) with the relative error less than 1%. Note that the predicted cost can be lower than the actual cost, so it is reasonable to consider $\epsilon$ in (6c) to ensure feasibility. Using the NN models trained for the setting load levels and the PGA algorithm, we can get the upper bounds in (5c) so as to conduct the sample-agnostic constraint screening. Then, this method is compared with the benchmark and KNN method, which is carried out on 118-bus system and the results are shown in Table I and Fig. 3. According to Table I where the cost error and the solution time of the reduced problems are relative to the result of the original problem (1), KNN methods can reduce more solution time than other methods. The total cost errors of the case K=5 are lower than that of the case K=10, while the situation of the solution time is on the contrary. Though requiring more solution time, the benchmark and our methods can promise the solution accuracy without cost error. Meanwhile, our method can screen more constraints and save more solution time than the benchmark in all cases of investigated power systems with load variation range from 0% to 50% according to Fig. 3. In the cases of 39-bus and 118-bus systems with 75% to 100% load range, the performance of the two methods are very close. This may be due to the increasing patterns of non- redundant constraints with larger load variation range, i.e., the percentage of the redundant constraints decreases when widening the load variation as shown in Fig. 3. Furthermore, according to Table II, the average percentage of redundant constraints can reach 92.4% to 97.3% for a specific load vector. Then, with the sample-aware constraint screening, most of the redundant constraints can be removed. Specifically, the benchmark method can screen 81.5% to 92.5% of total constraints as redundant, while our method defined in (6) can screen 85.7% to 95.1% with setting $\epsilon$ properly. The above results show the following positive effects of our method: 1. 1. Capturing the mapping between load vector and UC cost well at different load levels. 2. 2. Realizing the trade-off between computational efficiency and solution accuracy in the sample-agnostic case. 3. 3. Achieving higher screening efficiency in the sample-aware case. ## V Conclusion and Future Works In this paper, we introduce a novel usage of machine learning to help screen redundant constraints. The neural networks are trained to predict UC cost so as to integrate the cost constraints to original screening problem efficiently. With the cost constraints, the search space of constraint screening can be sufficiently tightened. Since our method does not necessarily yield a minimal set of active constraints for the underlying UC problem, in the future work we would like to seek theoretical understandings about the set of constraints and investigate how the proposed techniques can be generalized to multi-step UC problem with nonlinear constraints. We also plan to explore the potential of making sample-agnostic case serve as the warm-start for the sample-aware case. ## References * [1] J. D. Glover, M. S. Sarma, and T. Overbye, _Power system analysis & design, SI version_. Cengage Learning, 2012\. * [2] A. S. Xavier, F. Qiu, F. Wang, and P. R. Thimmapuram, “Transmission constraint filtering in large-scale security-constrained unit commitment,” _IEEE Transactions on Power Systems_ , vol. 34, no. 3, pp. 2457–2460, 2019. * [3] Q. Zhai, X. Guan, J. Cheng, and H. Wu, “Fast identification of inactive security constraints in scuc problems,” _IEEE Transactions on Power Systems_ , vol. 25, no. 4, pp. 1946–1954, 2010. * [4] A. J. Ardakani and F. Bouffard, “Identification of umbrella constraints in dc-based security-constrained optimal power flow,” _IEEE Transactions on Power Systems_ , vol. 28, no. 4, pp. 3924–3934, 2013. * [5] Á. Porras, S. Pineda, J. M. Morales, and A. Jiménez-Cordero, “Cost-driven screening of network constraints for the unit commitment problem,” _arXiv preprint arXiv:2104.05746_ , 2021. * [6] S. Pineda, J. M. Morales, and A. Jiménez-Cordero, “Data-driven screening of network constraints for unit commitment,” _IEEE Transactions on Power Systems_ , vol. 35, no. 5, pp. 3695–3705, 2020. * [7] A. Ardakani and F. Bouffard, “Prediction of umbrella constraints,” in _2018 Power Systems Computation Conference (PSCC)_. IEEE, 2018, pp. 1–7. * [8] Á. S. Xavier, F. Qiu, and S. Ahmed, “Learning to solve large-scale security-constrained unit commitment problems,” _INFORMS Journal on Computing_ , vol. 33, no. 2, pp. 739–756, 2021. * [9] L. A. Roald and D. K. Molzahn, “Implied constraint satisfaction in power system optimization: The impacts of load variations,” in _2019 57th Annual Allerton Conference on Communication, Control, and Computing (Allerton)_. IEEE, 2019, pp. 308–315. * [10] M. Awadalla and F. Bouffard, “Influence of stochastic dependence on network constraints screening for unit commitment,” _arXiv preprint arXiv:2208.12139_ , 2022. * [11] Y. Chen, L. Zhang, and B. Zhang, “Learning to solve dcopf: A duality approach,” _Electric Power Systems Research_ , vol. 213, p. 108595, 2022\. * [12] D. Bertsimas and J. N. Tsitsiklis, _Introduction to linear optimization_. Athena Scientific Belmont, MA, 1997, vol. 6. * [13] S. Diamond and S. Boyd, “Cvxpy: A python-embedded modeling language for convex optimization,” _Journal of Machine Learning Research_ , vol. 17, no. 83, pp. 1–5, 2016. [Online]. Available: http://jmlr.org/papers/v17/15-408.html * [14] A. Makhorin, “Glpk (gnu linear programming kit),” _http://www.gnu.org/s/glpk/glpk.html_ , 2008.
# CUNI Systems for the WMT 22 Czech-Ukrainian Translation Task Martin Popel11footnotemark: 1 Jindřich Libovický11footnotemark: 1 Jindřich Helcl11footnotemark: 1 Charles University, Faculty of Mathematics and Physics Institute of Formal and Applied Linguistics Malostranské náměstí 25, 118 00 Prague, Czech Republic <EMAIL_ADDRESS> ###### Abstract ††footnotetext: ∗ The author order was determined by a coin toss. We present Charles University submissions to the WMT 22 General Translation Shared Task on Czech-Ukrainian and Ukrainian-Czech machine translation. We present two constrained submissions based on block back-translation and tagged back-translation and experiment with rule-based romanization of Ukrainian. Our results show that the romanization only has a minor effect on the translation quality. Further, we describe Charles Translator, a system that was developed in March 2022 as a response to the migration from Ukraine to the Czech Republic. Compared to our constrained systems, it did not use the romanization and used some proprietary data sources. ## 1 Introduction How fast can the machine translation (MT) community react to a sudden need of a high-quality MT system which was previously under low demand? This question motivated the new task at the WMT this year, which is Czech-Ukrainian translation. Both languages belong to the Slavic language family (Czech is western Slavic, Ukrainian is eastern Slavic), and share some lexical and structural characteristics. Unlike Czech, which uses the Latin script, Ukrainian uses its variant of the Cyrillic alphabet. We submit three systems to the WMT 22 General Translation Shared Task for this language pair in each translation direction. The first system, CUNI-JL-JH, implemented in Marian (Junczys-Dowmunt et al., 2018), uses tagged back- translation and is a result of our experiments with romanization of Ukrainian. Our second system, CUNI-Transformer, implemented in Tensor2Tensor (Vaswani et al., 2018), uses block back-translation. Finally, we submit an unconstrained system, Charles Translator, implemented in Tensor2Tensor, which has been developed in spring 2022 as a response to the crisis caused by the Russian invasion of Ukraine and the following migration wave. ## 2 Constrained WMT Submissions We submitted two systems in each translation direction that use the same parallel and monolingual data, but different techniques and different toolkits. This section first describes the shared data processing steps and then the specifics of each of the submissions in separate subsections. ### 2.1 Training Data We use all parallel data allowed in the constrained task, along with 50 million Czech and 58 million Ukrainian sentences of monolingual data. In the following paragraphs we describe the data cleaning steps when preparing the training data. We further experiment with romanization of the Ukrainian Cyrillic alphabet and with artificial noising of the data. #### Parallel data. The data for the constrained translation task consist of OPUS corpora (Tiedemann, 2012) that have a Czech-Ukrainian part, WikiMatrix (Schwenk et al., 2021) and the ELRC EU acts in Ukrainian.111 https://elrc- share.eu/repository/search/?q=EU+acts+in+Ukrainian We clean the parallel data using rule-based filtering in the following way: 1. 1. Filter out non-printable and malformed UTF-8 characters. 2. 2. Detect language using FastText (Grave et al., 2018), only keep Czech and Ukrainian sentences on their respective source/target sides. 3. 3. Only keep sentence pairs with character length ratio between $0.67$ and $1.5$ if longer than 10 characters. 4. 4. Apply hand-crafted regular expressions to filter out the frequent errors, such that the system does not attempt to translate e-mail addresses, currencies, etc. In addition, regular expressions check translations of names of Czech222https://uk.wikipedia.org/wiki/ Міста_Чехії and Ukrainian333 https://cs.wikipedia.org/wiki/Seznam_měst_na_Ukrajině municipalities downloaded from Wikipedia. We omit steps 2 and 3 for the XLEnt corpus, which seems to be very clean and consist of short phrases (likely to get misclassified for language). The sizes of the used parallel data sources before and after cleaning are presented in Table 1. Source | Original | Filtered ---|---|--- bible-uedin | 8 k | 8 k CCMatrix | 3,992 k | 3,884 k EUbookshop | 2 k | 1 k GNOME | 150 | 81 KDE4 | 134 k | 64 k MultiCCAligned | 1,607 k | 1,199 k MultiParaCrawl | 1,773 k | 1,606 k OpenSubtitles | 731 k | 273 k QED | 161 k | 138 k Tatoeba | 3 k | 2 k TED2020 | 115 k | 106 k Ubuntu | 0.2k | 0.2k wikimedia | 2 k | 2 k XLEnt | 695 k | 695 k WikiMatrix | 105 k | 99 k ELRC EU Acts | 130 k | 108 k Total | 9,457 k | 8,186 k Table 1: Sizes of parallel data sources (number of sentence pairs). #### Monolingual data. The overview of the monolingual data sources is in Table 2. For Czech, we use the Czech monolingual portion of the CzEng 2.0 corpus (Kocmi et al., 2020). For Ukrainian, we used all resources, available for WMT, i.e., the NewsCrawl, the Leipzig Corpora (Biemann et al., 2007), UberText corpus (Khaburska and Tytyk, 2019) and Legal Ukrainian Crawling by ELRC. The Uber corpus and the Ukrainian Legal corpus are distributed tokenized with removed punctuation. We automatically restored the punctuation and detokenized the models using a lightweight Transformer model (Vaswani et al., 2017; Base model with 3 layers, 8k vocabulary) trained on the NewsCrawl corpus. For Ukrainian, we only keep sentences shorter than 300 characters. For Czech, we keep all sentence lengths from the CzEng corpus (up to 1400 characters). For both languages, we remove non-printable and malformed UTF-8 characters. Source | Original | Filtered ---|---|--- Czech | CzEng 2.0 | | 50.6 M Ukrainian | NewsCrawl | 0 2.3 M | 0 2.0 M Leipzig Corpora | 0 9.0 M | 0 7.6 M UberText Corpus | 47.9 M | 41.2 M ELRC Legal | 0 7.6 M | 0 7.2 M Total | 66.8 M | 58.1 M Table 2: Monolingual data sizes in number of sentences before and after filtering. #### Romanization. We develop a reversible romanization than transcribes between the Ukrainian and Czech alphabets. For example, Зараз у нас є 4-місячні миші is transcribed to Zaraz u nas je 4-misjačni myši. This way the model can better exploit the lexical similarities between the two languages (e.g. миші should be translated to Czech as myši), while keeping all the necessary information to reconstruct the original Cyrillic text. Note that the transcription of Cyrillic changes when changing the target language, reflecting the phonology of that language (e.g. ш transcribes to sh in English, but š in Czech). We introduce special tags for words and characters that are written in Latin script found in Cyrillic text. The romanization is specifically designed for Ukrainian (e.g. и transcribes to y, not i as would be the case in Russian), so its reversibility occasionally fails for Russian names. #### Artificial noise. We apply synthetic noise on the source side that should simulate the most frequent deviations from the standard orthography (missing capitalization, lower- or upper-casing parts of the sentences, missing or additional punctuation). All scripts for training data processing are available at https://github.com/ufal/uk-cs-data-scripts. We use Flores 101 (Goyal et al., 2022) development set for validation. ### 2.2 Tagged-back-translation-based System (CUNI-JL-JH) The CUNI-JL-JH submission is a constrained system and uses the data described in the paragraphs above. We train the system in 3 iterations of tagged back- translation (Caswell et al., 2019) with greedy decoding. Each iteration, we filter the back-translated data using Dual Cross-Entropy filtering (Junczys- Dowmunt, 2018) when keeping $40,930,735$ synthetic sentences, i.e., 5$\times$ the size of clean authentic parallel data. The first two back-translation iterations were done with the Cyrillic script on the Ukrainian side. In the final back-translation iteration, we performed romanization and noising of the source side. We train three models with random initialization and submit the ensemble. For all iterations, we used a Transformer Big model with tied embeddings and a shared SentencePiece vocabulary size of $32$k (fitted on 5M randomly sampled sentences; with sampling at the training time, $\alpha$=0.1; Kudo and Richardson, 2018). We set the learning rate to $0.0003$ and use $8,000$ warm- up steps. We initialize the models randomly in each back-translation iteration. For validation, we use greedy decoding. At test time, we decode with beam search with beam size of 4 and length normalization of 1.0 (estimated on validation data). The system is implemented using Marian (Junczys-Dowmunt et al., 2018). #### Negative results. We experimented with Dual-Cross-Entropy filtering (Junczys-Dowmunt, 2018) for parallel data selection and came to inconclusive results. Therefore, we used all parallel data after rule-based filtering.444 Note that we use Dual-Cross- Entropy for filtering the monolingual data, as described in the first paragraph of this section, but we have not done any experiments with keeping all the monolingual data. Additionally, we experimented with MASS-style (Song et al., 2019) pre-training using monolingual data only and continue with training on parallel data. We were not able to find a hyper-parameter setting where the pre-trained model would outperform the models trained from random initialization. Therefore, we only use model trained from random initialization. ### 2.3 Block back-translation System (CUNI-Transformer) The CUNI-Transformer submission is also constrained, trained on the same data as CUNI-JL-JH. The system was trained in the same way as the sentence-level English-Czech CUNI-Transformer systems submitted to previous years of WMT shared tasks (Popel, 2018, 2020; Gebauer et al., 2021). It uses Block back- translation (BlockBT) (Popel et al., 2020), where blocks of authentic (human- translated parallel) and synthetic (backtranslated) training data are not shuffled together, but checkpoint averaging is used to find the optimal ratio of checkpoints from the authentic and synthetic blocks (usually 5:3). The uk$\rightarrow$cs system was trained with a non-iterated BlockBT (i.e. cs-mono data was translated with an authentic-only trained baseline). The cs$\rightarrow$uk was trained with two iterations of BlockBT (i.e. the uk-mono data was translated with the above mentioned uk$\rightarrow$cs non-iterated BlockBT system). We had not enough time to train more iterations and apply noised training and romanization. The system was implemented using Tensor2Tensor (Vaswani et al., 2018). #### Inline casing. We experimented with Inline casing (InCa) pre-processing in the cs$\rightarrow$uk direction. The main idea is to lowercase all training data and insert special tags <titlecase> and <all-uppercase> before words in the respective case, so that the original casing can be reconstructed (with the exception of words like McDonald or iPhone, which use different casing patterns than all-lowercase, all-uppercase and titlecase). We improved this approach by remembering the most frequent casing variant of each (lowercased) word in the training data. The most frequent variant does not need to be prefixed with any tag, which makes the length of training sequences shorter. We also introduced a third tag <all-lowercase> for encoding all-lowercased words whose most frequent variant is different. For example, if the InCa vocabulary includes only two items: iPhone and GB, sentence My iPhone 64GB and iPod 64 GB or 32 gb will be encoded as <titlecase> my iphone <all-uppercase> 64gb and iPod 64 gb or 32 <all-lowercase> gb. Note that iPod was kept in the original case because it was not included in the InCa vocabulary and it does not match any of the three “regular” casing patterns. We applied InCa on both the source and target side and experimented with training the InCa vocabulary on the authentic data only or on authentic plus synthetic (monolingual backtranslated). Inline casing showed promising results in preliminary experiments (without backtranslation), especially when combined with romanization and artificial noise in training. Unfortunately, we had not enough time to train the backtranslated model long enough, so we submitted it only as a contrastive run and plan to explore it more in future. ## 3 Charles Translator for Ukraine Charles Translator for Ukraine is a free Czech-Ukrainian online translation service available for public at https://translator.cuni.cz and as an Android app. It was developed at Charles University in March 2022 to help refugees from Ukraine by narrowing the communication gap between them and other people in Czechia. Similarly to CUNI-Transformer, it is based on Transformer and iterated Block back-translation (Popel et al., 2020). The training used source-side artificial noising, but no romanization and no inline casing. It was trained on most (but not all) of the training data provided by WMT plus about one million uk-cs sentences from the InterCorp v14 corpus (Čermák and Rosen, 2012; Kotsyba, 2022), so this submission is unconstrained. ## 4 Results In this section, we report BLEU scores on the Flores 101 development set that we used to make our decisions about the system development and the final automatic scores. Note that the validation set is very different from the test set. The validation set consists of clean and rather complicated sentences from Wikipedia articles, whereas the WMT 22 test set is noisy user-generated text from the logs of the production deployment of Charles Translator.555 The test set only contains sentences from users who provided their consent for this usage and the sentences were pseudonymized. #### Tagged BT systems. Table 3 shows validation BLEU scores from the first three iterations of back- translation. The second and third iteration did not bring substantial improvements, so we decided not to further iterate. Model | cs$\rightarrow$uk | uk$\rightarrow$cs ---|---|--- Authentic only | 20.91 | 22.95 BT iteration 1 | 21.69 | 23.70 BT iteration 2 | 21.87 | 23.98 BT iteration 3 (seed 1) | 21.53 | 23.76 Table 3: Validation BLEU scores for the first two iterations of BT for the tagged BT systems. Table 4 shows validation BLEU scores from the last (third) BT iteration – three independently trained systems and their ensembles, and the Cyrillic and romanized versions of the data. In general, ensembling only brings a small improvement. Romanization does not bring a significant difference compared to using the Cyrillic script. In the Czech-to-Ukrainian direction, the best system was the ensemble of the romanized systems. However, in the Ukrainian- to-Czech direction, the best system was one of the Cyrillic systems that used accidentally 3 times higher batch size than the remaining ones. This result suggests that the batch size has a much stronger effect than most of the techniques that we experimented with and that we might have reached better results if we opted for higher batch size. Model | cs$\rightarrow$uk | uk$\rightarrow$cs ---|---|--- Cyrillic | Seed 1 | 21.53 | 23.76 Seed 2 | 22.28 | 25.10 Seed 3 | 21.96 | 24.39 Ensemble | 22.45 | 24.86 Romanized | Seed 1 | 21.42 | 23.99 Seed 2 | 21.76 | 23.91 Seed 3 | 22.37 | 24.18 Ensemble | 22.62 | 24.22 Table 4: Validation BLEU scores for the last (i.e., the third) iteration of BT comparing romanized and original script. | cs$\rightarrow$uk | uk$\rightarrow$cs ---|---|--- System | BLEU | chrF | COMET | BLEU | chrF | COMET Best constrained (HuaweiTSC/AMU) | 36.0 | 62.6 | 0.994 | 37.0 | 60.7 | 1.048 CUNI-Transformer | 35.0 | 61.6 | 0.873 | 35.8 | 59.0 | 0.885 CUNI-JL-JH | 34.8 | 61.6 | 0.900 | 35.1 | 58.7 | 0.890 Best unconstrained (Lan-Bridge/Online-B) | 38.1 | 64.0 | 0.942 | 36.5 | 60.4 | 0.965 Charles Translator | 34.3 | 61.5 | 0.908 | 35.9 | 59.0 | 0.901 Table 5: Final automatic results on the WTM22 test data compared to the best overall score achieved in each metric. #### Results on WMT test. Automatic evaluation on the WMT22 test set is presented in Table 5. Both the constrained systems and Charles Translator show comparable results. The tagged BT system reaches a slightly higher COMET score than the Block BT system, however, Czech-Ukrainian was not in the training data of the COMET score, which make the score unreliable for this particular language pair. For Czech- to-Ukrainian, Charles Translator reaches a slightly higher COMET score and slightly lower BLEU and chrF scores than both the constrained systems, but we do not consider such small differences of automatic metrics relevant. ## 5 Conclusions We presented Charles University submissions to the WMT 22 General Translation Shared Task on Czech-Ukrainian and Ukrainian-Czech machine translation. We present two constrained submissions based on block back-translation and tagged back-translation and experiment with rule-based romanization of Ukrainian. Further, we describe Charles Translator, a system that was developed in March 2022 as a response to the migration from Ukraine to the Czech Republic. Compared to our constrained systems, it did not use the romanization and used some proprietary data sources. Our results show that the romanization only has a minor effect on the translation quality, compared to machine-learning aspects that affect translation quality. Block back-translation seems to deliver slightly better results that tagged back-translation, however the differences are only small. ## Acknowledgements This work has been supported by the Ministry of Education, Youth and Sports of the Czech Republic, Project No. LM2018101 LINDAT/CLARIAH-CZ, by the Czech Science Foundation (GACR) grant 20-16819X (LUSyD), and by the European Commission via its Horizon 2020 research and innovation programme no. 870930 (WELCOME), Horizon Europe Innovation programme no. 101070350 (HPLT). ## References * Biemann et al. (2007) Chris Biemann, Gerhard Heyer, Uwe Quasthoff, and Matthias Richter. 2007. The leipzig corpora collection-monolingual corpora of standard size. _Proceedings of Corpus Linguistic_ , 2007. * Caswell et al. (2019) Isaac Caswell, Ciprian Chelba, and David Grangier. 2019. Tagged back-translation. In _Proceedings of the Fourth Conference on Machine Translation (Volume 1: Research Papers)_ , pages 53–63, Florence, Italy. Association for Computational Linguistics. * Čermák and Rosen (2012) František Čermák and Alexandr Rosen. 2012. The case of InterCorp, a multilingual parallel corpus. _International Journal of Corpus Linguistics_ , 17(3):411–427. * Gebauer et al. (2021) Petr Gebauer, Ondřej Bojar, Vojtěch Švandelík, and Martin Popel. 2021. CUNI systems in WMT21: Revisiting backtranslation techniques for English-Czech NMT. In _Proceedings of the Sixth Conference on Machine Translation_ , pages 123–129, Online. Association for Computational Linguistics. * Goyal et al. (2022) Naman Goyal, Cynthia Gao, Vishrav Chaudhary, Peng-Jen Chen, Guillaume Wenzek, Da Ju, Sanjana Krishnan, Marc’Aurelio Ranzato, Francisco Guzmán, and Angela Fan. 2022. The Flores-101 evaluation benchmark for low-resource and multilingual machine translation. _Transactions of the Association for Computational Linguistics_ , 10:522–538. * Grave et al. (2018) Edouard Grave, Piotr Bojanowski, Prakhar Gupta, Armand Joulin, and Tomas Mikolov. 2018. Learning word vectors for 157 languages. In _Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018)_ , Miyazaki, Japan. European Language Resources Association (ELRA). * Junczys-Dowmunt (2018) Marcin Junczys-Dowmunt. 2018. Dual conditional cross-entropy filtering of noisy parallel corpora. In _Proceedings of the Third Conference on Machine Translation: Shared Task Papers_ , pages 888–895, Belgium, Brussels. Association for Computational Linguistics. * Junczys-Dowmunt et al. (2018) Marcin Junczys-Dowmunt, Roman Grundkiewicz, Tomasz Dwojak, Hieu Hoang, Kenneth Heafield, Tom Neckermann, Frank Seide, Ulrich Germann, Alham Fikri Aji, Nikolay Bogoychev, André F. T. Martins, and Alexandra Birch. 2018. Marian: Fast neural machine translation in C++. In _Proceedings of ACL 2018, System Demonstrations_ , pages 116–121, Melbourne, Australia. Association for Computational Linguistics. * Khaburska and Tytyk (2019) Anastasiia Khaburska and Igor Tytyk. 2019. Toward language modeling for the ukrainian. _Advances in Data Mining, Machine Learning, and Computer Vision. Proceedings_ , pages 71–80. * Kocmi et al. (2020) Tom Kocmi, Martin Popel, and Ondrej Bojar. 2020. Announcing CzEng 2.0 parallel corpus with over 2 gigawords. _CoRR_ , abs/2007.03006. * Kotsyba (2022) Natalia Kotsyba. 2022. Ukrainian-Czech part of InterCorp v14. https://intercorp.korpus.cz. * Kudo and Richardson (2018) Taku Kudo and John Richardson. 2018. SentencePiece: A simple and language independent subword tokenizer and detokenizer for neural text processing. In _Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations_ , pages 66–71, Brussels, Belgium. Association for Computational Linguistics. * Popel (2018) Martin Popel. 2018. CUNI transformer neural MT system for WMT18. In _Proceedings of the Third Conference on Machine Translation: Shared Task Papers_ , pages 482–487, Belgium, Brussels. Association for Computational Linguistics. * Popel (2020) Martin Popel. 2020. CUNI English-Czech and English-Polish systems in WMT20: Robust document-level training. In _Proceedings of the Fifth Conference on Machine Translation_ , pages 269–273, Online. Association for Computational Linguistics. * Popel et al. (2020) Martin Popel, Marketa Tomkova, Jakub Tomek, Łukasz Kaiser, Jakob Uszkoreit, Ondřej Bojar, and Zdeněk Žabokrtský. 2020. Transforming machine translation: a deep learning system reaches news translation quality comparable to human professionals. _Nature Communications_ , 11(4381):1–15. * Schwenk et al. (2021) Holger Schwenk, Vishrav Chaudhary, Shuo Sun, Hongyu Gong, and Francisco Guzmán. 2021. WikiMatrix: Mining 135M parallel sentences in 1620 language pairs from Wikipedia. In _Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume_ , pages 1351–1361, Online. Association for Computational Linguistics. * Song et al. (2019) Kaitao Song, Xu Tan, Tao Qin, Jianfeng Lu, and Tie-Yan Liu. 2019. MASS: masked sequence to sequence pre-training for language generation. In _Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA_ , volume 97 of _Proceedings of Machine Learning Research_ , pages 5926–5936. PMLR. * Tiedemann (2012) Jörg Tiedemann. 2012. Parallel data, tools and interfaces in OPUS. In _Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC’12)_ , pages 2214–2218, Istanbul, Turkey. European Language Resources Association (ELRA). * Vaswani et al. (2018) Ashish Vaswani, Samy Bengio, Eugene Brevdo, Francois Chollet, Aidan Gomez, Stephan Gouws, Llion Jones, Łukasz Kaiser, Nal Kalchbrenner, Niki Parmar, Ryan Sepassi, Noam Shazeer, and Jakob Uszkoreit. 2018. Tensor2Tensor for neural machine translation. In _Proceedings of the 13th Conference of the Association for Machine Translation in the Americas (Volume 1: Research Track)_ , pages 193–199, Boston, MA. Association for Machine Translation in the Americas. * Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In _Advances in Neural Information Processing Systems 30_ , pages 6000–6010, Long Beach, CA, USA. Curran Associates, Inc.
# Real-Time High-Quality Stereo Matching System on a GPU Qiong Chang, Tsutomu Maruyama ###### Abstract In this paper, we propose a low error rate and real-time stereo vision system on GPU. Many stereo vision systems on GPU have been proposed to date. In those systems, the error rates and the processing speed are in trade-off relationship. We propose a real-time stereo vision system on GPU for the high resolution images. This system also maintains a low error rate compared to other fast systems. In our approach, we have implemented the cost aggregation (CA), cross-checking and median filter on GPU in order to realize the real- time processing. Its processing speed is 40 fps for $1436\times{992}$ pixels images when the maximum disparity is 145, and its error rate is the lowest among the GPU systems which are faster than 30 fps. ## I Introduction The aim of stereo vision systems is to reconstruct the 3-D geometry of a scene from images taken by two separate cameras. The computational complexity of the stereo vision is very high, and many acceleration systems with GPUs, FPGAs and dedicated hardware have been developed [1][4]. All of them succeeded in real- time processing of high resolution images, but their accuracy is not good enough because they simplify the algorithms to fit the hardware architecture. In this paper, we aim to construct a real-time GPU stereo vision system for the high resolution image set ($1436\times{992}\times{145}$ disparities) in the Middlebury Benchmark. GPUs have more than hundred cores which run faster than 1GHz, and drastic performance gain can be expected in many applications. However, the processing speed of the stereo vision by GPUs is much slower than FPGAs such as [2][3]. This is caused by the fact that data for several lines are intensively accessed at a time in the stereo vision, but the shared memory is too small to cache those lines. Therefore, many memory accesses to the global memory are required, and they limit the processing speed by GPUs. On the other hand, more sophisticated algorithms can be implemented on GPUs than FPGAs, and lower error rates have been achieved by many GPU systems[5][6]. The algorithms for the stereo vision can be categorized into two groups: local and global. In the local algorithms, only the local information around the target pixel is used to decide the disparity of the pixel, while the disparities of all pixels are decided considering the mutual effect of all pixels in the global algorithms. Thus, in general, the global algorithms achieve lower error rates, but require longer computation time. On many GPUs, algorithms which require only local information in each step, but propagate the mutual effect gradually by repeating the step many times are implemented. Their error rates are low enough, but their processing speed is far behind of the real-time requirement. We have implemented a local search algorithm, which is an improved version of the algorithm that we have implemented on FPGA[9]. In our algorithm, AD (absolute difference) and mini-Census transform [14] are used to calculate the matching costs of the pixels in the two images, and they are aggregated along the $x$\- and $y$-axes by using the Cross-based method [15][16], in order to compare the pixels as a block of the similar color. The matching costs are calculated twice using left and right image as the base, and two disparity maps are generated. Then, the disparity map is improved using the disparities which are common in both disparity maps. These operations are chosen to realize low error rates without repetitive computation. However, when we consider processing a high resolution image, its computational complexity is too high to achieve real-time processing. In order to achieve high performance, in this paper, we scale down the images to reduce the computational complexity. As shown in Fig.1, to reduce the computational complexity, we scale down the input images into a small size, and then scale up the disparity map to the original size. The images are scaled down to 1/4 by reducing the width and height by half, and the disparities are calculated on the scaled down images. The maximum disparity is also reduced to half, which means that the total computational complexity can be reduced to 1/8. This approach typically worsens the matching accuracy, but in our approach, an bilateral interpolation method is performed during the scaling up step, and a high matching accuracy can be maintained. This approach becomes possible because of the high quality of the high resolution images. Figure 1: Image Scaling ## II STEREO VISION In the stereo vision systems, the matching of pixels in the two images (left and right) taken by two separate cameras is searched to reconstruct the 3-D geometry of the scene. When the two cameras are calibrated properly, epipolar restriction can be used. Under this restriction, we can obtain a disparity map $D_{map}$ by finding the matching of pixels on the epipolar lines of the two images as shown in Fig.2. Fig.2 shows how to calculate a disparity map using the left image as the base. A pixel in the left image $L(x,y)$ (or a window centered by $L(x,y)$) is compared with $D$ pixels $R(x-d,y)d=[0,D-1]$ in the right image (or $D$ windows centered by those pixels), and the most similar pixel to $L(x,y)$ is searched. $D$ is the maximum disparity value. Suppose that $R(x-d,y)$ is the most similar to $L(x,y)$. Then, this means that $L(x,y)$ and $R(x-d,y)$ are the same point of an object, and the distance $Z$ to the object can be calculated from d and two parameters of the two cameras using the following equation: $Z=f{\frac{B}{d}}.$ (1) Here, $f$ is the focus length of both cameras, and $B$ is the distance between the two cameras. Smaller d means that the object is farther away from the cameras, and $d=0$ means that the object is at infinity. The main problem in the stereo vision is to find the matching pixels in the two images correctly. Another important problem in the stereo vision is the occlusion. When an object is taken by two cameras, some parts of the object appear in one image but do not appear in another image, depending on the positions and angles between the cameras and the object. These occlusions are the major source of errors in the stereo vision systems. In order to avoid those errors, $D_{map}$ is calculated twice: once using $L$, the left image, as the base ($D^{L}_{map}$), and another using $R$, the right image, as the base ($D^{R}_{map}$). Then, by using the same matching found in both $D^{L}_{map}$ and $D^{R}_{map}$ as the reliable disparities (called the ground control points, or GCPs), higher quality disparity maps can be obtained. Figure 2: Local matching under the epipolar restriction ## III OUR ALGORITHM In this section, we introduce the details of our algorithm. Our algorithm consists of the following steps. 1. 1. the two input images are gray-scaled 2. 2. scaling down the two images 3. 3. calculating matching cost of each pixel 4. 4. cost aggregation along the $x$\- and $y$-axes 5. 5. generating two disparity maps 6. 6. detecting GCPs (ground control pixels) by cross-checking the two disparity maps 7. 7. refinement by a median filter and filling the non-GCPs by using a bilateral estimation method 8. 8. scaling up the disparity map ### III-A Scaling Down In order to reduce the computational complexity, the two images are scaled down linearly in both horizontal and vertical directions using the mean- pooling method. Here, take the left image as an example: $\displaystyle L(x,y)=$ $\displaystyle\frac{1}{(2m+1)^{2}}\times{\sum_{j=-m}^{m}\sum_{i=-m}^{m}L_{org}(K\cdot{x}+i,K\cdot{y}+j)}$ (2) where $L_{org}(K\cdot{x}+i,K\cdot{y}+j)$ is the pixel in the original image, and $K$ is the factor for the scaling down (in our implementation, $K=2$). $L(x,y)$ is the pixel of the left scaled down image, and is smoothed by a mean-filter the size of which is $(2m+1)^{2}$. By choosing the block size carefully, we can avoid the loss of the matching accuracy, and can improve the processing speed. ### III-B Matching cost between two pixels The matching cost of each pixel is calculated using the absolute difference of the brightness and the mini-census transform. When the left image $L(x,y)$ is used as the base, the matching cost of the disparity $=d$ is given by $C^{L}(x,y,d)=C^{L}_{AD}(x,y,d)+C^{L}_{MC}(x,y,d).$ (3) $C^{L}_{AD}(x,y,d)$ is the cost by the absolute difference of the brightness of the two pixels, and given by $C^{L}_{AD}(x,y,d)=1-\exp(-\frac{|L(x,y)-R(x-d,y)|}{\lambda_{AD}})$ (4) where $\lambda_{AD}$ is a constant, In the same way, $C^{L}_{MC}(x,y,d)$, the cost by mini-census transform, is given by $C^{L}_{MC}(x,y,d)=1-\exp(-\frac{MC(L(x,y),R(x-d,y))}{\lambda_{MC}})$ (5) where $\lambda_{MC}$ is a constant, and MC($\alpha$, $\beta$) is the Hamming distance between the mini-census transform of $\alpha$ and $\beta$. Mini- census transform used in our approach is shown in Fig.3. The center pixel $L(x,y)$ is compared with its six neighbors, and a six bit sequence is generated as shown in Fig.3. This approach is based on the hypothesis that the relative values of the brightness are kept in both images. Figure 3: Mini-census Transform When the right image $R(x,y)$ is used as the base, the matching cost is given as follows. $\begin{split}C^{R}(x,y,d)&=C^{R}_{AD}(x,y,d)+C^{R}_{MC}(x,y,d)\\\ &=1-\exp(-\frac{|R(x,y)-L(x+d,y)|}{\lambda_{AD}})\hskip 5.0pt+\\\ &\hskip 13.00005pt1-\exp(-\frac{MC(R(x,y),L(x+d,y))}{\lambda_{MC}})\\\ &=C^{L}(x+d,y,d).\end{split}$ (6) This equation means that all $C^{R}(x,y,d)$ are already calculated when $C^{L}(x,y,d)$ are calculated, and $C^{L}(x,y,d)$ can be reused as $C^{R}(x-d,y,d)$. Figure 4: A cost aggregation method ### III-C Cost Aggregation The matching costs are aggregated as much as possible considering the similarity of the brightness of the pixels to compare the pixels as a block of the similar brightness. Fig.4 shows how the matching costs are aggregated. First, the matching costs are aggregated along the $x$-axis. $CA^{L}_{x}(x,y,d)=\sum^{+n}_{dx=-m}C^{L}(x+dx,y,d)$ (7) Here, m and n are the number of the continuous pixels with the similar brightness to $L(x,y)$ $(|L(x,y)-L(x+dx,y)|<\delta)$ on the left and right- side of $L(x,y)$. For example, in Fig.4, $m=3$ and $n=3$, because all pixels from $L(x-3,y)$ to $L(x+3,y)$ are similar to $L(x,y)$. Then, $CA^{L}_{x}(x,y,d)$ are aggregated along the $y$-axis as $CA^{L}(x,y,d)=\sum^{+N}_{dy=-M}CA^{L}_{x}(x,y+dy,d).$ (8) Here, $M$ and $N$ are the number of the continuous pixels with similar brightness to $L(x,y)$ on the upper and lower side of $L(x,y)$. In Fig.4, $M$ is 4 and $N$ is 4, because the pixels from $L(x,y-4)$ to $L(x,y+4)$ are similar to $L(x,y)$. Then, $d$ which minimizes $CA^{L}(x,y,d)$ is chosen as the disparity at $L(x,y)$, and disparity map $D^{L}_{map}(x,y)$ is obtained. $D^{L}_{map}(x,y)=\min_{d}{CA^{L}(x,y,d)}.$ (9) By enlarging the range for summing up along the $x$\- and $y$\- axes, we can obtain more accurate disparities, though it requires more computation time. ### III-D GCPs In our approach, $D^{R}_{map}(x,y)$ is also calculated in the same way as $D^{L}_{map}(x,y)$. Then, ground control points, or GCPs, are obtained by comparing them [7]. Suppose that $D^{L}_{map}(x,y)=k$. This means that $L(x,y)$ and $R(x-k,y)$ showed the best matching when the left image is used as the base, and they are the same point of the object in the images. Therefore, $D^{R}_{map}(x-k,y)$ should also be $k$. If this requirement $D^{L}_{map}(x,y)=D^{R}_{map}(x-k,y)=k$ (10) is satisfied, the point is called a GCP, and it is considered that GCPs have higher reliability. ### III-E Bilateral Estimation Ideally, all pixels except for those in the occluded regions should be GCPs, however, in actuality more pixels become non-GCPs because of the slight change of the brightness between the input images. To achieve more reliable disparities of non-GCPs, two approaches are often used [10]. In both approaches, for each non-GCP, the closest GCPs on the left and right hand-side along the $x$-axis are searched first. Then, in the first approach, as shown in Fig.5(a), the closer GCP in the distance is chosen as the disparity of the non-GCP because the non-GCP and the closer GCP can be considered to belong to the same object with a higher probability. In the second approach, as shown in Fig.5(b), the smaller disparity is chosen as the disparity of the non-GCP assuming that the non-GCP is caused by the occlusion. The disparity of the occluded region is smaller than that of the foreground object because the disparity of the closer object is larger, and the non-GCP should have a smaller disparity. Both of these two methods can be easily implemented on GPU because of their high parallelism. However, due to their single function, the overall accuracy is not good enough. Figure 5: Non-GCPs Filling In our system, we proposed a bilateral estimation methods to fill the non-GCPs as following steps: 1. 1. Define the disparities of the GCP of $L(x-i,y)$ and $L(x+j,y)$ as $D(x-i,y)$ and $D(x+j,y)$. 2. 2. If $|D(x-i,y)-D(x+j,y)|\leq{T}$, where $T$ is the threshold for the difference of disparity, it can be considered that the disparity is changing continuously in this range, and $D(x,y)$ is filled as: $D(x-i,y)+i\cdot((D(x-i,y)-D(x+j,y))/(i+j)).$ (11) 3. 3. If $|D(x-i,y)-D(x+j,y)|>T$, which means that the disparity changes rapidly in this range, it can be considered that an edge exists in this range. Thus $D(x,y)$ is chosen as the $D(x-i,y)$ if $L(x-i,y)$ is closer to $L(x,y)$ than $L(x+j,y)$ in color, and otherwise, $D(x,y)$ is chosen as $D(x+j,y)$. With this approach, we can fill the different areas in the different methods. Then, an accurate disparity map can be expected. ## IV IMPLEMENTATION ON GPU We have implemented the algorithm on Nvidia GTX780Ti. GTX780Ti has 15 streaming multi-processors (SMs). Each SM runs in parallel using 192 cores in it (2880 cores in total). GTX780Ti has two-layered memory system. Each SM has a 48KB shared memory, and one large global memory is shared among the SMs. The access delay to the shared memory is very short, but that to the global memory is very long. Therefore, the most important technique to achieve high performance on GPU is how to cache a part of the data on the shared memory, and to hide the memory access delay to the global memory. The shared and the global memory have the restriction of the access to them. In the CUDA, which is an abstracted architecture of Nvidia’s GPUs, 16 threads are managed as a set. When accessing to the global memory, 16 words can be accessed in parallel if the 16 threads access to continuous 16 words which start from 16 word- boundary. Otherwise, the bank conflict happens, and several accesses to the global memory happens. The shared memory consists of 16 banks, and in this case, the 16 words can be accessed in parallel if they are stored in the different memory banks (the addresses of the 16 words do not need to be continuous). For reducing the memory accesses to the global memory, the order of the calculation on GPU is different from the one described in the previous section. * • Step1 transfer the input images onto the GPU and scale down them * • Step2 compare the brightness of the pixels along the $x$-axis * • Step3 calculate the matching costs and aggregate them along the $x$-axis * • Step4 compare the brightness of the pixels along the $y$-axis * • Step5 aggregate the cost along the $y$-axis, and generate two disparity maps * • Step6 find GCPs by cross-checking * • Step7 apply median filter to remove noises and estimate the disparity map using bilateral method * • Step8 scale up the disparity map and transfer back to the CPU. In the following discussion, $X_{org}\times{Y_{org}}$ is the image size ($X_{org}$ is the width, and $Y_{org}$ is the height), and $L_{org}[y][x]$ and $R_{org}[y][x]$ are the pixels in the left and right images. $L[y][x]$ and $R[y][x]$ are the pixels in the scaled-down images, and $X\times{Y}$ is the image size of them ($X=X_{org}/2$, $Y=Y_{org}/2$). Fig.6 shows the task assignment and input/output of each step. The details are discussed in the following subsections. Figure 6: Task assignment of each step ### IV-A Step1 The inputs to this step are $L_{org}[y][x]$ and $R_{org}[y][x]$, and they are transferred onto the global memory of the GPU, and are processed in parallel using 15 SMs. 1. 1. $Y_{org}/15$ lines of both images are assigned to each SM as shown in Fig.7. 2. 2. $X_{org}$ columns in the $Y_{org}/15$ lines are processed using $X^{\prime}$ ($X_{org}\leq{X^{\prime}}$) threads in the SM ($X^{\prime}$ must be a multiple of 64 because of the reason described below). When $X^{\prime}>X_{org}$, $X^{\prime}-X_{org}$ threads work in the same way as the $X_{org}$ threads, but generate no outputs. 3. 3. When $X_{org}$ is larger than the maximum number of the threads in one SM (1024), each thread processes more than one columns. In our implementation, because the resolution of the input images are greater than 1024, each thread processes 2 columns during the scaling-down step. For each pixel $L_{org}(x_{even},y_{even})$, both of the vertical and horizontal coordinates of which are even, all of the surrounding pixels $L_{org}(x_{even}+dx,y_{even}+dy)$ ($dx\in[-1,1]$, $dy\in[-1,1]$) are added together. Then, as the pixel value of the scaled-image, the average of the summation is stored in the global memory. Figure 7: Mapping pixels to the threads ### IV-B Step2 For $Y/15$ pixels in one column (let the pixels be $L[y_{b}+k][x_{b}]$($k=0,14$)), each thread compares its pixel’s value with its neighbors along the $x$-axis ($L[y_{b}+k][x_{b}+dx]$ ($dx=1,W_{x}$) and $L[y_{b}+k][x_{b}-dx]$ ($dx=1,W_{x}$)). The data type of $L[y][x]$ is $8b$ (unsigned char). Therefore, four continuous pixels are packed in one $32b$ word, and stored in the same memory bank of the shared memory. This means that these four continuous pixels can not be accessed in parallel owing to the memory access restriction of the shared memory. In order to avoid the bank conflict, Threadi processes column ($i/(X/4)+(i\times{4})\%X$) as shown in Fig.7. In Fig.7, the first four pixels of each line ($L[y][0]$, $L[y][1]$, $L[y][2]$, $L[y][3]$) are stored in the first bank, and next four pixels ($L[y][4]$, $L[y][5]$, $L[y][6]$, $L[y][7]$) in the second bank. Thread0 processes $Y/15$ pixels on $x=0$ ($L[y_{b}+dy][0]$ ($dy=0,14$)) sequentially, and Thread1 processes $Y/15$ pixels on $x=4$ ($L[y_{b}+dy][4]$ ($dy=0,14$)) sequentially. By changing the order of the computation like this, bank conflict can be avoided. In our algorithm, all pixels can be processed independently, and the same results can be obtained regardless of the computation order. In this method, four continuous pixels are stored in the same bank, and 16 threads are executed at the same time in CUDA. Therefore, $X$ must be a multiple of 64 ($4\times{16}$). When, $X$ is not the multiple of 64, larger $X$ which is the multiple of 64 is chosen, and the computation results for the extended part are discarded. The outputs of this step are two integer values for each pixel, which show how many pixels are similar to the center pixel to the plus/minus direction of the $x$-axis. These values for $L[y][x]$ are stored in $W^{L}_{-}[y][x]$ and $W^{L}_{+}[y][x]$, and those for $R[y][x]$ are stored in $W^{R}_{-}[y][x]$ and $W^{R}_{+}[y][x]$. The data width of these arrays is $32b$, and the direct access to these values causes no bank conflict. $L[y][x]$ and $R[y][x]$ are transposed here, and stored in $L^{*}[x][y]$ and $R^{*}[x][y]$ respectively. ### IV-C Step3 In this step, first, two matching costs ($C^{L}(x,y,d)$ and $C^{R}(x,y,d)$) are calculated, and then, they are aggregated along the $x$-axis using the range information in $W^{L}_{-}[y][x]$, $W^{L}_{+}L[y][x]$, $W^{R}_{-}[y][x]$ and $W^{R}_{+}[y][x]$ to calculate $CA^{L}_{x}(x,y,d)$ and $CA^{R}_{x}(x,y,d)$. Here, actually, we do not need to calculate $C^{R}(x,y,d)$ as described in Section 3.C because $C^{L}(x+d,y,d)$ can be used as $C^{R}(x,y,d)$. Therefore, all SMs are used to calculate $C^{L}(x,y,d)$ as shown in Fig.6-$step3$, and each SM processes $Y/15$ lines as follows. 1. 1. For each of the Y/15 lines, repeat the following steps. 2. 2. Set $d=0$. 3. 3. Calculate $C^{L}(x,y,d)$ for all $x$ in the current line. $C^{L}(x,y,d)$ is stored in $C[x]$ (an array in the shared memory). For this calculation, 3 lines of $L[y][x]$ and $R[y][x]$ are cached in the shared memory for calculating the mini-census transform, and they are gradually replaced by the next line as the calculation progresses. 4. 4. Calculate $CA^{L}_{x}(x,y,d)$ as follows. 1. (a) Set $CA_{x}[x]=C[x]$. 2. (b) Add $C[x+dx]$ to $CA_{x}[x]$ starting from $dx=1$ to the position given by $W^{L}_{+}[y][x]$. 3. (c) Add $C[x-dx]$ to $CA_{x}[x]$ starting from $dx=1$ to the position given by $W^{L}_{-}[y][x]$. 4. (d) Store $CA_{x}[x]$ into $CA^{*L}_{x}[d][x][y]$ in the global memory (note that this array is transposed). 5. 5. Calculate $CA^{R}_{x}(x,y,d)$ as follows. 1. (a) Set $CA_{x}[x]=C[x+d]$. 2. (b) Add $C[x+d+dx]$ to $CA_{x}[x]$ starting from $dx=1$ to the position given by $W^{R}_{+}[y][x]$. 3. (c) Add $C[x+d-dx]$ to $CA_{x}[x]$ starting from $dx=1$ to the position given by $W^{R}_{-}[y][x]$. 4. (d) Store $CA_{x}[x]$ into $CA^{*R}_{x}[d][x][y]$ in the global memory (note that this array is transposed). 6. 6. Increment $d$ if $d<D$, and go to step 3. In this step, $D$ arrays are stored in the global memory. ### IV-D Step4 $L[y][x]$ and $R[y][x]$ have been transposed and stored as $L^{*}[x][y]$ and $R^{*}[x][y]$ in step2. By using these arrays, the brightness of the pixels are compared efficiently along the $y$-axis. In this case, the pixel data (for example $L^{*}[x][y]$) are compared horizontally (parallel memory accesses are allowed only in this direction), and this means that $L^{*}[x][y]$ are compared with $L^{*}[x][y+dy]$ ($dy=-W_{y},W_{y}$). The range of the similar pixels are stored in $W^{*L}_{-}[x][y]$ and $W^{*L}_{+}[x][y]$ for $L^{*}[x][y]$, and in $W^{*R}_{-}[x][y]$ and $W^{*R}_{+}[x][y]$ for $R^{*}[x][y]$. $L^{*}[x][y]$ and $R^{*}[x][y]$ are processed in parallel as shown in Fig.6-$step4$, and each SM processes $X/15$ columns. The $Y$ lines in each column are assigned to $Y$ threads in the same way shown in Fig.7 though $x$ and $y$ are transposed. ### IV-E Step5 In this step (Fig.6-$step5$), $CA^{*L}_{x}[d][x][y]$ and $CA^{*R}_{x}[d][x][y]$ are aggregated along the $y$-axis in parallel using $W^{*L}_{-}[x][y]$, $W^{*L}_{+}[x][y]$, $W^{*R}_{-}[x][y]$ and $W^{*R}_{+}[x][y]$, and then $D^{L}_{map}[y][x]$ and $D^{R}_{map}[y][x]$ (disparity maps when $L[y][x]$ and $R[y][x]$ are used as the base) are also generated as follows. 1. 1. Each SM processes $X/15$ columns. 2. 2. $Y$ threads in each SM processes $Y$ pixels in each of the $X/15$ columns. 3. 3. Each thread repeats the following steps (in the following, only the steps for the left image are shown). 1. (a) Read $W^{*L}_{-}[x][y]$ and $W^{*L}_{+}[x][y]$ from the global memory. 2. (b) Set $d=0$. 3. (c) $Min[y]=$MAX_VALUE and $D_{map}[y]=0$. 4. (d) Calculate $CA^{L}(x,y,d)$ as follow. 1. i. Set $CA[y]=CA^{*L}[d][x][y]$. 2. ii. Add $CA^{*L}_{x}[d][x][y+dy]$ to $CA[y]$ starting from $dy=1$ to the position given by $W^{*L}_{+}[x][y]$. 3. iii. Add $CA^{*L}_{x}[d][x][y-dy]$ to $CA[y]$ starting from $dy=1$ to the position given by $W^{*L}_{-}[x][y]$. 5. (e) If $CA[y]<Min[y]$ then $Min[y]=CA[y]$ and $D_{map}[y]=d$. 6. (f) Increment $d$ if $d<D$, and go to step 3(c). 7. (g) Store $D_{map}[y]$ in $D^{L}_{map}[y][x]$ in the global memory (note that his array is re-transposed). ### IV-F Step6 In this step (Fig.6-$step6$), $D^{L}_{map}[y][x]$ and $D^{R}_{map}[y][x]$ are read from the global memory line by line, and the condition for the GCP (described in Section 3.D) is checked. In this step, each SM processes $Y/15$ lines. Threadx first accesses $D^{L}_{map}[y][x]$, and then $D^{R}_{map}[y][+k]$ if $D^{L}_{map}[y][x]=k$. $k$ is different for each thread, and bank conflict happens in this step. ### IV-G Step7 The median filter is applied using 15 SMs for the left image at first. Then, the bilateral estimation is used to fill the non-GCPs. If $D^{L}_{map}[y][x]$ is not a GCP, $thread_{x}$ scans $D^{L}_{map}[y][x]$ to the $+/-$ direction of the $x$-axis in order, and finds two GCPs (the GCPs closest on the left- and right-hand side). Then, the difference of the disparities of the two GCPs is calculated. If the difference is smaller than the threshold, the $D^{L}_{map}[y][x]$ is filled linearly according to the disparities of the two GCPs. If the different is larger than the threshold, the brightness of the two GCP are compared with the target pixel, and the disparity of the target pixel is replaced by the one which has similar brightness. Then, the improved disparity map $D^{+L}_{map}[y][x]$ is stored in the global memory. ### IV-H Step8 Finally, the disparity map $D^{+L}_{map}[y][x]$ is scaled up by using the 15 SMs. In order to maintain a high accuracy, during the scaling-up along the $x$-axis, the bilateral estimation method described above is used again. On the other hand, the estimation along the $y$-axis is applied linearly. Then, the final disparity map $D^{fL_{org}}[y][x]$ is transferred back to the CPU. TABLE I: Error rate when the cost aggregation range is changed (average error rate (%)) $W_{L}$$\backslash$$W_{R}$ | $W_{y}=9$ | $W_{y}=11$ | $W_{y}=15$ | $W_{y}=21$ | $W_{y}=27$ | $W_{y}=31$ ---|---|---|---|---|---|--- $W_{x}=5$ | 25.10 | 24.98 | 24.83 | 24.68 | 24.62 | 24.61 $W_{x}=9$ | 24.67 | 24.55 | 24.4 | 24.3 | 24.26 | 24.26 $W_{x}=21$ | 24.39 | 24.34 | 24.21 | 24.13 | 24.1 | 24.09 $W_{x}=41$ | 24.53 | 24.38 | 24.29 | 24.21 | 24.19 | 24.17 $W_{x}=61$ | 24.53 | 24.5 | 24.41 | 24.32 | 24.31 | 24.28 $W_{x}=141$ | 24.6 | 24.55 | 24.48 | 24.42 | 24.38 | 24.28 TABLE II: Execution Time For The Middlebury Benchmark Set (ms) Image | Size | Dmax | SD | $W^{LR}_{\pm}$ | $W^{*LR}_{\pm}$ | $C+CA_{x}$ | CA | CC | Post | SU | Overall(GPU) | FPS ---|---|---|---|---|---|---|---|---|---|---|---|--- Adirondack(H) | $1436\times{992}$ | 145 | 0.035 | 0.149 | 0.317 | 10.443 | 13.943 | 0.437 | 0.181 | 0.09 | 25.595 | 40 Pipes(H) | $1482\times{994}$ | 128 | 0.038 | 0.135 | 0.289 | 19.8 | 11.44 | 0.506 | 0.238 | 0.09 | 32.536 | 31 Vintage(H) | $1444\times{960}$ | 380 | 0.038 | 0.143 | 0.315 | 26.142 | 36.394 | 0.092 | 0.221 | 0.087 | 63.432 | 16 Size: $W^{LR}_{\pm}=21$, $W^{*LR}_{\pm}=31$, $TC=13$ Dmax: Maximum Disparity SD: Scaling-Down. $W^{LR}_{\pm}$: Edge detection along the $x$-axis. $W^{*LR}_{\pm}$: Edge detection along the $y$-axis $C+CA_{x}$: cost calculation & Aggregation along the $x$-axis. CA: Aggregation along the $y$-axis. CC: Cross_check. Post: MedianFilter&Bilateral estimation. SU: Scaling-up. Overall: The overall time taken on GPU. TABLE III: COMPARISON WITH HIGH-SPEED STEREO VISION SYSTEMS System | Size | Dmax | Hardware | Benchmark | FPS | MDE/s ---|---|---|---|---|---|--- RT-FPGA[1] | $1920\times{1680}$ | 60 | Kintex 7 | Middlebury v2 | 30 | 5806 FUZZY[2] | $1280\times{1024}$ | 15 | Cyclone II | Middlebury v2 | 76 | 1494 Low-Power[13] | $1024\times{768}$ | 64 | Virtex-7 | Middlebury v2 | 30 | 1510 ETE[3] | $1242\times{375}$ | 256 | GTX TITAN X | KITTI 2015 | 29 | 3458 EmbeddedRT[8] | $640\times{480}$ | 128 | Tegra X1 | KITTI 2012 | 81 | 3185 MassP[4] | $1440\times{720}$ | 128 | GPU | Middlebury v3 | 128 | 3981 Our system | $1436\times{992}$ | 145 | GTX 780 Ti | Middlebury v3 | 40 | 7849 Figure 8: Processing results ## V experimental results We have implemented the algorithm on Nvidia GTX780Ti. The error rate and the processing speed are evaluated using Middlebury benchmark set [11]. In this evaluation, all parameters mentioned above affect the performance of the accuracy. According to our tuning results, we first set $\lambda_{AD}=0.3$, $\lambda_{MC}=2.3$ and $T=3$ to ensure a good accuracy. In the cost aggregation step, lower error rates can be expected by adding more cost along the $x$\- and $y$\- axes, though it requires more computation time, and makes the system slower. The maximum range of the cost aggregation $(W_{x},W_{y})$ can be changed when calculating $D^{L}_{map}$ and $D^{R}_{map}$. By changing them, the different criteria are used for the left and right image, and the GCPs can be more reliable. Table I shows the error rate (%) when the cost aggregation range is changed. In Table I, $W_{x}$ are the maximum aggregation range along the $x$-axis for the left and right images, and $W_{y}$ is the maximum aggregation range along the $y$-axis ($W_{y}$ is common to the left and right images). As shown in Table I, by enlarging $W_{y}$, the error rates can be improved when $W_{x}$ is small. We have fixed $W_{x}=21$ and $W_{y}=31$. To our best knowledge, the accuracy of our system is higher than other real-time systems (like [12]) which are listed in Middlebury Benchmark [11]. Additionally, we also compared our error rates with that obtained using the original size image set, which are processed using larger window ranges $W_{x}=41$ and $W_{y}=61$. According to our evaluation, the error rate (Bad 2.0) of our system (24.09%) is higher than that by the original size images (32.82%). One of the reasons is that we didn’t tuned the parameters $\lambda_{AD}$ and $\lambda_{MC}$ for the original image set. The other one is that in the scaling down images, some information that makes the matching difficult in the original size images, such as repetition of patterns and a serious of similar pixels, are discarded, and better matching becomes possible. The processing speed of our system is almost proportional to the window size and the maximum disparity. Therefore, the computation time for ’Vintage’ becomes the slowest. Table II shows the processing speed of our system and its details. We ignore the time for CPU-GPU data transfers (less than 3% of the total elapsed time) since it can be overlapped with the computation. As shown in Table II, most of the computation time is used for the cost calculation and its aggregation ($C+CA_{x}$ and $CA$). $W^{LR}_{\pm}$, $W^{*LR}_{\pm}$, $CC$, $Post$, $SD$, $SU$ show the computation time for finding the cost aggregation range along the $x$\- and $y$\- axes, cross checking, post-processing and the image scaling. Unfortunately, for the ’Vintage’ set, its processing speed is 16fps due to the large disparity, and we cannot achieve the real-time processing. Table III compares the processing speed of our system with other hardware systems. In Table III, all of the systems achieved a real-time processing as shown in FPS field, but their target image size (Size) and disparity range (Dmax) are different. According to the mega disparity evaluation per second (MDE/S), it can be noted that our system is much faster than other systems. Fig.7 shows the results of our system for the two benchmark sets: Adirondack, Pipes and Vintage. ## VI Conclusion In this paper, we have proposed a real-time stereo vision system for high resolution images on GPU. Its processing speed is much faster than previous one. At the same time, it also maintained a high accuracy. In our current implementation, the processing speed is still limited by the access delay to the global memory. To improve the processing speed so that the real-time processing can be achieved slower GPUs is our main future work. ## References * [1] Daolu Zha, Xi Jin, Tian Xiang: _A real-time global stereo-matching on FPGA_. Microprocessors and Microsystems - Embedded Hardware Design 47: 419-428 (2016) * [2] Madaín Pérez Patricio and Abiel Aguilar-González and Miguel O. Arias-Estrada and Héctor-Ricardo Hernandez-de Leon and Jorge-Luis Camas-Anzueto and J. A. de Jesús Osuna-Coutiño: _An FPGA stereo matching unit based on fuzzy logic_. Microprocessors and Microsystems - Embedded Hardware Design 42: 87-99 (2016) * [3] Alex Kendall, Hayk Martirosyan, Saumitro Dasgupta, Peter Henry, Ryan Kennedy, Abraham Bachrach, Adam Bry: _End-to-End Learning of Geometry and Context for Deep Stereo Regression_. CoRR abs/1703.04309 (2017) * [4] Wenbao Qiao and Jean-Charles Créput: _Stereo Matching by Using Self-distributed Segmentation and Massively Parallel GPU Computing_. ICAISC (2) 2016: 723-733 * [5] Jure Zbontar, Yann LeCun: _Stereo Matching by Training a Convolutional Neural Network to Compare Image Patches_. Journal of Machine Learning Research 17: 65:1-65:32 (2016) * [6] Xiaoqing Ye, Jiamao Li, Han Wang, Hexiao Huang, Xiaolin Zhang: _Efficient Stereo Matching Leveraging Deep Local and Context Information_. IEEE Access 5: 18745-18755 (2017) * [7] Aaron F. Bobick, Stephen S: _Intille: Large Occlusion Stereo_. International Journal of Computer Vision 33(3): 181-200 (1999) * [8] Daniel Hernández Juárez and Alejandro Chacón and Antonio Espinosa and David Vázquez and Juan Carlos Moure and Antonio M. López: _Embedded Real-time Stereo Estimation via Semi-Global Matching on the GPU_. ICCS 2016: 143-153 * [9] Minxi Jin, Tsutomu Maruyama: _Fast and Accurate Stereo Vision System on FPGA_. TRETS 7(1): 3:1-3:24 (2014) * [10] Qingxiong Yang and Liang Wang and Ruigang Yang and Henrik Stewénius and David Nistér: _Stereo Matching with Color-Weighted Correlation, Hierarchical Belief Propagation, and Occlusion Handling_. IEEE Trans. Pattern Anal. Mach. Intell. 31(3): 492-504 (2009) * [11] D.Scharstein and R.Szeliski: _http://vision.middlebury.edu/stereo/eval3/_. * [12] Leonid Keselman, John Iselin Woodfill, Anders Grunnet-Jepsen, Achintya Bhowmik: _Intel RealSense Stereoscopic Depth Cameras_. CoRR abs/1705.05548 (2017) * [13] Luca Puglia, Mario Vigliar, Giancarlo Raiconi: _Real-Time Low-Power FPGA Architecture for Stereo Vision_. IEEE Trans. on Circuits and Systems 64-II(11): 1307-1311 (2017) * [14] Lu Zhang, Ke Zhang, Tian Sheuan Chang, Gauthier Lafruit, Georgi Krasimirov Kuzmanov, Diederik Verkest: _Real-time high-definition stereo matching on FPGA._ FPGA 2011: 55-64 * [15] Ke Zhang, Jiangbo Lu, Gauthier Lafruit: _Cross-Based Local Stereo Matching Using Orthogonal Integral Images._ IEEE Trans. Circuits Syst. Video Techn. 19(7): 1073-1079 (2009) * [16] Edoardo Paone, Gianluca Palermo, Vittorio Zaccaria, Cristina Silvano, Diego Melpignano, Germain Haugou, Thierry Lepley: _An exploration methodology for a customizable OpenCL stereo-matching application targeted to an industrial multi-cluster architecture._ CODES+ISSS 2012: 503-512
# Principled Multi-Aspect Evaluation Measures of Rankings Maria Maistro<EMAIL_ADDRESS>University of CopenhagenDenmark , Lucas Chaves Lima<EMAIL_ADDRESS>University of CopenhagenDenmark , Jakob Grue Simonsen <EMAIL_ADDRESS>University of CopenhagenDenmark and Christina Lioma <EMAIL_ADDRESS>University of CopenhagenDenmark (2021) ###### Abstract. Information Retrieval evaluation has traditionally focused on defining principled ways of assessing the relevance of a ranked list of documents with respect to a query. Several methods extend this type of evaluation beyond relevance, making it possible to evaluate different aspects of a document ranking (e.g., relevance, usefulness, or credibility) using a single measure (multi-aspect evaluation). However, these methods either are (i) tailor-made for specific aspects and do not extend to other types or numbers of aspects, or (ii) have theoretical anomalies, e.g. assign maximum score to a ranking where all documents are labelled with the lowest grade with respect to all aspects (e.g., not relevant, not credible, etc.). We present a theoretically principled multi-aspect evaluation method that can be used for any number, and any type, of aspects. A thorough empirical evaluation using up to $5$ aspects and a total of $425$ runs officially submitted to $10$ TREC tracks shows that our method is more discriminative than the state-of-the-art and overcomes theoretical limitations of the state- of-the-art. Evaluation, ranking, multiple aspects, partial order ††journalyear: 2021††copyright: none††conference: Proceedings of the 30th ACM International Conference on Information and Knowledge Management; November 1–5, 2021; Virtual Event, QLD, Australia††booktitle: Proceedings of the 30th ACM International Conference on Information and Knowledge Management (CIKM ’21), November 1–5, 2021, Virtual Event, QLD, Australia††doi: 10.1145/3459637.3482287††isbn: 978-1-4503-8446-9/21/11††ccs: Information systems Information retrieval††ccs: Information systems Evaluation of retrieval results††ccs: Information systems Retrieval effectiveness ## 1\. Introduction Multi-aspect evaluation is a task in Information Retrieval (IR) evaluation where the ranked list of documents returned by an IR system in response to a query is assessed in terms of not only relevance, but also other aspects (or dimensions) such as credibility or usefulness. Generally, there are two ways to conduct multi-aspect evaluation: (1) evaluate each aspect separately using any appropriate single-aspect evaluation measure (e.g., AP, NDCG, F1), and then aggregate the scores across all aspects into a single score; or (2) evaluate all aspects at the same time using any appropriate multi-aspect evaluation measure (Amigó et al., 2018b; Lioma et al., 2017; Tang and Yang, 2017). An advantage of the aggregating option (1) is that it is easy to implement using evaluation measures that are readily available and well- understood in the community. Its disadvantage is that it is not guaranteed that all aspects will have similar distributions of labels, and aggregating across wildly different distributions can give odd results (Lioma et al., 2019). The second way of doing multi-aspect evaluation is to use a single multi-aspect evaluation measure. The problem here is that few such evaluation measures exist, and most of them are defined for specific aspects and do not generalise to other types/numbers of aspects (see $\S$2). Motivated by the above, we contribute a novel multi-aspect evaluation method that works with any type and number of aspects, and avoids the above problems. Given a ranked list, where documents are labelled with multiple aspects, our method, Total Order Multi-Aspect (TOMA) evaluation, first defines a preferential order (formally weak order relation) among documents with multiple aspect labels, and then aggregates the document labels across aspects to obtain a ranking of aggregated aspect labels, which can be evaluated by any single-aspect evaluation measure, such as Normalized Discounted Cumulated Gain (NDCG) or Average Precision (AP). Simply put, instead of evaluating each aspect separately and then aggregating their scores, we first aggregate the aspect labels and then evaluate the ranked list of documents. We do this in a way that provides several degrees of freedom: our method can be used with any number and type of aspects, can be instantiated with any binary or graded, set-based or rank-based evaluation measure, and can accommodate any granularity in the importance of each aspect or label, but still ensures, by definition, that the preference order among multi-aspect documents is not violated, and that the final measure score will meet some common requirements, i.e., the minimum (worst) score being $0$ and the maximum (perfect) score being $1$. We validate this empirically ($\S$4) and theoretically ($\S$4.2). ## 2\. Related Work Multi-aspect evaluation measures for IR have been studied for different tasks and aspects, starting from the INEX initiative with relevance and coverage (Kazai et al., 2004). Since then, measures have been proposed to evaluate relevance and novelty or diversity, such as $\alpha$-NDCG (Clarke et al., 2008), MAP-IA (Agrawal et al., 2009) and IA-ERR (Chapelle et al., 2009); relevance, novelty and the amount of user effort, such as nCT (Tang and Yang, 2017); relevance, redundancy and user effort, such as RBU (Amigó et al., 2018b); relevance and understandability, such as uRBP (Zuccon, 2016) and the Multidimensional Measure (MM) framework (Palotti et al., 2018); and relevance and credibility, such as NLRE, NGRE, nWCS, Convex Aggregating Measure (CAM) and WHAM (Lioma et al., 2017). All these measures have limitations; we describe these next. Firstly, except for RBU, none of the above measures are based on a formal framework. They are defined as stand-alone tools to assess the effectiveness of a ranked list of documents. This means that, even if the measure can assess the effectiveness of an input ranking, the order induced by the measure over the space of input rankings is not well-defined. Hence, there is no canonical _ideal ranking_ 111An _ideal ranking_ is the best ranking of all assessed documents for a given topic (Järvelin and Kekäläinen, 2002). that is well- defined or easy to compute, e.g., for $\alpha$-NDCG, the computation of the ideal ranking is equivalent to a minimal vertex covering problem (Clarke et al., 2008), an NP-complete problem, while for CT and nCT, computing the ideal ranking is equivalent to the minimum edge dominating set problem (Tang and Yang, 2017), an NP-hard problem. Computationally better ways of comparing to an ideal ranking can be devised using graded similarity—so-called _effectiveness levels_ to an ideal ranking using Rank-Biased Overlap (Clarke et al., 2020c, b, 2021), but this approach requires defining a (set of) ideal ranking(s), which has not appeared for multi-aspect ranking prior to the present paper. Evaluation measures that do not compare against an ideal ranking may be harder to interpret or problematic. DCG is not upper bounded by $1$, thus different topics are not weighted equally and scores are not comparable. Failing to compare against the ideal ranking is problematic in multi-aspect evaluation: $\alpha$-NDCG allows systems to reach scores greater than $1$, which is supposed to be the score of the perfect system. With NLRE and NGRE, a system that retrieves no relevant or credible documents has error =$0$, i.e., achieves the best score, because the relative order of pairs of documents is always correct (Lioma et al., 2019). Similarly, nWCS can reach the perfect score of $1$, even if no relevant or credible documents are retrieved, since the normalization is computed with a re-ranking of the input ranking, instead of the ideal ranking. Both uRBP and RBU have a different problem: to reach the perfect score of $1$, a system must retrieve an infinite number of relevant and understandable documents, even if those documents are not available in the collection. CAM and WHAM use the weighted arithmetic and weighted harmonic mean of any IR measure computed with respect to relevance and credibility independently. Therefore, depending on the distribution of labels across the aspects, it can be impossible for any system to reach the perfect score (see $\S$ 4.2). Secondly, most of the above multi-aspect evaluation measures are defined for specific contexts and with a limited set of aspects, e.g., novelty, diversity, credibility and understandability, thus they cannot deal with a more general scenario and a variable number of aspects. For RBU, even though a formal framework is defined, its formulation specifies only diversity and redundancy constraints, which cannot be applied to a general set of aspects. This inability to generalise to more/other types of aspects means that, if a system must be evaluated with respect to a new aspect, the measure needs to be properly adapted. This can be easily done for some measures, e.g., CAM, WHAM, and nWCS, but the lack of a formal framework behind them may lead to odd results, e.g., extending NLRE to $3$ aspects returns a score distribution compressed towards $0$, preventing the rankings to be evaluated in a fair way (Lioma et al., 2019). (a) Distance order. (b) Euclidean distance. (c) Manhattan distance. (d) Chebyshev distance. Figure 1. Example with two aspects $a_{1}$ and $a_{2}$. Each point is a tuple of labels. The best label $\mathbf{l^{\star}}$ is in the top right. The distance between tuples of labels and $\mathbf{l^{\star}}$ defines a weak order relation. Blue lines connect tuples of labels at the same distance from $\mathbf{l^{\star}}$. ## 3\. TOMA Framework We formalize the problem and our proposed methodology: we explain why reasoning in terms of multiple aspects leads to a partial order relation among documents ($\S$ 3.1); how we complete the partial order relation with the distance order ($\S$ 3.2); and how to use the distance order with state-of- the-art IR evaluation measures ($\S$ 3.3). ### 3.1. Formalization of the Problem Let $A=\\{a_{1},\ldots,a_{n}\\}$ be a set of _aspects_ ; each aspect $a\in A$ has a non-empty set of _labels_ $L_{a}=\\{l_{0}^{a},\ldots,l_{K_{a}}^{a}\\}$ and an order relation $\prec_{a}$ such that: $l_{0}^{a}\prec_{a}l_{1}^{a}\prec_{a}\cdots\prec_{a}l_{K_{a}}^{a}$, e.g., we may have $2$ aspects $A=\\{\text{relevance},\text{correctness}\\}$, with the set $L_{r}=\\{\texttt{nr},\texttt{mr},\texttt{fr},\texttt{hr}\\}$ (non- relevant, marginally relevant, fairly relevant, highly relevant) ordered as: $\texttt{nr}\prec_{r}\texttt{mr}\prec_{r}\texttt{fr}\prec_{r}\texttt{hr}$; and the set $L_{c}=\\{\texttt{nc},\texttt{pc},\texttt{c}\\}$ (non-correct, partially correct, correct) ordered as: $\texttt{nc}\prec_{c}\texttt{pc}\prec_{c}\texttt{c}$. Let $D$ be the set of _documents_ and $T$ the set of _topics_. Each document $d\in D$ is mapped to a _ground truth_ vector $\textrm{GT}(d,t)=(l_{1},\ldots,l_{n})\in L_{a_{1}}\times\cdots\times L_{a_{n}}$ that contains the “true” label of $d$ for each aspect, e.g., a document may have $\textrm{GT}(d,t)=(\textrm{highly relevant},\textrm{non-correct})$. In IR, given a topic $t$, the objective is to rank documents in $D$ such that for the documents $d^{\prime},d\in D$, if $d^{\prime}$ is ranked before $d$, then $\textrm{GT}(d,t)\preceq_{*}\textrm{GT}(d^{\prime},t)$ for a given order relation $\preceq_{*}$. When there is only one aspect $A=\\{a\\}$, one can use $\prec_{a}$, the order on the set of labels $L_{a}$, to induce a weak order on $D$ and decide if $d^{\prime}$ should be ranked before $d$. If only relevance is assessed, we consider the relation induced by relevance labels, i.e., documents labelled “highly relevant” should be ranked before “fairly/marginally relevant” and “non-relevant” documents. Applying this approach to multiple aspects requires reasoning about orderings of tuples of labels with different aspects, e.g., for documents $d^{\prime},d\in D$, such that $\textrm{GT}(d^{\prime},t)=(\text{highly relevant},\text{correct})$ and $\textrm{GT}(d,t)=($marginally relevant, correct$)$, it is reasonable to rank $d^{\prime}$ before $d$. Indeed, there is one _unequivocal_ way of deeming one document better than another, and this is if document $d^{\prime}$ has better labels than document $d$ for _every_ aspect: if for $\textrm{GT}(d,t)=(l_{1},\ldots,l_{n})$ and $\textrm{GT}(d^{\prime},t)=(l_{1}^{\prime},\ldots,l_{n}^{\prime})$ we have $l_{i}\preceq_{a_{i}}l_{i}^{\prime}$ for all $i\in\\{1,\ldots,n\\}$, then any document labeled $(l_{1}^{\prime},\ldots,l_{n}^{\prime})$ is better or equal than any document labelled $(l_{1},\ldots,l_{n})$ and should occur before it in a “good” ranking. We denote this order relation by $\textrm{GT}(d,t)\sqsubseteq\textrm{GT}(d^{\prime},t)$. The order relation $\sqsubseteq$ leads to a _partial_ instead of a _total_ order, i.e., there are documents that are _not comparable_ 222A partial order is reflexive, antisymmetric and transitive; a total order is a partial order where all items are comparable; a weak order is a total order without antisymmetry (Halmos, 1974)., e.g., if $d^{\prime}$ is now highly relevant and partially correct, the final ranking is not clear: should one promote $d^{\prime}$ (more relevant) or $d$ (more correct)? This is an example of documents that are not comparable, so we have $\textrm{GT}(d,t)\not\sqsubseteq\textrm{GT}(d^{\prime},t)$ and $\textrm{GT}(d^{\prime},t)\not\sqsubseteq\textrm{GT}(d,t)$, and the choice of whether $d^{\prime}$ is preferred to $d$ may lie on the intended application. A partial order relation and the presence of not comparable documents imply that it is not possible to univocally rank the documents in $D$. If we could “complete” the partial order with a total order, or at least a weak order, we could rank documents and define an ideal ranking, where for any $d^{\prime},d\in D$, the order relation determines the rank position of $d^{\prime}$ and $d$. So, before tackling the problem of evaluating a ranked list of documents in a multi-aspect way, we build such an order relation. This is detailed next. ### 3.2. The distance order We now explain how to obtain a weak order relation from the partial order relation $\sqsubseteq$. Consider the Cartesian product of all sets of labels $L=L_{a_{1}}\times\cdots\times L_{a_{n}}$. An element $\mathbf{l}\in L$ is a tuple of labels $\mathbf{l}=(l_{1},\dots,l_{n})$. The total order relation will be denoted by $\preceq_{*}$ and it will be a weak order relation on $L$, i.e., a total binary relation that is reflexive and transitive, but not necessarily anti-symmetric (Ferrante et al., 2015, 2017, 2019b). This weak order allows _all_ tuples of labels to be compared, i.e., for any two $\mathbf{l},\mathbf{l}^{\prime}\in L$ we will have $\mathbf{l}^{\prime}\preceq_{*}\mathbf{l}$ and/or $\mathbf{l}\preceq_{*}\mathbf{l}^{\prime}$. Consequently, all documents will be comparable through their tuple of labels. We require that the weak order relation $\preceq_{*}$ respects the partial order relation $\sqsubseteq$: (1) $\forall\ \mathbf{l},\mathbf{l}^{\prime}\in L\textrm{ we have }\mathbf{l}\sqsubseteq\mathbf{l}^{\prime}\Rightarrow\mathbf{l}\preceq_{*}\mathbf{l}^{\prime}$ This means that, for comparable documents, the partial order relation and the weak order relation rank documents in the same way. Moreover the weak order relation allows to rank even those documents that are not comparable with the partial order relation. To define $\preceq_{*}$, we embed the tuples of labels in the Euclidean space and derive the weak order $\preceq_{*}$ using known distance functions. Let $g$ be an embedding function that maps tuples of labels in Euclidean space $\mathcal{L}=\mathbb{R}^{n}$: $g(\mathbf{l})=g(l_{1},\ldots,l_{n})=(g_{a_{1}}(l_{1}),\ldots,g_{a_{n}}(l_{n}))$. We assume that for each $a\in A$, $g_{a}$ is a non-decreasing map, i.e., for any $l,l^{\prime}\in L_{a}$ if $l\preceq_{a}l^{\prime}$ then $g_{a}(l)\leq g_{a}(l^{\prime})$. Intuitively, $g_{a}$ assigns a number to each label, which allows to represent tuples of labels in the Euclidean space. We illustrate in $\S$3.4 how the embedding function $g$ affects the final ranking of documents. Through the embedding function $g$, each tuple of labels $\mathbf{l}$ is represented by a point in the Euclidean space $\mathcal{L}$ denoted by $\vec{l}=g(\mathbf{l})$. We define the _best label tuple_ as the tuple of labels $\mathbf{l^{\star}}$ whose coordinates are the best label for each aspect, $\mathbf{l^{\star}}=(l_{K_{a_{1}}},\ldots,l_{K_{a_{n}}})$. The idea is to treat $\mathbf{l^{\star}}$ as the maximum element and use the distance from this maximum element to define the desired weak order relation; e.g., for two aspects $a_{1}$ and $a_{2}$, each tuple of labels is represented as a point in the Euclidean plane, and the best label $\mathbf{l^{\star}}$ is represented by the topmost and right-most point (see Fig. 1(a)). Then, given two documents $d$ and $d^{\prime}$, $d$ is ranked before $d^{\prime}$ if $\textrm{GT}(d,t)$ is closer to the best label than $\textrm{GT}(d^{\prime},t)$. We formally define the _distance order_ as the following relation: (2) $\mathbf{l}\preceq_{*}\mathbf{l^{\prime}}\ \iff\ \text{Dist}(\vec{l},\vec{l^{\star}})\geq\text{Dist}(\vec{l^{\prime}},\vec{l^{\star}})$ where $\text{Dist}\colon\mathcal{L}\times\mathcal{L}\to[0,+\infty[$ is any function such that $\text{Dist}(\vec{l}^{\star},\vec{l}^{\star})=0$333Distance functions must be symmetric and satisfy the triangle inequality. Any such distance function satisfies our condition on Dist, and so do our example distances.. The relation $\preceq_{*}$ is a weak order: all $\mathbf{l}$, $\mathbf{l^{\prime}}$ are comparable because $\text{Dist}(\vec{l},\vec{l^{\star}})$ is defined for all $\mathbf{l}$, and as $\geq$ is reflexive and transitive on $[0,+\infty[$, the relation $\preceq_{*}$ is reflexive and transitive (but not necessarily antisymmetric). Since the distance order is a weak order, it allows to deem items “equally good” when it is impossible or undesirable to impose a strict total order444This is the reason that _weak_ orders (that are not necessarily anti- symmetric), rather than strict total orders, are typically used in the literature (Yao, 1995; Ferrante et al., 2019b).. Thus we write: (3) $\mathbf{l}=_{*}\mathbf{l^{\prime}}\ \iff\ \text{Dist}(\vec{l},\vec{l^{\star}})=\text{Dist}(\vec{l^{\prime}},\vec{l^{\star}})$ which means that $\vec{l}$ and $\vec{l^{\prime}}$ are at the same distance from $\vec{l}^{\star}$. Note that the distance order can be _tailored_ : we may instantiate Dist with any valid distance function. We illustrate this in Fig. 1(b)-1(d) with _Euclidean_ (order relation $\preceq_{2}$), _Manhattan_ (order relation $\preceq_{1}$), and _Chebyshev_ (order relation $\preceq_{\infty}$). With these choices of Dist, the distance order defined in Eq. (2)-(3) respects the partial order $\sqsubseteq$, which means that it satisfies the requirement in Eq. (1) because $g_{a}$ is a non decreasing map (a proof is provided in the online appendix555https://github.com/lcschv/TOMA/blob/1d562036c50f7ff0a6df00246195098d7282b1ac/CIKM2021_appendix.pdf). ### 3.3. Integration with IR measures Next we integrate the distance order with known IR measures such as AP or NDCG. The binary relation $=_{*}$ in Eq. (3) is an equivalence relation. Given a tuple of labels $\mathbf{l}\in L$, its equivalence class $[\mathbf{l}]_{*}$ is the set of all tuples of labels with equal distance from the best label $[\mathbf{l}]_{*}=\\{\mathbf{l^{\prime}}\in L\colon\text{Dist}(\vec{l^{\prime}},\vec{l^{\star}})=\text{Dist}(\vec{l},\vec{l^{\star}})\\}$. Inducing the relation defined in Eq. (2) on the set of documents $D$ allows to rank documents by their membership to each equivalence class, which corresponds to the distance of their tuple of labels to the best label. We place closest to the top of the ranking documents whose equivalence class is closest to the best label, and vice versa. To combine the distance order with IR measures we map each equivalence class (set of tuple of labels), to a non negative integer. This is similar to what happens with single-aspect evaluation, where each label is mapped to a weight: e.g., with $4$ relevance labels, we can compute NDCG with equi-spaced relevance weights $\\{0,1,2,3\\}$ (Järvelin and Kekäläinen, 2002) or exponential weights $\\{0,2,4,8\\}$ (Burges et al., 2005). We define a _weight function_ $W\colon L\to\mathbb{N}_{0}^{+}$ as a map such that the order relation $\preceq_{*}$ is preserved: (4) $\forall\ \mathbf{l},\mathbf{l^{\prime}}\in L\colon\mathbf{l}\preceq_{*}\mathbf{l^{\prime}}\implies W(\mathbf{l})\leq W(\mathbf{l^{\prime}})$ where the constraint in Eq. (4) entails that $W$ is a non-decreasing function with respect to the weak order $\preceq_{*}$ on the set of tuples of labels. This means that $W$ can return different integers for each equivalence class, but also the same integer for different equivalence classes, i.e., $0$ and $1$, whenever we need to compute a binary single-aspect IR measure as AP. To summarize, our TOMA method has $3$ steps: 1. (1) We embed tuples of labels into elements of Euclidean space, and we derive the weak order $\preceq_{*}$ using a distance function; 2. (2) We define an adjustable weight function $W$ that preserves $\preceq_{*}$ and maps each tuple of labels to a single integer weight (this allows to aggregate tuple of labels so that better documents can be given greater weight); 3. (3) Having such a weak order and the weight function $W$, any existing single- aspect IR evaluation measure can be used to assess the quality. Thus, we choose a single-aspect evaluation measure $\mu$ and compute the final evaluation score as $M=\mu\circ W\colon M(r_{t})=\mu(W(\textrm{GT}(d_{1},t)),\ldots,W(\textrm{GT}(d_{N},t)))$, where $r_{t}$ is a ranked list of documents. The above is compatible with any number and type of aspect. ### 3.4. Example Table 1. Final ordering of tuples of labels embedded in the Euclidean space. Relevance labels are always embedded in the same mapping (under Relevance). We use different mappings for correctness labels (under Correctness). Tuples that are relevant and not correct (high-traffic fake news) are in red. Relevance | Correctness | Distance | Order among Tuples of Labels ---|---|---|--- $\\{0,1,2,3\\}$ | $\\{0,3/2,3\\}$ | Euclidean | $(3,3)\preceq_{*}(2,3)\preceq_{*}(3,3/2)\preceq_{*}(2,3/2)\preceq_{*}(1,3)\preceq_{*}(1,3/2)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(3,0)}}\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(2,0)}}\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(1,0)}}\preceq_{*}(0,0)$ Manhattan | $(3,3)\preceq_{*}(2,3)\preceq_{*}(3,3/2)\preceq_{*}(1,3)\preceq_{*}(2,3/2)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(3,0)}}\preceq_{*}(1,3/2)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(2,0)}}\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(1,0)}}\preceq_{*}(0,0)$ Chebyshev | $(3,3)\preceq_{*}(2,3)\preceq_{*}(3,3/2)=_{*}(2,3/2)\preceq_{*}(1,3)=_{*}(1,3/2)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(3,0)}}=_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(2,0)}}=_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(1,0)}}=_{*}(0,0)$ $\\{0,1,2,3\\}$ | $\\{0,1,2\\}$ | Euclidean | $(3,2)\preceq_{*}(3,1)=_{*}(2,2)\preceq_{*}(2,1)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(3,0)}}=_{*}(1,2)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(2,0)}}=_{*}(1,1)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(1,0)}}\preceq_{*}(0,0)$ Manhattan | $(3,2)\preceq_{*}(3,1)=_{*}(2,2)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(3,0)}}=_{*}(2,1)=_{*}(1,2)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(2,0)}}=_{*}(1,1)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(1,0)}}\preceq_{*}(0,0)$ Chebyshev | $(3,2)\preceq_{*}(3,1)=_{*}(2,1)=_{*}(2,2)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(3,0)}}=_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(2,0)}}=_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(1,0)}}=_{*}(1,1)=_{*}(1,2)\preceq_{*}(0,0)$ $\\{0,1,2,3\\}$ | $\\{0,2,6\\}$ | Euclidean | $(3,6)\preceq_{*}(2,6)\preceq_{*}(1,6)\preceq_{*}(3,2)\preceq_{*}(2,2)\preceq_{*}(1,2)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(3,0)}}\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(2,0)}}\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(1,0)}}\preceq_{*}(0,0)$ Manhattan | $(3,6)\preceq_{*}(2,6)\preceq_{*}(1,6)\preceq_{*}(3,2)\preceq_{*}(2,2)\preceq_{*}(1,2)=_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(3,0)}}\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(2,0)}}\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(1,0)}}\preceq_{*}(0,0)$ Chebyshev | $(3,6)\preceq_{*}(2,6)\preceq_{*}(1,6)\preceq_{*}(3,2)=_{*}(2,2)=_{*}(1,2)\preceq_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(3,0)}}=_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(2,0)}}=_{*}\mathbin{{\color[rgb]{0.86328125,0.078125,0.234375}\definecolor[named]{pgfstrokecolor}{rgb}{0.86328125,0.078125,0.234375}(1,0)}}=_{*}(0,0)$ We present an example on the role of different choices of embedding, distance, and weight functions in TOMA with $4$ relevance labels $\\{\texttt{nr},\texttt{mr},\texttt{fr},\texttt{hr}\\}$ and $3$ correctness labels $\\{\texttt{nc},\texttt{pc},\texttt{c}\\}$. As in real scenarios (Lioma et al., 2019), we assume that not relevant documents are not correct: as they do not include information about the topic, they cannot be correct with respect to that topic. Tab. 1 shows $3$ different embeddings for correctness; the embedding for relevance is fixed. Note that the distance functions are invariant under translations and rotations, thus, rather than the actual values assigned from the embedding function $g$, it is important to consider the relation between different aspects. Independently of the choice of the embedding function and due to the definition of the selected distance functions, we see that: (i) Chebyshev generates the least number of equivalence classes and deems many tuples of labels as equal, since by taking the maximum it considers just the “furthest” or worst aspect to compute the distance; (ii) Manhatthan is somehow in-between Chebyshev and Euclidean and generates the equivalence classes by taking the sum across aspects; (iii) Euclidean generates the highest number of equivalence classes as it differentiates among tuples more than Manhatthan and is more sensitive to extreme cases, e.g., cases where one aspect has the best label and all other aspects have the lowest label. In the 1st scenario of Tab. 1 we map relevance and correctness to the same interval $[0,3]$ (i.e., a highly relevant document is as “important” as a correct document). All labels are equi-spaced in the given range (the difference between a fairly relevant and a marginally relevant document is the same as that between a highly relevant and a fairly relevant one). With the Euclidean distance all relevant and not correct documents will be deemed worse than all other documents, but will be placed before not relevant and not correct documents. On the other hand, Chebyshev places relevant and not correct documents in the same equivalence class as not relevant documents, so those documents do not provide any contribution and can be simply filtered out. Manhattan is a middle solution: highly relevant and not correct documents are deemed better than marginally relevant and partially correct documents, but worse than all other correct or partially correct documents. In the 2nd scenario of Tab. 1 relevance and correctness are mapped to different ranges, but all labels are equi-spaced with the same step of size $1$. Here, relevance is more important than correctness. This is reflected on the sorting of equivalence classes: for all distance functions, highly relevant and not correct documents do not belong to the worst equivalence classes, but they are somehow better than partially correct documents. Even Chebyshev, which can be seen as the “strictest” distance function, places all relevant and not correct documents in the same class, which is considered better than the class of not relevant and not correct documents. In the 3rd scenario of Tab. 1 correctness is mapped to a range twice the size as the relevance range and we do not use equi-spaced labels for correctness. We assign more importance to correctness than relevance, and among correctness labels we penalize not correct and partially correct documents. The result is that for all distance functions relevant and not correct documents are considered among the worst equivalence classes. This particular setting affects also the other equivalence classes: correctness is preferred over relevance, e.g., correct documents should be always ranked before partially correct documents, regardless of their relevance label. Note that TOMA requires a weight function satisfying the requirement in Eq. (4). If we wish to reward a system for sorting documents exactly as presented by the equivalence classes in Tab. 1, then the weight function should assign a different integer to each equivalence class. This choice of weight is similar to the choice of weights for relevance labels and its impact on the evaluation outcome is strictly tight to the evaluation measure used, as for example when one considers NDCG with different weighting schemes (Järvelin and Kekäläinen, 2002). Table 2. Experimental data. All aspects are labelled by TREC except popularity ($\dagger$ approximated by PageRank) and non-spamminess ($\ddagger$ approximated by Waterloo Spam Ranking). * means that the junk labels are merged with non relevant. | TREC tracks ---|--- | Web 2009 | Web 2010 | Web 2011 | Web 2012 | Web 2013 | Web 2014 | Task 2015 | Task 2016 | Decision 2019 | Misinfo2020 Collection | ClueWeb09 | ClueWeb12 | ClueWeb12-B13 | CommonCrawl News Topics | 50 | 48 | 50 | 50 | 50 | 50 | 35 | 50 | 50 | 46 Submitted runs | 71 | 56 | 61 | 48 | 61 | 30 | 6 | 9 | 32 | 51 | relevance (4) | relevance (5*) | relevance (4*) | relevance (5*) | relevance (3*) | relevance (3) | relevance (2) Aspects | popularity$\dagger$ (3) | popularity$\dagger$ (3) | popularity$\dagger$ (3) | popularity$\dagger$ (3) | usefulness (3) | credibility (2) | credibility (2) (label grades) | non-spam$\ddagger$ (3) | non-spam$\ddagger$ (3) | non-spam$\ddagger$ (3) | non-spam$\ddagger$ (3) | popularity$\dagger$ (3) | correctness (2) | correctness (2) | | | | | non-spam$\ddagger$ (3) | | ## 4\. Experimental Evaluation We evaluate TOMA on $425$ rankings that were submitted as official runs to $10$ TREC tracks (Voorhees and Ellis, 2016; Yilmaz et al., 2016; Lioma et al., 2019; Clarke et al., 2009; Clarke and Cormack, 2011; Clarke et al., 2011; Clarke et al., 2012; Collins-Thompson et al., 2014, 2015; Clarke et al., 2020a) (see Tab. 2). ### 4.1. Experimental Setup We use up to $5$ different aspects. All aspects are assessed by TREC assessors as part of the corresponding track, except popularity and non-spamminess. We approximate popularity by PageRank666http://www.lemurproject.org/clueweb12/PageRank.php, and non- spamminess by the Waterloo Spam Ranking777https://www.mansci.uwaterloo.ca/~msmucker/cw12spam/. We discretize the PageRank scores to generate $3$ grades of popularity (not popular, fairly popular, highly popular), while simulating a power law distribution of popular and not popular documents ($5\%$ highly popular, $10\%$ fairly popular, and $85\%$ not popular). For non-spamminess, we generate $3$ grades of labels (spam, fairly spam, not spam) from the Waterloo Spam Ranking. We treat any document with score $<80$ as spam ($77\%$), documents with score in $[80,89]$ ($14\%$) as fairly spam, and documents with score $\geq 90$ ($9\%$) as not spam (Petersen et al., 2015). For the Web 2010-2014 and Task 2015-2016 tracks, we merge the labels junk and non relevant into non relevant, as was done by the TREC track organisers. For Task 2015-2016, Decision 2019 and Misinformation 2020, _usefulness_ , _credibility_ , and _correctness_ were not assessed for not relevant documents, thus not relevant documents are assumed to be not useful, not credible, and not correct. We evaluate $3$ versions of our method TOMA, with Euclidean, Manhattan, and Chebyshev, as per the distance metric used in Eq. (2) (abbreviated as EUCL, MANH, and CHEB henceforth). We compare these to two state-of-the-art baselines, CAM (Lioma et al., 2017) and MM (Palotti et al., 2018). Given a set of aspects $A$888CAM was originally formulated for two aspects (Lioma et al., 2017)., CAM aggregates their scores through a weighted average: (5) $\text{CAM}(r_{t})=\sum_{a\in A}{p_{a}\times\mu(\hat{r}_{t,a})}$ where $\mu(\cdot)$ is the evaluation measure (e.g., NDCG), $\hat{r}_{t,a}$ is the ranking labelled with respect to aspect $a$, and $p_{a}$ is a parameter controlling the importance of each aspect: $p_{a}\in[0,1]$ and $\sum_{a\in A}p_{a}=1$. MM (Palotti et al., 2018) aggregates the evaluation measure scores computed for each aspect individually with a weighted harmonic mean: (6) $\text{MM}(r_{t})=\frac{\sum_{a\in A}p_{a}}{\sum\limits_{a\in A}\frac{p_{a}}{\mu(\hat{r}_{t,a})}}$ with the same notation as above. Out of the other multi-aspect methods presented in $\S$2, we do not use WHAM (Lioma et al., 2017) as baseline because it also uses the weighted harmonic mean to aggregate the evaluation measure scores. However, WHAM is defined only for relevance and credibility, and can therefore be seen as an instantiation of MM restricted to two aspects. All other multi aspect measures in $\S$2 need a predefined set and number of aspects, thus are not applicable in our scenario. We instantiate our method and the baselines using (1) NDCG (Kekäläinen and Järvelin, 2002) and graded labels (when available); and using (2) AP (Buckley and Voorhees, 2005) and binary labels (we convert all graded labels to binary by treating all grades above zero as one, and grades equal/below zero as zero). We consider all aspects equally important (all aspects are mapped to an integer scale with one unit separating each grade). All source code is publicly available999https://github.com/lcschv/TOMA.git. ### 4.2. Anomalies of CAM & MM Next we discuss anomalies of CAM and MM that TOMA overcomes. Table 3. CAM, MM and TOMA scores instantiated with AP & NDCG for all rankings in $D$. The highest scores are in bold. AP --- Length $3$ | CAM | MM | EUCL | MANH | CHEB | Length $2$ | CAM | MM | EUCL | MANH | CHEB | Length $1$ | CAM | MM | EUCL | MANH | CHEB $(d_{1},d_{2},d_{3})$ | $\mathbf{0.7917}$ | $\mathbf{0.3684}$ | $\mathbf{1}$ | $\mathbf{1}$ | $0.5$ | $(d_{1},d_{2})$ | $0.6250$ | $0.25$ | $\mathbf{1}$ | $\mathbf{1}$ | $0.5$ | $(d_{1})$ | $0.5$ | $0$ | $0.5$ | $0.5$ | $0$ $(d_{1},d_{3},d_{2})$ | $\mathbf{0.7917}$ | $\mathbf{0.3684}$ | $0.8333$ | $0.8333$ | $0.3333$ | $(d_{1},d_{3})$ | $0.6250$ | $0.25$ | $0.5$ | $0.5$ | $0$ | $(d_{2})$ | $0.25$ | $0$ | $0.5$ | $0.5$ | $\mathbf{1}$ $(d_{2},d_{1},d_{3})$ | $0.6667$ | $0.3125$ | $\mathbf{1}$ | $\mathbf{1}$ | $\mathbf{1}$ | $(d_{2},d_{1})$ | $0.5$ | $0.25$ | $\mathbf{1}$ | $\mathbf{1}$ | $\mathbf{1}$ | $(d_{3})$ | $0.25$ | $0$ | $0$ | $0$ | $0$ $(d_{2},d_{3},d_{1})$ | $0.6667$ | $0.25$ | $0.8333$ | $0.8333$ | $\mathbf{1}$ | $(d_{2},d_{3})$ | $0.5$ | $0$ | $0.5$ | $0.5$ | $\mathbf{1}$ | - | - | - | - | - | - $(d_{3},d_{1},d_{2})$ | $0.6667$ | $0.3125$ | $0.5833$ | $0.5833$ | $0.3333$ | $(d_{3},d_{1})$ | $0.5$ | $0.25$ | $0.25$ | $0.25$ | $0$ | - | - | - | - | - | - $(d_{3},d_{2},d_{1})$ | $0.6667$ | $0.25$ | $0.5833$ | $0.5833$ | $0.5$ | $(d_{3},d_{2})$ | $0.5$ | $0$ | $0.25$ | $0.25$ | $0.5$ | - | - | - | - | - | - NDCG Length $3$ | CAM | MM | EUCL | MANH | CHEB | Length $2$ | CAM | MM | EUCL | MANH | CHEB | Length $1$ | CAM | MM | EUCL | MANH | CHEB $(d_{1},d_{2},d_{3})$ | $\mathbf{0.9073}$ | $0.4489$ | $0.9367$ | $0.9711$ | $0.8597$ | $(d_{1},d_{2})$ | $0.7682$ | $0.3491$ | $0.8080$ | $0.8147$ | $0.8597$ | $(d_{1})$ | $0.4728$ | $0.1491$ | $0.4290$ | $0.4693$ | $0.3801$ $(d_{1},d_{3},d_{2})$ | $0.8824$ | $0.4386$ | $0.8917$ | $0.9404$ | $0.7602$ | $(d_{1},d_{3})$ | $0.6483$ | $0.3145$ | $0.5914$ | $0.6667$ | $0.3801$ | $(d_{2})$ | $0.4682$ | $0.2258$ | $0.6006$ | $0.5475$ | $0.7602$ $(d_{2},d_{1},d_{3})$ | $0.9056$ | $\mathbf{0.4516}$ | $\mathbf{1}$ | $\mathbf{1}$ | $\mathbf{1}$ | $(d_{2},d_{1})$ | $0.7665$ | $0.3776$ | $0.8713$ | $0.8436$ | $\mathbf{1}$ | $(d_{3})$ | $0.2781$ | $0$ | $0.2574$ | $0.3129$ | $0$ $(d_{2},d_{3},d_{1})$ | $0.8801$ | $0.4319$ | $0.9775$ | $0.9795$ | $0.9502$ | $(d_{2},d_{3})$ | $0.6437$ | $0.2679$ | $0.7630$ | $0.7449$ | $0.7602$ | - | - | - | - | - | - $(d_{3},d_{1},d_{2})$ | $0.8106$ | $0.3930$ | $0.8284$ | $0.8827$ | $0.6199$ | $(d_{3},d_{1})$ | $0.5765$ | $0.2801$ | $0.5281$ | $0.6089$ | $0.2398$ | - | - | - | - | - | - $(d_{3},d_{2},d_{1})$ | $0.8100$ | $0.3827$ | $0.8509$ | $0.8929$ | $0.6697$ | $(d_{3},d_{2})$ | $0.5735$ | $0.1897$ | $0.6364$ | $0.6583$ | $0.4796$ | - | - | - | - | - | - Figure 2. Box-plots for CAM and MM with NDCG on the Decision Track 2019. Topic numbers are on the $x$-axis and measures scores on the $y$-axis. The maximum value for CAM and MM is variable and depends on the topic and the aspects. ##### Problem 1: MM is ill-defined. As the harmonic mean is not defined with zero values, MM is not defined if $\exists a\in A$ such that $\mu(\hat{r}_{t,a})=0$, e.g., a ranking does not retrieve any correct or relevant document. To compute MM even in these cases, as the denominator in Eq. (6) tends to $+\infty$ if any $\mu(\hat{r}_{t,a})$ tends to zero, we set $\text{MM}(r_{t})=0$. For classification measures, this problem is called the Strong Definiteness Axiom (Sebastiani, 2015). It represents a serious issue for collections where there are a few documents with a positive label for certain aspects. For example, for the Task Tracks, since useful documents are very sparse, many systems are not able to retrieve any useful document and they all have a $0$ score, independently of the number of relevant documents they retrieve. TOMA does not have this problem, because we first assign a weight to each tuple of labels and then compute a single- aspect evaluation measure $\mu$, thus there is no division by $0$ and TOMA is well defined. ##### Problem 2: CAM and MM can range in different intervals. Given a set of documents $D$ and a set of aspects $A$, by definition CAM and MM are multi-aspect evaluation measures $M\colon D^{*}\to[0,X]$, where $D^{*}$ is the set of rankings and $X\leq 1$. Depending on $D$ and $A$, there exist cases with $X<1$. To prove this claim, we need to show that when $M$ is CAM or MM, $\exists\ D,A$ such that: (7) $\max_{r\in D^{*}}M(\hat{r}_{t})<1$ i.e., for each ranking of documents in $D^{*}$ the maximum measure score will be less than $1$. To build such an example, the set $D$ needs to contain documents with not comparable tuples of labels: (8) $\exists d_{1},d_{2}\in D\colon\textrm{GT}(d_{1})\not\sqsubseteq\textrm{GT}(d_{2})\textrm{ and }\textrm{GT}(d_{2})\not\sqsubseteq\textrm{GT}(d_{1})\iff\\\ \exists d_{1},d_{2}\in D\textrm{ and }\exists a_{1},a_{2}\in A\colon\\\ \textrm{GT}_{a_{1}}(d_{1})\prec_{a_{1}}\textrm{GT}_{a_{1}}(d_{2})\textrm{ and }\textrm{GT}_{a_{2}}(d_{2})\prec_{a_{2}}\textrm{GT}_{a_{2}}(d_{1})$ In this case, CAM and MM cannot achieve a score equal to $1$ as illustrated by the following example. Consider the example in $\S$3.4 with $A=\\{$relevance, correctness$\\}$. Let $D$ be a set with $3$ documents $D=\\{d_{1},d_{2},d_{3}\\}$ with labels $\textrm{GT}(d_{1})=(\texttt{mr},\texttt{c})$, $\textrm{GT}(d_{2})=(\texttt{hr},\texttt{pc})$ and $\textrm{GT}(d_{3})=(\texttt{hr},\texttt{nc})$. Documents $(d_{1},d_{2})$ and $(d_{1},d_{3})$ are not comparable and there is no unequivocal way of sorting them, e.g., it is not clear if $d_{1}$ should be ranked before $d_{2}$ or vice-versa. Let us consider CAM and MM instantiated with AP and NDCG. For AP we use a harsh mapping for relevance and correctness, i.e., $\\{\texttt{fr},\texttt{hr}\\}\mapsto 1$ and $\\{\texttt{mr},\texttt{nr}\\}\mapsto 0$, and $\texttt{c}\mapsto 1$ and $\\{\texttt{pc},\texttt{nc}\\}\mapsto 0$. For NDCG we map each category to a different integer, for relevance we have: $\texttt{hr}\mapsto 15$, $\texttt{fr}\mapsto 10$, $\texttt{mr}\mapsto 5$, $\texttt{nr}\mapsto 0$, and for correctness we have: $\texttt{c}\mapsto 10$, $\texttt{pc}\mapsto 5$, $\texttt{nc}\mapsto 0$. The NDCG ideal ranking (Järvelin and Kekäläinen, 2002) for relevance is: $(\texttt{hr},\texttt{hr},\texttt{mr})$ and for correctness is $(\texttt{c},\texttt{pc},\texttt{nc})$. The NDCG $\log$ base is set to $2$. For TOMA we use the embedding of the first row in Tab. 1 and as weight function we map each equivalence class to a different integer with step $1$. We instantiate TOMA with AP and NDCG with log base $2$ (Tab. 3). Since AP does not handle multi-graded weights, we map the top half of the equivalence classes to $1$ and the rest to $0$. Tab. 3 shows CAM, MM, and TOMA scores instantiated with AP NDCG for each possible ranking of documents in $D$. In Tab. 3 none of the rankings in $D^{*}$ can achieve a score equal to $1$ for CAM and MM, while TOMA has at least one ranking with score $1$. In CAM and MM this happens because, any way we sort the documents, either we penalize correctness, e.g., $(d_{2},d_{3},d_{1})$ or we penalize relevance, e.g., $(d_{1},d_{2},d_{3})$. TOMA does not have this problem, since it first defines how to sort tuples of labels, then weights them accordingly and computes the measure score. Thus, if we sort documents in the order induced by $\preceq_{*}$, we obtain a score equal to $1$ (proof in the appendix). Experiments on real data confirm this, as detailed next. ##### Estimating CAM and MM Upper Bound With the following experiment we show that with real data CAM and MM can be upper bounded by a value $X$ lower than $1$. To estimate the value $X$ with real data, we generate different ideal rankings of documents with different strategies. The intuition is that by ranking documents in the best possible way, we should achieve a score equal to $1$, as it happens for any single- aspect evaluation measure computed against the ideal ranking. Since CAM and MM do not define how to sort documents, i.e., a total order relation $\preceq_{*}$, we need to test different possible strategies to build these ideal rankings. First, we define the ideal rankings obtained with a recursive strategy: these are the ideal rankings for each aspect when considered separately, e.g., for $3$ aspects, $a_{1}$, $a_{2}$ and $a_{3}$, with a preference order where $a_{1}$ is followed by $a_{2}$, followed by $a_{3}$: (1) we sort the documents with decreasing label for $a_{1}$; (2) among the documents with the same label for $a_{1}$, we sort the documents with decreasing label for $a_{2}$; (3) among the documents with the same label for $a_{1}$ and $a_{2}$, we sort the documents with decreasing label for $a_{3}$. We generate these ideal rankings for each possible preference order among the aspects. We also generate $3$ additional ideal rankings: (1) we sum the weights across aspects and sort the documents by this sum; (2) we sum the squared weights across aspects and sort the documents by this sum; (3) we consider the highest weight across aspects and sort documents by their highest weight regardless of the aspect. Fig. 2 reports the distributions of CAM with AP scores for the ideal rankings for the Decision Track 2019. These distributions depend on the aspect and the topic. We see that the upper bound $X$ is variable and depends on the topic: just for $2\%$ of topics it is equal to $1$ and for $26\%$ topics it is lower than $0.9$. We obtain similar or even more extreme distributions of scores for all the other tracks (except for Misinformation 2020, reported in the online appendix). ##### Interpretability of CAM and MM scores Problem $2$ is especially important because it affects the interpretability of CAM and MM scores. When a measure is used to assess the quality of a single ranking in isolation, it should be intuitively interpretable (Kumar and Vassilvitskii, 2010), e.g., NDCG=$0.6$ has the intuitive interpretation that the ranking can be further improved by $0.4$. If TOMA is instantiated with NDCG, the intuitive interpretability of NDCG holds, but if CAM or MM are instantiated with NDCG, the intuitive interpretability of NDCG is lost: by the arguments above, CAM and MM may fail to obtain an optimal score of $1$, and the optimal score depends on $A$ and $D$, hence it cannot in general be known a priori. This issue is important for MM, which is affected also by Problem $1$, and therefore may have $X<<1$. Thus MM scores can be compressed towards $0$, and this can lead to cases with many ties, where it is hard to distinguish between different rankings. ### 4.3. Experimental Findings Table 4. Kendall’s $\tau$ correlation between rankings of systems and discriminative power (the higher, the better; best is in bold). Not all aspect combinations occur in all tracks (marked grey). | WEB2009 | WEB2010 | WEB2011 | WEB2012 | WEB2013 | WEB2014 | TASK15 | TASK16 | DECISION19 | MISINFO 2020 ---|---|---|---|---|---|---|---|---|---|--- | NDCG | AP | NDCG | AP | NDCG | AP | NDCG | AP | NDCG | AP | NDCG | AP | NDCG | AP | NDCG | AP | NDCG | AP | NDCG | AP | CORRELATION EUCL - CAM | 0.25 | 0.16 | 0.18 | 0.12 | 0.21 | 0.11 | 0.31 | 0.26 | 0.22 | 0.12 | 0.30 | 0.23 | 0.68 | 0.54 | 0.63 | 0.55 | 0.76 | 0.60 | 0.72 | 0.51 EUCL - MM | 0.07 | 0.04 | 0.05 | 0.00 | 0.06 | 0.03 | 0.16 | 0.13 | 0.12 | 0.04 | 0.17 | 0.04 | 0.36 | 0.17 | 0.02 | -0.10 | 0.46 | 0.31 | 0.45 | 0.27 MANH - CAM | 0.22 | 0.16 | 0.21 | 0.12 | 0.16 | 0.11 | 0.34 | 0.26 | 0.31 | 0.12 | 0.41 | 0.23 | 0.62 | 0.54 | 0.60 | 0.55 | 0.69 | 0.60 | 0.72 | 0.51 MANH - MM | 0.06 | 0.04 | 0.01 | 0.01 | 0.06 | 0.03 | 0.16 | 0.13 | 0.11 | 0.04 | 0.15 | 0.04 | 0.28 | 0.17 | -0.06 | -0.10 | 0.41 | 0.31 | 0.45 | 0.27 CHEB - CAM | 0.01 | 0.02 | 0.06 | 0.05 | 0.02 | 0.02 | 0.19 | 0.15 | 0.11 | 0.00 | 0.19 | 0.07 | 0.26 | 0.16 | -0.13 | -0.18 | 0.27 | 0.27 | 0.29 | 0.24 CHEB - MM | 0.06 | 0.09 | 0.00 | 0.03 | 0.09 | 0.01 | 0.19 | 0.16 | 0.12 | 0.08 | 0.14 | 0.05 | 0.88 | 0.86 | 0.52 | 0.53 | 0.58 | 0.60 | 0.54 | 0.52 EUCL - MANH | 0.36 | 1.00 | 0.19 | 1.00 | 0.21 | 1.00 | 0.34 | 1.00 | 0.20 | 1.00 | 0.34 | 1.00 | 0.87 | 1.00 | 0.71 | 1.00 | 0.72 | 1.00 | 1.00 | 1.00 EUCL - CHEB | 0.01 | 0.02 | 0.10 | 0.03 | 0.01 | 0.03 | 0.32 | 0.11 | 0.22 | 0.04 | 0.30 | 0.12 | 0.33 | 0.19 | -0.21 | -0.21 | 0.28 | 0.21 | 0.26 | 0.20 MANH - CHEB | 0.01 | 0.02 | 0.03 | 0.03 | 0.02 | 0.03 | 0.21 | 0.11 | 0.10 | 0.04 | 0.18 | 0.12 | 0.32 | 0.19 | -0.24 | -0.21 | 0.25 | 0.21 | 0.26 | 0.20 CAM - MM | 0.10 | 0.05 | 0.05 | 0.01 | 0.11 | -0.01 | 0.23 | 0.13 | 0.16 | 0.03 | 0.26 | 0.04 | 0.30 | 0.18 | 0.09 | 0.00 | 0.51 | 0.41 | 0.51 | 0.42 | CORRELATION Relevance - Popularity | 0.03 | 0.05 | 0.01 | 0.0 | 0.01 | 0.02 | 0.09 | 0.09 | 0.06 | 0.01 | 0.07 | 0.02 | 0.04 | 0.04 | -0.03 | 0.01 | | | | Relevance - Non-spam | 0.05 | 0.03 | 0.02 | 0.0 | 0.03 | 0.01 | 0.07 | 0.05 | -0.02 | -0.01 | 0.07 | -0.01 | 0.25 | 0.17 | -0.07 | -0.08 | | | | Popularity - Non-spam | 0.04 | 0.03 | 0.02 | -0.01 | 0.04 | 0.01 | 0.07 | 0.04 | -0.03 | -0.02 | 0.04 | 0.00 | 0.07 | 0.02 | 0.08 | 0.06 | | | | Relevance - Usefulness | | | | | | | | | | | | | 0.75 | 0.75 | 0.75 | 0.74 | | | | Usefulness - Popularity | | | | | | | | | | | | | 0.10 | 0.06 | -0.04 | 0.00 | | | | Usefulness - Non-spam | | | | | | | | | | | | | 0.40 | 0.33 | -0.16 | -0.19 | | | | Credibility - Correctness | | | | | | | | | | | | | | | | | 0.26 | 0.26 | 0.28 | 0.24 Relevance - Credibility | | | | | | | | | | | | | | | | | 0.33 | 0.33 | 0.29 | 0.25 Relevance - Correctness | | | | | | | | | | | | | | | | | 0.42 | 0.49 | 0.49 | 0.47 | DISCRIMINATIVE POWER OF MEASURES CAM | 75.98 | 64.43 | 66.32 | 61.23 | 75.14 | 61.64 | 68.71 | 56.74 | 76.89 | 57.05 | 85.06 | 78.85 | 53.33 | 33.33 | 72.22 | 55.56 | 72.58 | 70.56 | 71.53 | 70.90 MM | 75.61 | 50.58 | 72.89 | 67.79 | 67.32 | 67.81 | 62.68 | 56.12 | 80.71 | 46.99 | 74.25 | 53.56 | 0.00 | 0.00 | 0.00 | 0.00 | 60.08 | 53.23 | 68.31 | 62.20 EUCL | 75.29 | 72.64 | 62.96 | 66.75 | 75.14 | 70.33 | 66.13 | 64.10 | 75.14 | 59.45 | 80.92 | 78.85 | 66.67 | 66.67 | 69.44 | 75.00 | 73.59 | 73.99 | 72.86 | 75.14 MANH | 76.66 | 72.68 | 63.59 | 67.14 | 77.32 | 70.38 | 66.05 | 64.18 | 76.67 | 59.34 | 86.44 | 79.08 | 66.67 | 53.33 | 75.00 | 75.00 | 73.79 | 73.79 | 73.02 | 74.98 CHEB | 50.18 | 6.32 | 59.82 | 51.49 | 73.06 | 50.11 | 61.08 | 39.36 | 77.10 | 49.34 | 75.17 | 66.21 | 0.00 | 0.00 | 0.00 | 0.00 | 42.54 | 29.84 | 65.41 | 53.33 Empirically, evaluation measures are commonly assessed in terms of their correlation (Ferrante et al., 2019a), discriminative power (Sakai, 2007), informativeness (Aslam et al., 2005), intuitiveness (Sakai, 2012), and unanimity (Albahem et al., 2019). Out of these, we report only correlation and discriminative power because the rest does not apply: the informativeness test (Aslam et al., 2005) requires a precision recall-curve, which cannot be defined for multi-aspect evaluation; the intuitiveness test (Sakai, 2012) requires simple single-aspect measures (e.g. precision, recall), which do not apply to multi-aspect evaluation; the unanimity test (Albahem et al., 2019), which is defined for multi-aspect evaluation, requires that all the simple measures agree over all aspects, which happened extremely rarely in our data, especially as the number of aspects increased (see the low correlation among aspects in Tab. 4). #### 4.3.1. Correlation Analysis We use Kendall’s $\tau$ (Kendall, 1945) to estimate TOMA’s correlation to MM and CAM. Generally, if a new evaluation measure strongly correlates to an existing one, it is likely to represent redundant information (Webber et al., 2008). We use Kendall’s $\tau$ because it has better gross-error sensitivity than the Pearson correlation coefficient (Croux and Dehon, 2010), and because the Spearman correlation coefficient cannot handle ties. As per (Ferrante et al., 2019a), we compute the correlation topic-by-topic. For each topic we consider the Rankings of Submitted runs (RoS) corresponding to two different measures (one ranking per measure) and then compute Kendall’s $\tau$ between the two RoS. We report Kendall’s $\tau$ averaged across all topics. As per (Voorhees, 1998, 2001), we consider two rankings equivalent if Kendall’s $\tau$ is greater than $0.9$. Tab. 4 shows the findings, which are summarised as follows: * • The RoS corresponding to EUCL - MANH are equivalent ($\tau=1$) at all times for AP. This perfect correlation for AP happens because, by definition, when the sets of equivalence classes from these approaches are mapped to binary labels, they produce the exact same set of labels (see also Tab. 3). For NDCG, $\tau=0.19-1$, where higher correlations correspond to tracks where some aspects are not assessed for non relevant documents, thus there are less extreme cases and EUCL is more similar to MANH. * • The RoS corresponding to (EUCL, MANH) - CHEB are very weakly correlated ($\tau=0.01-0.32$), this is due to Chebyshev distance being very harsh, since many equivalence classes are considered equivalent to the class of non relevant documents. * • The RoS corresponding to EUCL - CAM and MANH - CAM are very weakly correlated ($\tau=0.11-0.41$) for the Web tracks, but moderately correlated ($\tau=0.54-0.76$) for the Task, Decision and Misinformation tracks. This happens because: (i) the runs submitted to the Web tracks were not designed to account for multiple aspects and (ii) for the Task, Decision and Misinformation tracks, usefulness, credibility and correctness are not assessed for non relevant documents. Therefore, since some of the values are missing, these methods generate a lower number of equivalence classes, which make them more similar to CAM. Whereas, for the Web tracks, popularity and non-spamminess are approximated for all documents, meaning that MANH and EUCL can possibly generate all the different equivalence classes, even for non relevant documents. This makes them less similar to CAM than on the Task or Decision tracks. * • For the Task, Decision and Misinformation tracks, the RoS corresponding to MM and CHEB are moderately correlated ($\tau=0.52-0.88$). The fact that, for these tracks, usefulness, credibility and correctness are not assessed for non relevant documents, means that all the documents that are mapped to a $0$ weight with CHEB, are also contributing as $0$ to MM. To contextualise these findings, the middle part of Tab. 4 shows the $\tau$ values of the RoS corresponding to evaluating a single aspect only. Overall, the resulting correlations are low to non-existent, meaning that considering multiple aspects affects the final evaluation outcome. The two exceptions where the correlation between RoS is not very low are: * • For Task 2015-2016, for relevance - usefulness, $\tau=0.74-0.75$. This happens because: (1) usefulness is not assessed for non relevant documents, thus non relevant documents are assumed to be not useful, and (2) usefulness is a very sparse signal (1.75% of documents are useful). * • For the Decision and Misinformation Tracks, for all aspects, $\tau=0.24-0.49$. Again here credibility and correctness are not assessed for non relevant documents ($6.89\%$ of documents are credible and $9.75\%$ are correct for the Decision Track; $13.73\%$ of documents are correct and $27.62\%$ are credible for the Misinformation Track), so the correlation is not as high as for the Task tracks. Overall, the most correlated RoS correspond to: EUCL - MANH ($\tau$ up to $1$), (EUCL, MANH)- CAM ($\tau$ up to $0.76$), and CHEB - MM ($\tau$ up to $0.88$). Intuitively, EUCL and MANH may be more similar to CAM (mean), while CHEB may be more similar to MM (harmonic mean). Thus TOMA proposes an alternative evaluation framework, which overcomes CAM and MM anomalies (see $\S$4.2). The fact that $\tau$ values between TOMA and the baselines are never above $0.9$ means that there are noticeable differences between the RoS generated by TOMA and by CAM or MM. Recall that all measures are instantiated with NDCG or AP, meaning that differences between them are due to how multi- aspect labels are treated. #### 4.3.2. Discriminative Power We use Bootstrap Hypothesis Test (Sakai, 2007) to estimate the discriminative power of TOMA, CAM and MM. Given a set of topics and a set of runs, we first generate subsets of topics by sampling with replacement the complete set of topics. We set the number of bootstrap samples to $10\,000$. To assess whether the measure scores for pairs of runs can be considered different at a given confidence level, we use a Paired Bootstrap Hypothesis Test. The confidence level is $1-\alpha$, where $\alpha$ is the Type I Error, i.e., the probability to consider two systems different even if they are equivalent. We set $\alpha=0.01$, requiring strong evidence for two systems to be different. Tab. 4 (bottom part) displays the results of the discriminative power analysis, where the higher the score, the more discriminative (i.e., the better) the approach. We see that $16/20$ times either MANH ($12/20$) or EUCL ($6/20$)101010Ties are included in these counts. is best. The remaining $4$ times, MM is best $3$ times, and CAM once. We also see that CHEB is never best, and for Task 2015-2016 it is actually zero. This is due to the very small amount of positive labels for usefulness in that track. For the same reason, MM is also zero for the same track. Overall, CHEB is the least discriminative measure, followed by MM; this is due to how these methods treat tuples of labels: the fact that if one aspect label is zero, then the whole score is zero, practically means that many runs are considered equal purely on that basis. #### 4.3.3. Zero-aspect documents Table 5. Number of times (%) that the labels of all aspects sum to $0$ for a document that is ranked at position $1$-$5$ (column 1) in a run that has been assessed as best per $\\{$topic, track, year$\\}$ separately with $\\{$CAM, MM, EUCL, MANH, CHEB$\\}$ using a retrieval cutoff of 5. The lower, the better. Rank | CAM | MM | EUCL | MANH | CHEB ---|---|---|---|---|--- 1 | 51 (1.18%) | 131 (3.02%) | 39 (0.90%) | 33 (0.76%) | 154 (3.55%) 2 | 65 (1.50%) | 159 (3.67%) | 50 (1.15%) | 48 (1.11%) | 179 (4.13%) 3 | 103 (2.38%) | 202 (4.66%) | 88 (2.03%) | 78 (1.80%) | 185 (4.17%) 4 | 102 (2.35%) | 173 (3.99%) | 86 (1.99%) | 74 (1.71%) | 183 (4.23%) 5 | 107 (2.47%) | 196 (4.53%) | 95 (2.19%) | 81 (1.87%) | 205 (4.73%) 1-5 | 428 (9.88%) | 861 (19.88%) | 358 (8.27%) | 314 (7.25%) | 906 (20.92%) Our next analysis is motivated by the empirical trash$@k$ measure often used in industry to mitigate the high cost of retrieving “trash” in high ranks. We count how often the labels of all aspects sum to zero for a document that has been ranked at position $1$-$5$ in a run that has been assessed as the best run per track year, on a per query basis, using a retrieval cutoff of $5$, separately with $\\{$CAM, MM, EUCL, MANH, CHEB$\\}$ when instantiated separately with NDCG and AP. When the labels of all aspects sum to zero, this means that the corresponding document is of the worst quality. Ideally, such documents should not be retrieved, but when they do, they should not be in the top 5. In Tab. 5 we see that MANH is associated with the lowest amount of zero-aspect documents, closely followed by EUCL. This happens because MANH is designed so that the higher the sum of a document’s labels across aspects, the better that document will be deemed. CHEB is overall worst, closely followed by MM. This closeness between EUCL-MANH and CHEB-MM agrees with the previous correlation and discriminative power analysis. Overall, MANH (and less so EUCL) penalise low quality documents the best. #### 4.3.4. Document quality @1-100 Table 6. Average sum of aspect labels for a document that is ranked at position $1$-$100$ (column 1) in a run that has been assessed as best per $\\{$topic, track, year$\\}$ separately with $\\{$CAM, MM, EUCL, MANH, CHEB$\\}$ using a retrieval cutoff of 100. The higher, the better. Ranks | CAM | MM | EUCL | MANH | CHEB ---|---|---|---|---|--- 1-25 | 1.70 | 1.49 | 1.67 | 1.69 | 1.39 26-50 | 0.85 | 0.78 | 0.91 | 0.94 | 0.70 51-75 | 0.57 | 0.53 | 0.63 | 0.64 | 0.48 76-100 | 0.40 | 0.39 | 0.43 | 0.44 | 0.36 We look at the quality of documents that have been ranked at positions 1-100 in a run that has been assessed as best per $\\{$topic, track, year$\\}$ separately with $\\{$CAM, MM, EUCL, MANH, CHEB$\\}$, when instantiated separately with NDCG and AP, using a retrieval cutoff of $100$. We split the ranks $1$-$100$ into four sets ($1$-$25$, $26$-$50$, $51$-$75$, $76$-$100$). For each document in each set, we sum the labels of its aspects, and we report the average of these sums, which can be seen as an approximation of the average document quality (the higher, the better). As expected, we see that the numbers in Tab. 6, and hence document quality, drop as we move down the ranking, at all times. Comparing across columns however, we see that, for the runs that were assessed as best by MANH, document quality is overall, albeit marginally, the best, at ranks 26-100. This illustrates that the design of MANH (the higher the sum of a document’s labels across aspects, the better that document will be considered) gives it the practical advantage of, not only reducing the amount of low quality documents in the top ranks (as seen in Tab. 5), but also of increasing the quality of documents further down the ranking, as we see now. Again, as previously, we observe that EUCL is a close second-best method, CHEB and MM are overall worst, and CAM is in between (although best, together with MANH, for the top ranks). ## 5\. Conclusion and Limitations Multi-aspect evaluation is a special case of IR evaluation where the ranked list of documents returned by an IR system in response to a query must be assessed in terms of not only relevance, but also other aspects (or dimensions) of the ranked documents, e.g., credibility or usefulness. We presented a principled multi-aspect evaluation approach, called TOMA, that is defined for any number and type of aspect, and that allows for (i) aspects having different gradings, (ii) any relative importance weighting for different aspects, and (iii) integration with any existing single-aspect evaluation measure, such as NDCG. We showed that TOMA has better discriminative power than prior approaches to multi-aspect evaluation, and that it is better at rewarding high quality documents across the ranking. One limitation of TOMA is represented by the arbitrary choices of the embedding function, the distance function and the weight function. The embedding function maps labels from a nominal or ordinal scale to an interval or ratio scale. This calls for a in-depth investigation of the theoretical properties of TOMA using the existing axiomatic treatments of IR effectiveness measures (Amigó et al., 2018a; Maddalena and Mizzaro, 2014; Busin and Mizzaro, 2013; Amigó and Mizzaro, 2020; Ferrante et al., 2021). This also motivates a deep analysis of the interactions between different aspects and/or documents and how to handle them with TOMA, for example by defining a proper embedding and distance function which account for aspects as diversity, novelty, and redundancy. Moreover, the embedding function combined with the distance function can generate a large number of tuples of labels, which can be mapped to different integers through the weight function. This might be a problem for gain based measures, thus a possible solution is to use TOMA to define the ideal ranking and then use effectiveness measures based on similarity to ideal rankings (Clarke et al., 2020c, b, 2021). Finally, the empirical impact of varying both distance and weight functions should also be investigated, as should the impact of employing further multi-graded measures as ERR (Chapelle et al., 2009), and the alignment of our current approach with real user preferences. Acknowledgments. This paper is partially supported by the EU Horizon 2020 research and innovation programme under the MSCA grant No. 893667. ## References * (1) * Agrawal et al. (2009) R. Agrawal, S. Gollapudi, A. Halverson, and S. Ieong. 2009\. Diversifying Search Results. In _Proc. 2nd ACM International Conference on Web Searching and Data Mining (WSDM 2009)_ , R. Baeza-Yates, P. Boldi, B. Ribeiro-Neto, and B. B. Cambazoglu (Eds.). ACM Press, New York, USA, 5–14. * Albahem et al. (2019) A. Albahem, , D. Spina, F. Scholer, and L. Cavedon. 2019. Meta-evaluation of Dynamic Search: How Do Metrics Capture Topical Relevance, Diversity and User Effort?. In _Advances in Information Retrieval. Proc. 41st European Conference on IR Research (ECIR 2019)_ , L. Azzopardi, B. Stein, N. Fuhr, P. Mayr, C. Hauff, and D. Hiemstra (Eds.). Lecture Notes in Computer Science (LNCS) 10772, Springer, Heidelberg, Germany, 607–620. * Amigó et al. (2018a) E. Amigó, F. Giner, S. Mizzaro, and D. Spina. 2018a. A Formal Account of Effectiveness Evaluation and Ranking Fusion. In _Proc. 4th ACM SIGIR International Conference on the Theory of Information Retrieval (ICTIR 2018)_ , D. Song, T.-Y. Liu, L. Sun, P. Bruza, M. Melucci, F. Sebastiani, and G. Hui Yang (Eds.). ACM Press, New York, USA, 123–130. * Amigó and Mizzaro (2020) E. Amigó and S. Mizzaro. 2020. On the Nature of Information Access Evaluation Metrics: a Unifying Framework. _Information Retrieval Journal_ 23, 3 (2020), 318–386. https://doi.org/10.1007/s10791-020-09374-0 * Amigó et al. (2018b) E. Amigó, D. Spina, and J. Carrillo-de Albornoz. 2018b. An Axiomatic Analysis of Diversity Evaluation Metrics: Introducing the Rank-Biased Utility Metric. In _Proc. 41th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2018)_ , K. Collins-Thompson, Q. Mei, B. Davison, Y. Liu, and E. Yilmaz (Eds.). ACM Press, New York, USA, 625–634. * Aslam et al. (2005) J. A. Aslam, E. Yilmaz, and V. Pavlu. 2005. The maximum entropy method for analyzing retrieval measures. In _SIGIR 2005: Proceedings of the 28th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Salvador, Brazil, August 15-19, 2005_ , Ricardo A. Baeza-Yates, Nivio Ziviani, Gary Marchionini, Alistair Moffat, and John Tait (Eds.). ACM, 27–34. https://doi.org/10.1145/1076034.1076042 * Buckley and Voorhees (2005) C. Buckley and E. M. Voorhees. 2005. Retrieval System Evaluation. In _TREC. Experiment and Evaluation in Information Retrieval_ , D. K. Harman and E. M. Voorhees (Eds.). MIT Press, Cambridge (MA), USA, 53–78. * Burges et al. (2005) C. Burges, T. Shaked, E. Renshaw, A. Lazier, M. Deeds, N. Hamilton, and G. Hullender. 2005. Learning to Rank using Gradient Descent. In _Proc. 22nd International Conference on Machine Learning (ICML 2005)_ , S. Dzeroski, L. De Raedt, and S. Wrobel (Eds.). ACM Press, New York, USA, 89–96. * Busin and Mizzaro (2013) L. Busin and S. Mizzaro. 2013. Axiometrics: An Axiomatic Approach to Information Retrieval Effectiveness Metrics. In _Proc. 4th International Conference on the Theory of Information Retrieval (ICTIR 2013)_ , O. Kurland, D. Metzler, C. Lioma, B. Larsen, and P. Ingwersen (Eds.). ACM Press, New York, USA, 22–29. * Chapelle et al. (2009) O. Chapelle, D. Metzler, Y. Zhang, and P. Grinspan. 2009\. Expected Reciprocal Rank for Graded Relevance. In _Proc. 18th International Conference on Information and Knowledge Management (CIKM 2009)_ , D. W.-L. Cheung, I.-Y. Song, W. W. Chu, X. Hu, and J. J. Lin (Eds.). ACM Press, New York, USA, 621–630. * Clarke et al. (2009) C. Clarke, N. Craswell, and I. Soboroff. 2009. Overview of the TREC 2009 Web Track. In _TREC_. National Institute of Standards and Technology (NIST). * Clarke et al. (2020a) Charles Clarke, Maria Maistro, Mark Smucker, and Guido Zuccon. 2020a. Overview of the TREC 2020 Health Misinformation Track (to appear). In _Proceedings of the Twenty-Nine Text REtrieval Conference, TREC 2020, Gaithersburg, Maryland, USA, November 16-19, 2020_ _(NIST Special Publication)_ , Ellen M. Voorhees and Angela Ellis (Eds.), Vol. -. National Institute of Standards and Technology (NIST). * Clarke and Cormack (2011) C. L. A. Clarke and G. V. Cormack. 2011. Overview of the TREC 2010 Web Track. In _TREC_. National Institute of Standards and Technology (NIST). * Clarke et al. (2011) C. L. A. Clarke, N. Craswell, I. Soboroff, and E. M. Voorhees. 2011. Overview of the TREC 2011 Web Track. In _TREC_. National Institute of Standards and Technology (NIST). * Clarke et al. (2012) C. L. A. Clarke, N. Craswell, and E. M. Voorhees. 2012\. Overview of the TREC 2012 Web Track. In _TREC_. National Institute of Standards and Technology (NIST). * Clarke et al. (2008) C. L. A. Clarke, M. Kolla, G. V. Cormack, O. Vechtomova, A. Ashkan, S. Büttcher, and I. MacKinnon. 2008. Novelty and Diversity in Information Retrieval Evaluation. In _Proc. 31st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2008)_ , T.-S. Chua, M.-K. Leong, D. W. Oard, and F. Sebastiani (Eds.). ACM Press, New York, USA, 659–666. * Clarke et al. (2020b) Charles L. A. Clarke, Mark D. Smucker, and Alexandra Vtyurina. 2020b. Offline Evaluation by Maximum Similarity to an Ideal Ranking. In _CIKM ’20: The 29th ACM International Conference on Information and Knowledge Management, Virtual Event, Ireland, October 19-23, 2020_ , Mathieu d’Aquin, Stefan Dietze, Claudia Hauff, Edward Curry, and Philippe Cudré-Mauroux (Eds.). ACM, 225–234. * Clarke et al. (2020c) Charles L. A. Clarke, Alexandra Vtyurina, and Mark D. Smucker. 2020c. Offline Evaluation without Gain. In _ICTIR ’20: The 2020 ACM SIGIR International Conference on the Theory of Information Retrieval, Virtual Event, Norway, September 14-17, 2020_ , Krisztian Balog, Vinay Setty, Christina Lioma, Yiqun Liu, Min Zhang, and Klaus Berberich (Eds.). ACM, 185–192. * Clarke et al. (2021) C. L. A. Clarke, A. Vtyurina, and M. D. Smucker. 2021\. Assessing Top-$k$ Preferences. arXiv:cs.IR/2007.11682 * Collins-Thompson et al. (2014) K. Collins-Thompson, P. Bennett, F. Diaz, C. L. A. Clarke, and E. M. Voorhees. 2014. TREC 2014 Web Track Overview. In _The Twenty-Second Text REtrieval Conference Proceedings (TREC 2013)_ , E. M. Voorhees (Ed.). National Institute of Standards and Technology (NIST), Special Publication 500-302, Washington, USA. * Collins-Thompson et al. (2015) K. Collins-Thompson, P. Bennett, F. Diaz, C. L. A. Clarke, and E. M. Voorhees. 2015. TREC 2014 Web Track Overview. In _The Twenty-Third Text REtrieval Conference Proceedings (TREC 2014)_ , E. M. Voorhees and A. Ellis (Eds.). National Institute of Standards and Technology (NIST), Special Publication 500-308, Washington, USA. * Croux and Dehon (2010) C. Croux and C. Dehon. 2010\. Influence Functions of the Spearman and Kendall Correlation Measures. _Statistical Methods & Applications_ 19 (2010), 497–515. https://doi.org/10.1007/s10260-010-0142-z * Ferrante et al. (2021) M. Ferrante, N. Ferro, and N Fuhr. 2021. Towards Meaningful Statements in IR Evaluation. Mapping Evaluation Measures to Interval Scales. _CoRR_ abs/2101.02668 (2021). arXiv:2101.02668 https://arxiv.org/abs/2101.02668 * Ferrante et al. (2019a) M. Ferrante, N. Ferro, and E. Losiouk. 2019a. How do Interval Scales Help us with Better Understanding IR Evaluation Measures? _Information Retrieval Journal_ (2019). https://doi.org/10.1007/s10791-019-09362-z * Ferrante et al. (2015) M. Ferrante, N. Ferro, and M. Maistro. 2015. Towards a Formal Framework for Utility-Oriented Measurements of Retrieval Effectiveness. In _Proceedings of the 2015 International Conference on The Theory of Information Retrieval_ _(ICTIR ’15)_. Association for Computing Machinery, New York, NY, USA, 21–30. https://doi.org/10.1145/2808194.2809452 * Ferrante et al. (2017) M. Ferrante, N. Ferro, and Silvia P. 2017. Are IR Evaluation Measures on an Interval Scale?. In _Proceedings of the ACM SIGIR International Conference on Theory of Information Retrieval_ _(ICTIR ’17)_. Association for Computing Machinery, New York, NY, USA, 67–74. https://doi.org/10.1145/3121050.3121058 * Ferrante et al. (2019b) M. Ferrante, N. Ferro, and S. Pontarollo. 2019b. A General Theory of IR Evaluation Measures. _IEEE Transactions on Knowledge and Data Engineering (TKDE)_ 31, 3 (2019), 409–422. * Halmos (1974) P. R. Halmos. 1974\. _Naive Set Theory_. Springer-Verlag, New York, USA. * Järvelin and Kekäläinen (2002) K. Järvelin and J. Kekäläinen. 2002. Cumulated Gain-Based Evaluation of IR Techniques. _ACM Trans. Inf. Syst._ 20, 4 (Oct. 2002), 422–446. https://doi.org/10.1145/582415.582418 * Kazai et al. (2004) G. Kazai, S. Masood, and M. Lalmas. 2004. A Study of the Assessment of Relevance for the INEX’02 Test Collection. In _Advances in Information Retrieval. Proc. 26th European Conference on IR Research (ECIR 2004)_ , S. McDonald and J. Tait (Eds.). Lecture Notes in Computer Science (LNCS) 2997, Springer, Heidelberg, Germany, 296–310. * Kekäläinen and Järvelin (2002) J. Kekäläinen and K. Järvelin. 2002. Using Graded Relevance Assessments in IR Evaluation. _Journal of the American Society for Information Science and Technology (JASIST)_ 53, 13 (November 2002), 1120—1129. * Kendall (1945) M. G. Kendall. 1945\. The Treatment of Ties in Ranking Problems. _Biometrika_ 33, 3 (November 1945), 239–251. * Kumar and Vassilvitskii (2010) Ravi Kumar and Sergei Vassilvitskii. 2010. Generalized distances between rankings. In _Proceedings of the 19th International Conference on World Wide Web, WWW 2010, Raleigh, North Carolina, USA, April 26-30, 2010_ , Michael Rappa, Paul Jones, Juliana Freire, and Soumen Chakrabarti (Eds.). ACM, 571–580. https://doi.org/10.1145/1772690.1772749 * Lioma et al. (2019) C. Lioma, M. Maistro, M. D. Smucker, , and D. Zuccon. 2019. Overview of the TREC 2019 Decision Track. In _The Twenty-Eighth Text REtrieval Conference Proceedings (TREC 2019)_ , E. M. Voorhees and A. Ellis (Eds.). National Institute of Standards and Technology (NIST), Special Publication 500-331, Washington, USA. * Lioma et al. (2017) C. Lioma, J. G. Simonsen, and B. Larsen. 2017. Evaluation Measures for Relevance and Credibility in Ranked Lists. In _Proc. 3rd ACM SIGIR International Conference on the Theory of Information Retrieval (ICTIR 2017)_ , J. Kamps, E. Kanoulas, M. de Rijke, H. Fang, and E. Yilmaz (Eds.). ACM Press, New York, USA, 91–98. * Maddalena and Mizzaro (2014) E. Maddalena and S. Mizzaro. 2014. Axiometrics: Axioms of Information Retrieval Effectiveness Metrics. In _Proc. 6th International Workshop on Evaluating Information Access (EVIA 2014)_ , S. Mizzaro and R. Song (Eds.). National Institute of Informatics, Tokyo, Japan, 17–24. * Palotti et al. (2018) J. Palotti, G. Zuccon, and A. Hanbury. 2018. MM: A New Framework for Multidimensional Evaluation of Search Engines. In _Proc. 27th International Conference on Information and Knowledge Management (CIKM 2018)_ , A. Cuzzocrea, J. Allan, N. W. Paton, D. Srivastava, R. Agrawal, A. Broder, M. J. Zaki, S. Candan, A. Labrinidis, A. Schuster, and H. Wang (Eds.). ACM Press, New York, USA, 1699–1702. * Petersen et al. (2015) C. Petersen, C. Lioma, J. G. Simonsen, and B. Larsen. 2015\. Entropy and Graph Based Modelling of Document Coherence Using Discourse Entities: An Application to IR. In _Proceedings of the 2015 International Conference on The Theory of Information Retrieval_ _(ICTIR ’15)_. ACM, New York, NY, USA, 191–200. https://doi.org/10.1145/2808194.2809458 * Sakai (2007) T. Sakai. 2007. Evaluating Information Retrieval Metrics Based on Bootstrap Hypothesis Tests. _IPSJ Digital Courier_ 3 (2007), 625–642. https://doi.org/10.2197/ipsjdc.3.625 * Sakai (2012) T. Sakai. 2012. Evaluation with Informational and Navigational Intents. In _Proc. 21st International Conference on World Wide Web (WWW 2012)_ , A. Mille, F. L. Gandon, J. Misselis, M. Rabinovich, and S. Staab (Eds.). ACM Press, New York, USA, 499–508. * Sebastiani (2015) F. Sebastiani. 2015\. An Axiomatically Derived Measure for the Evaluation of Classification Algorithms. In _Proc. 1st ACM SIGIR International Conference on the Theory of Information Retrieval (ICTIR 2015)_ , J. Allan, W. B. Croft, A. P. de Vries, C. Zhai, N. Fuhr, and Y. Zhang (Eds.). ACM Press, New York, USA, 11–20. * Tang and Yang (2017) Z. Tang and G. H. Yang. 2017. Investigating per Topic Upper Bound for Session Search Evaluation. In _Proceedings of the ACM SIGIR International Conference on Theory of Information Retrieval_ _(ICTIR ’17)_. Association for Computing Machinery, New York, NY, USA, 185–192. https://doi.org/10.1145/3121050.3121069 * Voorhees (1998) E. M. Voorhees. 1998\. Variations in relevance judgments and the measurement of retrieval effectiveness. In _Proc. 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 1998)_ , W. B. Croft, A. Moffat, C. J. van Rijsbergen, R. Wilkinson, and J. Zobel (Eds.). ACM Press, New York, USA, 315–323. * Voorhees (2001) E. M. Voorhees. 2001\. Evaluation by Highly Relevant Documents. In _Proc. 24th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR 2001)_ , D. H. Kraft, W. B. Croft, D. J. Harper, and J. Zobel (Eds.). ACM Press, New York, USA, 74–82. * Voorhees and Ellis (2016) E. M. Voorhees and A. Ellis (Eds.). 2016. _Proceedings of The Twenty-Fifth Text REtrieval Conference, TREC 2016, Gaithersburg, Maryland, USA, November 15-18, 2016_. Vol. Special Publication 500-321. National Institute of Standards and Technology (NIST). http://trec.nist.gov/pubs/trec25/trec2016.html * Webber et al. (2008) W. Webber, A. Moffat, J. Zobel, and T. Sakai. 2008\. Precision-at-ten Considered Redundant. In _Proceedings of the 31st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval_ _(SIGIR ’08)_. ACM, New York, NY, USA, 695–696. https://doi.org/10.1145/1390334.1390456 * Yao (1995) Y. Y. Yao. 1995\. Measuring Retrieval Effectiveness Based on User Preference of Documents. _Journal of the American Society for Information Science (JASIS)_ 46, 2 (1995), 133–145. https://doi.org/10.1002/(SICI)1097-4571(199503)46:2<133::AID-ASI6>3.0.CO;2-Z * Yilmaz et al. (2016) E. Yilmaz, M. Verma, R. Mehrotra, E. Kanoulas, B. Carterette, and N. Craswell. 2016\. Overview of the TREC 2015 Tasks Track. In _The Twenty-Fourth Text REtrieval Conference Proceedings (TREC 2015)_ , E. M. Voorhees and A. Ellis (Eds.). National Institute of Standards and Technology (NIST), Special Publication 500-319, Washington, USA. * Zuccon (2016) G. Zuccon. 2016\. Understandability Biased Evaluation for Information Retrieval. In _Advances in Information Retrieval. Proc. 38th European Conference on IR Research (ECIR 2016)_ , N. Ferro, F. Crestani, M.-F. Moens, J. Mothe, F. Silvestri, G. M. Di Nunzio, C. Hauff, and G. Silvello (Eds.). Lecture Notes in Computer Science (LNCS) 9626, Springer, Heidelberg, Germany, 280–292. *[IR]: Information Retrieval *[AP]: Average Precision *[NDCG]: Normalized Discounted Cumulated Gain *[TOMA]: Total Order Multi-Aspect *[INEX]: INitiative for the Evaluation of Retrieval *[$\alpha$-NDCG]: α-Normalized Discounted Cumulated Gain *[MAP-IA]: Intent Aware Mean Average Precision *[IA-ERR]: Intent Aware Expected Reciprocal Rank *[nCT]: Normalized Cube Test *[RBU]: Rank-Biased Utility *[uRBP]: Understandability-biased Rank-Biased Precision *[MM]: Multidimensional Measure *[NLRE]: Normalised Local Rank Error *[NGRE]: Normalised Global Rank Error *[nWCS]: Normalised Weighted Cumulative Score *[CAM]: Convex Aggregating Measure *[WHAM]: Weighted Harmonic Mean Aggregating Measure *[DCG]: Discounted Cumulated Gain *[RoS]: Rankings of Submitted runs *[ERR]: Expected Reciprocal Rank
# Compositional Covariance Shrinkage and Regularised Partial Correlations Suzanne Jin${}^{\leavevmode\nobreak\ ,2}$ Centre for Genomic Regulation (CRG), Dr Aiguader, 88, 08003 Barcelona, Spain Cédric Notredame1, Universitat Pompeu Fabra, Barcelona, Spain Ionas Erb1, Corresponding author<EMAIL_ADDRESS> ###### Abstract We propose an estimation procedure for covariation in wide compositional data sets. For compositions, widely-used logratio variables are interdependent due to a common reference. _Logratio uncorrelated_ compositions are linearly independent before the unit-sum constraint is imposed. We show how they are used to construct bespoke shrinkage targets for logratio covariance matrices and test a simple procedure for partial correlation estimates on both a simulated and a single-cell gene expression data set. For the underlying counts, different zero imputations are evaluated. The partial correlation induced by the closure is derived analytically. Data and code are available from GitHub. #### MSC: G2F30, G2H20, G2P99 #### Keywords: Compositional Covariance Structure, Logratio Analysis, Partial Correlation, James-Stein Shrinkage ## 1\. Introduction The study of data variability is at the heart of data analysis. A simple way of quantifying the variability around some central tendency is given by a bilinear function defined for each pair of variables: the covariance. The distribution that makes no assumptions beyond a finite covariance matrix is the multivariate Gaussian, and it is thus optimal in a maximum-entropy sense. Covariance matrices can capture the relevant aspects of variability in a wide variety of scenarios Whittaker (1990), and the Gaussian doesn’t have to be the “true” underlying distribution for this111For a discussion covering this aspect of maximum-entropy approaches see van Nimwegen (2016). A more general maximum-entropy approach to compositional data has recently been proposed Weistuch et al. (2022), but it is difficult to obtain analytic results from it.. Often the limitations of data acquisition make good covariance estimates difficult, and regularisation techniques are adopted to deal with small data sets. One such technique is covariance _shrinkage_ , which is a regularisation technique that lets us invert singular covariance matrices which arise from the presence of fewer samples than variables Schäfer and Strimmer (2005). It has recently been applied in the context of _compositional_ data in genomics Badri et al. (2020), where clear improvements over empirical measures were shown with a comprehensive and reproducible benchmark. Compositional data are constrained in that their samples sum to a constant, which causes biases that can be alleviated by logratio transformations Aitchison (1986). Badri et al. (2020) applied the standard shrinkage approach directly to logratio covariance matrices, which does not take into account the interdependence among logratio variables. Our main interest in this contribution is to provide and test an improved shrinkage procedure that is custom-made for logratio-transformed data. The properly regularized logratio covariance matrices can then be inverted to estimate logratio-based partial correlations, which we have proposed as a measure of interaction between compositional variables Erb (2020). On a more general note, we aim to contribute to finding a principled way of quantifying interactions in data sets consisting of compositions or relative counts. When inferring interactions between variables, a useful distinction (see, e.g., Werhli et al. (2006)) is made between relevance networks, which are based on pairwise interaction coefficients (such as Pearson correlation) and graphical models, which make use of all the variables to infer pairwise association (e.g., via partial correlation). While the former are susceptible to inferring indirect interactions that are mediated by variables that are unrelated to the pair of variables in question, the latter aim to infer the conditional independence structure of their joint distribution Whittaker (1990). This way they can in principle infer direct interactions between two variables. The distinction is highly relevant for compositional data, where variables are intrinsically connected due to the unit-sum constraint. Here, simple relevance networks are inadequate because of the negative bias coming from the constraint. This bias has been addressed by measures that use ratios of variables, such as Aitchison’s logratio variance Aitchison (1986), as well as other coefficients inferring proportionality Lovell et al. (2015); Erb and Notredame (2016). These can be used to infer relevance networks in compositional data, and Badri et al. (2020) have shown how shrinkage improves estimates of the proportionality coefficient. However, to define graphical models for compositions similar to Kurtz et al. (2015), logratio-based partial correlations will be instrumental. In what follows, in section 2 we briefly review how the covariance structure and partial correlations are described in the logratio framework and outline James-Stein shrinkage of general covariance matrices as proposed in Schäfer and Strimmer (2005). In Section 3, the concept of logratio uncorrelated compositions introduced by Aitchison is described, and it is shown how one can make use of it to define shrinkage for logratio covariance matrices. Two equivalent approaches are proposed: 1) shrinking a so-called _basis_ covariance matrix which is then transformed into a logratio covariance, and 2) shrinking logratio covariance matrices directly with the help of logratio uncorrelated targets. Due to its simplicity, we prefer approach 1, and both the synthetic and the experimental data benchmark of Section 4 will be based on it. In section 5, we discuss how the use of partial correlations relates to normalisation techniques in genomics and provide an expression for the remaining partial correlation in logratio-uncorrelated compositions. A summary section and an Appendix containing some of the more technical details conclude the paper. ## 2\. Preliminaries ### 2.1. Compositional Data and Logratio Covariance Let us denote $N$ compositional samples by vectors $\boldsymbol{p}_{i}$, $i=1,\dots,N$, whose $D$ positive components (called parts) sum to 1. Specifically, we consider data with a large number of parts $D$ compared with the number of samples $N$. Such wide compositional data are of interest in genomics Erb et al. (2020), where they appear mainly in the form of relative counts coming from sequencing experiments Quinn et al. (2018). In this case we would consider $\boldsymbol{p}$ the closed data obtained when dividing a count sample by its total, i.e., the compositions are the empirical parameter estimates of a multinomial model. The counts play the role of a so-called _basis_ of the composition, a concept that was introduced in Aitchison (1982), and which we will talk about in more detail in section 2.4. Considering the counts as _relative_ data is justified by assigning no importance to the variations in the count totals (which correspond to the _size_ of the basis). Compared with real numbers, relative counts constitute a _discrete_ basis, which implies some problems of its own, especially when counts are small Lovell et al. (2020) or vanish entirely. Such problems can be alleviated with suitable weights Law et al. (2014) or with finite-size corrections to the frequencies that can again be obtained via shrinkage Hausser and Strimmer (2009). The latter, known as frequency shrinkage, has the advantage that it imputes count zeros naturally. The promise that this has shown for compositional data Quinn and Erb (2021); Greenacre et al. (2022) will be confirmed here in our benchmarks. Let us now consider an $N\times(D-1)$ data matrix $\boldsymbol{X}$ with real- valued elements. These elements are obtained from the log-ratios we evaluate from our compositional samples $\boldsymbol{p}_{i}$ using the additive logratio transformation (ALR, introduced in Aitchison (1982)) $x_{ij}=\log\frac{p_{ij}}{p_{iD}},\qquad j=1,\dots,D-1.$ (1) The (unbiased) sample covariance matrix $\boldsymbol{S}$ is then defined by its elements $s_{ij}=\frac{1}{N-1}\sum_{k=1}^{N}(x_{ki}-\bar{x}_{i})(x_{kj}-\bar{x}_{j}),$ (2) where $\bar{x}_{i}$ are the components of the $N$-dimensional vector $\bar{\boldsymbol{x}}$ of column means of $\boldsymbol{X}$. In order to have a criterion allowing us to judge the quality of our covariance estimate based on $\boldsymbol{S}$, we need a distributional assumption. Let us assume that $\boldsymbol{X}$ was sampled from a normal distribution with population parameters $\boldsymbol{\mu}$ and $\boldsymbol{\Sigma}$. This can be written as $\boldsymbol{x}\sim\mathcal{N}(\boldsymbol{\mu},\boldsymbol{\Sigma}),$ (3) where $\boldsymbol{x}$ denotes a $D-1$-dimensional random vector taking values as in the rows of $\boldsymbol{X}$. In this case the random compositions $\boldsymbol{p}$ are distributed according to Aitchison’s celebrated logistic normal distribution Aitchison (1982). If we denote the density of the compositions themselves by $f_{\mathcal{L}}$, the Gaussian density $f_{\mathcal{N}}$ will be multiplied by an additional factor due to the resulting Jacobian: $f_{\mathcal{L}}(\boldsymbol{p}=\boldsymbol{p}_{i}|\boldsymbol{\mu},\boldsymbol{\Sigma})=\left(\prod_{j=1}^{D}p_{ij}\right)^{-1}f_{\mathcal{N}}(\boldsymbol{x}=\boldsymbol{x}_{i}|\boldsymbol{\mu},\boldsymbol{\Sigma}),$ (4) where $\boldsymbol{x}_{i}$ is a function of $\boldsymbol{p}_{i}$ defined by (1). ### 2.2. Covariance Shrinkage To improve estimates of $\boldsymbol{\Sigma}$ from data, adding a bit of bias to the unbiased estimator $\boldsymbol{S}$ can reduce its overall error. This is a consequence of the bias-variance decomposition of the mean-squared error and can be achieved by convexly combining $\boldsymbol{S}$ with (a.k.a. “shrinking it towards”) a suitable target matrix $\boldsymbol{T}$: $\hat{\boldsymbol{\Sigma}}=\lambda\boldsymbol{T}+(1-\lambda)\boldsymbol{S},$ (5) where the shrinkage intensity $\lambda$ is between zero and one. $\hat{\boldsymbol{\Sigma}}$ is a so-called James-Stein type estimator James and Stein (1961). If we forget for a moment that $\boldsymbol{X}$ was obtained from logratios of compositions, a typical target matrix could be $\mathrm{diag}(\boldsymbol{S})$. While (5) tells us nothing about suitable values of $\lambda$, its optimal value can be estimated from the data via an analytic expression that minimises a mean-squared error cost-function with respect to the population covariance Ledoit and Wolf (2003). The derivation of this expression makes use of the fact that $\boldsymbol{S}$ is unbiased. This optimum is given by $\lambda^{*}=\frac{\sum_{j=1}^{D}\sum_{i\neq j}\left[\mathrm{var}(s_{ij})-\mathrm{cov}(s_{ij},\tau_{ij})\right]}{\sum_{j=1}^{D}\sum_{i\neq j}\mathbbm{E}\left[(s_{ij}-\tau_{ij})^{2}\right]},$ (6) where $\tau_{ij}$ denotes the matrix entries of $\boldsymbol{T}$. Note that here, covariance and expectation of the matrix elements refer to parameters of a distribution which can be approximated by their empirical estimates for actual calculations, see Schäfer and Strimmer (2005). To provide further background, we also show a principled interpretation of covariance shrinkage in the Appendix. There, we sketch how covariance shrinkage is equivalent to optimising the posterior probability of $\boldsymbol{\Sigma}$ from data and prior information. Note that $\lambda^{*}$ is inferred from data, making this a quasi-empirical Bayes approach. The shrunk covariance estimator is useful in itself as it often by far outperforms the sample covariance unless the number of samples considerably exceeds the number of variables. Regularisation becomes a necessity for inverting the covariance matrix whenever the number of samples is smaller than the number of variables. Therefore, shrinkage improves and extends the range of applications of the analysis proposed in the next section. ### 2.3. Logratio-based Partial Correlations Partial correlations between two variables are evaluated by controlling for the linear dependence on all the remaining variables, e.g., Whittaker (1990). This is achieved by correlating the residuals of two variables, where the residuals are with respect to the linear least squares predictors obtained from the remaining variables. Intuitively, we only correlate those contributions to our variables that are orthogonal to a subspace spanned by the variables we control for. Partial correlations can therefore be used to extract direct dependencies between variables, as opposed to the ones based on Pearson correlation, which can be dominated by indirect interactions via other variables. Partial correlations between logratio variables can be defined independently of their reference part $D$, so they measure association directly between parts, provided their reference is controlled for Erb (2020). Let us recall the formula for the (logratio-based) partial correlation between two parts $i,j$ of the random composition $\boldsymbol{p}$ via their ALR covariance $\boldsymbol{\Sigma}$. Below, the backslash means taking the set difference, so the variables we control for are all the ones not belonging to the pair in question: $r_{ij}(\boldsymbol{p}):=\mathrm{corr}(x_{i},x_{j}|\left\\{x_{1},\dots,x_{D-1}\right\\}\backslash\\{x_{i},x_{j}\\})=\frac{-\sigma^{(-1)}_{ij}}{\sqrt{\sigma^{(-1)}_{ii}\sigma^{(-1)}_{jj}}},$ (7) where $i\neq j$ and $i,j=1,\dots,D-1$, and where $\sigma^{(-1)}_{ij}$ denotes the elements of the inverse of $\boldsymbol{\Sigma}$. It may appear as if the reference part $p_{D}$ is “sacrificed” here, and its partial correlations cannot be determined. Note, however, that $r_{ij}$ is invariant under permutation of the reference part: all $D$ choices of $\boldsymbol{\Sigma}$ lead to the same partial correlations (for the parts they have in common). It follows that we can obtain the partial correlation of any pair of parts by choosing an arbitrary reference part that doesn’t form part of the pair. We have stated our definition in terms of the ALR covariance matrix because it comes closest to what is known from unconstrained data. There is also a (computationally) more convenient representation using a covariance matrix involving all parts in form of centred logratios (the CLR covariance introduced in section 2.4), from which the partial correlations can be obtained via its pseudoinverse. The inverse can also be estimated with a sparsity assumption, which has been done to create ecological networks of microbes, see Kurtz et al. (2015). ### 2.4. Linear Independence in Logratio and Basis Covariance Matrices For covariance matrices of unconstrained data, vanishing off-diagonal elements mean linear independence of the variables. Logratio covariance matrices cannot be linearly independent in this sense. Although logratio transformations map compositional data from the simplex to a real-valued space, they cannot remove the dependence between the parts that is introduced by the loss of one dimension. The notion coming closest to linear independence for logratios is given by logratio uncorrelated compositions. In the following, we state some results from Aitchison (1986), section 5.9, where this concept is introduced. A composition $\boldsymbol{p}$ is said to be logratio uncorrelated (LU) if $\mathrm{cov}\left(\log\frac{p_{i}}{p_{k}},\log\frac{p_{j}}{p_{l}}\right)=0$ (8) for every selection of _four different_ indices $i,j,k,l$ from $\\{1,\dots,D\\}$. This case is shown in Figure 1. Figure 1: Dependence structure in a logratio uncorrelated composition $\boldsymbol{u}$ with four parts. Transformed parts can be visualized as centred vectors with sampled components (labels are in random-variable notation), where length corresponds to standard deviation. The reference part is indexed by $D=4$ and its basis $m_{4}=tu_{4}$ is usually unknown (resulting in an unknown origin $O$). The vectors from the origin to the points indexed by 1 to 4 are the log-transformed basis vectors, their squared lengths correspond to the $\alpha_{k}$ defined in (9). As $\boldsymbol{u}$ is LU, they are orthogonal to each other in four dimensions. In three dimensions, only the logratio vectors of equal colour are orthogonal (e.g., the orange vectors labelled with their logratios). These correspond to vectors where all indices are different, see (8). The covariance structure of LU compositions can be made explicit when defining a $D$-component vector $\boldsymbol{\alpha}$ with elements $\alpha_{k}=\mathrm{cov}\left(\log\frac{p_{i}}{p_{k}},\log\frac{p_{j}}{p_{k}}\right),$ (9) with the three indices $i$, $j$, $k$ in $\\{1,\dots,D\\}$ and all different from each other. For LU compositions, this definition makes sense because the value of $\alpha_{k}$ is constant regardless of the (unequal) indices $i,j$. The matrix $\boldsymbol{\Sigma}$ of an LU composition is now given by $\sigma_{ij}=\left\\{\begin{array}[]{c@{\quad}l}\alpha_{i}+\alpha_{D}&\mbox{if $i=j$,}\\\ \alpha_{D}&\mbox{if $i\neq j$.}\end{array}\right.\qquad i,j=1,\dots,D-1.$ (10) (we reproduce the proof of this in the Appendix). While these definitions may seem overly abstract, they appear much more compelling in the light of Aitchison’s concept of the _basis_ of a composition, see Aitchison (1986), chapter 9. Intuitively, the basis is one of many ways our compositions can look like before the closure operation is applied. It is defined by $\boldsymbol{m}=t\boldsymbol{p}$ for some positive real $t$ which is called the size of the basis and for which $t=\sum_{k=1}^{D}m_{k}$ holds (implying that it can change for each data sample). Of course, infinitely many bases of different size exist for a given unit-sum composition. Now the _basis covariance_ matrix $\boldsymbol{\Omega}$ is defined to have the elements $\omega_{ij}=\mathrm{cov}\left(\log m_{i},\log m_{j}\right),$ (11) from which the covariance structure of a composition can be obtained unambiguously. If we assume that $\boldsymbol{\Omega}$ is diagonal, i.e., the logged components of the basis vector are mutually uncorrelated, we have the result that $\boldsymbol{\alpha}$ coincides with $\mathrm{var}(\log\boldsymbol{m})$: $\alpha_{i}=\omega_{ii},\qquad i=1,\dots,D.$ (12) This goes to show that $\boldsymbol{\alpha}$ can be seen as the variance vector of an independent distribution in a sample space that has one additional dimension (and in which the logratio sample space is embedded), see Figure 1. As $\boldsymbol{\alpha}$ was defined for logratios, this statement may appear confusing because of the apparent dependence of $\mathrm{var}\log(t\boldsymbol{p})$ on the basis size $t$. However, the $t$-dependence drops out because $\boldsymbol{\Omega}$ is diagonal (see the proof of (12) in the Appendix). For completeness, let us also mention an equivalent expression to (10) in terms of the centred logratio transformation (CLR). Here the reference involves all the parts of the composition. It transforms the compositional data matrix as $y_{ij}=\log\frac{p_{ij}}{g(\boldsymbol{p}_{i})},$ (13) where $g(\boldsymbol{p}_{i})=\prod_{k=1}^{D}p_{ik}^{1/D}$ is the geometric mean of the $i$-th composition. Note that the resulting data matrix $\boldsymbol{Y}$ has a constraint $\sum_{k=1}^{D}y_{ik}=0$ acting on its rows. This constraint leads to a singular CLR covariance matrix of rank $D-1$. The covariance of $\boldsymbol{y}$ is denoted by $\boldsymbol{\Gamma}$. The two forms of log-ratio covariance are equivalent and can be transformed into each other (see the first two rows in Table 1). Using the transformation from $\boldsymbol{\Sigma}$ to $\boldsymbol{\Gamma}$ on (10), we obtain the elements of $\boldsymbol{\Gamma}$ for an LU composition as $\gamma_{ij}=\left\\{\begin{array}[]{c@{\quad}l}\alpha_{i}-(2\alpha_{i}-\frac{1}{D}\sum_{k}\alpha_{k})/D&\mbox{if $i=j$,}\\\ -(\alpha_{i}+\alpha_{j}-\frac{1}{D}\sum_{k}\alpha_{k})/D&\mbox{if $i\neq j$.}\end{array}\right.\qquad i,j=1,\dots,D.$ (14) $\boldsymbol{\Sigma}\rightarrow\boldsymbol{\Gamma}$ | $\gamma_{ij}=\sigma_{ij}-\frac{1}{D}\sum_{k=1}^{D}\sigma_{ik}-\frac{1}{D}\sum_{k=1}^{D}\sigma_{kj}+\frac{1}{D^{2}}\sum_{k,l=1}^{D}\sigma_{kl}$ ---|--- $\boldsymbol{\Gamma}\rightarrow\boldsymbol{\Sigma}$ | $\sigma_{ij}=\gamma_{ij}-\gamma_{iD}-\gamma_{Dj}+\gamma_{DD}$ 3.5 $\boldsymbol{\Omega}\rightarrow\boldsymbol{\Sigma}$ | $\sigma_{ij}=\omega_{ij}-\omega_{iD}-\omega_{Dj}+\omega_{DD}$ $\boldsymbol{\Sigma}\rightarrow\boldsymbol{\Omega}$ | $\omega_{ij}=\gamma_{ij}(\boldsymbol{\Sigma})+\beta_{i}+\beta_{j}$, where | $\beta_{j}=\mathrm{cov}\left(\mathrm{clr}_{j}(\boldsymbol{p}),\log g(\boldsymbol{m})\right)-\frac{1}{2}\mathrm{var}\left(\log g(\boldsymbol{m})\right)$. $\boldsymbol{\Omega}\rightarrow\boldsymbol{\Gamma}$ | $\gamma_{ij}=\omega_{ij}-\frac{1}{D}\sum_{k=1}^{D}\omega_{ik}-\frac{1}{D}\sum_{k=1}^{D}\omega_{kj}+\frac{1}{D^{2}}\sum_{k,l=1}^{D}\omega_{kl}$ $\boldsymbol{\Gamma}\rightarrow\boldsymbol{\Omega}$ | $\omega_{ij}=\gamma_{ij}+\beta_{i}+\beta_{j}$, see $\boldsymbol{\Sigma}\rightarrow\boldsymbol{\Omega}$. Table 1: Elementwise transformations between ALR, CLR, and basis covariance matrices (collected from Aitchison (1986)). Here, indices of $\sigma_{ij}$ are allowed to take the value $D$, in which case $\sigma_{ij}$ vanishes; clrj stands for the $j$-th element of the CLR transform, and $g$ denotes the geometric mean. All these transformations can also be expressed as simple matrix operations that are not shown here. ## 3\. Logratio Covariance Shrinkage The specific covariance structure of compositional data would suggest to use log-ratio uncorrelated shrinkage targets instead of the diagonal targets used for unconstrained data. We will show that it is straightforward to construct such targets. However, it is usually more convenient in practice to work with diagonal targets, for which we have to do the shrinkage on a basis of the compositions. Once the shrinkage estimate for the basis covariance is obtained, it can then be back-transformed to a logratio covariance matrix. This strategy is described in the following section. ### 3.1. Basis Covariance Shrinkage Using Diagonal Targets A diagonal shrinkage target may often be preferred for practical reasons. As an example, the corpcor R package Schäfer and Strimmer (2005) uses the diagonal of the sample covariance as a target and does not allow for alternative targets. However, we have seen that logratio covariance matrices do not allow for this simple independence structure. Therefore, we now define the shrinkage estimator directly for the _basis_ covariance matrix $\boldsymbol{\Omega}$: $\hat{\boldsymbol{\Omega}}=\lambda\boldsymbol{T}_{C}+(1-\lambda)\boldsymbol{C},$ (15) where $\boldsymbol{C}$ is the empirical basis covariance matrix, and the shrinkage target $\boldsymbol{T}_{C}$ could be simply its diagonal. Thus, given some compositional data, our first step to construct an uncorrelated shrinkage target is to obtain an empirical basis covariance matrix from the data. For this, it is necessary to fix the size of the basis $t$. Note that doing this specifies the variability of the totals across samples. Given that we will back-transform our estimator to a logratio covariance matrix, this variability is irrelevant. The logratio covariance structure will not contain information about the basis size, so it is convenient to use $t_{i}=1$ for all samples $i=1,\dots,N$, giving a basis $\boldsymbol{m}_{i}=\log\boldsymbol{p}_{i}$. In the case of relative count data, the counts themselves can provide the basis, and $\boldsymbol{C}$ can be estimated from their logarithms (provided zeros are taken care of). Yet another possibility is to start from an empirical logratio covariance matrix and obtain $\boldsymbol{C}$ via a transformation given in Table 1. These transformations simplify for a constant basis size. The diagonal target has the additional advantage that the expression for the optimal value of $\lambda$ simplifies Schäfer and Strimmer (2005) compared to the more general expression (6). For diagonal targets we have $\hat{\lambda}^{*}=\frac{\sum_{j=1}^{D}\sum_{i\neq j}\widehat{\mathrm{var}}(c_{ij})}{\sum_{j=1}^{D}\sum_{i\neq j}c_{ij}^{2}}.$ (16) To obtain the shrinkage estimator of a logratio covariance matrix, all we have to do now is back-transform $\hat{\boldsymbol{\Omega}}$ to an appropriate logratio covariance matrix as specified in Table 1. This will also be the strategy we use for our benchmark in section 4. ### 3.2. Logratio Uncorrelated Targets While we will not use them in our benchmarks, it can be of general interest to construct LU targets for direct logratio covariance shrinkage. A simple way of constructing these targets would be as follows: pick the diagonal elements of the empirical basis covariance $\boldsymbol{C}$ as our $\alpha_{i}$ and then obtain the ALR target covariance by (10) and the CLR target covariance by (14). However, we can also skip the construction of a basis covariance entirely and express the targets directly in terms of the logratio sample covariance matrices. For the ALR target given the empirical covariance $\boldsymbol{S}$ we find $\tau_{ij}=\left\\{\begin{array}[]{c@{\quad}l}s_{ii}-\frac{2}{D}\sum_{k=1}^{D-1}s_{ik}+\frac{2}{D^{2}}\sum_{k,l}s_{kl}&\mbox{if $i=j$,}\\\ \frac{2}{D^{2}}\sum_{k,l}s_{kl}&\mbox{if $i\neq j$.}\end{array}\right.\quad i,j=1,\dots,D-1.$ (17) Similarly, using an empirical CLR covariance matrix $\boldsymbol{G}$, the corresponding target elements $t_{ij}$ are given via the diagonal elements of $\boldsymbol{G}$ by $t_{ij}=\left\\{\begin{array}[]{c@{\quad}l}g_{ii}-(2g_{ii}-\frac{1}{D}\sum_{k}g_{kk})/D&\mbox{if $i=j$,}\\\ -(g_{ii}+g_{jj}-\frac{1}{D}\sum_{k}g_{kk})/D&\mbox{if $i\neq j$.}\end{array}\right.\qquad i,j=1,\dots,D.$ (18) ## 4\. A Benchmark of Logratio Covariance Shrinkage We verify our estimators of covariance and partial correlation on both simulated data as well as single-cell gene expression data. The latter are subsampled from a “ground truth” of greater sample size (this strategy has also been used for microbiome data in Badri et al. (2020)). Throughout, the corpcor R package from Schäfer and Strimmer (2005) is used for covariance matrix shrinkage. Note that we use the default implemented there, which consists in separately shrinking the diagonal elements of the covariance matrix with another $\lambda$ parameter, see Opgen-Rhein and Strimmer (2007). The aim of the benchmark is the evaluation of our proposed procedure to first shrink a basis covariance and then transform the result to a log-ratio covariance. This is compared with a naive approach of direct log-ratio covariance shrinkage (with a diagonal target containing variances that inevitably mix parts) as well as no shrinkage at all. In the first part of this section, the logratios of our compositions are taken for granted and we just sample them from a normal distribution of rather moderate dimension ($D=40$). In this case we do not have to be concerned about zeros in the basis samples. A scenario typical for genomics data sets is treated in the second part of this section, where the basis of our compositions are high dimensional relative counts taken from single-cell gene expression data ($D=500$). There we also test various ways of dealing with zeros as well as a finite-size correction for low counts in form of frequency shrinkage. Figure 2: Mean squared error of partial correlation matrix for sample sizes $N=200$, $N=40$, and $N=8$. Colours indicate different estimation procedures (no shrinkage, naive shrinkage of CLR / ALR covariance and basis covariance shrinkage). Each boxplot contains the results of 200 simulations. Whenever estimates without shrinkage are not shown, their median value is given in the legend. ### 4.1. Synthetic Logistic Normal Data To have reasonably realistic population parameters for the normal distribution, they are inferred from a single-cell gene expression data set Riba et al. (2022), where a subset of 240 genes that are non-zero everywhere across 5637 cells are selected. From these, in each of the 200 repetitions of the simulation, in step 1) $D=40$ parts are chosen randomly, from which ALR and CLR covariance matrices as well as the partial correlation matrix (via the pseudoinverse of the CLR covariance) are constructed across the 5K+ cells. This is our ground truth. Then 2) the ALR mean vector $\boldsymbol{\mu}$ and the ALR covariance matrix $\boldsymbol{\Sigma}$ are used to produce $N$ multivariate normal samples, where $N=8$, 40, and 200. 3) The $N$ samples are backtransformed to compositions. From these, sample estimates of CLR, ALR and (constant-size) basis covariance matrices as well as the partial correlation estimates are produced. This is done in three different ways: without shrinkage, with direct shrinkage of the logratio covariance matrices (using a diagonal target for the logratios), and via shrinkage of the basis covariance matrix (i.e., using a diagonal target for the log-transformed basis). 4) These sample estimates are compared to the ground truth established in the first step using an element-wise mean squared error. Note that after basis covariance shrinkage, partial correlations are identical if the basis is transformed to an ALR or a CLR covariance matrix (but it is convenient to use the CLR to obtain them). The results of this procedure are seen in Figure 2 for partial correlation and Supplementary Figure 6 for covariance. There is no doubt that shrinkage leads to substantial improvements when estimating covariation, even for $N=5D$. In this regime, naive CLR shrinkage fares worse than using no shrinkage at all when estimating partial correlations. Interestingly, the error is small for the covariance matrix itself but seems to be compounded for inference of its inverse. In all cases we see a clear advantage when using shrinkage of the basis covariance, especially when comparing with naive ALR covariance shrinkage. We conclude that we achieve substantial improvements of our estimators when tailoring the shrinkage to the needs of logratio covariance matrices. Figure 3: Benchmark procedure for single-cell gene expression data. ### 4.2. Single-cell gene expression data Instead of simulating samples from a (realistic) covariance matrix, we now draw our covariance matrices directly from the data. Since these are sample covariance matrices, the benchmark now consists in recovering them as well as possible with undersampled data. While this is not ideal (the test case could in principle be as close or closer to the population covariance than the sample covariance that serves as ground truth), it seems a reasonable assumption that the degradation effect due to undersampling is greater than the deviation of the full sample covariance from the population covariance. In the previous section we could see that shrinking the covariance still has positive effects even for $N=5D$. To avoid circularity, however, we prefer to not shrink the ground truth covariance matrix here. Figure 4: Mean squared error of partial correlation matrices computed on three different sizes of subsamples ($N$=2500, 1000, 100) of single-cell gene expression data. Boxplots contain 200 samples each using different estimation procedures (no shrinkage, naive ALR/CLR shrinkage, basis covariance shrinkage). Their colour indicates the type of zero imputation used. Our benchmark procedure is summarized in Figure 3. Single-cell gene expression data usually contain a very large number of zero counts, which would require strategies (e.g., pooling several cells into a single sample) that are not explored here. Instead, we concentrate on a core set of genes that are highly expressed and thus yield only a few zero counts. These can still be used to test a number of imputation strategies. We test two basic procedures implemented in the R package zCompositions Palarea-Albaladejo et al. (2015) (Count Zero Multiplicative replacement, or CZM, and Geometric Bayesian Multiplicative replacement, or GBM), as well as one more sophisticated strategy that takes covariance structure into account (lrSVD), which was added to zCompositions recently Palarea-Albaladejo et al. (2022). We also test frequency shrinkage as implemented in the R package entropy Hausser and Strimmer (2009). To separate the effects of shrinkage and imputation, we defined two scenarios: 1) imputation is avoided by using a large subset of cells with only nonzero counts, and 2) the effect of shrinkage and imputation are jointly evaluated by including all the cells. More precisely, we used 770 genes for which 3986 cells (out of a total of 5637 cells) have no zeros, from which a ground-truth can be constructed. For each of the 200 resamplings, 500 genes are chosen randomly from the 770 genes. Covariance and partial correlation matrices computed on the 3986 $\times$ 500 nonzero matrix define our ground truth. Then, in scenario 1, this matrix is subsampled to have $N$=2500, $N$=1000, and $N$=100 cells. Covariance and partial correlations computed on each of these matrices, and under the different shrinkage schemes are compared to the ground truth using mean squared error. In the second scenario, the same ground truth as defined before is used, while the subsampled data ($N$=2500, $N$=1000, and $N$=100) are drawn from the extended matrix of 5637 cells x 500 genes. Here, on average about 1500 zeros are imputed for the 2500 samples, 600 zeros are imputed for the 1000 samples and 60 for the 100 samples. Having additional samples that contain zeros is not ideal because these samples could in principle induce a slightly different covariance structure. The benefit, however, is that it provides a natural set of zero counts when otherwise zeros would have to be artificially introduced. Now the subsampled matrix is first imputed using four different methods, and then covariance and partial correlations are again computed under the different shrinkage schemes. The results for partial correlations are shown in Figure 4, and for the covariance matrices in Supplementary Figure 7. Two main conclusions can be drawn from the results for such high-dimensional data. First, the effect of using the basis covariance shrinkage is strong compared with naive ALR covariance shrinkage, but it is much weaker when comparing with naive CLR covariance shrinkage. Naive shrinkage of the CLR covariance only leads to an important deterioration for the partial correlations at higher sample sizes, in concordance with what we observed in the synthetic benchmark. Apart from that, the improvement is mainly observed when no zeros are imputed, which hints at a flattening of the effect caused by the imputation. The advantage of basis covariance shrinkage becomes entirely invisible when directly evaluating the covariance matrices, whose inference seems to be less prone to error than their inverse. A second main conclusion is that imputation of zeros leads to considerable increase in mean squared error. All imputation schemes lead to a comparable loss in accuracy, with only slight differences between methods. ## 5\. Logratio-based Partial Correlations for Wide Data We have addressed some questions regarding covariance structure that remained unanswered in Erb (2020), especially with respect to the application to wide data sets like the ones occurring in genomics. We think it is beneficial to also discuss some questions regarding the interpretation of logratio-based partial correlations both in CoDA and genomics. In genomics, partial correlation analysis is common, but it has rarely been based on logratio analysis Kurtz et al. (2015). In CoDA, on the other hand, logratio analysis is the dominant paradigm, but partial correlations have been introduced quite recently Erb (2020). In the following sections we will discuss the relationship between classical partial correlations and their logratio counterparts. ### 5.1. Normalisations to Avoid Compositional Bias We can think of two strategies to avoid compositional bias when analyzing covariance structure: trying to recover the original (absolute) signal (as contained in one particular basis covariance $\boldsymbol{\Omega}$) or using only that part of the signal that absolute and relative data have in common (which is contained in the logratio covariance). The first (“normalisation”) strategy relies on assumptions because the closure operation discards a part of the signal and to recover the original totals up to a constant factor, additional information is needed. The second (“CoDA”) strategy voluntarily discards this information and doing so introduces dependencies between the variables, but it is assumption-free. Interestingly, the distinction between the two approaches is not all that clear-cut because the logratio transformations Aitchison introduced can be used to pursue the “normalisation” strategy of re-opening the data with additional assumptions. This was discussed in (Quinn et al., 2018), see especially the Supplementary material. In the following, we revisit these arguments from the perspective of logratio covariance matrices and their basis covariance $\boldsymbol{\Omega}$. Let us start with the ALR covariance $\boldsymbol{\Sigma}$. It can be obtained from the basis covariance by $\sigma_{ij}=\omega_{ij}-\omega_{iD}-\omega_{jD}+\omega_{DD}$ (19) (see Table 1). Note that if the reference variable $m_{D}$ remains unchanged across samples in the basis, the three last terms on the right-hand side vanish, and the two covariances coincide. This is the reason behind the widespread normalisation procedure in genomics where (absolutely) unchanged reference genes are used to “normalize” the data to avoid compositional bias. Similarly, for the CLR covariance $\boldsymbol{\Gamma}$ we have the relationship $\omega_{ij}=\gamma_{ij}+\beta_{i}+\beta_{j}$ (20) (see again Table 1). From the dependence of $\beta_{i}$ on $g(\boldsymbol{m})$ we see that this implies that if the geometric mean of the basis remains unchanged across samples, the $\beta$ terms vanish and the CLR covariance is identical with the basis covariance. This is in agreement with the notion of effective library size normalisation, where a reference that remains unchanged on the absolute data is constructed to avoid compositional bias. The geometric mean can remain unchanged across samples before closure when the majority of genes is not subject to systematic change. Under this assumption the CLR transformed data have an identical correlation structure as the log- transformed absolute data. We can see that the CoDA strategy has the advantage that, while it is on the safe side if the assumptions aren’t met, it also recovers the original signal if they happen to be fulfilled. In the latter case, however, a specific reference and thus logratio covariance would be picked out because it recovers the basis covariance. While this is necessary for evaluating both covariance and correlation, it is unnecessary for partial correlation. The results remain invariant under permuting the reference in $\boldsymbol{\Sigma}$ and they are also identical when using the pseudoinverse of $\boldsymbol{\Gamma}$ for their evaluation, see also Erb (2020). These benefits of logratio-based partial correlation are lost when using more naive approaches, like deriving partial correlations from the covariance matrix of $\log\boldsymbol{p}$, or from a relative count basis using $\log\boldsymbol{m}$ directly. Formally, partial correlations derived when not specifying a reference do not coincide with their logratio counterparts. Figure 5: A worst-case scenario for spurious logratio-based partial correlation of a fixed pair of parts in artificially uncorrelated data (see text). Successive increase of the number of parts from three (PC=0.69) to 770 (PC=0.001) shows how the closure-induced spurious signal tends to zero with the number of parts. Inset: The case $D=3$. The vanishing correlation between the first two parts could be recovered from the logratios with respect to the third part (indicated in red) by correlating their projections into the plane that is orthogonal to $\log m_{3}$. This would just be the correlation between residuals when controlling for $\log m_{3}$. However, the information about $\log m_{3}$ is lost, so partial correlation between the logratios remains biased. ### 5.2. What is the effect of the closure? Partial correlations based on logratios can emulate their absolute counterparts extremely well (a benchmark of which will be shown in another contribution). However, the closure operation makes us lose one dimension which cannot be recovered. So there remains a difference between partial correlations derived from absolute (using logarithms only) versus relative data (using logratios). Or, to put it differently, partial correlations can change when we derive them from a logratio covariance as opposed to the original (usually unknown) basis covariance. Using the relation between $\boldsymbol{\Omega}$ and $\boldsymbol{\Gamma}$ of Table 1, the respective covariance inverses are related by $\boldsymbol{\Omega}^{-1}=\left(\boldsymbol{\Gamma}+\boldsymbol{B}\right)^{-1},$ (21) where the $D\times D$ matrix $B$ has elements $\beta_{i}+\beta_{j}$. We see that we can neatly separate the covariance signal introduced by the variability of the basis size in the matrix $\boldsymbol{B}$, but taking the inverse will mix these signals again. To further simplify the relationship of the inverses, we need a more specific expression for the covariance. A simple way to extract the partial correlation that is only due to the closure operation is to consider a diagonal basis covariance, as it has no partial correlation by definition. We know from section 2.4 that linear independence corresponds to a logratio uncorrelated (LU) covariance in the embedded logratio sample space (where one dimension is lost). In other words, the closure applied to an uncorrelated basis leads to the dependence structure of LU compositions222Dirichlet compositions are LU [Aitchison 1986], and this (often undesired) property gave rise to Aitchsion’s introduction of the logistic normal. The richer covariance structure of the latter enables modelling that goes beyond unit-sum induced interactions. Of course, a logistic normal with an LU covariance matrix also models LU compositions.. The elements of $\boldsymbol{\Sigma}$ for an LU composition are given by (10) and (9). The fact that $\boldsymbol{\Sigma}$ has nonzero off-diagonal elements for LU compositions immediately shows that the partial correlations of LU compositions do not vanish. It is not difficult to determine the precise expression. For $\boldsymbol{u}$ an LU composition, its (logratio-based) partial correlations are given by $r_{ij}(\boldsymbol{u})=\sqrt{\frac{\alpha^{-1}_{i}\alpha^{-1}_{j}}{\left(\sum_{k\neq i}\alpha^{-1}_{k}\right)\left(\sum_{k\neq j}\alpha^{-1}_{k}\right)}},$ (22) for all $i\neq j$, and where $\boldsymbol{\alpha}$ is given by (9) (see Appendix for a proof). To get an idea how strong this “spurious” signal is, we make the following experiment. To obtain an (artificial) system of linearly independent variables, we take the diagonal of the basis covariance matrix of the 770 genes considered in the previous section. This diagonal corresponds to $\boldsymbol{\alpha}$, and we can calculate the logratio-based partial correlations between all pairs using (22). For a worst-case scenario, we select the pair of parts for which we obtain the strongest partial correlation. With this pair fixed, we repeat calculating its partial correlation after eliminating an arbitrary part (and thus one component from the vector $\boldsymbol{\alpha}$). We repeat this process until we have only one additional part left with our original pair. The result of this process is shown in Figure 5 (going from right to left along the horizontal axis). We see a similar “dilution” of the compositional effect as for the negative correlations between multinomial counts, see Greenacre et al. (2022). In our case, the spurious correlation turns out to be positive, but it is evaluated between residuals of logratios, not compositional parts. ## 6\. Conclusion We have proposed a simple way of obtaining shrinkage-based estimators for log- ratio covariance matrices and their inverse. As shown in our benchmark, these estimators clearly improve on the (still widely used) empirical estimators and also outperform estimators that shrink log-ratio covariance matrices towards a diagonal target as proposed in Badri et al. (2020). The use of Aitchison’s notion of log-ratio uncorrelated compositions is crucial to obtain the improved estimates. With our contribution we also aim to promote the use of partial correlations for the analysis of compositional data sets. We see three advantages in using them: 1) They take all parts into account for the evaluation of pairwise interaction (and thus can factor out indirect interactions). 2) Their evaluation is reference-independent. 3) Negative correlations can be meaningfully evaluated with respect to the set of variables that are controlled for. While partial correlations change under taking subcompositions, this is expected because it reflects the effect of the removal of variables that are partialled out. However, we have also shown that a “spurious” positive signal remains that is due to the loss of the $D$-th dimension in compositional data sets. We have derived the expression for this remaining partial correlation induced by the closure operation and have shown how it “dilutes out” with a growing number of parts. In this sense, partial correlations between parts are expected to show little spurious signal whenever there are sufficiently many parts, a common occurrence in genomic data. For genomic data sets, we have also drawn some connections with the problem of their normalisation. Sequencing data are relative counts, and as such their analysis within the scale-free logratio framework can be beneficial. The use of various normalisation schemes to control for sequencing depth can essentially be circumvented in this way. We see much promise in the use of logratio-based partial correlations especially for situations when normalisation assumptions fail. ## 7\. Appendix ### 7.1. Bayesian Interpretation of Covariance Shrinkage In the Bayesian view of statistical inference, the parameters of a distribution are considered to follow distributions themselves. The posterior distribution of the parameter in question incorporates information from both the data (via the likelihood) and the prior distribution of the parameter. Shrinkage estimators also can incorporate prior information, and while a suitable empirical estimator can maximise likelihood, optimization of shrinkage is related to maximizing the posterior of the parameter333As an example, in the case of shrinkage estimates of multinomial frequencies, the target frequencies are renormalized pseudocounts to the count data, and the shrinkage intensity $\lambda$ is the prior sample size of the conjugate Dirichlet prior that has the pseudocounts as its parameters Hausser and Strimmer (2009).. In the following we want to make this more precise for Gaussian covariance shrinkage. Let us start with the Gaussian conjugate prior distribution of the precision matrix $\boldsymbol{\Sigma}^{-1}$ in the case of a fixed mean $\boldsymbol{\mu}$. This is known as the Wishart (or multivariate Gamma) distribution with density $f_{\mathcal{W}}(\boldsymbol{\Sigma}^{-1}|\boldsymbol{V},\nu)=\frac{|\boldsymbol{\Sigma}^{-1}|^{(\nu-D)/2}e^{-\frac{1}{2}\mathrm{tr}\left(\boldsymbol{\boldsymbol{V}^{-1}\Sigma}^{-1}\right)}}{2^{\nu(D-1)/2}|\boldsymbol{V}|^{\nu/2}\Gamma_{D-1}(\frac{\nu}{2})},$ (23) where the number of degrees of freedom $\nu$ is a positive integer, the parametric matrix $\boldsymbol{V}$ is positive definite of order $(D-1)\times(D-1)$, and $\Gamma$ denotes the multivariate Gamma function. The particular form of this density can be understood better when decomposing the precision matrix as $\boldsymbol{\Sigma}^{-1}=\boldsymbol{U}\boldsymbol{U}^{T},$ (24) where $\boldsymbol{U}$ is of order $(D-1)\times\nu$. If the $\nu$ vectors in $\boldsymbol{U}$ are independently drawn from a normal distribution with mean zero and covariance $\boldsymbol{V}$ (where $\nu\geq D-1$), then the precision matrix follows a Wishart distribution with density (23). To estimate a Gaussian covariance matrix, we need more than a single sample of data. Let us now consider a data matrix $\boldsymbol{X}$ where the $N$ rows were sampled according to (3). The joint likelihood of these samples can be written in terms of their sample parameter estimates $\boldsymbol{S}$ and $\bar{\boldsymbol{x}}$ as $f_{\mathcal{N}}(\boldsymbol{X}|\boldsymbol{\mu},\boldsymbol{\Sigma})=\prod_{i=1}^{N}f_{\mathcal{N}}(\boldsymbol{x}_{i}|\boldsymbol{\mu},\boldsymbol{\Sigma})=\\\ \left(\frac{(2\pi)^{D-1}}{|\boldsymbol{\Sigma}|}\right)^{\frac{N}{2}}\mathrm{exp}\left\\{-\frac{1}{2}\mathrm{tr}\left[\boldsymbol{\Sigma}^{-1}\left((N-1)\boldsymbol{S}+N(\bar{\boldsymbol{x}}-\boldsymbol{\mu})^{T}(\bar{\boldsymbol{x}}-\boldsymbol{\mu})\right)\right]\right\\}.$ (25) Let us now assign the conjugate priors as follows: $\displaystyle\boldsymbol{\mu}\leavevmode\nobreak\ |\leavevmode\nobreak\ \boldsymbol{\Sigma}^{-1}$ $\displaystyle\sim$ $\displaystyle\mathcal{N}\left(\boldsymbol{\mu}_{0},\frac{1}{\kappa}\boldsymbol{\Sigma}\right),$ (26) $\displaystyle\boldsymbol{\Sigma}^{-1}$ $\displaystyle\sim$ $\displaystyle\mathcal{W}(\boldsymbol{V},\nu).$ (27) The joint posterior density then factorizes as $P(\boldsymbol{\mu}|\boldsymbol{\Sigma}^{-1})P(\boldsymbol{\Sigma}^{-1})$, where the marginal $P(\boldsymbol{\Sigma}^{-1})$ is again Wishart. More precisely $P(\boldsymbol{\Sigma}^{-1}|\boldsymbol{S},\bar{\boldsymbol{x}},\boldsymbol{\mu}_{0},\kappa,\boldsymbol{V},\nu)=f_{\mathcal{W}}(\boldsymbol{\Sigma}^{-1}|\boldsymbol{V}^{*},\nu+N),$ (28) where $\boldsymbol{V}^{*}=(N-1)\boldsymbol{S}+\boldsymbol{V}+\frac{\kappa N}{\kappa+N}(\bar{\boldsymbol{x}}-\boldsymbol{\mu}_{0})^{T}(\bar{\boldsymbol{x}}-\boldsymbol{\mu}_{0})$ (29) (see DeGroot (2004), p.178)444For the sake of completeness, it should perhaps be mentioned that $P(\boldsymbol{\mu}|\boldsymbol{\Sigma}^{-1})=\\\ f_{\mathcal{N}}\left((\kappa\boldsymbol{\mu}_{0}+N\bar{\boldsymbol{x}})/(\kappa+N),\boldsymbol{\Sigma}/(\kappa+N)\right)$.. It follows that under the Bayesian model, an improved covariance estimator will involve an additive correction to the sample covariance. We can now write (29) in terms of the shrinkage estimator when taking into account the posterior sample size: $(\nu+N)\hat{\boldsymbol{\Sigma}}=\boldsymbol{V}^{*}.$ (30) Identifying (30) with (5), we obtain the identities $\displaystyle(1-\lambda)$ $\displaystyle=$ $\displaystyle(N-1)/(\nu+N),$ (31) $\displaystyle\lambda\boldsymbol{T}$ $\displaystyle=$ $\displaystyle\frac{1}{\nu+N}\left(\boldsymbol{V}+\frac{\kappa N}{\kappa+N}(\bar{\boldsymbol{x}}-\boldsymbol{\mu}_{0})^{T}(\bar{\boldsymbol{x}}-\boldsymbol{\mu}_{0})\right),$ (32) from which it follows that $\displaystyle\lambda$ $\displaystyle=$ $\displaystyle\frac{\nu+1}{\nu+N},$ (33) $\displaystyle\boldsymbol{T}$ $\displaystyle=$ $\displaystyle\frac{1}{\nu+1}\left(\boldsymbol{V}+\frac{\kappa N}{\kappa+N}(\bar{\boldsymbol{x}}-\boldsymbol{\mu}_{0})^{T}(\bar{\boldsymbol{x}}-\boldsymbol{\mu}_{0})\right).$ (34) Optimization of $\lambda$ thus boils down to a tuning of the prior degrees of freedom $\nu$ relative to the sample size $N$, while the target $\boldsymbol{T}$ defines a prior covariance structure. ### 7.2. Proof of (10) for LU compositions We start with the case $i\neq j$ (see [Aitchison 1986], section 5.9): $\sigma_{ij}=\mathrm{cov}\left(\log\frac{p_{i}}{p_{D}},\log\frac{p_{j}}{p_{D}}\right)=\mathrm{cov}\left(\log\frac{p_{i}}{p_{k}}+\log\frac{p_{k}}{p_{D}},\log\frac{p_{j}}{p_{D}}\right)\\\ \ =0+\sigma_{kj}$ (35) because of (8). It is therefore clear that $\sigma_{ij}$ is constant for unequal indices and the definition (9) makes sense for LU compositions. So $\sigma_{ij}=\alpha_{D}$ if $i\neq j$. Similarly, in the case $i=j$ we have $\sigma_{ii}=\mathrm{var}\left(\log\frac{p_{i}}{p_{D}}\right)\\\ =\mathrm{cov}\left(\log\frac{p_{i}}{p_{D}},\log\frac{p_{i}}{p_{D}}\right)=\mathrm{cov}\left(\log\frac{p_{i}}{p_{k}}+\log\frac{p_{k}}{p_{D}},\log\frac{p_{i}}{p_{D}}\right)\\\ =\mathrm{cov}\left(\log\frac{p_{k}}{p_{i}},\log\frac{p_{D}}{p_{i}}\right)+\mathrm{cov}\left(\log\frac{p_{k}}{p_{D}},\log\frac{p_{i}}{p_{D}}\right)=\alpha_{i}+\alpha_{D},$ (36) which concludes the proof. ### 7.3. Proof of (12) for an Uncorrelated Basis Let the composition $\boldsymbol{u}$ have a basis $t\boldsymbol{u}$ with $\mathrm{cov}(\log(tu_{i}),\log(tu_{j}))=0$ for all $i\neq j$. We show that it follows that $\mathrm{var}\left(\log(tu_{k})\right)=\mathrm{cov}\left(\log\frac{u_{i}}{u_{k}},\log\frac{u_{j}}{u_{k}}\right)$ (37) where none of the indices $i$, $j$, $k$ are equal. We write the left-hand side as $\mathrm{cov}\left(\log(tu_{k}),\log(tu_{k})\right)=\mathrm{cov}\left(\log\frac{tu_{k}}{tu_{i}}+\log(tu_{i}),\log\frac{tu_{k}}{tu_{j}}+\log(tu_{j})\right)\\\ =\mathrm{cov}\left(\log\frac{tu_{k}}{tu_{i}},\log\frac{tu_{k}}{tu_{j}}\right)+\mathrm{cov}\left(\log\frac{tu_{k}}{tu_{i}},\log(tu_{j})\right)\\\ +\mathrm{cov}\left(\log(tu_{i}),\log\frac{tu_{k}}{tu_{j}}\right)+\mathrm{cov}\left(\log(tu_{i}),\log(tu_{j})\right)\\\ =\mathrm{cov}\left(\log\frac{u_{i}}{u_{k}},\log\frac{u_{j}}{u_{k}}\right)+\mathrm{cov}\left(\log(tu_{k})-\log(tu_{i}),\log(tu_{j})\right)+\dots{}$ (38) The terms after the first one are all zero because of the condition that $\mathrm{cov}(\log(tu_{i}),\log(tu_{j}))=0$ for any indices $i\neq j$, which concludes the proof. ### 7.4. Proof of (22) for LU compositions Let $\boldsymbol{\Sigma}$ be given by the elements (10). The following are expressions for its determinant (here given without proof): $|\boldsymbol{\Sigma}|=\prod_{i=1}^{D}\alpha_{i}\sum_{i=1}^{D}\frac{1}{\alpha_{i}}=\sum_{i=1}^{D}\prod_{k\neq i}\alpha_{k},$ (39) where the second equality is found by evaluating the greatest common denominator of the $\alpha_{i}^{-1}$. We show that the matrix inverse of $\boldsymbol{\Sigma}$ is given by the elements $\sigma_{ij}^{(-1)}=\frac{1}{\sum_{k=1}^{D}\prod_{l\neq k}\alpha_{l}}\left\\{\begin{array}[]{c@{\quad}l}\sum_{k\neq i}\prod_{l\neq k,i}\alpha_{l}&\mbox{if $i=j$,}\\\ -\prod_{k\neq i,j}\alpha_{k}&\mbox{if $i\neq j$.}\end{array}\right.\quad i,j=1,\dots,D-1.$ (40) That this is the inverse can be easily proven when multiplying the resulting matrix with the matrix whose elements are given by (10). The off-diagonal elements (where $i\neq j$) of the result are $\sum_{m=1}^{D-1}\sigma_{im}\sigma^{(-1)}_{mj}=\sum_{m\neq i,j,D}\sigma_{im}\sigma_{mj}^{(-1)}+\sigma_{ii}\sigma_{ij}^{(-1)}+\sigma_{ij}\sigma_{jj}^{(-1)}=\\\ \frac{1}{\sum_{k=1}^{D}\prod_{l\neq k}\alpha_{l}}\left[-\alpha_{D}\sum_{m\neq i,j,D}\prod_{k\neq m,j}\alpha_{k}-(\alpha_{i}+\alpha_{D})\prod_{k\neq i,j}\alpha_{k}+\alpha_{D}\sum_{k\neq j}\prod_{l\neq k,j}\alpha_{l}\right]$ (41) The square bracket evaluates to $-\alpha_{D}\sum_{m\neq j,D}\prod_{k\neq m,j}\alpha_{k}-\alpha_{i}\prod_{k\neq i,j}\alpha_{k}+\alpha_{D}\sum_{k\neq j}\prod_{l\neq k,j}\alpha_{l}\\\ =-\alpha_{D}\sum_{m\neq j,D}\prod_{k\neq m,j}\alpha_{k}-\alpha_{i}\prod_{k\neq i,j}\alpha_{k}+\alpha_{D}\sum_{k\neq j,D}\prod_{l\neq k,j}\alpha_{l}+\alpha_{D}\prod_{l\neq D,j}\alpha_{l}=0.$ (42) Thus the off-diagonal elements vanish. For the diagonal elements we have $\sum_{m=1}^{D-1}\sigma_{im}\sigma^{(-1)}_{mi}=\sum_{m\neq i,D}\sigma_{im}\sigma_{mi}^{(-1)}+\sigma_{ii}\sigma_{ii}^{(-1)}=\\\ \frac{1}{\sum_{k=1}^{D}\prod_{l\neq k}\alpha_{l}}\left[-\alpha_{D}\sum_{m\neq i,D}\prod_{k\neq m,i}\alpha_{k}+(\alpha_{i}+\alpha_{D})\sum_{k\neq i}\prod_{l\neq k,i}\alpha_{l}\right].$ (43) The square bracket evaluates to $-\alpha_{D}\sum_{m\neq i,D}\prod_{k\neq m,i}\alpha_{k}+\alpha_{i}\sum_{k\neq i}\prod_{l\neq k,i}\alpha_{l}+\alpha_{D}\sum_{k\neq i,D}\prod_{l\neq k,i}\alpha_{l}+\alpha_{D}\prod_{l\neq D,i}\alpha_{l}\\\ =\alpha_{i}\sum_{k\neq i}\prod_{l\neq k,i}\alpha_{l}+\alpha_{D}\prod_{l\neq D,i}\alpha_{l}=\sum_{k\neq i}\prod_{l\neq k}\alpha_{l}+\prod_{l\neq i}\alpha_{l}=\sum_{k}\prod_{l\neq k}\alpha_{k}.$ (44) Inserting this back into (43) shows that the diagonal elements are 1, proving the inverse. We can now evaluate the partial correlation coefficient (7) for an LU composition $\boldsymbol{u}$ by inserting the expression for the inverse (40) and (to obtain the third equality below) making use of the second equality in (39): $r_{ij}(\boldsymbol{u})=\frac{-\sigma^{(-1)}_{ij}}{\sqrt{\sigma^{(-1)}_{ii}\sigma^{(-1)}_{jj}}}=\frac{\prod_{k\neq i,j}\alpha_{k}}{\sqrt{\left(\sum_{k\neq i}\prod_{l\neq k,i}\alpha_{l}\right)\left(\sum_{k\neq j}\prod_{l\neq k,j}\alpha_{l}\right)}}\\\ =\frac{\prod_{k\neq i,j}\alpha_{k}}{\sqrt{\left(\prod_{l\neq i}\alpha_{l}\sum_{k\neq i}\alpha_{l}^{-1}\right)\left(\prod_{l\neq j}\alpha_{l}\sum_{k\neq j}\alpha_{l}^{-1}\right)}}\\\ =\frac{\prod_{k\neq i,j}\alpha_{k}}{\left(\prod_{k\neq i,j}\alpha_{k}\right)\sqrt{\left(\alpha_{j}\sum_{k\neq i}\alpha_{l}^{-1}\right)\left(\alpha_{i}\sum_{k\neq j}\alpha_{l}^{-1}\right)}}\\\ =\sqrt{\frac{\alpha^{-1}_{i}\alpha^{-1}_{j}}{\left(\sum_{k\neq i}\alpha^{-1}_{k}\right)\left(\sum_{k\neq j}\alpha^{-1}_{k}\right)}},$ (45) which is the expression given in (22). ### 7.5. Supplementary Figures Figure 6: Mean squared error of CLR and ALR covariance matrices for different sample sizes ($N$=200, 40, 8) and estimation procedures (no shrinkage, naive shrinkage of CLR/ALR covariance matrix, and basis covariance shrinkage) computed on synthetic data. Each boxplot contains the results of 200 simulations. Whenever estimates without shrinkage are not shown, their median value is given in the legend. Figure 7: Mean squared error of CLR and ALR covariance matrices for different sample sizes ($N$=2500, 1000, 100) and estimation procedures (no shrinkage, naive shrinkage of CLR/ALR covariance matrix, and basis covariance shrinkage) computed on single-cell gene expression data. Each boxplot contains the results of 200 resamplings from data. Colours indicate the type of zero imputation used. ### 7.6. Code and Data Availability All the code used to perform the benchmark and reproduce the results of this paper, as well as the subset of single-cell gene expression data, are available on GitHub under https://github.com/suzannejin/pcor.bshrink.git. The R package propr Quinn et al. (2017) enables compositional data analysis on relative gene expression data. Originally designed for efficient calculations of proportionality indices, it has been updated several times, e.g., to include differential proportionality across groups Erb et al. (2017). Until the full implementation of partial correlations in propr, an R implementation for the computation of partial correlations with basis shrinkage is available on the repository above (bShrink function in file bin/rlib/shrink.R). ## References * Whittaker (1990) Whittaker, J: Graphical models in applied multivariate statistics. Wiley (1990) * van Nimwegen (2016) van Nimwegen E (2016). Inferring Contacting Residues within and between Proteins: What Do the Probabilities Mean? PLoS Comput Biol 12(5), e1004726. * Weistuch et al. (2022) Weistuch, C, Zhu, J, Deasy, JO and Tannenbaum, AR (2022). The maximum entropy principle for compositional data. BMC Bioinformatics 23:449. * Schäfer and Strimmer (2005) Schäfer, J and Strimmer, K (2005). A Shrinkage Approach to Large-Scale Covariance Matrix Estimation and Implications for Functional Genomics. Statistical Applications in Genetics and Molecular Biology 4(1), 32. * Badri et al. (2020) Badri, M, Kurtz, ZD, Bonneau, R, Müller, CL (2020). Shrinkage improves estimation of microbial associations under different normalization methods, NAR Genomics and Bioinformatics 2(4), lqaa100. * Aitchison (1986) Aitchison, J: The statistical analysis of compositional data. Chapman and Hall (1986) * Erb (2020) Erb, I (2020). Partial correlations in compositional data analysis Applied Computing and Geosciences 6, 100026. * Werhli et al. (2006) Werhli, AV, Grzegorczyk, M, Husmeier, D (2006). Comparative evaluation of reverse engineering gene regulatory networks with relevance networks, graphical gaussian models and bayesian networks, Bioinformatics, 22(20), 2523–2531. * Lovell et al. (2015) Lovell, D. et al. (2015). Proportionality: a valid alternative to correlation for relative data. PLoS Comp Biol 11, e1004075. * Erb and Notredame (2016) Erb, I and Notredame, C (2016). How should we measure proportionality on relative gene expression data? Theory in Biosciences 135(1-2), 21–36. * Kurtz et al. (2015) Kurtz, ZD, Müller, CL, Miraldi, ER, Littman, DR, Blaser, MJ, Bonneau, RA (2015) Sparse and Compositionally Robust Inference of Microbial Ecological Networks. PLoS Comput Biol 11(5):e1004226. * Erb et al. (2020) Erb, I, Gloor, GB, Quinn, TP (2020). Editorial: Compositional data analysis and related methods applied to genomics – a first special issue from NAR Genomics and Bioinformatics. NAR Genomics and Bioinformatics 2(4) lqaa103. * Quinn et al. (2018) Quinn, TP, Erb, I, Richardson, MF, Crowley, TM (2018). Understanding sequencing data as compositions: an outlook and review. Bioinformatics 34(16), 2870–2878. * Aitchison (1982) Aitchison, J (1982). The Statistical Analysis of Compositional Data. J Royal Stat Soc B 44(2), 139–160. * Lovell et al. (2020) David R Lovell, Xin-Yi Chua, Annette McGrath (2020). Counts: an outstanding challenge for log-ratio analysis of compositional data in the molecular biosciences, NAR Genomics and Bioinformatics, 2(2), lqaa040. * Law et al. (2014) Law, CW, Chen, Y, Shi, W, and Smyth, GK (2014). voom: precision weights unlock linear model analysis tools for RNA-seq read counts Genome Biology, 15:R29. * Hausser and Strimmer (2009) Hausser, J and Strimmer, K (2009). Entropy Inference and the James-Stein Estimator, with Application to Nonlinear Gene Association Networks Journal of Machine Learning Research 10, 1469-1484. * Quinn and Erb (2021) Quinn, TP, Erb, I (2021). Amalgams: data-driven amalgamation for the dimensionality reduction of compositional data NAR Genomics and Bioinformatics 2(4) Iqaa076. * Greenacre et al. (2022) Greenacre, M Grunsky, E Bacon-Shone, J, Erb, I and Quinn, T (2022). Aitchison’s Compositional Data Analysis 40 years On: A Reappraisal. arXiv:2201.05197 * James and Stein (1961) James, W and Stein, C. (1961) Estimation with quadratic loss. In: Proc. Fourth Berkeley Symp. Math. Statist. Probab. 1 361–379, Univ. California Press * Ledoit and Wolf (2003) Ledoit, O. and Wolf, M (2003). Improved estimation of the covariance matrix of stock returns with an application to portfolio selection. Journal of Empirical Finance 10, 603–621. * Opgen-Rhein and Strimmer (2007) Opgen-Rhein, R and Strimmer, K (2007). Accurate Ranking of Differentially Expressed Genes by a Distribution-Free Shrinkage Approach. Statistical Applications in Genetics and Molecular Biology 6(1). * Palarea-Albaladejo et al. (2015) Palarea-Albaladejo, J, Josep Antoni MartÃn-Fernández, JA (2015). zCompositions – R package for multivariate imputation of left-censored data under a compositional approach, Chemometrics and Intelligent Laboratory Systems 143, 85–96. * Palarea-Albaladejo et al. (2022) Palarea-Albaladejo, J, Antoni MartÃn-Fernández, J, Ruiz-Gazen, A, Thomas-Agnan, C. (2022). lrSVD: An efficient imputation algorithm for incomplete high-throughput compositional data. Journal of Chemometrics e3459. * Riba et al. (2022) Riba, A., Oravecz, A., Durik, M. et al. (2022). Cell cycle gene regulation dynamics revealed by RNA velocity and deep-learning. Nat Commun 13, 2865. * DeGroot (2004) DeGroot, MH: Optimal Statistical Decisions, Wiley (2004) * Quinn et al. (2017) Quinn, TP, Richardson, MF, Lovell, D and Crowley, TL (2017). propr: An R-package for Identifying Proportionally Abundant Features Using Compositional Data Analysis. Scientific Reports 7(1) 16252. * Erb et al. (2017) Erb, I, Quinn, TP, Lovell, D and Notredame, C (2017). Differential Proportionality – A Normalization-Free Approach To Differential Gene Expression. Proceedings of CoDaWork 2017, The 7th Compositional Data Analysis Workshop; available under bioRxiv, 134536.
# HIKE: High Intensity Kaon Experiments at the CERN SPS Matthew Moulson for the HIKE Collaboration INFN Laboratori Nazionali di Frascati, 00044 Frascati RM, Italy<EMAIL_ADDRESS> ###### Abstract The availability of high intensity kaon beams at the CERN SPS North Area gives rise to unique possibilities for sensitive tests of the Standard Model in the quark flavor sector. Precise measurements of the branching ratios for the flavor-changing neutral current decays $K\to\pi\nu\bar{\nu}$ can provide unique constraints on CKM unitarity and, potentially, evidence for new physics. Building on the success of the NA62 experiment, plans are taking shape at CERN for a comprehensive program that will include experimental phases to measure the branching ratio for $K^{+}\to\pi^{+}\nu\bar{\nu}$ to $\sim$5% and to $K_{L}\to\pi^{0}\nu\bar{\nu}$ to $\sim$20% precision. These planned experiments would also carry out lepton flavor universality tests, lepton number and flavor conservation tests, and perform other precision measurements in the kaon sector, as well as searches for exotic particles in kaon decays. We overview the physics goals, detector requirements, and project status for the next generation of kaon physics experiments at CERN. ## 1 Introduction Rare kaon decays provide information on the unitary triangle, as illustrated in Figure 1. These are flavor-changing neutral current processes (FCNC) that probe the $s\to d\nu\bar{\nu}$ or $s\to d\ell^{+}\ell^{-}$ transitions. They are highly GIM-suppressed and their SM rates are very small. The $K\to\pi\nu\bar{\nu}$ decays are the least affected by long-distance physics. The branching ratios (BRs) for the $K\to\pi\nu\bar{\nu}$ decays are among the observable quantities in the quark-flavor sector most sensitive to new physics. Figure 1: Determination of the unitary triangle with rare kaon decays. In the SM, the uncertainties on the $K\to\pi\nu\bar{\nu}$ rates are entirely dominated by the uncertainties on the CKM matrix elements $|V_{ub}|$ and $|V_{cb}|$ and the angle $\gamma$. Using values for these parameters from the analysis of tree-level observables, Buras et al. obtain [1] $\begin{split}{\rm BR}(K_{L}\to\pi^{0}\nu\bar{\nu})&=(3.4\pm 0.6)\times 10^{-11},\\\ {\rm BR}(K^{+}\to\pi^{+}\nu\bar{\nu})&=(8.4\pm 1.0)\times 10^{-11}.\end{split}$ (1) Assuming no new-physics effects in $\epsilon_{K}$ and $\sin{2\beta}$ from $B\to J/\psi K_{S}$, the BRs can be determined independently of $|V_{cb}|$ as ${\rm BR}(K_{L}\to\pi^{0}\nu\bar{\nu})=(2.94\pm 0.15)\times 10^{-11}$ and ${\rm BR}(K^{+}\to\pi^{+}\nu\bar{\nu})=(8.60\pm 0.42)\times 10^{-11}$ [2]. The intrinsic theoretical uncertainties, if the CKM parameters are taken to be exact, are about 1.5% and 3.5%, respectively. Because the amplitude for $K^{+}\to\pi^{+}\nu\bar{\nu}$ has both real and imaginary parts, while the amplitude for $K_{L}\to\pi^{0}\nu\bar{\nu}$ is purely imaginary, the decays have different sensitivity to new sources of $CP$ violation. Measurements of both BRs would therefore be extremely useful not only to uncover evidence of new physics in the quark-flavor sector, but also to distinguish between new physics models. More generally, measurement of all of the FCNC kaon decays would allow the unitarity triangle to be overconstrained as illustrated in Figure 1, potentially providing evidence of new physics independently of and in comparison to results from $B$ and $D$ meson decays and providing important information about the flavor structure of that physics. ## 2 The NA62 experiment NA62 is a fixed-target experiment at the CERN SPS, the goal of which is to measure ${\rm BR}(K^{+}\to\pi^{+}\nu\bar{\nu})$ to within 10% [3]. The secondary positive beam, consisting of 6% $K^{+}$ with a total rate of 750 MHz, is derived from the 400-GeV primary proton beam from the SPS incident at zero angle on a beryllium target at a rate of $3\times 10^{12}$ protons per pulse (ppp). The $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay is detected in flight. The signature is a $K^{+}$ entering the experiment and a $\pi^{+}$ leaving the decay vertex, with missing momentum at the vertex and no other particles observed in the final state. Principal backgrounds include those from the abundant decays $K^{+}\to\mu^{+}\nu$ and $K^{+}\to\pi^{+}\pi^{0}$, as well as backgrounds from upstream decays and interactions. The signal decay is identified via selection in the $(p_{\pi},m^{2}_{\rm miss})$ plane to exclude the abundant two-body decays, where $p_{\pi}$ is the momentum of the candidate pion track and $m^{2}_{\rm miss}$ is the squared missing mass at the vertex. The distinguishing features of the experiment include high-rate, precision tracking for both the beam and secondary particles, redundant particle identification systems and muon vetoes, and hermetic photon vetoes, including a high-performance EM calorimeter. Between 2016 and 2018, NA62 observed more than $4\times 10^{12}$ $K^{+}$ decays in its fiducial volume, with the expectation of observing 10 signal events and 7 background events, principally from upstream decays and interactions. A total of 20 events were observed, establishing the $K^{+}\to\pi^{+}\nu\bar{\nu}$ decay with $3.4\sigma$ significance and providing the most precise measurement to date for the branching ratio [4]: ${\rm BR}(K^{+}\to\pi^{+}\nu\bar{\nu})=(10.6^{+4.0}_{-3.4}|_{\rm stat}\pm 0.9_{\rm syst})\times 10^{-11}.$ NA62 resumed data taking in July 2021 with a number of key modifications to the beamline and detector to reduce background from upstream decays and interactions and to allow data to be taken at the full nominal beam intensity. The experiment is approved for data taking up to LHC Long Shutdown 3 (LS3), currently foreseen for the end of 2025, and is expected to measure ${\rm BR}(K^{+}\to\pi^{+}\nu\bar{\nu})$ to 10% precision by then. In the longer term, fixed-target runs in the SPS North Area are foreseen at least through 2040. There is therefore an opportunity at the SPS for an integrated program to pin down new physics in the kaon sector via measurement of all rare kaon decay modes—both charged and neutral. ## 3 The HIKE physics program The HIKE program (High Intensity Kaon Experiments at the CERN SPS) [5] is foreseen to include three experimental phases for the comprehensive, high- precision study of rare kaon decays during the period from the end of LS3 to the FCC era: * • Phase 1: A $K^{+}$ experiment running at four times the intensity of NA62 ($1.2\times 10^{13}$ ppp) to measure ${\rm BR}(K^{+}\to\pi^{+}\nu\bar{\nu})$ to 5% precision. Phase 1 will also focus on studies of lepton universality/number/flavor violation through observables such as $R_{K}=\Gamma(K^{+}\to e^{+}\nu)/\Gamma(K^{+}\to\mu^{+}\nu)$ and $K^{+}\to\pi^{+}\ell\ell$, and searches for decays such as $K^{+}\to\pi^{-}\ell^{+}\ell^{+}$ and $K^{+}\to\pi^{+}\mu e$. The experiment will also make precision measurements of leptonic and semileptonic, radiative, and Dalitz decays and chiral parameters of the kaon system. * • Phase 2: An experiment with a neutral beam and tracking and PID systems optimized for the measurement of decays like $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ and $K_{L}\to\mu^{+}\mu^{-}$, as well as studies of lepton universality/number/flavor violation in $K_{L}$ decays, radiative $K_{L}$ decays and precision measurements, and measurements of $K_{L}$, $n$, and $\Lambda$ fluxes in the neutral beam and halo to prepare for the final phase. * • Phase 3: An experiment to measure ${\rm BR}(K_{L}\to\pi^{0}\nu\bar{\nu})$ to 20%, also known as KLEVER [6]. In addition, periodic runs will be taken in beam-dump mode, with the target out and the collimator in the secondary beam line closed, to allow sensitive searches for the decays of exotic, long-lived particles produced in the decays of mesons from interactions in the beam dump, with the goal of collecting $10^{19}$ protons on target (pot) in dump mode in Phase 1 and up to $5\times 10^{19}$ pot by the end of Phase 3. The experimental setup for all three phases will rely to the maximum extent possible on the reuse of the same detectors in different configurations. In particular, when a detector for HIKE is newly built or extensively upgraded, if at all possible, it is designed to meet the performance requirements for all successive phases. ### 3.1 HIKE Phase 1 Figure 2: Experimental setup for HIKE Phase 1 The four-fold increase in primary intensity needed for the measurement of ${\rm BR}(K^{+}\to\pi^{+}\nu\bar{\nu})$ to 5% will require major upgrades of the primary and secondary beamlines, as discussed in [7]. From the standpoint of the experiment, the success of NA62 has validated the measurement technique and proves that the background can be handled. The key challenge from the intensity increase is that the time resolution of the detectors must be improved across the board by a factor of four in order to maintain the loss of events from accidental coincidence (random veto) to acceptable levels ($\lesssim 25\%$), which must be achieved while maintaining other key performance specifications such as space-time reconstruction performance, low material budget, high single photon efficiencies, etc. The experimental setup, shown in Figure 2, is not very different from that of NA62, but most detectors will need to be rebuilt or extensively overhauled to secure the needed performance. Of particular interest, the NA62 beam tracker (Gigatracker, GTK), consisting of three stations of silicon pixel detectors, will need to be upgraded to track at 3 GHz. A time resolution of better than 50 ps will be required, and the detector will have to be able to handle rates of 8 MHz/mm2 and be radiation resistant up for particle fluences of more than $2\times 10^{15}$ $n$ eq/cm2/yr. An excellent candidate technology is provided by the timeSPOT project [8, 9], which is developing hybrid 3D-trenched pixels in which the pixel electrode geometry is optimized for timing performance. The experiment’s rate capability for secondary particles must be improved as well. New straw-tube designs are being developed at CERN, based on past collaboration with Dubna, that will allow straw chambers for use in vacuum to be developed with 5-mm diameter straws with wall thickness of 20 $\mu$m. It is natural to inquire whether the NA48 liquid-krypton (LKr) calorimeter [10] used in NA62 can be reused for any of the HIKE phases. NA48-era studies suggest and NA62 experience confirms that the photon detection efficiency is sufficient for at least Phases 1 and 2. The time resolution, however, is insufficient for the high-intensity program and would require improvement by at least a factor of four. Two directions are being pursued. The first is to examine whether the LKr can be used for HIKE Phase 1. In addition to necessary consolidation work, this would require upgrades to make the calorimeter faster, including an increase in the operating voltage to increase the drift velocity and faster signal shaping and digitizers for the readout system. For the $K_{L}$ phases, the diameter of the LKr inner bore limits the solid angle of the beam that can be used, so a new calorimeter would be necessary in any case. The new calorimeter could also be used in Phase 1, if it is ready in time. An ideal choice of technologies appears to be the fine-sampling shashlyk design used for the KOPIO and PANDA calorimeters [11], which has been shown to provide excellent energy and time resolution. For HIKE, PID capability could be added by including independently read out, 1-cm-thick “spy tiles” at key points in the shashlyk stack (for example, at the front of the calorimeter for charged-particle identification, at shower maximum, and deep in the stack). Prototypes with this design have been tested at Protvino and DESY. Another option under investigation is to construct the calorimeter with new- generation, nanocomposite scintillators [12], which offer high light output, fast response, and good radiation robustness. ### 3.2 HIKE Phase 2 Figure 3: Experimental setup for HIKE Phase 2 For HIKE Phase 2, a new neutral beamline is required. The baseline design is the original 120-m beamline for KLEVER [13], featuring four collimation stages, including an active final collimator that is incorporated into the experiment and an oriented-crystal-metal photon converter at the center of the dump collimator to reduce the flux of prompt photons in the beam [14]. The collimation system defines a beam opening angle of 0.4 mrad. The neutral beam is produced at $\theta=2.4$ mrad; $K_{L}$ mesons in the beam have an average momentum of 79 GeV, while those decaying in the fiducial volume (FV) have an average momentum of 46 GeV. Relative to the Phase-1 experiment, in addition to the changes to the beamline and the removal of the beamline detectors for charged particles, the RICH is removed and the spectrometer is moved further downstream to increase the acceptance for decays such as $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$. The beamline and experimental setup are illustrated in Figure 3. At a primary intensity of $2\times 10^{13}$ ppp (a six-fold increase with respect to NA62), nearly $2\times 10^{14}$ kaon decays will be observed in the FV in five years of running. This will allow single-event sensitivities for $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ to be improved by two orders of magnitude with respect to the current best limits from KTeV [15, 16]. There are topologically identical backgrounds to these channels from $K_{L}\to\gamma\gamma\ell^{+}\ell^{-}$, with BRs that are 3–4 orders of magnitude greater than for $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ [17]. Suppression of these backgrounds relies on the excellent energy resolution of the HIKE EM calorimeter for the reconstruction of the $\pi^{0}$ mass peak for the signal decay. HIKE Phase 2 will be well positioned to make the first observation of the $K_{L}\to\pi^{0}\ell^{+}\ell^{-}$ decay, as well as to measure ${\rm BR}(K_{L}\to\mu^{+}\mu^{-})$ with a statistical precision of 0.2%, improving on the current best result from BNL-E871 [18], and should also achieve BR sensitivities of $O(10^{-12})$ for a broad range of rare and forbidden $K_{L}$ decays, such as lepton-flavor violating processes, representing improvement on current best limits from BNL-E871 by up to a factor of 60. ### 3.3 HIKE Phase 3 Figure 4: Experimental setup for HIKE Phase 3 (KLEVER) HIKE Phase 3 is a dedicated experiment, known as KLEVER, to measure ${\rm BR}(K_{L}\to\pi^{0}\nu\bar{\nu})$ to 20%. Specifically, with a total exposure of $6\times 10^{19}$ pot in five years at an intensity of $2\times 10^{13}$ ppp, the KLEVER goal is to detect 60 signal events at the Standard Model branching ratio, with a signal-to-background ratio $S/B\sim 1$. The experiment is complementary to KOTO, in the sense that the beam energy is significantly higher. As a result, photons from $K_{L}$ decays receive a significant boost, which makes photon vetoing easier. On the other hand, the length of the experiment is much greater, and a very long beamline is required to allow $\Lambda$ baryons and $K_{S}$ mesons to decay upstream of the fiducial volume. Relative to the Phase-2 beamline, the KLEVER beamline needs to be extended by an additional 150 m from target to final collimator. This in turn requires a downstream extension of the ECN3 hall by a similar amount. For KLEVER running, the production angle for the neutral beam will be increased from 2.4 to 8.0 mrad: this decreases the $n/K_{L}$ and $\Lambda/K_{L}$ ratios for the beam and softens the momentum spectra so that the average $K_{L}$ momentum is 39 GeV at production (26 GeV for $K_{L}$ mesons that decay in the FV). This also decreases the absolute $K_{L}$ flux, which is further reduced by the need to collimate the beam more tightly ($\Delta\theta=0.256$ mrad) for the extended beamline. The need for additional construction to extend the ECN3 hall is a major factor in scheduling KLEVER towards the end of the HIKE program. Cost estimates are in progress. From the standpoint of the experiment, the layout of the vacuum tank and fiducial volume is roughly the same as for the other HIKE phases. The spectrometer will be removed and the number of large-angle photon veto detectors will be increased from 12 to 25 to extend the polar angle coverage out to 100 mrad (from 50 mrad for Phases 1 and 2). The HIKE large-angle vetoes themselves will be fine-segmented lead/scintillating tile detectors similar to the VVS detectors for the planned but never realized CKM experiment [19]. One particularly challenging detector for KLEVER is the small-angle calorimeter (SAC), which sits in the neutral beam at the downstream end of the experiment and which must reject photons from background decays such as $K_{L}\to\pi^{0}\pi^{0}$ that would otherwise escape via the beam exit. The SAC must have good photon detection efficiency, especially for high energy photons (e.g., the inefficiency must be $<10^{-4}$ for photons with $E>30$ GeV) while being as insensitive as possible to the accidental coincidence of nearly 600 MHz of neutral hadrons ($n$ and $K_{L}$) in the beam. The SAC must also have $\sigma_{t}<100$ ps, be able to separate two pulses a few ns apart, and be radiation hard to $10^{14}$ $n$/cm2 and $10^{5}$–$10^{6}$ Gy. Our proposed solution is an ultra-fast, high-$Z$ crystal calorimeter based on a Cerenkov radiator like PbF2 or an ultra-fast scintillator such as PWO-UF [20], which has a dominant emission component with $\tau<0.7$ ns. The SAC will have transverse and longitudinal segmentation for $\gamma/n$ discrimination. From an engineering standpoint, it will be very similar to the CRILIN calorimeter [21], and R&D is proceeding in concert between the HIKE and CRILIN collaborations. An additional possibility under investigation is to exploit the coherent interactions of high-energy photons in oriented crystals to stimulate early pair conversion, allowing the calorimeter to be realized with reduced thickness, increasing the transparency to neutral hadrons while maintaining high detection efficiency for photons [22]. ## 4 Conclusions The HIKE project consists of a three-phase experimental program for the comprehensive study of flavor physics in the kaon sector. The experimental apparatus changes over time with a staged approach, allowing HIKE to evolve and adapt its physics scope, an important feature for a project that embraces a time scale of more than decade, during which the physics landscape could change. Thanks to the successful experience of NA62 and its predecessor NA48, the experimental techniques are well established and robust expectations of sensitivity can be obtained from the extrapolation of existing data. The HIKE Letter of Intent was submitted to the CERN SPSC at the beginning of November 2022 [5]. A formal proposal is in preparation for submission in fall 2023. ## References ## References * [1] Buras A J, Buttazzo D, Girrbach-Noe J and Knegjens R 2015 JHEP 11 033 (Preprint 1503.02693) * [2] Buras A J and Venturini E 2021 Acta Phys. Polon. B 53 A1 (Preprint 2109.11032) * [3] Cortina Gil E et al. (NA62) 2017 JINST 12 P05025 (Preprint 1703.08501) * [4] Cortina Gil E et al. (NA62) 2021 JHEP 06 093 (Preprint 2103.15389) * [5] Cortina Gil E et al. (HIKE) 2022 HIKE, High Intensity Kaon Experiments at the CERN SPS: Letter of Intent Tech. Rep. SPSC-2022-031/SPSC-I-257 CERN (Preprint 2211.16586) URL https://cds.cern.ch/record/2839661 * [6] Ambrosino F et al. (KLEVER Project) 2019 (Preprint 1901.03099) * [7] Gatignon L, Bernhard J, Brugger M, Doble N et al. 2018 Report from the Conventional Beams Working Group to the Physics Beyond Collider Study and to the European Strategy for Particle Physics Tech. rep. CERN Geneva URL https://cds.cern.ch/record/2650989 * [8] Lai A 2018 A system approach towards future trackers at high luminosity colliders: the timespot project (Sydney, NSW, Australia: IEEE) pp 1–3 ISBN 978-1-5386-8495-5 ISSN 1082-3654 * [9] Lai A et al. 2022 Timespot project testbeam results and simulations talk given at ”1st MONOLITH Workshop on silicon sensors for timing and their applications”, Geneve, September 2022 URL https://indico.cern.ch/event/1179742/contributions/4961955/attachments/2502408/4298914/ Timing_ALai_UniGe.pdf * [10] Fanti V et al. (NA48) 2007 Nucl. Instrum. Meth. A574 433–471 * [11] Atoian G S et al. 2008 Nucl. Instrum. Meth. A584 291–303 (Preprint 0709.4514) * [12] Gandini M et al. 2020 Nat. Nanotechnol. 15 462–468 * [13] Van Dijk M et al. 2019 The K12 beamline for the KLEVER experiment URL doi.org/ 10.18429/JACoW-IPAC2019-THPGW061 * [14] Soldani M et al. 2022 (Preprint 2203.07163) * [15] Alavi-Harati A et al. (KTeV) 2004 Phys. Rev. Lett. 93 021805 (Preprint hep-ex/0309072) * [16] Alavi-Harati A et al. (KTEV) 2000 Phys. Rev. Lett. 84 5279–5282 (Preprint hep-ex/0001006) * [17] Greenlee H B 1990 Phys. Rev. D 42 3724–3731 * [18] Ambrose D et al. (E871) 2000 Phys. Rev. Lett. 84 1389–1392 * [19] Ramberg E, Cooper P and Tschirhart R 2004 IEEE Trans. Nucl. Sci. 51 2201–2204 * [20] Korzhik M et al. 2022 Nucl. Instrum. Meth. A 1034 166781 * [21] Ceravolo S et al. 2022 JINST 17 P09033 (Preprint 2206.05838) * [22] Bandiera L, Haurylavets V and Tikhomirov V 2019 Nucl. Instrum. Meth. A 936 124–126
# Study of radio transients from the quiet Sun during an extremely quiet time Surajit Mondal Center for Solar Terrestrial Research, New Jersey Institute of Technology, Newark, New Jersey, 07102, United States of America Divya Oberoi National Centre for Radio Astrophysics, Tata Institute of Fundamental Research, Pune-411007, India Ayan Biswas National Centre for Radio Astrophysics, Tata Institute of Fundamental Research, Pune-411007, India (Received TBD ; Revised ; Accepted TBD ) ###### Abstract In this work we study a class of recently discovered metrewave solar transients referred to as Weak Impulsive Narrowband Quiet Sun Emission (WINQSEs, Mondal et al., 2020). Their strength is a few percent of the quiet Sun background and are characterised by their very impulsive, narrow-band and ubiquitous presence in quiet Sun regions. Mondal et al. (2020) hypothesised that these emissions might be the radio counterparts of the nanoflares and their potential significance warrants detailed studies. Here we present an analysis of data from an extremely quiet time and with an improved methodology over the previous work. As before, we detect numerous WINQSEs, which we have used for their further characterisation. Their key properties, namely, their impulsive nature and ubiquitous presence on the quiet Sun are observed in these data as well. Interestingly, we also find some of the observed properties to differ significantly from the earlier work. With this demonstration of routine detection of WINQSEs, we hope to engender interest in the larger community to build a deeper understanding of WINQSEs. Quiet solar corona (1992); Solar corona (1483); Solar coronal radio emission (1993); Solar coronal heating (1989) ††facilities: Murchison Widefield Array (MWA) (Lonsdale et al., 2009; Tingay et al., 2013)††software: astropy (Astropy Collaboration et al., 2013, 2018), matplotlib (Hunter, 2007), Numpy (Harris et al., 2020), SciPy (Virtanen et al., 2020), CASA (McMullin et al., 2007) ## 1 Introduction Radio transients from the Sun were discovered at very the dawn of solar radio astronomy (e.g. Wild et al., 1963; Wild & Smerd, 1972, etc.). Since then numerous studies have analysed and studied their various properties ranging from emission mechanisms, spectro-temporal characteristics, relationships with other solar phenomenon, signatures in other wavebands and so on (e.g. Takakura, 1967; Sturrock, 1964; Melrose, 1980; McLean & Labrum, 1985; Pick & Vilmer, 2008; Kontar et al., 2017; Reid & Kontar, 2017; Reid, 2020, and many others). Most of these transients, particularly those happening in the meter wavelengths regime, are associated with active regions and violent solar activity like X-ray flares and coronal mass ejections (e.g. Pick & Vilmer, 2008; Bain & Fletcher, 2009; Krucker et al., 2011; Reid, 2020; Mondal & Oberoi, 2021, etc.). However, there are hypotheses which suggest that such radio transients, albeit much weaker, should also be present in the quiet solar corona (Che, 2018). The primary reasoning behind this expectation comes from the Parker’s nanoflare hypothesis (Parker, 1988) which states that the solar corona maintains its million K temperature because of the energy continuously deposited by numerous tiny nanoflares happening everywhere in the corona. If this hypothesis is true, it implies that small scale reconnections are happening everywhere and all the time in the corona. These small scale reconnections, like their larger counterparts, will also accelerate electrons, which in turn, on their interaction with the thermal plasma, should emit plasma emission, observable in the radio bands (Che, 2018). To the best of our knowledge, the first detection of the ubiquitous presence of these radio transients from the quiet sun came from Mondal et al. (2020) (henceforth referred to as M20). They found that these transients are highly impulsive (duration $\lesssim 1$s) and narrowband in nature. Based on these properties they also concluded that these emissions are generated by nonthermal processes. They also found these emissions to be present throughout the quiet sun. Mondal (2021) has demonstrated, although in a single instance, that the energy associated with a group of these transients is $\sim 10^{25}$ergs. These observations suggest the possibility that these transients are the radio counterparts of the long hypothesised nanoflares. Henceforth we refer to these emissions as Weak Impulsive Narrowband Quiet Sun Emissions (WINQSEs). Due to the large potential significance of this discovery, it is crucial to verify the presence of WINQSEs in other datasets obtained under different solar conditions, using independent detection techniques and especially during quiet Sun times. As a part of this exercise, recently Sharma et al. (2022) (henceforth referred to as S22) have reported detection of WINQSEs using a completely independent technique and analysis pipeline using a different dataset. Interestingly S22 also identified a very weak type III solar radio burst like drifting emission feature, which supports the hypothesis that WINQSEs are the weaker cousins of the stronger type III bursts. Continuing with the same motivation, here we present an analysis of data from an extremely quiet time on 20th June, 2020. The solar activity levels on this day are much lower than those prevailing during the observations of M20 and S22. We also take this opportunity to improve the analysis methodology beyond what was used in M20. This paper is structured as follows. In Section 2 we present the observation details, Section 3 discusses the details of data analysis including the improvements over M20. In Section 4 we present the results, in Section 5 the discussion and finally Section 6 gives the concluding remarks. ## 2 Observation details The data presented here come from a solar observation with the Murchison Widefield Array Phase II (MWA, Lonsdale et al., 2009; Tingay et al., 2013; Wayth et al., 2018). These data were acquired on 20 June, 2020 between 03:35:00–04:40:00 UT under the project ID G0002. The MWA was in its extended configuration, with a maximum baseline of $\sim 5$km. The observing frequency spanned the band from 119.68–150.40 MHz. The data was acquired with a time and frequency resolution of 0.5 s and 10 kHz respectively. Here we present the analysis at four selected spectral slices, each of width 160 kHz. These slices are centered around 120.22, 127.9, 135.58 and 139.9 MHz. The Sun was extremely quiet on this day111https://solarmonitor.org/index.php?date=20200620. No X-ray flares were reported by the Geostationary Operational Environmental Satellites (GOES) and no events of any kind were reported by the Space Weather Prediction Center event list for the day. The Learmonth Spectrograph operating between 25–180 MHz also did not report any radio flare on this day. No NOAA active region was present on the visible solar disc. ## 3 Data analysis Making the large number of images needed for this investigation necessarily requires a robust automated pipeline. Such a tool has recently been developed and christened the Automated Imaging Routine for Compact Arrays for Radio Sun (AIRCARS; Mondal et al., 2019). The credentials of this self-calibration based pipeline have already been established in multiple recent works requiring high quality spectroscopic snapshot imaging (e.g. Mohan et al., 2019; Mondal et al., 2020; Mohan, 2021; Mondal & Oberoi, 2021, etc.). It is known from past experience that the presence of strong compact sources on the Sun makes calibration easier and the data corresponding to a large but featureless quiet Sun is harder to calibrate (Mondal et al., 2019). While the imaging of the data used here was done using AIRCARS, to improve the accuracy of calibration for these quiet Sun data, an improved strategy from what was followed in M20 was employed. For this work, the calibration solutions were determined using data averaged over 9 s. This time integration helps reduce the thermal noise substantially and is especially useful in improving the signal-to-noise of visibilities from longer baselines. The calibration solutions obtained were linearly interpolated in time and applied to the 0.5 s resolution data which were then imaged. The analysis for each of the four frequencies presented here was carried out independently following this procedure. This approach relies on the assumption that the solar variability over the integration period is small enough to be ignored. We find this to be a good assumption for this extremely quiet time and is evidenced by the higher imaging dynamic ranges of the images produced using solutions derived from time-averaged visibilities. On its own, AIRCARS does not ensure a common flux density scale across independent observations. For the present analysis, this can lead to discontinuities in the fluxscale at the boundaries of observations, typically of a duration of 4-5 minutes, though this issue has now been resolved in its next incarnation, named, Polarimetry using Automated Imaging Routine for Compact Arrays for the Radio Sun (P-AIRCARS; Kansabanik et al., 2022b). A precise absolute flux density is not required for this work and we have adopted the strategy described below for relative flux density calibration between all of the observations at a given frequency. It also takes care of the amplitude variability of the order of 4-5% of the MWA instrumental response which otherwise is not possible to take care of by using the method described in Kansabanik et al. (2022a). While under most circumstances, this level of variability is too small to be of any consequence, in the present context it is important to correct for it as it is of a magnitude comparable to the strengths of WINQSEs. A correction for the instrumental primary beam is first applied to the solar maps and then it is enforced that the median disc integrated flux density of the Sun be the same for all five-minute chunks of data used. The underlying assumption is that as the Sun was very quiet during our observing period, its disc-integrated flux density should also be stable in this period. The characteristics of WINQSEs - flux densities of $\sim$ mSFU (as compared to a few SFU from the Sun) and short durations ($\lesssim$1 s) - imply that they are too weak to influence the disc-integrated flux density in any significant manner. We also found that the point-spread-function (psf) corresponding to different time slices for a given frequency can change slightly across time. Flux densities in radio maps are measured in units of $Jy/beam$, where $beam$ refers to the area of the psf. Hence, slight changes in the psf area can give rise to low-level jumps in flux density estimates. To avoid contamination due to this, we have smoothed all of the images used here to a common coarser resolution of $280^{"}\times 280^{"}$. A key step for identifying WINQSEs in M20 was to calculate the median image, which requires accurate alignment of images relative to each other. M20 achieved this by aligning the radio peak of the active region present in their dataset across images. Absence of a clearly identifiable compact feature forces us to look for other alternatives to achieve this objective. We compute a center of intensity ($\hat{X}$, $\hat{Y}$) for each 0.5 s solar image, defined as: $\hat{X}=\frac{\sum_{i}x_{i}}{\sum_{i}1},\quad\hat{Y}=\frac{\sum_{i}y_{i}}{\sum_{i}1},$ (1) where ($x_{i}$, $y_{i}$) are the coordinates of the pixels which exceed the threshold chosen visually such that no noise contours appear in the image and the region it demarcates is reasonably symmetric. The centers of intensity for all the images were aligned to lie at the same pixel. The aligned maps were used to generate the median map, one for each five-minute data chunk and frequency. An example 0.5 s map at 127.9 MHz and the corresponding median map are shown in Fig. 1. Figure 1: Left panel: An example 0.5 s map at 127.9 MHz. Right panel: The corresponding median map at 127.9 MHz. ## 4 Results For detection of WINQSEs in the large number of images used in this work, we have modified the technique used in M20. First, all 0.5 s solar images were smoothed using a circular kernel of size 280". Let the images prior and post this smoothing operation be denoted by $I_{init}$ and $I_{flux}$ respectively. As mentioned in the previous section, this smoothing operation ensured that the numerical value of each pixel in the radio image $I_{flux}$ is equal to the flux density for the resolution element corresponding to the circular smoothing kernel centered on that pixel in $I_{init}$. The median map obtained using $I_{init}$ images was used to obtain $I_{median,flux}$ using the same technique as that used to obtain $I_{flux}$. $I_{median,flux}$ was then subtracted from each $I_{flux}$ to obtain the residual maps for each 0.5 s timeslice, henceforth referred to as $I_{res}$. The pixel values in $I_{res}$ is equivalent to $\Delta F_{i,\nu,t}=F_{i,\nu,t}-<F_{i,\nu}>$, where $F_{i,\nu,t}$ is the flux density of a psf sized region $i$ at frequency $\nu$ some time $t$ and $<F_{i,\nu}>$ is the median flux density for region $i$ at frequency $\nu$. A WINQSE is said to be present when the pixel value in $I_{res}$ exceeds a given threshold. In this work, the threshold is set to $3\sigma$, where $\sigma$ is the rms in the corresponding $I_{flux}$ in a region away from the Sun. To ensure that only independent pixels are searched for WINQSEs, the search algorithm steps through the pixels in steps of 7 pixels in both directions which translates to $280^{"}$ in each direction and is equal to the effective image resolution. Figure 2 shows the contours of an example residual image with a bright WINQSE, overlaid on a AIA 193Å image. The compact feature seen in the image is the WINQSE detected in this snapshot and it is evident that it has been detected with good signal-to-noise. The following subsections discuss the various properties of the WINQSEs detected in these data. Figure 2: Radio contours of an example WINQSE at 136 MHz are overlaid on a 193 Å image from AIA/SDO. The contour levels are at 0.7, 0.85 and 0.99 times the peak of the median subtracted map. ### 4.1 Flux density distribution For convenience and consistency with M20, for a given frequency $\nu$, we refer to the collection of $\Delta F_{i,\nu,t}$ for all psf sized regions $i$ s and all times $t$ s as $\Delta F$ and similarly to $<F_{i,\nu}>$ as $F$. We define $\Delta F/F$ to be the ratio of the flux density in residual images ($I_{res}$) to that in the corresponding median image ($I_{median,flux}$) at the same pixel location. Figure 3 shows the histogram of $\Delta F/F$ in detected WINQSEs. These distributions are strikingly different from those found by M20. In Sec. 5.2 we try to understand and reconcile these differences. The red line in Fig. 3 shows a lognormal fit to the data shown with blue points. It is evident that the lognormal distribution is a good fit to the data. However, it is worth noting that the part of the distribution where $\Delta F/F\lesssim 0.1$ is affected by incompleteness due to a combination of multiple factors including sensitivity, and spatial and temporal resolution. These factors are not expected to significantly affect the data at higher $\Delta F/F$ values and hence this part of the histogram is robust. The fact that such lognormal distributions have been observed in earlier works when studying weak impulsive emissions, though at EUV wavelengths and with rather different temporal and spatial sampling (Pauluhn & Solanki, 2007; Tajfirouze & Safari, 2012; Upendran & Tripathi, 2021), suggests that WINQSEs too might follow a lognormal distribution. Testing this hypothesis in a more robust manner requires an instrument with better sensitivity and resolution. We plan to explore this with the next phase of the MWA (Phase III) which will offer a factor of two larger effective collecting area. Figure 3: Histograms of $\Delta F/F$ are shown using the blue circles. The red line shows the lognormal fit to these data. The observation frequencies are mentioned in each panel. ### 4.2 Temporal width distribution The temporal width of a WINQSE is defined as the time duration for which the value of a pixel in $I_{res}$ continuously exceeds the 3$\sigma$ threshold. Figure 4 shows the temporal width distribution of the detected WINQSEs. The powerlaw index of the distributions is close to -2, very similar to that found in M20. The bursts are highly impulsive with a steep powerlaw like distribution. The peak of the distribution lies at the instrumental resolution of 0.5 s, implying that most of the WINQSEs are unresolved in time. Figure 4: Temporal width distribution of the detected WINQSEs. ### 4.3 Wait time distribution The wait time between WINQSEs is defined as the time interval between the successive times when the pixel value at a given pixel in $I_{res}$ exceeds the 3$\sigma$ threshold. The wait time distribution of detected WINQSEs is shown in Fig. 5. The wait time distribution is very similar to that found in M20. During the interpretation of the wait time distribution, a key point was however missed by M20. Kivelä & Porter (2015) pointed out that the sharp cutoff at high wait times, like what is observed in Fig. 5, can arise because of data gaps and hence is not reliable. To take this into account, in Fig. 5, we draw red and black dashed lines to indicate the 10% and 20% error probabilities respectively. This implies that the data is sufficient to say that at low wait times the distribution follows a powerlaw, but the nature of the distribution at higher wait times cannot be constrained with these data. The powerlaw nature of wait time distribution is observed for solar flares and has been explained using a non-stationary Poisson model for solar flares (Aschwanden & McTiernan, 2010). It is possible that similar behaviour is also present for WINQSEs. M20 mention this aspect and alluded to the possibility that the observed wait time distribution might be severely affected by a lack of data. Based on the analysis here we conclude that this is indeed the case for the high wait time regime. Figure 5: Wait time distribution of the detected WINQSEs. The red and black lines show the wait time till which the uncertainties due to incomplete sampling is less than 10% and 20% respectively. ### 4.4 Spatial distribution A key result of M20 was that the WINQSEs are ubiquitous in nature. Figure 6 shows the fractional occupancy of the detected WINQSEs over the radio Sun. The fractional occupancy is defined as the fraction of observation time for which WINQSEs were detected in a resolution element. There is no patch on the Sun where WINQSEs are absent. Figure 6: Fractional occupancy distribution of the detected WINQSEs. Contours of the median map at respective frequencies has been overlaid. The contour levels correspond to 0.2, 0.4, 0.6, 0.8, 0.9, 0.95 times the peak in the median map. ## 5 Discussion ### 5.1 Improving the data analysis technique over M20 We have taken this opportunity to improve the analysis methodology to account for two low-level effects which were realized during the course of this work and are listed below: 1. 1. In M20 each five minute datachunk was processed independently, this can lead to small differences in absolute fluxscale between data chunks. At a few percent level, these differences are smaller than the uncertainty in the absolute flux density and usually too small to be of any consequence. In the present context, however, it is important to correct for them as they are comparable in strength to flux densities of WINQSEs. As described in Sec. 3, an additional step of forcing the median disc integrated solar flux density to a constant value has been introduced in the calibration process to ensure a constant fluxscale throughout the observing duration. 2. 2. During the calibration process, a varying number of antennas can get flagged for different time slices even at the same frequency. This can lead to small differences between the psfs for different images. To ensure that this does not affect the flux density estimated, all of the images were smoothed to the same coarser resolution for this work, as described in Sec. 3. ### 5.2 Comparison with M20 The results from this improved methodology confirm the key properties of WINQSEs reported by M20 – weak flux densities, impulsive nature and ubiquitous presence on the quiet Sun during this period of very low solar activity. However, there are some significant quantitative differences which need some discussion. #### 5.2.1 $\Delta F/F$ distribution of WINQSEs This work highlights that the distribution of $\Delta F/F$, can change significantly between observations. At the low $\Delta F/F$ end, the distributions found by M20 tend to flatten out rather than drop, as seen in Fig. 3. At the high $\Delta F/F$ end they show clear power law tails spanning more than an order of magnitude which are clearly missing in this work. For a given frequency, M20 plotted all $\Delta F/F$ values for all times and all the psf sized regions when presenting these histograms, irrespective of the signal-to-noise of the detection. Hence the low end of the $\Delta F/F$ distribution was dominated primarily by noise fluctuations. In contrast, we use an image noise based threshold on $\Delta F$ to limit ourselves only to reliable detections of WINQSEs. This filters out the low signal-to-noise detections and explains the discrepancy between these works at the low end of the $\Delta F/F$ histograms. To substantiate this, we plot in Fig. 7 the data from an example frequency from M20 with the same thresholding technique as used here. It is evident that the low $\Delta F/F$ end of the histogram is now similar to the ones shown in Fig. 3, though it retains the powerlaw behaviour shown in M20 at the large $\Delta F/F$ end. The powerlaw index estimated is consistent with that reported in M20. Similar results were obtained at other frequencies as well. Though M20 had attempted to avoid contamination from the only active region present on the Sun during their observations, we find that despite their efforts the high $\Delta F/F$ tail observed in the histograms obtained in M20 arises due to this lone active region. To substantiate this, the left panel of Fig. 8 shows the location of all WINQSEs detected in M20 at 132 MHz, overlaid on an example radio image at the same frequency. The regions are colored blue (red) if they host at least one WINQSE with $\Delta F/F>0$ ($>1$). It is evident that while WINQSEs are distributed all over the Sun, those with $\Delta F/F>1$ are clustered around the active region and not a single one of them is present far away from it. Active regions have very complex magnetic fields and are sites of continuous magnetic reconnections (e.g. Mondal & Oberoi, 2021). The complexity of the magnetic field is expected to decrease smoothly as the distance from the active region increases. Consequently, the strength and the number of magnetic reconnections taking place would also decrease in a similar smooth manner. We hypothesise this to be the reason for regions showing the presence of strong WINQSEs in M20 to lie close to the active region. Additionally, such a region is expected not only to produce strong WINQSEs, but also to give rise to a larger number of weaker WINQSEs, as compared to regions far away from the active region. Hence we hypothesise this to be the reason behind the presence (absence) of powerlaw behaviour in the $\Delta F/F$ distribution of the detected WINQSEs in M20 (this work). To corroborate this hypothesis, the right panel of Fig. 8 presents the $\Delta F/F$ distribution after removing the contribution of the regions shown in red in the left panel. The red line shows the straight line fitted to the same range of $\Delta F/F$ used for the fit in Fig. 7. It is evident that the powerlaw is no longer a good representation of the distribution. It is also observed that there is hump-like feature around $\Delta F/F\approx 0.5$. The signature of this hump is also observed at a similar location in Fig. 7 in this work and Fig 4 of M20. Understanding the exact details of this feature is beyond the scope of the present work. This observation of the transition from a powerlaw behaviour towards a more lognormal-like nature suggests that as the solar activity increases, more active regions are observed, leading to a larger number of stronger WINQSEs. Hence with increasing solar activity a transition from a lognormal nature to a powerlaw behaviour in the $\Delta F/F$ distribution can be expected. This prediction should be tested in future with data spanning a large part of a solar cycle. Note that the peaks in the $\Delta F/F$ histograms lie at lower values in the current data as compared to M20. We attribute this to a combination of exceptionally quiet solar conditions and lower image noise characteristics delivered by the improved calibration employed here. Figure 7: The data used here correspond to the 132 MHz data presented in M20. The $\Delta F/F$ histogram using data from M20 following the same procedure as that followed here. The red line shows the powerlaw fit to the data beyond the peak. The powerlaw index is mentioned in the figure. Figure 8: Left panel: The blue filled circles show the locations of all the psf sized regions used by M20 with WINQSEs detections. The regions which were determined by M20 to be affected by active region emission and excluded from their analysis have been excluded here as well. The red filled circles show the locations where the $\Delta F/F$ of the detected WINQSEs exceeds 1. The contours correspond to 0.01, 0.02, 0.04, 0.08, 0.16, 0.32, 0.64 times the peak of an example image at 132 MHz. Histogram made using the same data as presented in Fig. 7 after removing contribution of regions which show at least one $\Delta F/F>1$. #### 5.2.2 Fractional occupancy of WINQSEs The median fractional occupancies obtained from the data presented here are 0.017, 0.021, 0.024 and 0.045 at 120, 128, 136 and 146 MHz respectively. In comparison, at similar frequencies of 120 and 132 MHz, M20 had found higher median fractional occupancies of 0.07 and 0.06 respectively. There are at least four known factors which contribute to a higher fractional occupancy for M20. The first is the improved analysis methodology used here which rigorously rejects low signal-to-noise detections. This was not the case with M20 and hence some spurious noise peaks can get counted as WINQSEs inflating the true fractional occupancy number. Secondly, M20 computed a single median for the entire duration, in contrast to the piece-wise median estimated here. The appendix shows the results obtained using a single median and it is evident that the results thus obtained are closer to that obtained by M20 (3-5%). The third important reason is the more active state of the Sun during M20, some evidence and arguments for contributions from which have already been presented. ### 5.3 Comparison with S22 As mentioned earlier, S22 have recently reported the detection of WINQSEs using a very different technique, namely the visibility subtraction technique. For completeness, we briefly describe the technique first and then present a comparison of their technique and the results with the present work. Similar to the technique used here, S22 also assumes that there is a slowly varying background solar radio emission on which are the superposed the impulsive WINQSEs. While we separate the slowly varying and impulsive parts of emission in the image domain by estimating and subtracting a median image, S22 does so in the visibility domain. For each baseline, they compute the median complex visibility over a 15 second running window and assume that the median represents the visibility from the slowly varying Sun for that baseline for the instant centered on that time interval. In principle, subtraction in the visibility domain and subtraction in the image domain should be identical due to the Fourier relationship between them. However, in practice, there can be differences due to non-linear nature of deconvolution which is used to produce the radio images. The subtraction in visibility domain is expected to perform better if there are significant deconvolution errors in the images. A disadvantage of the visibility subtraction technique, however, is that the Fourier component sampled by a given baseline changes as a function of time. For a source with the complex morphology like the Sun, the emission morphology can itself give rise to variations in the observed visbilities on a given baseline even if the source were time invariant. To limit the magnitudes of these intrinsic changes in measured visibilities, S22 used a rather short duration of 15 s over which to compute the median, while our technique was able to use the full 4 min duration of the scan to compute the median image. An implication is that S22 would miss the longer duration events. While this will bias any studies of the temporal width distribution done using this technique, given the very steep slope of this distribution, however, the number of events missed will be a very small fraction of the total number. The solar images used in this work were of high quality and no deconvolution errors or artefacts were evident. In addition to the excellent uv-coverage, precise calibration performed using AIRCARS (Mondal et al., 2019) and the absence of any strong source on the solar disc, all have contributed to ensuring a high fidelity images. This has allowed us to use the image subtraction technique with a high confidence and enabled the detection of these WINQSEs. ### 5.4 Need for a better WINQSE detection technique A shortcoming of both this work and M20 stems from the way WINQSEs are detected. We have searched for WINQSEs in discrete psf-sized regions and classified the transients as true detections if their flux density exceeds some threshold. If a WINQSE is larger than the psf, then it will contribute flux density to multiple regions and hence can lead to over-counting of the number of WINQSEs. On the other hand, since the psf is circular/elliptical in nature, the regions used here and in M20 do not form a closed packed structure. Hence some part of the Sun is not covered by these regions and could lead to some WINQSEs being missed. Additionally, when the spatial location of a WINQSE does not coincide with the centre of some psf sized region, the observed increase in flux density of the region due to the WINQSE is lower than its true flux density. This will lead to instances when the observed flux density will fall short of the detection threshold due to this artificial reduction. Due to these effects, it is non-trivial to estimate the true number of WINQSEs present in the data. To address this shortcoming, we have recently developed a machine learning based technique to reliably identify and characterize WINQSEs in solar radio images. This work will be presented in an upcoming paper (Bawaji et al., 2022). An additional shortcoming of both these works stems from the way the median is computed. M20 had computed a single median map for the entire time duration studied, whereas here we have computed a median map for every five minutes of data. Both of these methods have their own strengths and weaknesses. For example, while calculating an independent median for every five minutes of data, can reduce some systematic calibration artefacts, if present, it can also produce spurious WINQSEs at the five minute data boundaries if for that data chunk and pixel, the lightcurve has a smoothly decreasing or increasing trend. We have repeated the entire analysis using a single median map and, as expected, we find that the number of WINQSEs found in this analysis is significantly larger than when done using independent medians for each data chunk. We are not aware of any systematic effects in the data which can change the flux densities in different regions of the Sun while maintaining the integrated flux at all times. Choosing to err on the side of caution, we have presented the results corresponding to the medians computed every five minutes in the preceding sections. The statistical properties of WINQSEs obtained from the single median analysis are very similar and are presented in the Appendix. While the results are consistent, we realize the need to improve WINQSEs detection strategy beyond what has been employed here. Work on this front is already underway. ### 5.5 A dynamic middle corona Radio observations at the small range of frequencies used here probe coronal height of $\sim 1.4\ (2.0)\ R_{\odot}$, under the standard assumption that the emission arises due to plasma emission mechanism at the fundamental (harmonic) and using well established coronal electron density models. The observations in M20, Sharma et al. (2022) and this work imply that energetically weak reconnections are ever-present even in the quiet Sun regions at these coronal heights. These reconnections accelerate electrons which then emit in the radio band. Due to instrumental limitations at high energies (EUV and X-rays), these coronal heights are typically accessible only at metrewave radio frequencies. The bulk of the radio studies in the past have focused on strong events, again primarily due to the instrumental limitations in radio bands. So there have been rather few investigations of weak or level emissions from the middle coronal heights. As more capable instruments are becoming available, the evidence for middle coronal heights to also be a dynamic place is slowly building. Recently, gradual bulk flows originating in the middle corona have been observed using the Solar Ultraviolet Imager (SUVI) onboard the Geostationary Operational Environmental Satellite (Seaton et al., 2021). The authors suggest that this can happen due to magnetic reconnection processes higher in the corona. While a direct comparison between these EUV emissions and WINQSEs is beyond the scope of this work, these disparate observations paint a consistent picture of a highly dynamic middle solar corona even during extremely quiet times. ## 6 Conclusion M20 was the first work to present ubiquitous detection of WINQSEs. Due to the significant potential implications of this work on coronal heating, and presence of nonthermal particles in the solar corona in general, it is very important to verify their presence under a variety of solar conditions, take a closer look at the analysis and characterise them better. With this motivation, this work forms the next step in this progression. We have examined data from extremely quiet solar conditions, improved on the methodology used in M20 and presented robust detection of large numbers of WINQSEs. This work corroborates the key findings of M20 – their ubiquity and impulsive nature. While significant differences between the distribution of $\Delta F/F$ presented here and M20 are observed, we show that they can be justifiably attributed to differences in methodology and variations in the level of solar activity. Parallel to this effort, independent techniques have been developed to identify and characterise WINQSEs (Bawaji et al., 2022; Sharma et al., 2022) and they all paint a consistent picture. Together these works place the presence of WINQSEs on a firm pedestal. We hope that this will provide the motivation to the wider community to understand them in greater detail. This scientific work makes use of the Murchison Radio-astronomy Observatory, operated by CSIRO. We acknowledge the Wajarri Yamatji people as the traditional owners of the Observatory site. Support for the operation of the MWA is provided by the Australian Government (NCRIS), under a contract to Curtin University administered by Astronomy Australia Limited. We acknowledge the Pawsey Supercomputing Centre which is supported by the Western Australian and Australian Governments. We thank the anonymous referee for their comments, which have improved this paper significantly. SM acknowledges partial support by USA NSF grant AGS-1654382 to the New Jersey Institute of Technology. AB and DO acknowledge support of the Department of Atomic Energy, Government of India, under the project no. 12-R&D-TFR-5.02-0700. As discussed in Sec. 5.4, the number of WINQSEs detected depends on the choice of the median map used. Results presented in the paper are based on the use of an independent median computed for each five minute dataset. For completeness and comparison, here we present results (analogs of Figs. 3–6) obtained using a single median computed over the entire observing duration as was done by Mondal et al. (2020). As expected the number of WINQSEs detected here are larger than using a median computed every five minutes. Figure 9: Histogram of $\Delta F/F$ using a median computed over the entire duration of the data is shown using the blue circles. The red line shows the lognormal fit to these data. The observation frequencies are written in each panel. Figure 10: Temporal width distribution of the detected WINQSEs using a median computed over the entire duration of the data. Figure 11: Wait time distribution of the detected WINQSEs using a median computed over the entire duration of the data. The red and black lines show the wait time till which the uncertainties due to incomplete sampling is less than 10% and 20% respectively. Figure 12: Fractional occupancy distribution of the detected WINQSEs using a median computed over the entire duration of the data. ## References * Aschwanden & McTiernan (2010) Aschwanden, M. J., & McTiernan, J. M. 2010, ApJ, 717, 683, doi: 10.1088/0004-637X/717/2/683 * Astropy Collaboration et al. (2013) Astropy Collaboration, Robitaille, T. P., Tollerud, E. J., et al. 2013, A&A, 558, A33, doi: 10.1051/0004-6361/201322068 * Astropy Collaboration et al. (2018) Astropy Collaboration, Price-Whelan, A. M., Sipőcz, B. M., et al. 2018, AJ, 156, 123, doi: 10.3847/1538-3881/aabc4f * Bain & Fletcher (2009) Bain, H. M., & Fletcher, L. 2009, A&A, 508, 1443, doi: 10.1051/0004-6361/200911876 * Bawaji et al. (2022) Bawaji, S., Alam, U., Mondal, S., & Oberoi, D. 2022, in Astronomical Society of the Pacific Conference Series, Vol. 532, Astronomical Society of the Pacific Conference Series, ed. J. E. Ruiz, F. Pierfedereci, & P. Teuben, 211. https://arxiv.org/abs/2103.05371 * Che (2018) Che, H. 2018, in Journal of Physics Conference Series, Vol. 1100, Journal of Physics Conference Series, 012005, doi: 10.1088/1742-6596/1100/1/012005 * Harris et al. (2020) Harris, C. R., Millman, K. J., van der Walt, S. J., et al. 2020, Nature, 585, 357, doi: 10.1038/s41586-020-2649-2 * Hunter (2007) Hunter, J. D. 2007, Computing in Science & Engineering, 9, 90, doi: 10.1109/MCSE.2007.55 * Kansabanik et al. (2022a) Kansabanik, D., Mondal, S., Oberoi, D., Biswas, A., & Bhunia, S. 2022a, ApJ, 927, 17, doi: 10.3847/1538-4357/ac4bba * Kansabanik et al. (2022b) Kansabanik, D., Oberoi, D., & Mondal, S. 2022b, ApJ, 932, 110, doi: 10.3847/1538-4357/ac6758 * Kivelä & Porter (2015) Kivelä, M., & Porter, M. A. 2015, Phys. Rev. E, 92, 052813, doi: 10.1103/PhysRevE.92.052813 * Kontar et al. (2017) Kontar, E. P., Yu, S., Kuznetsov, A. A., et al. 2017, Nature Communications, 8, 1515, doi: 10.1038/s41467-017-01307-8 * Krucker et al. (2011) Krucker, S., Kontar, E. P., Christe, S., Glesener, L., & Lin, R. P. 2011, ApJ, 742, 82, doi: 10.1088/0004-637X/742/2/82 * Lonsdale et al. (2009) Lonsdale, C. J., Cappallo, R. J., Morales, M. F., et al. 2009, Proceedings of the IEEE, 97, 1497, doi: 10.1109/JPROC.2009.2017564 * McLean & Labrum (1985) McLean, D. J., & Labrum, N. R. 1985, Solar radiophysics : studies of emission from the sun at metre wavelengths * McMullin et al. (2007) McMullin, J. P., Waters, B., Schiebel, D., Young, W., & Golap, K. 2007, in Astronomical Society of the Pacific Conference Series, Vol. 376, Astronomical Data Analysis Software and Systems XVI, ed. R. A. Shaw, F. Hill, & D. J. Bell, 127 * Melrose (1980) Melrose, D. B. 1980, Space Sci. Rev., 26, 3, doi: 10.1007/BF00212597 * Mohan (2021) Mohan, A. 2021, A&A, 655, A77, doi: 10.1051/0004-6361/202142029 * Mohan et al. (2019) Mohan, A., Mondal, S., Oberoi, D., & Lonsdale, C. J. 2019, ApJ, 875, 98, doi: 10.3847/1538-4357/ab0ae5 * Mondal (2021) Mondal, S. 2021, Sol. Phys., 296, 131, doi: 10.1007/s11207-021-01877-3 * Mondal et al. (2019) Mondal, S., Mohan, A., Oberoi, D., et al. 2019, The Astrophysical Journal, 875, 97, doi: 10.3847/1538-4357/ab0a01 * Mondal & Oberoi (2021) Mondal, S., & Oberoi, D. 2021, ApJ, 920, 11, doi: 10.3847/1538-4357/ac1076 * Mondal et al. (2020) Mondal, S., Oberoi, D., & Mohan, A. 2020, The Astrophysical Journal, 895, L39, doi: 10.3847/2041-8213/ab8817 * Parker (1988) Parker, E. N. 1988, ApJ, 330, 474, doi: 10.1086/166485 * Pauluhn & Solanki (2007) Pauluhn, A., & Solanki, S. K. 2007, Astronomy and Astrophysics, 462, 311, doi: 10.1051/0004-6361:20065152 * Pick & Vilmer (2008) Pick, M., & Vilmer, N. 2008, A&A Rev., 16, 1, doi: 10.1007/s00159-008-0013-x * Reid (2020) Reid, H. A. S. 2020, Frontiers in Astronomy and Space Sciences, 7, 56, doi: 10.3389/fspas.2020.00056 * Reid & Kontar (2017) Reid, H. A. S., & Kontar, E. P. 2017, A&A, 606, A141, doi: 10.1051/0004-6361/201730701 * Seaton et al. (2021) Seaton, D. B., Hughes, J. M., Tadikonda, S. K., et al. 2021, Nature Astronomy, 5, 1029, doi: 10.1038/s41550-021-01427-8 * Sharma et al. (2022) Sharma, R., Oberoi, D., Battaglia, M., & Krucker, S. 2022, ApJ, 937, 99, doi: 10.3847/1538-4357/ac87fc * Sturrock (1964) Sturrock, P. A. 1964, Type III Solar Radio Bursts, Vol. 50, 357 * Tajfirouze & Safari (2012) Tajfirouze, E., & Safari, H. 2012, ApJ, 744, 113, doi: 10.1088/0004-637X/744/2/113 * Takakura (1967) Takakura, T. 1967, Sol. Phys., 1, 304, doi: 10.1007/BF00151359 * Tingay et al. (2013) Tingay, S. J., Goeke, R., Bowman, J. D., et al. 2013, Publications of the Astronomical Society of Australia, 30, doi: 10.1017/pasa.2012.007 * Upendran & Tripathi (2021) Upendran, V., & Tripathi, D. 2021, ApJ, 916, 59, doi: 10.3847/1538-4357/abf65a * Virtanen et al. (2020) Virtanen, P., Gommers, R., Oliphant, T. E., et al. 2020, Nature Methods, 17, 261, doi: 10.1038/s41592-019-0686-2 * Wayth et al. (2018) Wayth, R. B., Tingay, S. J., Trott, C. M., et al. 2018, PASA, 35, e033, doi: 10.1017/pasa.2018.37 * Wild & Smerd (1972) Wild, J. P., & Smerd, S. F. 1972, ARA&A, 10, 159, doi: 10.1146/annurev.aa.10.090172.001111 * Wild et al. (1963) Wild, J. P., Smerd, S. F., & Weiss, A. A. 1963, ARA&A, 1, 291, doi: 10.1146/annurev.aa.01.090163.001451
[a,b]Kai Yi # Recent CMS results on exotic resonances Jingqing Zhang and on behalf of the CMS Collaboration ###### Abstract Many exotic resonances have been recently observed at the LHC and other experiments. In this report, CMS studies of exotic multiquark states are reported using the data collected in pp collisions at $\sqrt{s}$ = 13 TeV. ## 1 Selected CMS contributions to heavy exotic states Quantum chromodynamics (QCD) is an important part of the standard model (SM) in particle physics and has gotten much support from experimental results as has the rest of the SM. In the QCD framework, hadrons beyond the quark configurations of $q\bar{q}$ and $qqq$ ($\bar{q}\bar{q}\bar{q}$) are ‘exotic’ and are allowed in the QCD theory. Experimental studies on the exotic hadrons will help deepen our understanding of QCD. The CMS experiment [1] at the LHC has performed many important studies in hadron spectroscopy and the exotic hadron sector [2, 3, 4, 5, 6, 7, 8, 9], which include the measurement of the $X(3872)$ production cross section [2], the confirmation of the $Y(4140)$ in $B^{\pm}\rightarrow J/\psi\phi K^{\pm}$ decays [4], and observation of the $B^{0}_{s}\rightarrow X(3872)\phi$ decay [9] —all in proton-proton collisions. Here we present recent results on exotic resonances from the CMS experiment: evidence for the $X(3872)$ in heavy-ion collisions [10]; observation of the $B^{0}_{s}\rightarrow\psi(2S)K^{0}_{S}$ and the $B^{0}\rightarrow\psi(2S)K^{0}_{S}\pi^{+}\pi^{-}$ decays [11], and observation of new structures in the $J/\psi J/\psi$ mass spectrum [12]. ## 2 Evidence for $X(3872)$ in PbPb collisions The $X(3872)$ was first observed by the Belle Collaboration [13]. Although its quantum numbers have been determined to be $J^{PC}=1^{++}$ by the LHCb collaboration [14], its nature is still not fully understood. The production and survival of the $X(3872)$ in relativistic heavy ion collisions is expected to depend on the $X(3872)$’s internal structure [15, 16]. Therefore, study of the $X(3872)$ production in relativistic heavy-ion collisions provides new opportunities to probe the nature of the $X(3872)$. The CMS Collaboration performed a study of $X(3872)$ production in Pb-Pb collisions at $\sqrt{S_{NN}}=5.02$ TeV using 1.7 nb-1 sample collected in 2018 [10]. The candidates for the $X(3872)$ and $\psi(2S)$ are reconstructed via their decays into $J/\psi\pi^{+}\pi^{-}$, where the $J/\psi$ decays into $\mu^{+}\mu^{-}$. Figure 1 shows the observed $m_{\mu\mu\pi\pi}$ distribution for the $X(3872)$ and $\psi(2S)$ candidates, where the upper plot shows the inclusive sample and the bottom one shows the b-enriched (nonprompt dominated, i.e. transverse decay length $l_{xy}>0.1$ mm) sample. The significance of the inclusive $X(3872)$ signal is 4.2 standard deviations. The prompt $X(3872)$ to $\psi(2S)$ yield ratio is found to be $1.08\pm 0.49(stat)\pm 0.52(syst)$, whereas the typical value is around 0.1 in pp collisions. Figure 1: Invariant mass distribution of $m_{\mu\mu\pi\pi}$ in Pb-Pb collisions, for the inclusive (upper) and $b$-enriched (bottom) samples [10]. The vertical lines on points represent statistical uncertainties in the data. The results of the unbinned maximum-likelihood fits for the signal + background, and background alone, are also shown by the solid and dashed lines, respectively. The pull distribution is represented by the shaded bars. The $X(3872)$ peak mass resolution, $\sigma_{X(3872)}$, calculated at the half-maximum of the signal peak, is also listed for reference. ## 3 Observation of the $B^{0}_{s}\rightarrow\psi(2S)K^{0}_{S}$ and the $B^{0}\rightarrow\psi(2S)K^{0}_{S}\pi^{+}\pi^{-}$ decays Multibody decays of the B mesons are well suited to the search for, and study of, exotic resonances. For example, the discovery of $X(3872)$ was in $B\rightarrow KJ/\psi\pi\pi$ decays [13], and that of the first charged tetraquark candidate, $Z(4430)^{+}$, was in $B\rightarrow\psi(2S)K\pi^{\pm}$ [17]. The CMS experiment performed the first measurement of the $B^{0}_{s}\rightarrow\psi(2S)K^{0}_{S}$ and $B^{0}\rightarrow\psi(2S)K^{0}_{S}\pi^{+}\pi^{-}$ decays, using a data sample of proton-proton collisions at $\sqrt{s}=13$ TeV, and an integrated luminosity of 103 fb-1, collected in 2017 and 2018 [11]. The $\psi(2S)$ and $K^{0}_{S}$ mesons are reconstructed using their decays into $\mu^{+}\mu^{-}$ and $\pi^{+}\pi^{-}$, respectively. The observed invariant mass distribution of $\psi(2S)K^{0}_{S}$ (left) and $\psi(2S)K^{0}_{S}\pi^{+}\pi^{-}$ (right) are shown in Fig. 2. Using the $B^{0}\rightarrow\psi(2S)K^{0}_{S}$ as a reference channel, the relative branching fractions of $B^{0}_{s}\rightarrow\psi(2S)K^{0}_{S}$ and $B^{0}\rightarrow\psi(2S)K^{0}_{S}\pi^{+}\pi^{-}$ decays are measured to be $\mathcal{B}(B^{0}_{s}\rightarrow\psi(2S)K^{0}_{S})/\mathcal{B}(B^{0}\rightarrow\psi(2S)K^{0}_{S})=(3.33\pm 0.69(stat)\pm 0.11(syst)\pm 0.34(f_{s}/f_{d}))\times 10^{-2}$, and $\mathcal{B}(B^{0}\rightarrow\psi(2S)K^{0}_{S}\pi^{+}\pi^{-})/\mathcal{B}(B^{0}\rightarrow\psi(2S)K^{0}_{S})=0.480\pm 0.013(stat)\pm 0.032(syst)$, where the last uncertainty in the first ratio corresponds to the uncertainty in the ratio of the production cross sections of $B^{0}_{s}$ and $B^{0}$ mesons. With the currently available, statistics- limited data, the 2- and 3- body invariant mass distributions of the $B^{0}\rightarrow\psi(2S)K^{0}_{S}\pi^{+}\pi^{-}$ decay products do not show significant exotic narrow structures in addition to the known light meson resonances. Figure 2: Measured mass distribution of $\psi(2S)K^{0}_{S}$ (left) and $\psi(2S)K^{0}_{S}\pi^{+}\pi^{-}$ (right) candidates [11]. ## 4 Observation of new structures in the $J/\psi J/\psi$ mass spectrum in pp collisions The $X(3872)$ and many other exotic candidates contain two heavy quarks ($c\bar{c}$). An analogue to heavy quarkonia would be fully heavy tetraquarks, which have been explored in theoretical models and are expected to be experimentally observable. The recent observation of the $X(6900)$ decaying into $J/\psi J/\psi$ has been reported by the LHCb Collaboration [18]. The CMS experiment performed a study of the low-mass region of the $J/\psi J/\psi$ mass spectrum in pp collisions, using a data sample corresponding to an integrated luminosity of 135 fb-1 at a center-of-mass energy of 13 TeV [12]. The two $J/\psi$ candidates are reconstructed using their $\mu^{+}\mu^{-}$ mode, and the final $J/\psi J/\psi$ mass distribution is shown in Fig. 3, where three signal Breit-Wigner structures and a background component are used to fit the distribution. The statistical significance of the three structures are $6.5\sigma$, $9.4\sigma$, and $4.1\sigma$ for $X(6600)$, $X(6900)$ and $X(7300)$, respectively. The measured masses and widths of three structures are summarized in Table 1. Figure 3: The CMS $J/\psi J/\psi$ mass spectrum with a fit consisting of three signal BW functions and a background model [12]. The left plot shows the fit over the full mass range, and on the right is the same fit expanded by only displaying masses below 9 GeV. | BW1 | BW2 | BW3 ---|---|---|--- $m$ | $6552\pm 10\pm 12$ | $6927\pm 9\pm 5$ | $7287\pm 19\pm 5$ $\Gamma$ | $124\pm 29\pm 34$ | $122\pm 22\pm 19$ | $95\pm 46\pm 20$ $N$ | $474\pm 113$ | $492\pm 75$ | $156\pm 56$ Table 1: Summary of the fit results of the CMS $m(J/\psi J/\psi)$ distribution: the mass $m$ and natural width $\Gamma$, in MeV, and the signal yields $N$ are given for three signal structures [12]. The first uncertainties are statistical and the second systematic. Our $X(6900)$ parameters are in a good agreement with LHCb’s non-interference result, while the $X(6600)$ and $X(7300)$ are new structures. In order to remove potential model dependencies in a comparison of the $X(6900)$ results, we also apply the principal two LHCb fit models to the CMS data, but using CMS-specific background shapes. Figure 4 shows the application of LHCb’s Model I (left, non-interference) and Model II (right, non-resonant single parton scattering (NRSPS) interfering with a Breit-Wigner structure – the $X(6700)$ in our application). LHCb’s Model I consists of the $X(6900)$ signal, NRSPS, non-resonant double parton scattering (NRDPS) and two more BWs – around 6300 (BW0) and 6500 MeV (BW1) – to account for the threshold enhancement. LHCb’s Model II consists of the $X(6900)$ signal, NRDPS, and the interference contribution of a Breit-Wigner structure $X(6700)$ and NRSPS. In both models, the $X(6900)$ parameters are in a good agreement with LHCb’s measurements, while our $X(6700)$ in Model II has a much larger amplitude and width compared to the LHCb’s interfering Breit-Wigner, and none of the LHCb models provide a satisfactory description of our data. Figure 4: The CMS $J/\psi J/\psi$ mass spectrum [12]. The data are fit using LHCb models: Model I (non-interference) on the left and Model II (interference) on the right. ## 5 Summary In summary, recent CMS results are presented, including the first evidence for the $X(3872)$ production in heavy ion collisions, the first observation of the $B^{0}\rightarrow\psi(2S)K^{0}_{S}\pi^{+}\pi^{-}$ and $B^{0}_{s}\rightarrow\psi(2S)K^{0}_{S}$, and the observation of new structures in $J/\psi J/\psi$ mass spectrum in pp collisions. ## References * [1] CMS collaboration, S. Chatrchyan et al., _The CMS Experiment at the CERN LHC_ , _JINST_ 3 (2008) S08004. * [2] CMS collaboration, S. Chatrchyan et al., _Measurement of the $X$(3872) Production Cross Section Via Decays to $J/\psi\pi^{+}\pi^{-}$ in $pp$ collisions at $\sqrt{s}$ = 7 TeV_, _JHEP_ 04 (2013) 154 [1302.3968]. * [3] CMS collaboration, S. Chatrchyan et al., _Search for a New Bottomonium State Decaying to $\Upsilon(1S)\pi^{+}\pi^{-}$ in pp Collisions at $\sqrt{s}$ = 8 TeV_, _Phys. Lett. B_ 727 (2013) 57 [1309.0250]. * [4] CMS collaboration, S. Chatrchyan et al., _Observation of a Peaking Structure in the $J/\psi\phi$ Mass Spectrum from $B^{\pm}\to J/\psi\phi K^{\pm}$ Decays_, _Phys. Lett. B_ 734 (2014) 261 [1309.6920]. * [5] CMS collaboration, A. M. Sirunyan et al., _Search for the X(5568) state decaying into $\mathrm{B}^{0}_{\mathrm{s}}\pi^{\pm}$ in proton-proton collisions at $\sqrt{s}=$ 8 TeV_, _Phys. Rev. Lett._ 120 (2018) 202005 [1712.06144]. * [6] CMS collaboration, A. M. Sirunyan et al., _Study of the ${\mathrm{B}}^{+}\to\mathrm{J}/\psi\overline{\Lambda}\mathrm{p}$ decay in proton-proton collisions at $\sqrt{s}$ = 8 TeV_, _JHEP_ 12 (2019) 100 [1907.05461]. * [7] CMS collaboration, A. M. Sirunyan et al., _Observation of the $\Lambda_{\mathrm{b}}^{0}\to$ J/$\psi\Lambda\phi$ decay in proton-proton collisions at $\sqrt{s}=$ 13 TeV_, _Phys. Lett. B_ 802 (2020) 135203 [1911.03789]. * [8] CMS collaboration, A. M. Sirunyan et al., _Measurement of the $\Upsilon$(1S) pair production cross section and search for resonances decaying to $\Upsilon$(1S)$\mu^{+}\mu^{-}$ in proton-proton collisions at $\sqrt{s}=$ 13 TeV_, _Phys. Lett. B_ 808 (2020) 135578 [2002.06393]. * [9] CMS collaboration, A. M. Sirunyan et al., _Observation of the B ${}^{0}_{\mathrm{s}}\to$X(3872)$\phi$ decay_, _Phys. Rev. Lett._ 125 (2020) 152001 [2005.04764]. * [10] CMS collaboration, A. M. Sirunyan et al., _Evidence for X(3872) in Pb-Pb Collisions and Studies of its Prompt Production at $\sqrt{s_{NN}}$=5.02 TeV_, _Phys. Rev. Lett._ 128 (2022) 032001 [2102.13048]. * [11] CMS collaboration, A. Tumasyan et al., _Observation of B 0$\to$$\psi$(2S)K${}^{0}_{\mathrm{S}}\pi^{+}\pi^{-}$ and B${}^{0}_{\mathrm{S}}$$\to$$\psi$(2S)K${}^{0}_{\mathrm{S}}$ decays_, _Eur. Phys. J. C_ 82 (2022) 499 [2201.09131]. * [12] CMS collaboration, _Observation of new structures in the $\mathrm{J}/\psi\mathrm{J}/\psi$ mass spectrum in $\mathrm{p}\mathrm{p}$ collisions at $\sqrt{s}=13$ TeV_, _CMS-PAS-BPH-21-003_ (2022) . * [13] Belle collaboration, S. K. Choi et al., _Observation of a narrow charmonium-like state in exclusive $B^{\pm}\to K^{\pm}\pi^{+}\pi^{-}J/\psi$ decays_, _Phys. Rev. Lett._ 91 (2003) 262001 [hep-ex/0309032]. * [14] LHCb collaboration, R. Aaij et al., _Determination of the X(3872) meson quantum numbers_ , _Phys. Rev. Lett._ 110 (2013) 222001 [1302.6269]. * [15] H. Zhang, J. Liao, E. Wang, Q. Wang and H. Xing, _Deciphering the Nature of X(3872) in Heavy Ion Collisions_ , _Phys. Rev. Lett._ 126 (2021) 012301 [2004.00024]. * [16] B. Wu, X. Du, M. Sibila and R. Rapp, _$X(3872)$ transport in heavy-ion collisions_, _Eur. Phys. J. A_ 57 (2021) 122 [2006.09945]. * [17] Belle collaboration, S. K. Choi et al., _Observation of a resonance-like structure in the $\pi^{\pm}\psi^{\prime}$ mass distribution in exclusive $B\to K\pi^{\pm}\psi^{\prime}$ decays_, _Phys. Rev. Lett._ 100 (2008) 142001 [0708.1790]. * [18] LHCb collaboration, R. Aaij et al., _Observation of structure in the $J/\psi$ -pair mass spectrum_, _Sci. Bull._ 65 (2020) 1983 [2006.16957].
# Subquadratic Weighted Matroid Intersection Under Rank Oracles††thanks: To appear in the _33rd International Symposium on Algorithms and Computation (ISAAC 2022)_. Ta-Wei Tu Department of Computer Science and Information Engineering, National Taiwan University, Taipei, Taiwan. Email<EMAIL_ADDRESS> ###### Abstract Given two matroids $\mathcal{M}_{1}=(V,\mathcal{I}_{1})$ and $\mathcal{M}_{2}=(V,\mathcal{I}_{2})$ over an $n$-element integer-weighted ground set $V$, the weighted matroid intersection problem aims to find a common independent set $S^{*}\in\mathcal{I}_{1}\cap\mathcal{I}_{2}$ maximizing the weight of $S^{*}$. In this paper, we present a simple deterministic algorithm for weighted matroid intersection using $\widetilde{O}(nr^{3/4}\log{W})$ rank queries, where $r$ is the size of the largest intersection of $\mathcal{M}_{1}$ and $\mathcal{M}_{2}$ and $W$ is the maximum weight. This improves upon the best previously known $\widetilde{O}(nr\log{W})$ algorithm given by Lee, Sidford, and Wong [FOCS’15], and is the first subquadratic algorithm for polynomially-bounded weights under the standard independence or rank oracle models. The main contribution of this paper is an efficient algorithm that computes shortest- path trees in weighted exchange graphs. ## 1 Introduction #### Matroid Intersection. A matroid is an abstract structure that models the notion of independence on a given ground set $V$. In particular, a subset $S\subseteq V$ is either _independent_ or _dependent_ , such that the family of independent sets is well-structured (see Section 2 for a complete definition). Matroids model many fundamental combinatorial objects, and examples of independent sets of a matroid include acyclic subgraphs of an undirected graph and linearly independent rows of a matrix. One of the most important optimization problems related to matroids is _matroid intersection_ : Given two matroids, we would like to find a set with the largest cardinality that is independent in both matroids. Similarly, in the weighted case, each element in the ground set is associated with an integer weight, and the weighted matroid intersection problem is to find the maximum-weight common independent set. These problems have been extensively studied in the past since they capture many combinatorial optimization problems such as bipartite matching and colorful spanning trees. #### Oracle Model. Since we are dealing with general matroids without additional constraints, we have to specify a way of reading the description of the two matroids. One way is to express them directly by reading the truth table of independence. However, that would require an exponentially-sized input. Instead, we are given oracle access to the matroids, which gives us information about a queried set $S\subseteq V$. Standard oracles include the _independence oracle_ , which returns whether $S$ is independent in $O(\mathcal{T}_{\mathsf{ind}})$ time, and the _rank oracle_ , which returns the rank, i.e., the size of the largest independent subset, of $S$ in $O(\mathcal{T}_{\mathsf{rank}})$ time. In this paper, we focus on the stronger rank oracle model. #### Prior Work. Polynomial-time algorithms for both the weighted and unweighted matroid intersection problems have long been designed and improved. For the unweighted case, Edmonds [Edm10, Edm01, Edm79], Lawler [Law75], and also Aigner and Dowling [AD71] gave algorithms that run in $O(nr^{2}\cdot\mathcal{T}_{\mathsf{ind}})$ time. Here, $n$ denotes the number of elements in $V$ and $r$ denotes the size of the largest intersection of the two matroids. Cunningham [Cun86] obtained an $O(nr^{3/2}\cdot\mathcal{T}_{\mathsf{ind}})$ algorithm using the “blocking- flow” idea. Lee, Sidford, and Wong [LSW15] gave quadratic algorithms using the cutting-plane method, running in $\widetilde{O}(nr\cdot\mathcal{T}_{\mathsf{rank}}+n^{3})$ and $\widetilde{O}(n^{2}\cdot\mathcal{T}_{\mathsf{ind}}+n^{3})$ times111 For function $f(n)$, $\widetilde{O}(f(n))$ denotes $O(f(n)\operatorname{polylog}{f(n)})$. , respectively. This also gives rise to the “quadratic barrier” of matroid intersection: most previous algorithms involve building exchange graphs that contain $\Theta(nr)$ edges explicitly and therefore cannot go beyond quadratic time. Chakrabarty, Lee, Sidford, Singla, and Wong [CLS+19] were the first to partially break the barrier. They obtained a $(1-\epsilon)$-approximation algorithm running in $\widetilde{O}(n^{3/2}/\epsilon^{3/2}\cdot\mathcal{T}_{\mathsf{ind}})$ time and an $\widetilde{O}(n\sqrt{r}\cdot\mathcal{T}_{\mathsf{rank}})$ exact algorithm. One of the major components of Chakrabarty et al.’s improvements is to show that edges in exchange graphs can be efficiently discovered using binary search (this was discovered independently by Nguyễn [Ngu19]). This technique also allows them to obtain improved $\widetilde{O}(nr\cdot\mathcal{T}_{\mathsf{ind}})$ exact algorithms. Combining the approximation algorithm and a faster augmenting-path algorithm, Blikstad, van den Brand, Mukhopadhyay, and Nanongkai [BvdBMN21] broke the quadratic barrier completely by giving an $\widetilde{O}(n^{9/5}\cdot\mathcal{T}_{\mathsf{ind}})$ exact algorithm. This result was later optimized by Blikstad [Bli21] to $\widetilde{O}(nr^{3/4}\cdot\mathcal{T}_{\mathsf{ind}})$ by improving the approximation algorithm to run in $\widetilde{O}(n\sqrt{r}/\epsilon\cdot\mathcal{T}_{\mathsf{ind}})$ time. For the weighted case, the blocking flow idea does not seem to apply anymore. Frank [Fra81] obtained an $O(nr^{2}\cdot\mathcal{T}_{\mathsf{ind}})$ algorithm by characterizing the optimality of a common independent set using weight splitting. Fujishige and Zhang [FZ95] improved the running time to $\widetilde{O}(nr^{3/2}\log{W}\cdot\mathcal{T}_{\mathsf{ind}})$ by solving a more general _independent assignment_ problem using a scaling framework. The same bound was achieved by Shigeno and Iwata [SI95] and also by Gabow and Xu [GX96]. Lee, Sidford, and Wong’s [LSW15] algorithms work for the weighted case as well, albeit with an extra factor of $\operatorname{polylog}{W}$, in $\widetilde{O}(n^{2}\log{W}\cdot\mathcal{T}_{\mathsf{ind}}+n^{3}\operatorname{polylog}{W})$ and $\widetilde{O}(nr\log{W}\cdot\mathcal{T}_{\mathsf{rank}}+n^{3}\operatorname{polylog}{W})$ times. Huang, Kakimura, and Kamiyama [HKK16] obtained a generic framework that transforms any algorithm that solves the unweighted case into one that solves the weighted case with an extra $O(W)$ factor. Plugging in the state-of-the- art algorithms of [CLS+19] and [Bli21], we get $\widetilde{O}(n\sqrt{r}\cdot W\cdot\mathcal{T}_{\mathsf{rank}})$ and $\widetilde{O}(nr^{3/4}\cdot W\cdot\mathcal{T}_{\mathsf{ind}})$ algorithms. Chekuri and Quanrud [CQ16] also gave an $\widetilde{O}(n^{2}/\epsilon^{2}\cdot\mathcal{T}_{\mathsf{ind}})$ approximation algorithm which, according to [BvdBMN21], can be improved to subquadratic by applying more recent techniques. A similar $\widetilde{O}(nr^{3/2}/\epsilon\cdot\mathcal{T}_{\mathsf{ind}})$ approximation algorithm was obtained independently by Huang et al. [HKK16]. #### Our Result. The question of whether weighted matroid intersection can be solved in subquadratic time with polylogarithmic dependence on $W$ under either oracle model remained open. We obtain the first subquadratic algorithm for exact weighted matroid intersection under rank oracles. The formal statement of Theorem 1.1 is presented as Theorem 2.4 in Section 2. ###### Theorem 1.1. Weighted matroid intersection can be solved in $\widetilde{O}(nr^{3/4}\log{W}\cdot\mathcal{T}_{\mathsf{rank}})$ time. Our algorithm relies on the framework of Fujishige-Zhang [FZ95] and Shigeno- Iwata [SI95], where they first obtain an approximate solution by adjusting weights of some elements (similar to the “auction“ algorithms for bipartite matching [OA92]) and then refine it by augmenting the solution iteratively. We obtain efficient algorithms for these two phases, leading to the final subquadratic algorithm. ## 2 Preliminaries #### Notation. For a set $S$, let $|S|$ denote the cardinality and $2^{S}$ the power set of $S$. Let $S\setminus R$ consist of elements of $S$ which are not in $R$. Let $e=(u,v,w)$ denote a weighted directed edge directing from $u$ to $v$ with weight $w=w(e)$ and $(u,v)$ be its unweighted counterpart. Let $\operatorname{head}(e)=v$ and $\operatorname{tail}(e)=u$. For an edge set $E$, let $\operatorname{head}(E)=\\{\operatorname{head}(e)\mid e\in E\\}$ and $\operatorname{tail}(E)=\\{\operatorname{tail}(e)\mid e\in E\\}$. For functions $f,g$ mapping from a set $V$ to $\mathbb{R}$, let $f+g$, $f-g$, and $f+c$ for $c\in\mathbb{R}$ denote functions from $V$ to $\mathbb{R}$ with $(f+g)(x)=f(x)+g(x)$, $(f-g)(x)=f(x)-g(x)$, and $(f+c)(x)=f(x)+c$ for each $x\in V$. We often abuse notation and use $f$ to denote the function from $2^{V}$ to $\mathbb{R}$ with $f(S)=\sum_{x\in S}f(x)$ for each $S\subseteq V$. #### Matroid. Let $V$ be a finite set and $w:V\to\mathbb{Z}$ be a given weight function. For $S\subseteq V$, let $\overline{S}=V\setminus S$. Let $n=|V|$ and $W=\max_{x\in V}|w(x)|$. An ordered pair $\mathcal{M}=(V,\mathcal{I})$ with _ground set_ $V$ and a non-empty family $\emptyset\in\mathcal{I}\subseteq 2^{V}$ is a _matroid_ if M1. for each $S\in\mathcal{I}$ and $R\subseteq S$, it holds that $R\in\mathcal{I}$, and M2. for each $R,S\in\mathcal{I}$ with $|R|<|S|$, there exists an $x\in S\setminus R$ such that $R\cup\\{x\\}\in\mathcal{I}$. Sets in $\mathcal{I}$ are _independent_ ; sets not in $\mathcal{I}$ are _dependent_. A _basis_ is a maximal independent set. A _circuit_ is a minimal dependent set. It is well-known from the definition of matroid that all bases are of the same cardinality. For an independent set $S$ and $x\not\in S$, $S\cup\\{x\\}$ contains at most one circuit $C$ and if it does, then $x\in C$ (see [Pri15, Lemma 1.3.3]). The _rank_ of $S\subseteq V$, denoted by $\mathsf{rank}(S)$, is the size of the largest $S^{\prime}\subseteq S$ such that $S^{\prime}\in\mathcal{I}$. The rank of $\mathcal{M}$ is the rank of $V$, i.e., the size of the bases of $\mathcal{M}$. Given two matroids $\mathcal{M}_{1}=(V,\mathcal{I}_{1})$ and $\mathcal{M}_{2}=(V,\mathcal{I}_{2})$ over the same ground set, the _weighted matroid intersection_ problem is to find an $S^{*}\in\mathcal{I}_{1}\cap\mathcal{I}_{2}$ maximizing $w(S^{*})$. Let $r=\max_{S\in\mathcal{I}_{1}\cap\mathcal{I}_{2}}|S|$. In this paper, the two matroids are accessed through _rank oracles_ , one for each matroid. Specifically, let $\mathsf{rank}_{1}(\cdot)$ and $\mathsf{rank}_{2}(\cdot)$ denote the rank functions of $\mathcal{M}_{1}$ and $\mathcal{M}_{2}$, respectively. We assume that given pointers to a linked list containing elements of $S$ (see, e.g., [CLSW17]), the rank oracles compute $\mathsf{rank}_{1}(S)$ and $\mathsf{rank}_{2}(S)$ in $O(\mathcal{T}_{\mathsf{rank}})$ time. With the $O(n\sqrt{r}\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ unweighted matroid intersection algorithm of Chakrabarty et al. [CLS+19], we also assume that $\mathcal{M}_{1}$ and $\mathcal{M}_{2}$ are of the same rank and share a common basis $S^{(0)}$ of size $r$ by adjusting the given rank oracles properly.222We can compute $r$ via the unweighted matroid intersection algorithm and regard all sets of size greater than $r$ as dependent. By adding $r$ zero-weight elements to $V$, we may also assume that each common independent set $S\in\mathcal{I}_{1}\cap\mathcal{I}_{2}$ is contained in a common basis of the same weight.333In particular, let $Z=\\{z_{1},\ldots,z_{r}\\}$ be the set of newly added zero-weight elements. For each $i\in\\{1,2\\}$, instead of working with $\mathcal{M}_{i}$, we now work with $\widetilde{\mathcal{M}}_{i}=(V\cup Z,\widetilde{\mathcal{I}}_{i})$ such that for each $\widetilde{S}\subseteq V\cup Z$, $\widetilde{S}\in\widetilde{\mathcal{I}}_{i}$ if and only if $|\widetilde{S}|\leq r$ and $\widetilde{S}\setminus Z\in\mathcal{I}_{i}$. This change is reflected in the new rank function $\widetilde{\mathsf{rank}}_{i}(\widetilde{S})=\min(\mathsf{rank}_{i}(\widetilde{S}\setminus Z)+|\widetilde{S}\cap Z|,r)$, which can be implemented via the given oracle $\mathsf{rank}_{i}$. Therefore, it suffices to find a common basis $S^{*}$ maximizing $w(S^{*})$. Note that elements with negative weights can be safely discarded from $V$. #### Weight-Splitting. For weight function $f:V\to\mathbb{R}$, a basis $S$ of matroid $\mathcal{M}$ is _$f$ -maximum_ if $f(S)\geq f(R)$ holds for each basis $R$ of $\mathcal{M}$. Let $w^{\epsilon}=(w^{\epsilon}_{1},w^{\epsilon}_{2})$ with $w^{\epsilon}_{i}:V\to\mathbb{R}$ being a weight function for each $i\in\\{1,2\\}$. Let $w^{\epsilon}(x)=w^{\epsilon}_{1}(x)+w^{\epsilon}_{2}(x)$. We say that $w^{\epsilon}$ is an _$\epsilon$ -splitting_ (see, e.g., [SI95]) of $w$ with $\epsilon>0$ if $w(x)\leq w^{\epsilon}(x)\leq w(x)+\epsilon$ holds for each $x\in V$. If $w^{\epsilon}$ is an $\epsilon$-splitting of $w$ and $S_{i}$ is a $w^{\epsilon}_{i}$-maximum basis of $\mathcal{M}_{i}$ for each $i\in\\{1,2\\}$, then we call $(w^{\epsilon},S)$ with $S=(S_{1},S_{2})$ an _$\epsilon$ -partial-solution_ of $w$. Note that by M1., $S_{1}\cap S_{2}$ is a common independent set. If $S_{1}=S_{2}$, then $(w^{\epsilon},S)$ is an _$\epsilon$ -solution_ of $w$. In this case, we may abuse notation and refer to $S_{1}$ as simply $S$. #### Matroid Algorithms. The unweighted version of the following lemma was shown in [CLS+19] (it is also mentioned in [Ngu19]), and it was extended to the weighted case implicitly in [BvdBMN21]. ###### Lemma 2.1 ([CLS+19, Lemma 13], [Ngu19], and [BvdBMN21]). For $i\in\\{1,2\\}$, given $S\in\mathcal{I}_{i}$, $B\subseteq S$ (respectively, $B\subseteq\overline{S}$), $x\in\overline{S}$ (respectively, $x\in S$), and weight function $f:V\to\mathbb{R}$, it takes $O(\log{|B|}\cdot\mathcal{T}_{\mathsf{rank}})$ time to either obtain a $b\in B$ minimizing/maximizing $f(b)$ such that $(S\setminus\\{b\\})\cup\\{x\\}\in\mathcal{I}_{i}$ (respectively, $(S\setminus\\{x\\})\cup\\{b\\}\in\mathcal{I}_{i}$) or report that such an element does not exist in $B$. The main idea of Lemma 2.1 is to perform binary search on $B$ ordered by $f$. Throughout this paper, we will maintain such an ordered set in a balanced binary search tree where each element holds pointers to its successor and predecessor and each node holds pointers to the first and the last elements in its corresponding subtree. This allows us to perform binary search on the tree and obtain pointers to the linked list containing elements in a consecutive range efficiently. The following greedy algorithm for finding a maximum-weight basis is folklore. ###### Lemma 2.2 (See, e.g., [Edm71]). It takes $O(n\log{n}+n\mathcal{T}_{\mathsf{rank}})$ time to obtain a $f$-maximum basis $S$ of a given matroid $\mathcal{M}$ and weight function $f:V\to\mathbb{R}$. ### 2.1 The Framework The core of our algorithm is the following subroutine. ###### Theorem 2.3. Given a $2\epsilon$-solution $(w^{2\epsilon},S^{\prime})$ of $w$, it takes $O(nr^{3/4}\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time to obtain an $\epsilon$-solution $(w^{\epsilon},S)$. With Theorem 2.3, the weighted matroid intersection algorithm follows from the standard weight-scaling framework (see, e.g, [FZ95, SI95]). Recall that our goal is to find a maximum-weight common basis. ###### Theorem 2.4 (Weighted Matroid Intersection). Given two matroids $\mathcal{M}_{1}=(V,\mathcal{I}_{1})$ and $\mathcal{M}_{2}=(V,\mathcal{I}_{2})$, it takes $O(nr^{3/4}\log{n}\log{(rW)}\cdot\mathcal{T}_{\mathsf{rank}})$ time to obtain an $S^{*}\in\mathcal{I}_{1}\cap\mathcal{I}_{2}$ maximizing $w(S^{*})$. ###### Proof. Let $w^{W}=(w^{W}_{1},w^{W}_{2})$ with $w^{W}_{i}(x)=\frac{W}{2}$ for each $x\in V$ and the initial common basis $S^{(0)}$ obtained via the unweighted matroid intersection algorithm be a $W$-solution of $w$. Repeatedly apply Theorem 2.3 for $O(\log{rW})$ iterations to obtain a $\frac{1}{2r}$-solution $(w^{\frac{1}{2r}},S^{*})$. For each $S\in\mathcal{I}_{1}\cap\mathcal{I}_{2}$, we have $w(S)\leq w^{\frac{1}{2r}}(S)\leq w^{\frac{1}{2r}}(S^{*})\leq w(S^{*})+r\cdot\frac{1}{2r}<w(S^{*})+1.$ Since $w(S)$ and $w(S^{*})$ are integers, $S^{*}$ is a maximum-weight common basis. The algorithm runs in $O(nr^{3/4}\log{n}\log{(rW)}\cdot\mathcal{T}_{\mathsf{rank}})$ time. The theorem is proved. ∎ The rest of the paper proves Theorem 2.3. ## 3 The Algorithm As in [FZ95] and [SI95], the algorithm of Theorem 2.3 consists of the following two parts. ### 3.1 Weight Adjustment The first part of the algorithm is the following subroutine which computes two bases $S_{1}$ and $S_{2}$ with a large enough intersection. This part is essentially the same as Shigeno and Iwata’s algorithm [SI95], except that we replace the fundamental (co-)circuit queries in it with calls to Lemma 2.1. ###### Lemma 3.1. Given a $2\epsilon$-solution $(w^{2\epsilon},S^{\prime})$ and a parameter $1\leq k\leq r$, it takes $O(nk\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time to obtain an $\epsilon$-partial-solution $(w^{\epsilon},S)$ with $|S_{1}\cap S_{2}|\geq\left(1-\frac{O(1)}{k}\right)r$. Since the algorithm and analysis are essentially the same as in [SI95], here we only describe how we can obtain $S_{1}$ and $S_{2}$ in the desired time bound. Please refer to [SI95] or Lemma A.1 in Appendix A for the proof of $|S_{1}\cap S_{2}|\geq\left(1-\frac{O(1)}{k}\right)r$. #### Algorithm of Lemma 3.1. Let $w^{\epsilon}=(w^{2\epsilon}_{1},w-w^{2\epsilon}_{1}+\epsilon)$ be the initial $\epsilon$-splitting and $S_{i}$ be the $w^{\epsilon}_{i}$-maximum basis of $\mathcal{M}_{i}$ obtained by Lemma 2.2 in $O(n\log{n}+n\mathcal{T}_{\mathsf{rank}})$ time for each $i\in\\{1,2\\}$. Let $p(x)=0$ for each $x\in V$. Repeat the following _weight adjustment_ for an arbitrary $x\in S_{1}\setminus S_{2}$ with $p(x)<k$ until such an $x$ becomes non-existent. * • If $w^{\epsilon}(x)=w(x)+\epsilon$, then set $w^{\epsilon}_{1}(x)\leftarrow w^{\epsilon}_{1}(x)-\epsilon$. Apply Lemma 2.1 to obtain a $y\in V\setminus S_{1}$ maximizing $w^{\epsilon}_{1}(y)$ such that $(S_{1}\setminus\\{x\\})\cup\\{y\\}\in\mathcal{I}_{1}$. If $w^{\epsilon}_{1}(x)<w^{\epsilon}_{1}(y)$, then set $S_{1}\leftarrow(S_{1}\setminus\\{x\\})\cup\\{y\\}$. * • Otherwise, set $p(x)\leftarrow p(x)+1$ and $w^{\epsilon}_{2}(x)\leftarrow w^{\epsilon}_{2}(x)+\epsilon$. Apply Lemma 2.1 to obtain a $y\in S_{2}$ minimizing $w^{\epsilon}_{2}(y)$ such that $(S_{2}\setminus\\{y\\})\cup\\{x\\}\in\mathcal{I}_{2}$. If $w^{\epsilon}_{2}(x)>w^{\epsilon}_{2}(y)$, then set $S_{2}\leftarrow(S_{2}\setminus\\{y\\})\cup\\{x\\}$. Since $p(x)$ is only incremented when $x\in S_{1}\setminus S_{2}$, we have $p(x)\leq k$ for each $x\in V$ when the procedure terminates. Apparently, $w^{\epsilon}(x)$ oscillates between $w(x)$ and $w(x)+\epsilon$, and thus the number of weight adjustments for $x$ is bounded by $2p(x)$. We also have that $S_{i}$ remains $w^{\epsilon}_{i}$-maximum for each $i\in\\{1,2\\}$ due to the potential exchange of $x$ and $y$ after the adjustment. Each weight adjustment takes $O(\mathcal{T}_{\mathsf{rank}}\log{n})$ time by Lemma 2.1, hence the total running time is $O(nk\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$. ### 3.2 Augmentation With $S_{1}$ and $S_{2}$ obtained from Lemma 3.1, we then run “few” augmentations to make these two bases equal. To do so, we need the following notion of exchange graphs, which is slightly different compared to previous algorithms for unweighted matroid intersection (e.g., [BvdBMN21, CLS+19, Cun86, Law75]). #### Exchange Graph. Let $(w^{\epsilon},S)$ be an $\epsilon$-partial-solution of $w$ with $S_{1}\neq S_{2}$. The _exchange graph_ with respect to $(w^{\epsilon},S)$ is a weighted directed multi-graph $G_{w^{\epsilon},S}=(V\cup\\{s,t\\},E)$ with $s,t\not\in V$ and $E=E_{1}\cup E_{2}\cup E_{s}\cup E_{t}$, where $\displaystyle E_{1}$ $\displaystyle=\\{(x,y,w^{\epsilon}_{1}(x)-w^{\epsilon}_{1}(y))\mid x\in S_{1},y\not\in S_{1},\;\text{and}\;(S_{1}\setminus\\{x\\})\cup\\{y\\}\in\mathcal{I}_{1}\\},$ $\displaystyle E_{2}$ $\displaystyle=\\{(y,x,w^{\epsilon}_{2}(x)-w^{\epsilon}_{2}(y))\mid x\in S_{2},y\not\in S_{2},\;\text{and}\;(S_{2}\setminus\\{x\\})\cup\\{y\\}\in\mathcal{I}_{2}\\},$ $\displaystyle E_{s}$ $\displaystyle=\\{(s,x,0)\mid x\in S_{1}\setminus S_{2}\\},\;\text{and}$ $\displaystyle E_{t}$ $\displaystyle=\\{(x,t,0)\mid x\in S_{2}\setminus S_{1}\\}.$ Since $S_{i}$ is $w^{\epsilon}_{i}$-maximum for each $i\in\\{1,2\\}$, all edge weights are non-negative. Note that this definition of exchange graph is a simplified version of the _auxiliary graph_ defined by Fujishige and Zhang [FZ95] to solve the more generalized _independent assignment_ problem444Specifically, given a bipartite graph $G=(V_{1}\cup V_{2},E)$ with $V_{1}$ and $V_{2}$ being copies of $V$ and two matroids $\mathcal{M}_{1}=(V,\mathcal{I}_{1})$, $\mathcal{M}_{2}=(V,\mathcal{I}_{2})$ on $V$, the independent assignment problem aims to find the largest $S_{1}\in\mathcal{I}_{1}$ and $S_{2}\in\mathcal{I}_{2}$ such that $G$ admits a perfect matching between $S_{1}\subseteq V_{1}$ and $S_{2}\subseteq V_{2}$. Analogously, the weighted version of the problem wants to find $S_{1}$ and $S_{2}$ such that the weight of the maximum-weight perfect matching between $S_{1}$ and $S_{2}$ is maximized. Clearly, the (weighted) matroid intersection problem is a special case of the (weighted) independence assignment problem with $E=\\{(v,v)\mid v\in V\\}$.. We have the following properties of the exchange graph, for which we also provide simplified and more direct proofs for self-containedness in Appendix A. ###### Lemma 3.2 ([FZ95]; See Appendix A). $G_{w^{\epsilon},S}$ admits an $st$-path. Let $d(x)$ be the $sx$-distance in $G_{w^{\epsilon},S}$ for each $x\in V$ (set $d(x)$ to a large number if $x$ is unreachable from $s$; see Section 3.3 for the exact value) and $P$ be the shortest $st$-path with the least number of edges. ###### Lemma 3.3 ([FZ95]; See Appendix A). $\widehat{S}_{1}=(S_{1}\setminus\operatorname{tail}(P\cap E_{1}))\cup\operatorname{head}(P\cap E_{1})$ and $\widehat{S}_{2}=(S_{2}\setminus\operatorname{head}(P\cap E_{2}))\cup\operatorname{tail}(P\cap E_{2})$ are a $\widehat{w}^{\epsilon}_{1}$-maximum and $\widehat{w}^{\epsilon}_{2}$-maximum basis of $\mathcal{M}_{1}$ and $\mathcal{M}_{2}$, respectively, where $\widehat{w}^{\epsilon}_{1}(x)=w^{\epsilon}_{1}(x)+d(x)$ and $\widehat{w}^{\epsilon}_{2}(x)=w^{\epsilon}_{2}(x)-d(x)$ for each $x\in V$. In particular, $(\widehat{w}^{\epsilon},\widehat{S})$ with $\widehat{w}^{\epsilon}=(\widehat{w}^{\epsilon}_{1},\widehat{w}^{\epsilon}_{2})$ and $\widehat{S}=(\widehat{S}_{1},\widehat{S}_{2})$ is an $\epsilon$-partial- solution. Moreover, we have $|\widehat{S}_{1}\cap\widehat{S}_{2}|>|S_{1}\cap S_{2}|$. With the above properties and Lemma 3.1, we finish our algorithm with the following shortest-path procedure. Note that in order to make the algorithm subquadratic, we do not construct the exchange graphs explicitly. Nevertheless, we show that a partial construction suffices to compute the shortest-path trees in them. ###### Lemma 3.4. It takes $O(n\sqrt{r}\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time to obtain $d(x)$ for each $x\in V$ and the shortest $st$-path with the least number of edges in $G_{w^{\epsilon},S}$. We are now ready to prove Theorem 2.3. ###### Proof of Theorem 2.3. Apply Lemma 3.1 with $k=r^{3/4}$ to obtain an $\epsilon$-partial-solution $(w^{\epsilon},S)$ of $w$ such that $|S_{1}\cap S_{2}|\geq r-O(r^{1/4})$ in $O(nr^{3/4}\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time. For $O(r^{1/4})$ iterations, apply Lemmas 3.4 and 3.3 to obtain $(\widehat{w}^{\epsilon},\widehat{S})$ with $\widehat{S}_{1}$ and $\widehat{S}_{2}$ having a larger intersection than $S_{1}$ and $S_{2}$ do, and set $(w^{\epsilon},S)\leftarrow(\widehat{w}^{\epsilon},\widehat{S})$ until $S_{1}=S_{2}$. This takes overall $O(nr^{3/4}\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time as well. Note that $w^{\epsilon}(x)=w^{\epsilon}_{1}(x)+w^{\epsilon}_{2}(x)$ remains the same, completing the proof. ∎ The remainder of this section proves Lemma 3.4. For the ease of notation, we abbreviate $G_{w^{\epsilon},S}$ as $G$. Intuitively, we would like to run Dijkstra’s algorithm on $G$ to build a shortest-path tree. However, naïve implementation takes $O(nr)$ time since we might need to relax $O(nr)$ edges. This is unlike the BFS algorithm of Chakrabarty et al. [CLS+19] for the unweighted case, where we can immediately mark all out-neighbors of the current vertex as “visited”, leading to a near- linear running time. To speed things up, note that using Lemma 2.1, for a vertex $x$, we can efficiently find the vertex which is “closest” to $x$. Let $F$ denote the set of visited vertices whose exact distances are known. The closest unvisited vertex to $F$ must be closest to some $x\in F$. Therefore, in each iteration, it suffices to only relax the “shortest” edge from each $x\in F$. This can be done efficiently by maintaining a set of “recently visited” vertices $B$ of size roughly $\sqrt{n}$ and computing the distance estimate from $F\setminus B$ to all unvisited vertices555In the actual algorithm, we maintain two buffers instead of one to further improve the running time to $\widetilde{O}(n\sqrt{r})$ from $\widetilde{O}(n\sqrt{n})$. This makes our weighted matroid intersection algorithm $o(nr)$ as opposed to just $o(n^{2})$.. In each iteration, we relax the shortest edge from each $x\in B$, and now the vertex with the smallest distance estimate is closest to $F$ and therefore we include it into $B$ (and thus $F$). When $B$ grows too large, we clear $B$ and recompute the distance estimates from $F$ in $\widetilde{O}(n)$ queries. This leads to a subquadratic algorithm. We now prove the lemma formally. ###### Proof of Lemma 3.4. The algorithm builds a shortest-path tree of $G$ using Dijkstra’s algorithm. We maintain a distance estimate $\widehat{d}(x)$ for each $x\in V\cup\\{s,t\\}$. Initially, $\widehat{d}(x)=0$ for each $x\in(S_{1}\setminus S_{2})\cup\\{s\\}$ and $\widehat{d}(x)=\infty$ for other vertices. Edge set $E_{t}$ is only for the convenience of defining an $st$-path and thus we may ignore it here. Let $F$ be the set of _visited_ vertices whose distance estimates are correct, i.e., $d(x)=\widehat{d}(x)$ holds for each $x\in F$. Initially, $F=\\{s\\}$. The algorithm runs in at most $n$ iterations, and in the $t$-th iteration, we visit a new vertex $v_{t}$ such that $d(v_{t})=\widehat{d}(v_{t})$ and $d(v_{t})\leq d(v)$ for each $v\not\in F$. We maintain two _buffers_ $B_{1}\subseteq F\cap S_{1}$ and $B_{2}\subseteq F\cap\overline{S_{2}}$ containing vertices in $S_{1}$ and $\overline{S_{2}}$ that are visited “recently”. That is, after the $t$-th iteration, we have $B_{1}=\\{v_{i},\ldots,v_{t}\\}\cap S_{1}$ or $B_{1}=\emptyset$ and $B_{2}=\\{v_{j},\ldots,v_{t}\\}\cap\overline{S_{2}}$ or $B_{2}=\emptyset$ for some $i,j\leq t$. Recall that $E_{1}$ and $E_{2}$ are the edges in $G$ that correspond to exchange relations in $\mathcal{M}_{1}$ and $\mathcal{M}_{2}$, respectively. For each $i\in\\{1,2\\}$ and edge $e=(x,y)$, let $w_{i}(x,y)=|w^{\epsilon}_{i}(x)-w^{\epsilon}_{i}(y)|$ be the edge weight of $e$ in $E_{i}$ if $e\in E_{i}$ and $w_{i}(x,y)=\infty$ otherwise. Let $E({B_{i}})=\\{(x,y)\in E_{i}\mid x\in B_{i}\;\text{and}\;y\not\in F\\}$ be edges in $E_{i}$ directing from $B_{i}$ to $V\setminus F$ and $E(B)=E({B_{1}})\cup E({B_{2}})$. Let $E({F})=\\{(x,y)\in E_{1}\cup E_{2}\mid x\in F\;\text{and}\;y\not\in F\\}$. For $v\in V\setminus F$ and edge set $E^{\prime}$ such that $\operatorname{tail}(E^{\prime})\subseteq F$, let $\widetilde{d}(v,E^{\prime})=\min_{(x,v)\in E^{\prime}}\\{d(x)+w(x,v)\\}$ be the shortest distance to $v$ “relaxed” by edges in $E^{\prime}$ (recall that $w(x,v)$ is the weight of the edge $(x,v)$). We maintain the following invariants after each iteration of the algorithm except the last one. 1. (i) $d(v)\leq\widehat{d}(v)\leq\widetilde{d}(v,E(F)\setminus E(B))$ holds for each $v\in V\setminus F$. 2. (ii) There exists a $v\in V\setminus F$ such that $\widehat{d}(v)=d^{*}_{F}:=\min_{u\in V\setminus F}\\{\widetilde{d}(u,E(F))\\}$. Intuitively, Invariant (i) asserts that all edges in $E(F)\setminus E(B)$ are “relaxed” while Invariant (ii) ensures that the distance estimate of the target vertex, i.e., one with the shortest distance from $s$, is correct. Initially, both invariants are satisfied since $\widehat{d}(x)=0$ holds for each $x\in S_{1}\setminus S_{2}$. We maintain a priority queue $Q$ containing vertices in $V\setminus F$ ordered by $\widehat{d}(\cdot)$. In the $t$-th iteration, let $v_{t}$ be the vertex $v$ with the smallest $\widehat{d}(v)$. By Invariants (i) and (ii), we have $\widehat{d}(v_{t})=d^{*}_{F}$ and thus $d(v_{t})\leq d(v^{\prime})$ holds for each $v^{\prime}\in V\setminus F$ according to Dijkstra’s algorithm. As such, we push $v_{t}$ into $F$ and update $B_{1}$, $B_{2}$ appropriately by checking if $v_{t}$ belongs to $S_{1}$ and $\overline{S_{2}}$. Now, we would like to modify $\widehat{d}(v)$ for some $v\in V\setminus F$ so that both invariants remain true. For each $i\in\\{1,2\\}$, depending on the size of $B_{i}$, we perform one of the following. 1. 1. If $|B_{i}|\geq\sqrt{r}$, then we compute $\widetilde{d}_{i}(v)=\widetilde{d}(v,E(B_{1}))$ and set $\widehat{d}(v)\leftarrow\min(\widehat{d}(v),\widetilde{d}_{i}(v))$ for each $v\in V\setminus F$ using Lemma 3.5 below. For $i=1$, by definition of $G$, $\operatorname{head}(E_{1})\subseteq V\setminus S_{1}$ and thus we only need to compute $\widetilde{d}_{i}(v)$ for $v\in V\setminus S_{1}$, and therefore Lemma 3.5 takes $O(n\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time. For $i=2$, similarly, $\operatorname{head}(E_{2})\subseteq S_{2}$ and thus we only need to compute $\widetilde{d}_{i}(v)$ for $v\in S_{2}$, taking $O(r\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time. Then, we set $B_{i}\leftarrow\emptyset$, and the above modification ensures that Invariant (i) holds since $\widetilde{d}(v,E(F))=\min(\widetilde{d}(v,E(F)\setminus E(B)),\widetilde{d}(v,E(B)))$. 2. 2. If $|B_{i}|<\sqrt{r}$, then we do not clear $B_{i}$ and therefore Invariant (i) trivially holds. For each $b\in B_{i}$, we find a $v_{b}\in V\setminus F$ minimizing $d(b)+w_{i}(b,v_{b})$ via Lemma 2.1 as follows. If $i=1$, then we have $w_{1}(b,v_{b})=w^{\epsilon}_{1}(b)-w^{\epsilon}_{1}(v_{b})$, and thus we find the $v_{b}$ maximizing $w^{\epsilon}_{1}(v_{b})$ such that $(S_{1}\setminus\\{b\\})\cup\\{v_{b}\\}\in\mathcal{I}_{1}$. If $i=2$, then $w_{2}(b,v_{b})=w^{\epsilon}_{2}(v_{b})-w^{\epsilon}_{2}(b)$, and thus we find the $v_{b}$ minimizing $w^{\epsilon}_{2}(v_{b})$ such that $(S_{2}\setminus\\{v_{b}\\})\cup\\{b\\}\in\mathcal{I}_{2}$. Then, we set $\widehat{d}(v_{b})\leftarrow\min(\widehat{d}(v_{b}),d(b)+w_{i}(b,v_{b}))$ and update $v_{b}$’s position in $Q$ appropriately. This takes $O(\sqrt{r}\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time. In both cases, as argued above, Invariant (i) holds. We argue that Invariant (ii) holds after the iteration as well. Let $B_{1}^{(t)}$ be $B_{1}$ after the $t$-th iteration and define $B_{2}^{(t)}$ and $F^{(t)}$ similarly. Let $E(B^{(t)})$ denote $E(B_{1}^{(t)})\cup E(B_{2}^{(t)})$. Let $v^{*}=\operatorname*{arg\,min}_{v\in V\setminus F^{(t)}}\\{\widetilde{d}(v,E(F^{(t)}))\\}$ be an unvisited vertex after the $t$-th iteration with the smallest distance from $s$ and let $e^{*}=(u,v^{*})$ be the edge such that $u\in F^{(t)}$ and $d(v^{*})=d(u)+w(e^{*})$. That is, $e^{*}$ is the edge connecting $v^{*}$ and its parent in the shortest-path tree. If $e^{*}\in E(F^{(t-1)})\setminus E(B^{(t-1)})$, then Invariant (ii) trivially follows from the end of the $(t-1)$-th iteration. Otherwise, we must have either $e^{*}\in E(B_{1}^{(t-1)})$ or $e^{*}\in E(B_{2}^{(t-1)})$. Without loss of generality, let’s assume $e^{*}\in E(B_{1}^{(t-1)})$. If $|B_{1}^{(t-1)}|+1\geq\sqrt{r}$ (i.e., Case 1), then after setting $B_{1}^{(t)}\leftarrow\emptyset$, Invariant (ii) follows from the fact the Invariant (i) holds for $v^{*}$ and $\widetilde{d}(v^{*},E(F^{(t)})\setminus E(B^{(t)}))\leq d(u)+w(e^{*})$ since $e^{*}\in E(F^{(t)})\setminus E(B^{(t)})$. If $|B_{1}^{(t-1)}|+1<\sqrt{r}$ (i.e., Case 2), then there must exists a $b\in B_{1}^{(t)}$ such that $d(b)+w_{1}(b,v^{*})=\min_{v}\\{d(b)+w_{1}(b,v)\\}$ and thus we have at least one $v_{b}\in V\setminus F^{(t)}$ such that $\widehat{d}(v_{b})\leq d(b)+w_{1}(b,v_{b})=d(v^{*})$. This shows that Invariant (ii) indeed holds after the $t$-th iteration. The correctness of the algorithm follows from the two invariants and the analysis of Dijkstra’s algorithm. To bound the total running time, observe that for $B_{1}$, Case 1 happens at most $O(r/\sqrt{r})=O(\sqrt{r})$ times since $|S_{1}|=r$. Thus, it takes $O(n\sqrt{r}\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time in total. Similarly, for $B_{2}$, Case 1 happens at most $O(n/\sqrt{r})$ time, taking $O(n/\sqrt{r}\cdot r\log{n}\cdot\mathcal{T}_{\mathsf{rank}})=O(n\sqrt{r}\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time in total as well. For Case 2, each iteration takes $O(\sqrt{r}\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time, contributing a total of $O(n\sqrt{r}\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time. As a result, the algorithm runs in $O(n\sqrt{r}\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time, as claimed. Finally, it is easy to maintain balanced binary search trees of elements ordered by $u_{1}$, $u_{2}$, $\widehat{d}+u_{1}$, and $\widehat{d}-u_{2}$ in $O(n\log{n})$ time throughout the procedure so that Lemma 2.1 can be applied without overhead. The shortest $st$-path can also be easily recovered by maintaining the optimal parent in the shortest-path tree for each vertex. This proves the lemma. ∎ ###### Lemma 3.5. For each $i\in\\{1,2\\}$, given $B_{i}\subseteq F$ and $R\subseteq V\setminus F$, it takes takes $O(|R|\log{n}\cdot\mathcal{T}_{\mathsf{rank}})$ time to compute $\widetilde{d}(v,E(B_{i}))$ for all $v\in R$. ###### Proof. For $i=1$ and $e=(b,v)\in E(B_{i})$, we have $w(e)=w^{\epsilon}_{1}(b)-w^{\epsilon}_{1}(v)$. Therefore, $d(v,E(B_{1}))$ can be computed by finding the $b\in B$ with the smallest $d(b)+w^{\epsilon}_{1}(b)$ such that $(S_{1}\setminus\\{b\\})\cup\\{v\\}\in\mathcal{I}_{1}$ via Lemma 2.1. Similarly, for $i=2$, we have $w(e)=w^{\epsilon}_{2}(v)-w^{\epsilon}_{2}(b)$, and thus $d(v,E(B_{2}))$ can be computed by finding the $b\in B$ with the smallest $d(b)-w^{\epsilon}_{2}(b)$. The lemma simply follows by calling Lemma 2.1 once for each $v\in R$. ∎ ### 3.3 Bounding the Numbers Finally, to conclude the analysis of our algorithm, we argue that the numbers such as $w^{\epsilon}_{1}(x)$ and $w^{\epsilon}_{2}(x)$ are bounded by $\widetilde{O}(\operatorname{poly}(nW))$ so that the number of bits needed to store them and the time for a single arithmetic operation only grow by a constant factor. In the weight adjustment stage, each number is adjusted at most $O(r)$ times and each adjustment changes the number by at most $O(W)$ since $\epsilon$ is at most $W$. Therefore, the accumulative change to a number via weight adjustments is at most $O(\operatorname{poly}(nW))$. For growth incurred by augmentations, we first assume that all vertices are reachable from $s$ in $G_{w^{\epsilon},S}$. Consider a single run of Lemma 3.4 and fix an $x\in V$. Let $P_{x}=\\{s,v_{1},\ldots,v_{k}\\}$ with $v_{k}=x$ be the shortest $sx$-path in $G_{w^{\epsilon},S}$. Suppose that $(v_{1},v_{2}),(v_{k-1},v_{k})\in E_{1}$, then by definition, we have $\displaystyle d(x)$ $\displaystyle=w^{\epsilon}_{1}(v_{1})-w^{\epsilon}_{1}(v_{2})+w^{\epsilon}_{2}(v_{3})-w^{\epsilon}_{2}(v_{2})+\cdots+w^{\epsilon}_{1}(v_{k-1})-w^{\epsilon}_{1}(v_{k})$ $\displaystyle\leq w^{\epsilon}_{1}(v_{1})-w(v_{2})+(w(v_{3})+\epsilon)+\cdots+(w(v_{k-1})+\epsilon)-w^{\epsilon}_{1}(v_{k})$ $\displaystyle\leq w^{\epsilon}_{1}(v_{1})-w^{\epsilon}_{1}(v_{k})+\left(\sum_{i=2}^{k-1}(-1)^{i+1}w(v_{i})\right)+nW.$ Since $\widehat{w}^{\epsilon}_{1}(x)=w^{\epsilon}_{1}(x)+d(x)$ and $\widehat{w}^{\epsilon}_{2}(x)=w^{\epsilon}_{2}(x)-d(x)$ as defined in Lemma 3.3, we have $\displaystyle|\widehat{w^{\epsilon}}_{1}(x)|\leq|w^{\epsilon}_{1}(v_{1})|+2nW$ and $\displaystyle|\widehat{w}^{\epsilon}_{2}(x)|\leq|w^{\epsilon}_{1}(v_{1})|+2nW.$ (1) Similarly, if $(v_{k-1},v_{k})\in E_{2}$, then $\displaystyle d(x)$ $\displaystyle=w^{\epsilon}_{1}(v_{1})-w^{\epsilon}_{1}(v_{2})+w^{\epsilon}_{2}(v_{3})-w^{\epsilon}_{2}(v_{2})+\cdots+w^{\epsilon}_{2}(v_{k})-w^{\epsilon}_{2}(v_{k-1})$ $\displaystyle\leq w^{\epsilon}_{1}(v_{1})-w(v_{2})+(w(v_{3})+\epsilon)+\cdots+(w(v_{k-2})+\epsilon)-w(v_{k-1})+w^{\epsilon}_{2}(v_{k})$ $\displaystyle\leq w^{\epsilon}_{1}(v_{1})+w^{\epsilon}_{2}(v_{k})+\left(\sum_{i=2}^{k-1}(-1)^{i+1}w(v_{i})\right)+nW,$ implying (1) as well. The case when $(v_{1},v_{2})\in E_{2}$ holds similarly, except now we have $\displaystyle|\widehat{w}^{\epsilon}_{1}(x)|\leq|w^{\epsilon}_{2}(v_{1})|+2nW$ and $\displaystyle|\widehat{w}^{\epsilon}_{2}(x)|\leq|w^{\epsilon}_{2}(v_{1})|+2nW.$ (2) Since the number of augmentations is $\widetilde{O}(r^{1/4})$, we indeed have that $|w^{\epsilon}_{1}(x)|=|w^{\epsilon}_{2}(x)|=O(\operatorname{poly}(nW))=\Theta((nW)^{k})$ for some constant $k$. For the case where some vertex $x$ is not reachable from $s$, we can simply set $d(x)$ to some $c(nW)^{k+1}$ for a large enough constant $c$ and the desired bound still holds. ## 4 Concluding Remarks We present a simple subquadratic algorithm for weighted matroid intersection under the rank oracle model, providing a partial yet affirmative answer to one of the open problems raised by Blikstad et al. [BvdBMN21]. Whether the same is achievable under the independence oracle model remains open. It seems that our techniques for computing shortest-path trees do not solely result in a subquadratic augmenting-path algorithm under the independence oracle. Removing the dependence on $\log{W}$ and making the algorithm run in strongly- polynomial time is also of interest. Finally, as noted in [BvdBMN21], there were very few non-trivial lower bound results for matroid intersection. It would be helpful to see if there is any super-linear lower bound on the number of queries for these problems or even for computing shortest-path trees in the exchange graphs under either oracle model. ## Acknowledgements I would like to thank Prof. Hsueh-I Lu for advising the project and helpful suggestions on the writing and notation, Brian Tsai for proof-reading an initial draft of this paper, and the anonymous reviewers of ISAAC 2022 for their useful comments. ## References * [AD71] Martin Aigner and Thomas A Dowling. Matching theory for combinatorial geometries. Transactions of the American Mathematical Society, 158(1):231–245, 1971. doi:10.2307/1995784. * [Bli21] Joakim Blikstad. Breaking o(nr) for matroid intersection. In 48th International Colloquium on Automata, Languages, and Programming, ICALP 2021, volume 198, pages 31:1–31:17. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2021. doi:10.4230/LIPIcs.ICALP.2021.31. * [BvdBMN21] Joakim Blikstad, Jan van den Brand, Sagnik Mukhopadhyay, and Danupon Nanongkai. Breaking the quadratic barrier for matroid intersection. In Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing, STOC 2021, pages 421–432. ACM, 2021. doi:10.1145/3406325.3451092. * [CLS+19] Deeparnab Chakrabarty, Yin Tat Lee, Aaron Sidford, Sahil Singla, and Sam Chiu-wai Wong. Faster matroid intersection. In 60th IEEE Annual Symposium on Foundations of Computer Science, FOCS 2019, pages 1146–1168. IEEE Computer Society, 2019. doi:10.1109/FOCS.2019.00072. * [CLSW17] Deeparnab Chakrabarty, Yin Tat Lee, Aaron Sidford, and Sam Chiu-wai Wong. Subquadratic submodular function minimization. In Proceedings of the 49th Annual ACM SIGACT Symposium on Theory of Computing, STOC 2017, pages 1220–1231. ACM, 2017. doi:10.1145/3055399.3055419. * [CQ16] Chandra Chekuri and Kent Quanrud. A fast approximation for maximum weight matroid intersection. In Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2016, pages 445–457. SIAM, 2016. doi:10.1137/1.9781611974331.ch33. * [Cun86] William H. Cunningham. Improved bounds for matroid partition and intersection algorithms. SIAM J. Comput., 15(4):948–957, 1986. doi:10.1137/0215066. * [Edm71] Jack Edmonds. Matroids and the greedy algorithm. Math. Program., 1(1):127–136, 1971. doi:10.1007/BF01584082. * [Edm79] Jack Edmonds. Matroid intersection. Annals of Discrete Mathematics, 4:39–49, 1979. doi:10.1016/S0167-5060(08)70817-3. * [Edm01] Jack Edmonds. Submodular functions, matroids, and certain polyhedra. In Combinatorial Optimization - Eureka, You Shrink!, Papers Dedicated to Jack Edmonds, 5th International Workshop, Aussois, France, March 5-9, 2001, Revised Papers, volume 2570 of Lecture Notes in Computer Science, pages 11–26. Springer, 2001. doi:10.1007/3-540-36478-1\\_2. * [Edm10] Jack Edmonds. Matroid partition. In 50 Years of Integer Programming 1958-2008 - From the Early Years to the State-of-the-Art, pages 199–217. Springer, 2010. doi:10.1007/978-3-540-68279-0\\_7. * [Fra81] András Frank. A weighted matroid intersection algorithm. J. Algorithms, 2(4):328–336, 1981. doi:10.1016/0196-6774(81)90032-8. * [FZ95] Satoru Fujishige and Xiaodong Zhang. An efficient cost scaling algorithm for the independent assignment problem. Journal of the Operations Research Society of Japan, 38(1):124–136, 1995. doi:10.15807/jorsj.38.124. * [GX96] Harold N. Gabow and Ying Xu. Efficient theoretic and practical algorithms for linear matroid intersection problems. J. Comput. Syst. Sci., 53(1):129–147, 1996. doi:10.1006/jcss.1996.0054. * [HKK16] Chien-Chung Huang, Naonori Kakimura, and Naoyuki Kamiyama. Exact and approximation algorithms for weighted matroid intersection. In Proceedings of the Twenty-Seventh Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2016, pages 430–444. SIAM, 2016. doi:10.1137/1.9781611974331.ch32. * [Law75] Eugene L. Lawler. Matroid intersection algorithms. Math. Program., 9(1):31–56, 1975. doi:10.1007/BF01681329. * [LSW15] Yin Tat Lee, Aaron Sidford, and Sam Chiu-wai Wong. A faster cutting plane method and its implications for combinatorial and convex optimization. In IEEE 56th Annual Symposium on Foundations of Computer Science, FOCS 2015, pages 1049–1065. IEEE Computer Society, 2015. doi:10.1109/FOCS.2015.68. * [Ngu19] Huy L. Nguyen. A note on cunningham’s algorithm for matroid intersection. CoRR, abs/1904.04129, 2019. URL: http://arxiv.org/abs/1904.04129, arXiv:1904.04129. * [OA92] James B. Orlin and Ravindra K. Ahuja. New scaling algorithms for the assignment and minimum mean cycle problems. Math. Program., 54:41–56, 1992. doi:10.1007/BF01586040. * [Pri15] Christopher Price. Combinatorial algorithms for submodular function minimization and related problems. Master’s thesis, University of Waterloo, 2015. URL: http://hdl.handle.net/10012/9356. * [S+03] Alexander Schrijver et al. Combinatorial optimization: polyhedra and efficiency, volume 24. Springer, 2003. * [SI95] Maiko Shigeno and Satoru Iwata. A dual approximation approach to weighted matroid intersection. Oper. Res. Lett., 18(3):153–156, 1995. doi:10.1016/0167-6377(95)00047-X. ## Appendix A Omitted Proofs ### A.1 Proofs of Lemmas in Section 3.1 For self-containedness, we include the proof that the two bases obtained in the weight adjustment phase have a large intersection by Shigeno and Iwata [SI95] here. ###### Lemma A.1 ([SI95]). Let $S_{1}$ and $S_{2}$ be obtained from the procedure described in Lemma 3.1. Then, $|S_{1}\cap S_{2}|\geq\left(1-\frac{O(1)}{k}\right)r$. ###### Proof. Let $p(S)$ denote $\sum_{x\in S}p(x)$. Observe that an element is never moved to $S_{2}\setminus S_{1}$ during weight adjustments, and therefore we have $p(S_{2}\setminus S_{1})=0$ and $p(S_{1}\setminus S_{2})=p(S_{1})-p(S_{2})$. Recall that $S^{\prime}$ is a common basis such that $(w^{2\epsilon},S^{\prime})$ is a $2\epsilon$-solution. Since $p(x)$ equals the number of adjustments of $w^{\epsilon}_{2}(x)$ and each such adjustment is preceded by an adjustment of $w^{\epsilon}_{1}(x)$, we have $p(x)\cdot\epsilon=w^{\epsilon}_{2}(x)-(w(x)-w^{2\epsilon}_{1}(x))\leq w^{2\epsilon}_{1}(x)-w^{\epsilon}_{1}(x)$ for each $x\in V$. Thus, $\displaystyle p(S_{1}\setminus S_{2})\cdot\epsilon$ $\displaystyle=\left(p(S_{1})-p(S_{2})\right)\cdot\epsilon$ $\displaystyle\leq\left(w^{2\epsilon}_{1}(S_{1})-w^{\epsilon}_{1}(S_{1})\right)-\left(w^{\epsilon}_{2}(S_{2})-w(S_{2})+w^{2\epsilon}_{1}(S_{2})\right)$ $\displaystyle\stackrel{{\scriptstyle\text{(a)}}}{{\leq}}w^{2\epsilon}_{1}(S_{1})-w^{\epsilon}_{1}(S^{\prime})-w^{\epsilon}_{2}(S^{\prime})+w(S_{2})-w^{2\epsilon}_{1}(S_{2})$ $\displaystyle\stackrel{{\scriptstyle\text{(b)}}}{{\leq}}w^{2\epsilon}_{1}(S_{1})-w(S^{\prime})+w(S_{2})-w^{2\epsilon}_{1}(S_{2}),$ where (a) is because $S_{i}$ is $w^{\epsilon}_{i}$-maximum for each $i\in\\{1,2\\}$ and (b) is because $w(S^{\prime})\leq w^{\epsilon}(S^{\prime})$ as $w^{\epsilon}$ is an $\epsilon$-splitting. Since $(w^{2\epsilon},S^{\prime})$ is a $2\epsilon$-solution, $w^{2\epsilon}_{2}(S)-2\epsilon r\leq w(S)-w^{2\epsilon}_{1}(S)\leq w^{2\epsilon}_{2}(S)$ holds for each basis $S$. This combined with the fact that $S^{\prime}$ is $w^{2\epsilon}_{i}$-maximum for each $i\in\\{1,2\\}$ implies $p(S_{1}\setminus S_{2})\cdot\epsilon\leq 2\epsilon r-w^{2\epsilon}_{2}(S^{\prime})+w^{2\epsilon}_{2}(S_{2})\leq 2\epsilon r\implies p(S_{1}\setminus S_{2})\leq 2r.$ When the algorithm terminates, we have $p(x)=k$ for all $x\in S_{1}\setminus S_{2}$, implying $p(S_{1}\setminus S_{2})=|S_{1}\setminus S_{2}|\cdot k\leq 2r\implies|S_{1}\setminus S_{2}|\leq\frac{2r}{k}.$ As a result, $|S_{1}\cap S_{2}|=r-|S_{1}\setminus S_{2}|\geq\left(1-\frac{O(1)}{k}\right)r.$ ∎ ### A.2 Proofs of Lemmas in Section 3.2 In this section, we prove the properties of the exchange graphs. The proofs for the more generalized auxiliary graph given by Fujishige and Zhang can be found in [FZ95]. To prove Lemma 3.2, it would be more convenient to refer to the following definition of a directed bipartite graph based on exchange relationships, which is heavily used in unweighted matroid intersection algorithms. For $S\in\mathcal{I}_{1}\cap\mathcal{I}_{2}$, let $\widetilde{G}_{S}=(V\cup\\{s,t\\},\widetilde{E})$ with $s,t\not\in V$ denote the directed graph with $\widetilde{E}=\widetilde{E}_{1}\cup\widetilde{E}_{2}\cup\widetilde{E}_{s}\cup\widetilde{E}_{t}$, where $\displaystyle\widetilde{E}_{1}$ $\displaystyle=\\{(x,y)\mid x\in S,y\not\in S,\;\text{and}\;(S\setminus\\{x\\})\cup\\{y\\}\in\mathcal{I}_{1}\\},$ $\displaystyle\widetilde{E}_{2}$ $\displaystyle=\\{(y,x)\mid x\in S,y\not\in S,\;\text{and}\;(S\setminus\\{x\\})\cup\\{y\\}\in\mathcal{I}_{2}\\},$ $\displaystyle\widetilde{E}_{s}$ $\displaystyle=\\{(s,x)\mid S\cup\\{x\\}\in\mathcal{I}_{1}\\},\;\text{and}$ $\displaystyle\widetilde{E}_{t}$ $\displaystyle=\\{(x,t)\mid S\cup\\{x\\}\in\mathcal{I}_{2}\\}.$ ###### Lemma A.2 ([Law75]). $\widetilde{G}_{S}$ for $|S|<r$ admits an $st$-path. We will use the existence of an $st$-path in $\widetilde{G}_{\widetilde{S}}$ to prove that such a path exists in $G_{w^{\epsilon},S}$, for $\widetilde{S}=S_{1}\cap S_{2}$. The following claims certify that $\widetilde{G}_{\widetilde{S}}$ and $G_{w^{\epsilon},S}$ are almost the same. ###### Claim A.3. Let $\mathcal{M}=(V,\mathcal{I})$ be a matroid, $S\subseteq S^{\prime}\in\mathcal{I}$, $x\in S$, and $y\not\in S^{\prime}$ such that $(S\setminus\\{x\\})\cup\\{y\\}\in\mathcal{I}$ but $S\cup\\{y\\}\not\in\mathcal{I}$, then $(S^{\prime}\setminus\\{x\\})\cup\\{y\\}\in\mathcal{I}$. ###### Proof. Let $C$ be the unique circuit in $S\cup\\{y\\}$. Since $C\subseteq S^{\prime}\cup\\{y\\}$ and $S^{\prime}\cup\\{y\\}$ has only one circuit, $C$ is the unique circuit in $S^{\prime}\cup\\{y\\}$ as well. Moreover, $(S\setminus\\{x\\})\cup\\{y\\}\in\mathcal{I}$ if and only if $x\in C$ and therefore $(S^{\prime}\setminus\\{x\\})\cup\\{y\\}\in\mathcal{I}$. ∎ ###### Claim A.4. Let $\mathcal{M}=(V,\mathcal{I})$ be a matroid, $S\subseteq S^{\prime}\in\mathcal{I}$ where $S^{\prime}$ is a basis of $\mathcal{M}$, and $x\not\in S^{\prime}$ such that $S\cup\\{x\\}\in\mathcal{I}$. Then, there exists a $y\in S^{\prime}\setminus S$ such that $(S^{\prime}\setminus\\{y\\})\cup\\{x\\}\in\mathcal{I}$. ###### Proof. Let $S^{\prime\prime}$ be an arbitrary basis of $\mathcal{M}$ that contains $S\cup\\{x\\}$. Since $x\in S^{\prime\prime}\setminus S^{\prime}$, by the strong exchange property (see, e.g., [S+03, Theorem 39.6]) of bases, there exists a $y\in S^{\prime}\setminus S^{\prime\prime}\subseteq S^{\prime}\setminus S$ such that $(S^{\prime}\setminus\\{y\\})\cup\\{x\\}\in\mathcal{I}$, completing the proof. ∎ We are now ready to prove Lemma 3.2. ###### Proof of Lemma 3.2. Let $\widetilde{P}=\\{s,v_{1},\ldots,v_{k},t\\}$ be the shortest $st$-path in $\widetilde{G}_{\widetilde{S}}$ for $\widetilde{S}=S_{1}\cap S_{2}$. The existence of such a path is guaranteed by Lemma A.2. We have $\widetilde{S}\cup\\{v_{i}\\}\not\in\mathcal{I}_{1}$ and $\widetilde{S}\cup\\{v_{i}\\}\not\in\mathcal{I}_{2}$ for each $1<i<k$ since $\widetilde{P}$ is the shortest path. For an odd $1\leq i<k$, we have $v_{i}\not\in S$ and $v_{i+1}\in S$. If $v_{i}\not\in S_{2}\setminus S_{1}$, then by A.3, we have $(v_{i},v_{i+1})\in E(G_{w^{\epsilon},S})$. Similarly, for an even $1\leq i<k$, if $v_{i+1}\not\in S_{1}\setminus S_{2}$, then we have $(v_{i},v_{i+1})\in E(G_{w^{\epsilon},S})$. Suppose that $v_{1}\not\in S_{1}\setminus S_{2}$, then by A.4, we can find a $v_{0}\in S_{1}\setminus S_{2}$ such that $(S_{1}\setminus\\{v_{0}\\})\cup\\{v_{1}\\}\in\mathcal{I}_{1}$. Similarly, if $v_{k}\not\in S_{2}\setminus S_{1}$, then we can find a $v_{k+1}\in S_{2}\setminus S_{1}$ such that $(S_{2}\setminus\\{v_{k+1}\\})\cup\\{v_{k}\\}\in\mathcal{I}_{2}$. Therefore, without loss of generality, we may assume that there exists the last vertex $v_{i}\in S_{1}\setminus S_{2}$ and the first vertex $v_{j}\in S_{2}\setminus S_{1}$ after $v_{i}$. Now, for each $i<k<j$, we have $v_{k}\not\in(S_{1}\setminus S_{2})\cup(S_{2}\setminus S_{1})$. Therefore, $(v_{k},v_{k+1})\in E(G_{w^{\epsilon},S})$ holds for each $i\leq k<j$, and we obtain an $st$-path in $G_{w^{\epsilon},S}$ as $P=\\{s,v_{i},\ldots,v_{j},t\\}$. This concludes the proof. ∎ Finally, to prove Lemma 3.3, we need the following results. ###### Lemma A.5 ([Pri15, Proposition 2.4.1]). Given a matroid $\mathcal{M}=(V,\mathcal{I})$ and an $S\in\mathcal{I}$. Suppose that $(a_{1},\ldots,a_{p})\subseteq V\setminus S$ and $(b_{1},\ldots,b_{p})\subseteq S$ are two sequences satisfying the following conditions: 1. 1. $(S\setminus\\{b_{i}\\})\cup\\{a_{i}\\}\in\mathcal{I}$ for each $1\leq i\leq p$ and 2. 2. $(S\setminus\\{b_{j}\\})\cup\\{a_{i}\\}\not\in\mathcal{I}$ for each $1\leq j<i\leq p$. Then, $(S\setminus\\{b_{1},\ldots,b_{p}\\})\cup\\{a_{1},\ldots,a_{p}\\}\in\mathcal{I}$ holds. ###### Lemma A.6 ([Pri15, Lemma 2.4.2]). Let $\mathcal{M}$, $S$, $(a_{1},\ldots,a_{p})$, and $(b_{1},\ldots,b_{p})$ be the same as in Lemma A.5. Let $S^{\prime}=(S\setminus\\{b_{1},\ldots,b_{p}\\})\cup\\{a_{1},\ldots,a_{p}\\}$. For $x\in S^{\prime}$ and $y\in V\setminus S^{\prime}$, if $(S^{\prime}\setminus\\{x\\})\cup\\{y\\}\in\mathcal{I}$ but either $y\in S$ or $(S\setminus\\{x\\})\cup\\{y\\}\not\in\mathcal{I}$, then there exists $1\leq\ell\leq k\leq p$ such that $(S\setminus\\{x\\})\cup\\{a_{k}\\}\in\mathcal{I}$ and either $b_{\ell}=y$ or $(S\setminus\\{b_{\ell}\\})\cup\\{y\\}\in\mathcal{I}$. In essence, Lemma A.5 captures the validity of an augmentation while Lemma A.6 models the condition in which new exchange relationships emerge in the augmented independent set. The following three claims imply Lemma 3.3. Recall that $P=\\{s,v_{1},\ldots,v_{k},t\\}$ is the shortest $st$-path with the least number of edges and $d(x)$ is the $sx$-distance in $G_{w^{\epsilon},S}$. ###### Claim A.7. $|\widehat{S}_{1}\cap\widehat{S}_{2}|>|S_{1}\cap S_{2}|$ holds. ###### Proof. If $k=2$, then either $v_{1}\in\widehat{S}_{1}\cap\widehat{S}_{2}$ or $v_{k}\in\widehat{S}_{1}\cap\widehat{S}_{2}$ must hold, depending on whether $(v_{1},v_{2})\in E_{1}$ or $(v_{1},v_{2})\in E_{2}$, and the claim trivially holds in this case. Thus, in the following, we assume that $k>2$. Since $P$ is the shortest, we may assume that $v_{i}\in(S_{1}\cap S_{2})\cup(\overline{S_{1}}\cap\overline{S_{2}})$ holds for each $1<i<k$. Also, for $1<i<k-1$, if $v_{i}\in S_{1}\cap S_{2}$, then $v_{i+1}$ must be in $\overline{S_{1}}\cap\overline{S_{2}}$ due to the way $G_{w^{\epsilon},S}$ is constructed. Similarly, if $v_{i}\in\overline{S_{1}}\cap\overline{S_{2}}$, then we must have $v_{i+1}\in S_{1}\cap S_{2}$. Let $P_{\text{mid}}=\\{v_{2},\ldots,v_{k-1}\\}$, $I=S_{1}\cap S_{2}$, and $O=\overline{S_{1}}\cap\overline{S_{2}}$. Clearly, we have $|\widehat{S}_{1}\cap\widehat{S}_{2}|-|S_{1}\cap S_{2}|=|P_{\text{mid}}\cap O|-|P_{\text{mid}}\cap I|+\llbracket v_{1}\in\widehat{S}_{1}\cap\widehat{S}_{2}\rrbracket+\llbracket v_{k}\in\widehat{S}_{1}\cap\widehat{S}_{2}\rrbracket.$ (3) We prove the claim by considering the following four possible cases. $k$ is even and $v_{2}\in I$. We have $(v_{1},v_{2})\in E_{2}$, $(v_{k-1},v_{k})\in E_{2}$, and $|P_{\text{mid}}\cap I|=|P_{\text{mid}}\cap O|$. Also, $v_{1}\in\operatorname{tail}(P\cap E_{2})$ and therefore $v_{1}\in\widehat{S}_{1}\cap\widehat{S}_{2}$. $k$ is even and $v_{2}\in O$. We have $(v_{1},v_{2})\in E_{1}$, $(v_{k-1},v_{k})\in E_{1}$, and $|P_{\text{mid}}\cap I|=|P_{\text{mid}}\cap O|$. Also, $v_{k}\in\operatorname{head}(P\cap E_{1})$ and therefore $v_{k}\in\widehat{S}_{1}\cap\widehat{S}_{2}$. $k$ is odd and $v_{2}\in I$. We have $(v_{1},v_{2})\in E_{2}$, $(v_{k-1},v_{k})\in E_{1}$, and $|P_{\text{mid}}\cap I|=|P_{\text{mid}}\cap O|+1$. Also, $v_{1}\in\operatorname{tail}(P\cap E_{2})$, $v_{k}\in\operatorname{head}(P\cap E_{1})$ and therefore $v_{1},v_{k}\in\widehat{S}_{1}\cap\widehat{S}_{2}$. $k$ is odd and $v_{2}\in O$. We have $(v_{1},v_{2})\in E_{1}$, $(v_{k-1},v_{k})\in E_{2}$, and $|P_{\text{mid}}\cap I|=|P_{\text{mid}}\cap O|-1$. In all cases, we have $|\widehat{S}_{1}\cap\widehat{S}_{2}|>|S_{1}\cap S_{2}|$ via Equation 3, concluding the proof. ∎ We prove the following claims for $i=1$. The proofs for $i=2$ follow analogously. ###### Claim A.8. $\widehat{S}_{i}\in\mathcal{I}_{i}$ holds for each $i\in\\{1,2\\}$. ###### Proof. Let $P_{1}=P\cap E_{1}=\\{(b_{1},a_{1}),(b_{2},a_{2}),\ldots(b_{p},a_{p})\\}$, where $(S_{1}\setminus\\{b_{i}\\})\cup\\{a_{i}\\}\in\mathcal{I}_{1}$ holds for each $1\leq i\leq p$. Since $P$ is the shortest path, $d(b_{i})+u_{1}(b_{i})-u_{1}(a_{i})=d(a_{i})\implies d(b_{i})+u_{1}(b_{i})=d(a_{i})+u_{1}(a_{i})$ holds for each $i$. Reorder $P_{1}$ so that $d(b_{1})+u_{1}(b_{1})\leq d(b_{2})+u_{1}(b_{2})\leq\cdots\leq d(b_{p})+u_{1}(b_{p})$. Moreover, if $d(b_{i})+u_{1}(b_{i})=d(b_{j})+u_{1}(b_{j})$ for some $i,j$, then $(b_{i},a_{i})$ precedes $(b_{j},a_{j})$ in $P_{1}$ if and only if $(b_{i},a_{i})$ precedes $(b_{j},a_{j})$ in $P$. It follows that for each $1\leq j<i\leq p$, it holds that $(S_{1}\setminus\\{b_{j}\\})\cup\\{a_{i}\\}\not\in\mathcal{I}_{1}$ since otherwise we would have $d(b_{j})+u_{1}(b_{j})-u_{1}(a_{i})\geq d(a_{i})\implies d(b_{j})+u_{1}(b_{j})\geq d(a_{i})+u_{1}(a_{i}).$ (4) Because $j<i$, (4) must take equality, but this would contradict with the fact the $P$ has the least number of edges since the edge $(b_{j},a_{i})$ “jumps” over vertices $a_{j},b_{j+1},\ldots,b_{i}$ in $P$ and has the same weight as the subpath $b_{j},a_{j},\ldots,b_{i},a_{i}$. As such, by Lemma A.5, the claim is proved. ∎ ###### Claim A.9. $\widehat{S}_{i}$ is $\widehat{w}^{\epsilon}_{i}$-maximum for each $i\in\\{1,2\\}$. ###### Proof. Let $P_{1}=P\cap E_{1}=\\{(b_{1},a_{1}),\ldots,(b_{p},a_{p})\\}$ be ordered the same way as in the proof of A.8. It suffices to show that $\widehat{w}^{\epsilon}_{1}(x)\geq\widehat{w}^{\epsilon}_{1}(y)$ holds for each $x\in\widehat{S}_{1}$ and $y\not\in\widehat{S}_{1}$ with $(\widehat{S}_{1}\setminus\\{x\\})\cup\\{y\\}\in\mathcal{I}_{1}$. Consider the following two cases. 1. 1. $(S_{1}\setminus\\{x\\})\cup\\{y\\}\in\mathcal{I}_{1}$: Since $(x,y)\in E_{1}$, it follows that $d(x)+u_{1}(x)-u_{1}(y)\geq d(y)\implies\widehat{w}^{\epsilon}_{1}(x)=d(x)+u_{1}(x)\geq d(y)+u_{1}(y)=\widehat{w}^{\epsilon}_{1}(y).$ 2. 2. $(S_{1}\setminus\\{x\\})\cup\\{y\\}\not\in\mathcal{I}_{1}$: By Lemma A.6, there exists $1\leq\ell\leq k\leq p$ such that (1) $(S_{1}\setminus\\{x\\})\cup\\{a_{k}\\}\in\mathcal{I}_{1}$ and either (2.1) $b_{\ell}=y$ or (2.2) $(S_{1}\setminus\\{b_{\ell}\\})\cup\\{y\\}\in\mathcal{I}_{1}$. (1) implies that $\widehat{w}^{\epsilon}_{1}(x)\geq\widehat{w}^{\epsilon}_{1}(a_{k})$. If (2.1) holds, then $\widehat{w}^{\epsilon}_{1}(x)\geq\widehat{w}^{\epsilon}_{1}(a_{k})\geq\widehat{w}^{\epsilon}_{1}(b_{\ell})=\widehat{w}^{\epsilon}_{1}(y)$. If (2.2) holds, then $\widehat{w}^{\epsilon}_{1}(x)\geq\widehat{w}^{\epsilon}_{1}(a_{k})\geq\widehat{w}^{\epsilon}_{1}(b_{\ell})\geq\widehat{w}^{\epsilon}_{1}(y)$. The claim is proved. ∎
Magnetic storms during the space age: Occurrence and relation to varying solar activity Kalevi Mursula1, Timo Qvick1, Lauri Holappa1, Timo Asikainen1 1Space Climate Group, Space Physics and Astronomy Research Unit, University of Oulu, Finland Kalevi<EMAIL_ADDRESS> * We explain the occurrence of magnetic storms in space age by their relation to the varying solar activity and solar magnetic structure. * The occurrence of large, moderate and weak HSS/CIR storms follows the decrease of the HCS tilt in the declining phase of solar cycle. * Maxima of HSS/CIR storms have shifted from late declining phase in cycles 20-22 to earlier times in cycles 23-24 due to recent HCS widening. We study the occurrence of magnetic storms in space age (1957–2021) using Dst and Dxt indices. We find 2526/2743 magnetic storms in the Dxt/Dst index, out of which 45% are weak, 40% moderate, 12% intense and 3% major storms. Occurrence of storms in space age follows the slow decrease of sunspot activity and the related change in solar magnetic structure. We quantify the sunspot - CME storm relation in the five cycles of space age. We explain how the varying solar activity changes the structure of the heliospheric current sheet (HCS), and how this affects the HSS/CIR storms. Space age started with a record number of storms in 1957–1960, with roughly one storm per week. Solar polar fields attained their maximum in cycle 22, which led to an exceptionally thin HCS, and a space age record of large HSS/CIR storms in 1990s. In the minimum of cycle 23, for the only time in space age, CME storm occurrence reduced below that predicted by sunspots. Weak sunspot activity since cycle 23 has weakened solar polar fields and widened the HCS, which has decreased the occurrence of large and moderate HSS/CIR storms. Because of a wide HCS, the Earth has spent 50% of its time in slow solar wind since cycle 23. The wide HCS has also made large and moderate HSS/CIR storms occur in the early declining phase in recent cycles, while in the more active cycles 20–22 they occurred in the late declining phase. § INTRODUCTION Magnetic storms <cit.>for a review see, e.g.,>[]Dessler_Parker_1959, Gonzalez_1994,Daglis_1999, Daglis_2003, Daglis_SSR_2006 are the largest disturbances of the near-Earth space, driven by enhanced interaction between the solar wind (SW), including the heliospheric magnetic field (HMF; also called the interplanetary magnetic field, IMF), and the Earth's magnetosphere. Certain structures in the solar wind and the HMF contain favourable conditions for enhanced interaction that last typically from one to several days. The two most important structures leading to storms are the interplanetary manifestations of coronal mass ejections (CME) <cit.> and the corotating interaction regions (CIR) related to high-speed solar wind streams (HSS) <cit.>. CMEs arise from the eruption of large coronal flux tubes and lead to increased solar wind density and pressure <cit.>. They often have the internal structure of a magnetic cloud which includes a systematic rotation of magnetic field lines and a long period of southward directed magnetic field <cit.>. This leads to enhanced reconnection at the dayside magnetopause, producing a magnetic storm. Coronal flux tubes that precede CMEs arise from the solar convection layer and have a close connection with sunspots. Accordingly, the occurrence of CMEs varies closely in phase with sunspot activity, reflecting the appearance of new magnetic flux on the solar surface <cit.>. High-speed solar wind streams emanate from large solar coronal holes <cit.>, where the magnetic field is typically unipolar and experiences less of super-radial expansion than in the neighborhood of closed field regions <cit.>. Accordingly, solar wind flow can more freely escape from large coronal holes and is accelerated to higher speeds than elsewhere. On the other hand, the wind originating at the streamer belt and neighboring regions is slower but more dense. Since the streamer belt also carries the heliospheric current sheet (HCS) and its center, the magnetic neutral line (NL), the slow wind is commonly equated with the HCS. When the HCS is inclined with respect to the solar equator, the slow and fast solar wind regions can be at the same heliographic latitude and be emitted successively in the same direction. When the fast stream catches the slow wind ahead, an interaction region called the corotating interaction region, also called the stream interaction region (SIR), is formed <cit.>. Due to compressed plasma density and increased magnetic field intensity, as well as to the following fast solar wind, the HSS/CIR structure is very effective in producing weak and moderate storms <cit.>. Solar magnetic fields experience a dramatic structural change over the sunspot cycle. During solar minima the global solar magnetic field is mainly dipolar, with a few active regions around the solar equator and large coronal holes of unipolar field with opposite polarities around the two solar poles <cit.>. As activity increases in the ascending phase of the cycle, surges of magnetic flux with polarity opposite to the prevailing polar field are transported to each pole <cit.>. These surges reduce the size of the old-polarity coronal holes and, eventually, reverse the polarity of the polar field around solar maxima. Subsequent surges in the declining phase increase the new-polarity field at the poles and extend the area of polar coronal holes. Moreover, surges can form contiguous unipolar regions from the low-latitude origin of surges to the solar pole. This leads to the formation of longitudinally asymmetric extensions of polar coronal holes, so called elephant trunks, that can cover a wide range of latitudes and emit fast solar wind even at low latitudes <cit.>. Such extensions of polar coronal holes to low latitudes are an important source of high-speed streams reaching the Earth. Smaller-scale coronal holes can be formed between active regions at any latitude and at any time of the cycle, especially at solar maxima. Related HSS streams can affect (e.g., accelerate) CMEs bursting from neighboring active regions <cit.>. Later in the declining phase, as the emergence of magnetic flux and related surge production subside, the global field slowly returns back to its minimum-time dipolar structure with a polarity structure opposite to the previous minimum. The lower boundary of polar coronal holes becomes more symmetric, which reduces the tilt of the HCS and the occurrence of HSSs at the Earth. In addition to this solar cycle evolution of solar magnetic fields, there are also longer-term changes, e.g., in the overall sunspot activity, leading to the varying height of sunspot cycles. It is known that sunspot activity in the 20th century reached a record level at least for a couple of thousand years <cit.>, culminating at the maximum of solar cycle 19 in 1957. This highly active time of the mid-20th century is commonly called the Modern Grand Maximum (MGM). After solar cycle 19 (to be called SC19), solar activity remained at a fairly high level for 3–4 cycles whereafter, since the maximum of SC23, solar activity has considerably subsided. This is evidenced by an exceptionally long and deep minimum between SC23 and SC24, and a considerably low cycle 24. We note that, curiously, the start of space age, as marked by the flight of the first satellite in 1957, coincides with the maximum of the MGM. Thus, the space age, at least until the recent times, is characterized and affected by slowly, but unsteadily reducing solar activity. The varying level of sunspot activity during the space age directly affects, e.g., the occurrence of CMEs and magnetic storms produced by CMEs. In addition, there is also a related long-term change in the solar magnetic field structure, which affects coronal holes and, thereby, the occurrence of HSS/CIRs and magnetic storms produced by them. In fact, while in the earlier, active cycles there were only small and short-lived coronal holes at low latitudes <cit.>, the declining phase of cycle 23 was characterized by rather large, persistent low-latitude coronal holes <cit.>, which led to a space age record activity of HSSs and geomagnetic activity in 2003 <cit.>. Another long-term change is the recent weakening of solar polar fields <cit.> as the result of reduced emergence of new flux on the solar surface. These changes have important consequences to the long-term occurrence of magnetic storms. The enhanced solar wind-magnetosphere interaction during CMEs and HSS/CIRs accelerates particles and drives several current systems that lead to magnetic disturbances in different parts of the Earth's surface. However, rather than measuring the overall disturbance level like, e.g., the Kp index of geomagnetic activity, magnetic storms are defined and quantified in terms of one current system only, the ring current <cit.>. The storm-time ring current consists mainly of (positively charged) hydrogen, helium and oxygen ions of some 10–300 keV energy, drifting around the Earth at the distance of about 3–7 Earth radii <cit.>. Drifting westward around the Earth, they produce a negative deflection in the horizontal magnetic field on the ground, which is a direct measure of the energy content of the ring current <cit.>. This deflection has been measured from 1957 onward by four ground-based magnetometer stations located at low latitudes, roughly equidistantly in longitude. A dedicated recipe was developed <cit.> in order to remove the secular, seasonal and daily quiet-time variations from the locally observed magnetic field, and to quantify the ring current in terms of an index called the Dst index. Year 1957 was the International Geophysical Year (IGY), when several international programs on solar-terrestrial research were started, developing both ground-based and flying instrumentation. Thus, it is not surprising that the Dst index is available since the start of space age. However, as noted above, it is curious, but also quite appropriate, that the Sun reached its all-time maximum activity during the International Geophysical Year. Since continuous monitoring of space by satellites started soon after the IGY, we also have satellite observations of the solar wind and the HMF available for most of the time of the Dst index. This allows us to study the solar wind drivers of magnetic storms almost over the whole space age. During the several years of storm studies some errors and inconsistencies have been found in the Dst index <cit.>. Therefore Karinen_Mursula_2005 recalculated the Dst index using the original recipe but correcting the noticed problems. This revised Dst index is called the Dxt index. We will use here both the Dst and the Dxt index in order to study magnetic storms during the space age in 1957–2021. Even though there are considerable differences between the two ring-current indices leading, e.g., to somewhat different numbers of magnetic storms, we find that the storm occurrences and their implications about the long-term change of the Sun, remain the same. This paper is organized as follows. In Section <ref>, we present the Dst and Dxt indices and discuss their differences. Section <ref> describes the solar wind classification into the three main structures (flow types). Section <ref> explains how storms are derived from the two indices. Section <ref> gives the total numbers and mean intensities of storms of different intensity for the whole space age, and Section <ref> presents their yearly numbers. We assign the storms into their solar wind drivers in Section <ref> and present their yearly numbers in Section <ref>. In Section <ref> we discuss large and moderate CME storms and their relation to sunspots, separately during the five solar cycles. Section <ref> presents the connection between HSS/CIR storms and the structure of the heliospheric current sheet, and discusses the implications of our results to the long-term evolution of the Sun. Finally, in Section <ref> we discuss the obtained results and give our conclusions. § DXT AND DST INDICES As noted above, in our quest for long-term homogeneity, we have found earlier that the Dst index depicts some errors and inconsistencies. Correcting these problems eventually led to the recalculation of the index and to the development of the corrected Dst index, the Dxt index <cit.>. The first error found was an erroneous diurnal variation of the Dst index. While the Dst index includes a rather small diurnal UT variation, it is artificially enlarged in 1971 <cit.>. We suggested <cit.> that this erroneous UT variation in 1971 is most likely due to an erroneous treatment of the diurnal variation of SJG station data. Unfortunately, the WDC-C2 at Kyoto does not provide the individual disturbances of the four Dst stations, i.e., the local Dst indices, which would clarify this problem. Therefore the cause of this error remains without final clarification. Nonetheless, the Dst index is erroneous in 1971 and the Dxt index corrects this error. The Dst index is some 2 nT more negative, on an average, than the Dxt index <cit.>. However, there are four consecutive years in 1963–1966 when the Dst index is considerably above (less negative than) the Dxt index <cit.>see Figure 5 in>[]Karinen_Mursula_2005. Moreover, it was found that the year 1965 is unique in that it is the only year for which the annual Dst index is positive, far above any other year. In comparison, annual means of the Dxt index remain negative in all years, even in 1965. Again, at the lack of local Dst indices, the cause of the exceptionally high Dst index in 1963–1966 remains without final explanation, but the great consistency of global and local Dxt indices suggests that the Dst index is slightly flawed in these years. The latitudinal normalization of the Dst index is questionable. Originally the mean of local disturbances was normalized by the cosine of the mean of geomagnetic latitudes <cit.>. However, later the mean of local disturbances was normalized by the mean of the cosines of geomagnetic latitudes <cit.>. Alas, since the Dst stations are at different latitudes, each local disturbance should be normalized by the cosine of the respective geomagnetic latitude in order to find the local disturbance of the same equatorial (horizontal) electric current. Only then the different stations measure the same ring current intensity, and their normalized disturbances (local Dst indices) can be averaged to find the global mean of the ring current <cit.>. Note that this physically correct way of normalization has never been adopted in the Dst index recipe, nor can the Dst index be properly normalized because of the lack of local disturbances. Note also that, although later studies have shown that other magnetospheric current systems <cit.> and induced earth-currents also contribute to the Dst index, their contribution is not subtracted from the Dst index when identifying storms or estimating their intensity. Here we mostly follow the Dxt reconstruction presented in Karinen and Mursula (2005). However, meanwhile, we have adopted a few new practices that slightly modify the original Dst recipe and further improve the Dxt index. In order to reduce the effect of seasonal variation and solar activity upon the baseline, we now calculate the secular variation using local midnight values (23 and 00 in local time) of the international quiet days. We also use a more sophisticated smoothing of the annual values by a 5-point quadratic Savitzky-Golay filter <cit.>, instead of using the second-order polynomial of the original recipe. We now calculate the hourly values of the baseline by interpolating the smoothed annual values to hourly resolution by a piecewise cubic Hermite interpolating polynomial (PCHIP). We also smooth the daily quiet-time curves using a 60-day Gaussian window, which removes the steps between successive monthly values that exist in the Dst index and in the earlier version of the Dxt index. The top and middle panels of Figure <ref> show the hourly and annual values of the Dxt and Dst indices, and the bottom panel depicts the Dxt-Dst difference of annual indices. One can see that the two indices differ even at annual resolution where their mean difference (mean absolute difference) is about 2.5 nT (3.3 nT, respectively). Note that the Dxt-Dst difference depicts a systematic long-term variation, with larger positive differences (relatively more disturbed Dst) during the active times from 1970s to early 2010s, and larger negative differences (relatively more disturbed Dxt) during the less active times (mainly in 1960s). This inconsistent long-term difference between the two indices is mainly due to the inconsistent quiet-time level of the Dst index, which follows the varying level of solar activity, being too low in the weak decade of 1960s and too high in the subsequent, more active decades. While a more complete analysis of the differences between the two indices will be presented elsewhere, here we will only discuss the practical consequences of these differences to the number and strength of magnetic storms. In particular, because the Dst index is, on an average, 2.5 nT lower than the Dxt index, it produces a somewhat larger number of magnetic storms than the Dxt index. However, as we will show in this paper, despite these differences, our results on storm occurrences and their implications about the long-term change of the Sun, remain the same for the two indices. Hourly (top panel) and yearly (middle panel) values of the Dxt (red) and Dst (blue) indices, and the Dxt-Dst difference (bottom panel) of yearly values. We would also like to note that there are currently a few slightly different versions of the Dst index in the different servers and databases. E.g., the Dst index in the ISGI data server and in the OMNI2 database of the NSSDC differ from the Kyoto WDC Dst index and from each other during long time intervals in the last few years, at least by the time of this writing (April 2022). This is most likely due to different (provisional or final) versions of the Dst index of WDC being implemented in the other databases at different times. (We have notified the respective institutions of these differences. A detailed list of these differences can be obtained from the authors of this paper at request.) Finally, we note that there are a number of more recent developments aiming to an improved estimate of the ring current, either using a novel reconstruction method <cit.>, improved temporal accuracy <cit.> or increased spatial accuracy by an extended station network <cit.>. While all of these developments are motivated and have their own applications, we will use here only the Dst/Dxt indices since they are the only indices to cover the whole space age. § SOLAR WIND FLOW TYPES Richardson_JGR_2000,Richardson_etal_2002a and Richardson_GA_2012 developed an hourly list of solar wind flow types using the measured values of the near-Earth solar wind parameters and some auxiliary data, like the sudden storm commencements, magnetospheric energetic particles, and cosmic rays. They classified the solar wind into three different main flow types: * CME-related flows that consist of interplanetary CMEs, including their possible upstream shocks and sheath regions * high-speed streams and the related corotating interaction regions * slow solar wind that can be related with the streamer belt and the HCS. The hourly list of solar wind flow types is very extensive but not complete, covering about 91.9% of all hours in 1964–2021. The annual coverages of solar wind flow type data are given in the bottom panel of Figure <ref>. Flow type data coverage is almost complete since 1995 when the ACE and WIND satellites started operation. It was very good also in 1965–1970 and 1973–1981, when several satellites were flying in the solar wind. However, in 1982–1994, when solar wind was measured mostly by only one Earth-orbiting satellite (IMP-8), gaps in the flow type data cover almost one third of time. Note also that, due to the auxiliary data, the overall coverage of solar wind flow type data is clearly better than the overall coverage of solar wind data (about 76.5%). Here we use an updated version of the flow type list extending from November 1963 to January 2022. (For more details on flow types and auxiliary data, see Richardson_GA_2012). The top panel of Figure <ref> shows the annual fractions of the three solar wind types. Here the fractions are calculated as a ratio to all available flow type values in the respective year. Thus, they exclude the data gaps, and could, therefore, also be called relative flow type fractions. The overall (relative) coverage of CME fraction is about 18.0%. The top panel of Figure <ref> shows that the CME flow fraction varies closely in phase with the sunspot cycle. The correlation coefficient between yearly sunspots and CME fractions is excellent, cc = 0.88, with an almost vanishing p-value of $1.7*10^{-19}$. Even the overall level of CME fraction during each cycle closely follows the changing height of sunspot cycles, although the exact timing of cycle peaks of CME fractions may differ from sunspot maxima by 1–2 years. The good agreement between CME fractions and sunspot activity is also seen, e.g., in the considerably lower CME fraction during the low sunspot cycle 24, compared to the level of CME fractions during the earlier, higher sunspot cycles. Top: Annual fractions of CME (red), HSS/CIR (blue) and slow solar wind (cyan) solar wind flows in 1964–2021. Scaled yearly sunspot numbers are depicted as a shaded area. Bottom: Annual coverages of the solar wind flow data. The HSS/CIR flow fraction maximizes in the declining phase of the sunspot cycle, when solar polar coronal holes extend to low latitudes. This timing difference leads to a significant negative correlation between yearly sunspots and HSS/CIR fractions (cc = -0.42; p = 0.0012). Note that the (relative) HSS/CIR fraction is, on an average, 41.3%, i.e., twice larger than the average CME fraction. In fact, the HSS/CIR fraction is larger than the CME fraction in almost all years except for a few sunspot maximum years. Interestingly, the cycle maxima of the HSS/CIR fraction seem to shift earlier in the sunspot cycle during the time interval included in Fig. <ref>. The HSS/CIR maxima are found in the pre-minimum to late declining phase in SC20–SC22, but in the early to mid-declining phase in SC23–SC24. Frequent high-speed streams in 1974 were found to come from persistent polar coronal hole extensions <cit.>, while in 2003 most HSSs originated from low-latitude coronal holes <cit.>. These differences reflect the systematic long-term evolution in the structure of solar magnetic fields (to be discussed later in more detail). Note also that, similarly to the CME fraction, the HSS/CIR fraction is smaller during SC24 than in earlier cycles. The slow wind fraction depicts a small solar cycle variation with maxima typically during or soon after sunspot minima, and minima around sunspot maxima. Accordingly, there is a significant negative correlation between yearly sunspots and slow wind fractions (cc = -0.59; p = $1.1*10^{-6}$). However, the most prominent feature in the slow wind fraction is its recent increase. Until 2002 the average slow wind fraction was 36.0%, twice larger than the CME fraction but clearly smaller than the HSS/CIR fraction. However, after 2003 the average slow wind fraction is 50.5%, making the slow solar wind the most common solar wind flow type in the last two decades. This recent increase in slow wind fraction during weakening sunspot activity contributes to the above-mentioned negative correlation between the two parameters. The two highest peaks in the slow wind fraction occur around the two sunspot minima in 2009 and 2019, and attain values slightly above and below 70%, respectively. This increase in the slow wind fraction during the last two decades reflects the widening of the streamer belt due to the weakening of solar polar magnetic fields <cit.>. Polar fields, again, are slowly weakening since cycle 22 as a consequence of the slow decrease in the overall solar activity that characterizes the whole space age. § STORM IDENTIFICATION We use here a fairly similar procedure to identify geomagnetic storms as adopted earlier by Yakovchouk_2012. We apply this procedure here separately both to the Dxt index and to the Dst index in order to study how the storms of different classes differ between these two indices. According to the storm identification procedure, we first locate all local minima in the Dxt/Dst index in 1964–2021. A storm is then identified as the deepest index minimum within a 2-day (48-hour) interval from any other minimum. Accordingly, all minima within a 2-day interval are counted to belong to the same storm as separate (smaller) intensifications. This definition correctly joins together to one storm, e.g., the two possible intensifications of a CME storm, one due to the sheath and the other due to the core (ejecta), which typically have a smaller time separation of about 12–24 hours. However, using this definition we miss one storm in those cases where two separate CMEs, each of which exceeds the storm threshold, follow each other within two days. While a more detailed estimate will be made in a separate study later, we have made a preliminary estimate that this leads to an underestimate of storms by less than about 10–15%. This is too small to have an effect on our main results and, therefore, no change due to possibly omitted storms is made here. However, we note that a shorter time interval of 2 days was applied here, in difference to the 3-day interval used by Yakovchouk_2012, because the shorter separation alleviates the problem of possibly omitted storms. This also leads to somewhat higher numbers of storms here than in Yakovchouk_2012. Figure <ref> depicts the Dxt index (red line) and the Dst index (blue line) in March-April 1980, as well as the storms identified by the Dxt index (magenta squares) and the Dst index (blue dots). The Dst index identifies six storms during this time interval, but the Dxt index only five storms. As discussed above, the Dxt index is, on an average, slightly higher than the Dst index and misses the weak storm in April 7 (min Dxt = -24 nT; min Dst = -32 nT). Overall, the selected index minima are quite appropriate to denote the peaks of storm main phases. Dxt index (red line) and Dst index (blue line) in March–April 1980, as well as storms identified by the Dxt index (magenta dots) and the Dst index (blue dots). We have classified the storms identified by the Dxt index (the Dxt storms) and, separately, by the Dst index (the Dst storms) according to the minimum value of the respective index into four storm classes or categories using the following definition: * Weak: -50 nT $<$ Dxt, Dst $\leq$ -30 nT * Moderate: -100 nT $<$ Dxt, Dst $\leq$ -50 nT * Intense: -200 nT $<$ Dxt, Dst $\leq$ -100 nT * Major: Dxt, Dst $\leq$ -200 nT. § TOTAL STORM NUMBERS DURING SPACE AGE Table <ref> shows the storm numbers in 1957–2021 both in total and when classified into the four intensity categories, separately for the two indices. There are in total 2526 geomagnetic storms according to the Dxt index and 2743 storms, i.e., some 8.6% more, according to the Dst index. When dividing these by the number of years (65), one finds that there have been 39/42 storms per year according to the Dxt/Dst index. Thus, as a rule of thumb, one can say that there have been, on an average, three storms per solar rotation during the space age. The largest category of storms, about 45% of all storms are weak storms, while 40% are moderate storms, 12% intense storms and 3% major storms. Almost the same percentages are found for both indices, which gives evidence for the robustness of these results. So, roughly speaking, almost a half of the storms were weak storms and three quarters of the rest were moderate storms. As seen in Table <ref>, There are 8.0% more of weak storms, 10.1% more of moderate storms, 6.9% more of intense storms, and 4.4% more of major storms in the Dst index than in the Dxt index. Table <ref> also lists the means of the storm minimum Dxt/Dst (storm peak) values (mean intensities), for all storms and separately for the four intensity categories. One can see that the two indices give very closely similar storm peak mean values of about -38 nT, -68 nT, -131 nT and -277/-276 nT for the weak, moderate, intense and major storms. Even though the Dst index produces more storms than the Dxt index, the distribution of storm numbers as a function of storm intensity is very similar for the two indices. Therefore the storm mean intensities also remain quite similar. The mean storm peak value for all storms is almost the same as for the moderate storms. Storm numbers (with respective percentages) and mean intensities for four storm classes in 1957–2021. Weak Moderate Intense Major All Dxt storms 1142 (45.2%) 1012 (40.1%) 304 (12.0%) 68 (2.7%) 2526 Dst storms 1233 (45.0%) 1114 (40.6%) 325 (11.8%) 71 (2.6%) 2743 Mean Dxt-min -38.1 nT -67.5 nT -130.5 nT -277.4 nT -67.5 nT Mean Dst-min -38.3 nT -67.7 nT -130.6 nT -275.8 nT -67.3 nT § YEARLY STORM NUMBERS DURING SPACE AGE Figure <ref> depicts the yearly numbers of Dxt storms (red lines and dots) and Dst storms (blue lines and dots) during the space age. The top panel of Fig. <ref> shows the yearly numbers of all storms, with yearly sunspots depicted as gray background to indicate sunspot cycles. One can see that in most years the number of Dst storms is slightly larger than Dxt storms. Only in 8 years the Dxt index includes more storms than the Dst index. Yearly numbers of Dxt and Dst storms. Top: all storms with yearly sunspots depicted as gray background; Second: weak storms; Third: moderate storms; Bottom: yearly sums of intense and major storms. Despite these differences in the total number of storms, the two indices agree on the two years when the number of storms is largest: 1960 and 2003, as well as on the two years when it is smallest: 1965 and 2009. The two indices also depict almost the same number of storms in these four extreme years. The storm numbers experienced a dramatic decline in the declining phase of SC23 from 60/61 storms in 2003 to 7/8 storms in 2009. A similar but slightly less dramatic decline occurred in the declining phase of SC19 from 58/58 storms in 1960 to 11/17 storms in 1965. After 1960, the two storm number series agree quite well with each other until the declining phase of SC20. The differences between Dxt and Dst storm numbers are particularly large in the declining phases of cycles 20 and 21 and around the maxima of cycles 22 and 23. They agree very well again during the last two, rather quiet decades since 2003. The larger differences in yearly storm numbers during the more active cycles correspond well to the time evolution of the Dxt-Dst difference depicted in the bottom panel of Figure <ref>. The second panel of Fig. <ref> shows the yearly numbers of weak storms during the space age. Weak storms follow a fairly systematic solar cycle variation with maxima in the declining phase of the cycle. During SC19–SC22 these maxima are located in the late declining phase or even at sunspot minimum, but during the last two cycles SC23–SC24 they have shifted to the early to mid-declining phase of the cycle. We will study this shift later in this paper when discussing the long-term evolution of solar magnetic fields and storm drivers. During the active cycles SC20–SC22 the number of weak storms had cycle minima typically around sunspot maxima, while during SC19 and the last two cycles the weak storm minima were around sunspot minima. Years 1965 and 2009, which are exceptionally low in the number of all storms (see top panel of Fig. <ref>), are the lowest also in weak storms but do not stand out so dramatically among other cycle minima of weak storms. Moreover, the number of weak storms has been at a quite high level even during the weak cycle 24, around the minimum thereafter and in the first years of cycle 25. The third panel of Fig. <ref> shows the yearly numbers of moderate storms, which have a fairly similar long-term evolution as the number of all storms. Here the minimum of 2009 also stands out quite dramatically. Cycle maxima of moderate storms are mostly found slightly earlier in the declining phase than for weak storms. We will discuss this difference in more detail later in the connection with storm drivers. Note that, excluding some years like 2003, there is a rather steady long-term decline in the number of moderate storms since the maximum in SC21. The latest minimum between SC24 and SC25 also remained quite low in the number of moderate storms. The bottom panel of Fig. <ref> shows the yearly numbers of intense and major storms combined together (to be called here large storms), because the yearly number of major storms is very low and vulnerable to be dominated by random fluctuations. The cyclic evolution of large storms, as well as their cycle maximum numbers, follow the sunspot cycles and their heights very well. This is due to the fact that most large storms are caused by CMEs, which follow sunspots <cit.>see, e.g.,>[]Webb_and_Howard_1994, Gopalswamy_2004, Cremades_2007, Robbrecht_ASR_2006, Webb_LRSP_2012. Note also the steady decline in the cycle maxima of large storms since the maximum of SC22 in 1989. Moreover, the 4-year period in 2007–2010 with no large storms is unique in the 65-year time interval, masking out even the latest minimum with two years (2019–2020) devoid of large storms. Figure <ref> also gives information on the differences between Dxt and Dst storm numbers. There are 36 years when the number of weak Dst storms is larger than the number of weak Dxt storms. However, there are also 19 years when the number of weak Dxt storms is larger than the number of weak Dst storms. This leads to the perhaps somewhat surprising result that the relative difference between the Dst and Dxt storm numbers is not in weak storms but in moderate storms. Note also that the large storm numbers are the same for Dxt and Dst in 41 years. § DRIVER CONTRIBUTIONS TO STORM NUMBERS We have identified the solar wind driver (flow type) for each storm in 1964–2021 (when there is solar wind flow type data available) by comparing the time of the storm peak (Dxt/Dst minimum) with the corresponding time in the list of solar wind flow types. A storm is assigned as a CME, HSS/CIR or slow wind storm, if the storm peak was inside the corresponding solar wind stream. Since, as discussed above, Richardson's classification has gaps, some storms remain unclassified. We call the latter no-SW storms. (Corresponding hours are called “unclear” in Richardson list.) The number of storms for each intensity category and solar wind driver, separately for Dxt and Dst storms, are summarized in Table <ref>. Out of a total of 2202/2416 storms in 1964–2021, 2059/2244 storms (93.5%$/$92.9%) could be assigned to a solar wind driver. This large fraction guarantees that the results on the relative fractions of storms of different solar wind drivers are representative and reliable. Using the storm numbers of Table <ref> we have also calculated various fractions of storms and listed them in Table <ref> and Table <ref>. Table <ref> gives the fraction (in percentage) of storms of a certain intensity category and solar wind driver out of all solar wind-classified storms, separately for Dxt and Dst storms. Accordingly, the sum of all numbers in the four central columns (or only in the 'All' column) of Table <ref> add up to 100%. Note that the relative fraction of all weak storms has increased from Table <ref> due to the shorter time interval and the neglect of the active cycle 19 which has relatively more of moderate and large storms than in the whole time interval. On the other hand, Table <ref> lists the fraction (in percentage) of Dxt/Dst storms of certain intensity and driver out of all storms of certain intensity. Thus, the sum of each column in Table <ref> is 100%. Dxt/Dst storm numbers in 1964–2021 classified according to storm intensity and storm driver. Weak Moderate Intense Major All CME 162/154 371/377 204/218 48/51 785/800 HSS 591/640 386/452 35/37 0/0 1012/1129 Slow 203/237 56/76 3/2 0/0 262/315 No SW 75/98 63/66 5/8 0/0 143/172 All 1031/1129 876/971 247/265 48/51 2202/2416 Fractions (in percentage) of Dxt/Dst storms of certain intensity and driver out of all Dxt/Dst storms in 1964–2021 covered by solar wind classification. Weak Moderate Intense Major All CME 7.9%/6.9% 18.0%/16.8% 9.9%/9.7% 2.3%/2.3% 38.3%/35.7% HSS 28.7%/28.5% 18.7%/20.1% 1.7%/1.6% 0.0%/0.0% 49.2%/50.3% Slow 9.9%/10.6% 2.7%/3.4% 0.1%/0.1% 0.0%/0.0% 12.7%/14.0% All 46.5%/46.0% 39.4%/40.3% 11.7%/11.4% 2.3%/2.3% 100%/100% Fractions (in percentage) of Dxt/Dst storms of certain intensity and driver out of all Dxt/Dst storms of certain intensity in 1964–2021. Accordingly, the sum of each column is 100%. Weak Moderate Intense Major CME 16.9%/14.9% 45.6%/41.7% 84.3%/84.8% 100%/100% HSS 61.8%/62.1% 47.5%/49.9% 14.5%/14.4% 0%/0% Slow 21.3%/23.0% 6.9%/8.4% 1.2%/0.8% 0%/0% Tables <ref> and <ref> show that the HSS/CIR streams produced 1012/1129 storms according to the Dxt/Dst index, almost exactly one half (49.2%$/$50.3%) of all solar wind-classified storms in 1964–2021. As a rule of thumb, HSS/CIR streams produced some 1.5 storms per solar rotation. There were 785/800 CME storms, making a good third (38.3%$/$35.7%) of all solar wind-classified storms. This is very close to one CME storm per solar rotation, on an average. Finally, there were 262/315 (12.7%$/$14.0%) slow wind storms, roughly one slow wind storm in a couple of rotations. The three solar wind streams contributed very differently to the three storm intensity categories. This is most notable for the largest storms, with CME streams producing all 48/51 major Dxt/Dst storms and 204/218 (84.3%/84.8%) out of the 242/257 solar wind-classified intense storms. Altogether, CME streams were responsible for 87% of all large (intense and major) solar wind classified storms, while HSS/CIR streams produced only 12% and slow wind streams only 1% of them. This solar wind driver division (87%$/$12%$/$1%) is the same for the large storms of both indices. The strongest storm in this 65-year interval occurred on 14 March 1989. It was, naturally, driven by a CME and reached the maximum intensity of -594/-589 nT according to Dxt/Dst index. This storm was by far larger than all other storms, since the second strongest storm on 20 November 2003 reached the intensity of only -427/ -422 nT, and all other storms remained above -400 nT. No HSS/CIR storm reached even close to the threshold of a major storm, since the strongest HSS/CIR storm on 13 September 1993 had the intensity only of -155/-161 nT. As seen in Tables <ref> and <ref>, there were altogether 35/37 (14.5%/14.4%) intense HSS/CIR driven storms. Slow wind streams produced only 3/2 (1.2%/0.8%) intense storms. § DRIVER CONTRIBUTIONS YEARLY Figures <ref> and <ref> show the yearly numbers of Dxt/Dst magnetic storms of different categories in 1964–2021, separately for the three solar wind drivers: CME streams (red line and dots), HSS/CIR streams (blue) and slow solar wind streams (cyan). Figures <ref> and <ref> depict all storms (second panel), weak storms (third panel), moderate storms (fourth panel) and large (intense and major) storms (fifth panel) related to the three drivers. Yearly and monthly sunspot numbers are depicted in top panel for reference. Top panel: Monthly (blue) and yearly (red) sunspot numbers for reference. Panels 2–5: Yearly storm numbers of all (panel 2), weak (panel 3), moderate (panel 4) and large (intense and major; panel 5) storms of the Dxt index, separated into CME (red), HSS/CIR (blue), and slow solar wind (cyan) storms. Same as Figure <ref>, but using storms of the Dst index. The third panel of Figures <ref> and <ref> verifies that the weak storms are mainly driven by HSS/CIR streams. The number of weak HSS/CIR storms greatly exceeds the number of weak CME storms or weak slow wind storms in most years, and maximizes in the declining phase of each cycle. Cycle maxima of the yearly number of weak HSS/CIR storms vary from 24 in SC20 to 12 in SC24. There is an interesting long-term shift in the occurrence of weak HSS storms and in the location of their cycle maxima. While during SC20, SC21 and SC22 most weak HSS/CIR storms are found in the late declining to minimum phase, during SC23 and SC24 they occur clearly earlier in the cycle, with cycle maxima in the early to mid-declining phase. This shift agrees with a similar shift in the location of cycle maxima of all weak storms (see Fig. <ref>), and proves that this shift is related to the evolution of HSS/CIR storms. In fact, a similar shift is also seen in the fraction of HSS/CIR streams depicted in Figure <ref>. On the other hand, in accordance with the close relation between sunspots and CMEs <cit.>see, e.g.,>[]Webb_and_Howard_1994, Gopalswamy_2004, Cremades_2007, Robbrecht_ASR_2006, Webb_LRSP_2012, the number of weak CME storms also maximizes in sunspot maximum years, but remains mostly below the number of weak HSS/CIR storms even then. As shown in Table <ref>, slow wind streams produce more weak storms than CME streams. In fact, as seen in Figures <ref> and <ref>, the yearly numbers of weak slow wind storms even surpass the corresponding numbers of weak HSS/CIR storms in a couple of years. The number of weak slow wind storms has notably increased during cycle 24. The average annual number of weak slow wind storms since 2010 is some 70% larger than before 2000. This increase is in agreement with the increasing fraction of slow wind streams in solar wind seen in Figure <ref>. The occurrence of large CME storms (see bottom panel of Figs. <ref> and <ref>) follows closely the sunspot cycle. Note, however, that the number of large CME storms decreases in SC24 relatively faster than, e.g., the number of weak or moderate CME storms. As discussed above (see Table <ref>), only 12% of large storms are HSS/CIR storms. Interestingly, the number of large HSS/CIR storms maximized in the declining phase of cycle 22, in early to mid-1990s, when their number even exceeded the number of large CME storms in four successive years (1993–1996). In fact, about 43% of all large HSS/CIR storms occurred in six years from 1991–1996. This is very similarly reproduced in both Dxt and Dst storms. We will discuss this development in Section <ref> in more detail. Moderate storms (fourth panel in Figs. <ref> and <ref>) are caused roughly equally by CME and HSS/CIR streams. In fact, as shown in Table <ref>, out of the 813/905 moderate Dxt/Dst storms with solar wind data, 371/377 are driven by CMEs and 386/452 by HSS/CIRs. Moderate CME storms peak around the sunspot maximum, while moderate HSS/CIR storms maximize in the declining phase. As in the case of weak HSS/CIR storms, moderate HSS/CIR storms peak somewhat earlier in the declining phase in the two recent, less active solar cycles (SC23 and SC24) than in the earlier, more active cycles. The all-time maximum number of moderate HSS/CIR storms occurred in 2003, in the mid-declining phase of cycle 23. This year also marks the all-time maximum of all HSS/CIR storms during the 65-year time interval, as seen in the second panel of Figures <ref> and <ref>. This year was the third most frequent year in high-speed streams (see Fig. <ref>) that mostly emanated from persistent, isolated low-latitude coronal holes <cit.>. § CME STORMS DURING SPACE AGE We have studied magnetic storms during the space age, which is characterized by the decline of the Modern Grand Maximum when solar activity has been slowly but unsteadily decreasing. As discussed above, large storms (bottom panel in Figs. <ref> and <ref>) are almost exclusively CME storms, whose occurrence is related to the new magnetic flux emerging on the solar surface. Accordingly, the temporal evolution of the yearly numbers of large storms closely follows the corresponding evolution of sunspot numbers. Figure <ref> shows that the first years 1957–1960 include the largest average number of large storms in any sequence of four consecutive years since 1957. This agrees very well with the fact these are the years of the maximum and early declining phase of SC19, and form the stormy peak of the MGM. The yearly numbers of large storms follow not only the sunspot cycle but also the longer-term evolution of sunspot activity during the whole space age. Large storms are, similarly to sunspots, greatly depleted from SC19 to SC20. Then their number increases until SC22 and, thereafter, systematically decreases during cycles 23 and 24. The variation of sunspot cycle amplitudes (see top panel of Fig. <ref>) roughly corresponds to the variation of cycle maxima of large storms, with the exception that the peak of SC22 is slightly higher in large storms than SC21, while the two cycles are roughly equally high in sunspots. In order to better quantify the relation between large storms and sunspots, we have calculated the correlation between yearly sunspot numbers and yearly numbers of large storms using different (yearly) lags between the two parameters. The maximum correlation is obtained at zero lag and has a correlation coefficient of 0.74 (p $< 2*10^{-10}$). Accordingly, the relation between sunspots and large storms is extremely significant, and sunspots explain more than a half of the variability of yearly numbers of large storms. This result supports the intimate connection between the two parameters. A much higher correlation between sunspots and large storms is not even expected since, in addition to sunspots, also other, much longer-lived active regions and filaments can produce CMEs <cit.>. We also note that the correlation between sunspots and large storms is considerably reduced already at one-year lag. Assuming that the fractions of Table <ref> are valid for the whole time period of Figure <ref> (so, also for the first 7 years 1957–1963 not covered by solar wind flow type data), a slightly smaller fraction than one half of all moderate storms are CME storms. On the other hand, almost exactly one half of moderate storms are HSS/CIR storms. As discussed above, the moderate HSS/CIR storms have their cycle maxima in the early to mid-declining phase of the solar cycle. Accordingly, we find that the maximum correlation between the yearly numbers of sunspots and all moderate storms is obtained at one-year delay. This correlation (cc = 0.58) is lower than for large storms but is still extremely significant (p $< 5*10^{-6}$). §.§ Large CME storms in five separate solar cycles Scatter plot of yearly sunspots and yearly numbers of large CME storms in 1965–2019, together with the corresponding best fit line (black thick line) and the two 95% CL lines (black dashed lines). Years of the five sunspot cycles and their separate best fit lines with sunspots are marked with separate colors and markers as indicated in legend. Slopes, intercepts, correlation coefficients and p-values for the correlation between yearly sunspots and large CME storm numbers for five solar cycles (years in second column) and for all years in 1965–2019. Cycle Years slope intercept cc p SC20 1965–1975 0.040 0.30 0.72 0.013 SC21 1976–1986 0.024 2.94 0.55 0.082 SC22 1987–1996 0.067 -1.16 0.97 2.3$*10^{-6}$ SC23 1997–2008 0.047 1.31 0.79 0.002 SC24 2009–2019 0.029 0.67 0.53 0.092 All 1965–2019 0.044 0.68 0.76 2.3$*10^{-9}$ Using data from the time interval (1964–2021) that is covered with information on solar wind flow types, we find a slightly higher correlation coefficient (cc = 0.75) between the yearly numbers of sunspots and the (flow type certified) large CME storms. However, due to the slightly smaller number of years, the significance is slightly reduced but remains extremely high (p $<4*10^{-10}$). In view of conducting here a study of individual, complete solar cycles, we limit the time to 1965–2019, from the start year of SC20 to the end year of SC24. Calculating the same correlation for these years (1965–2019), we find a similar change: the correlation coefficient is slightly increased to cc = 0.76, and the p-value is reduced to p $<2.3*10^{-9}$. Thus, sunspots explain, on an average, about 60% of the variability of the yearly number of large CME storms. We have depicted these latter values, as well as the slope and intercept of the best fit line of the corresponding linear correlation in the last row of Table <ref>. The decline in the yearly number of large CME storms from SC22 to SC24 depicted in the bottom panel of Figures <ref> and <ref> is indeed quite dramatic and may even look exceptional compared to the corresponding decline in sunspot activity. In order to study the relation between sunspots and large CME storms we have plotted the scatter of all yearly sunspots and yearly numbers of large CME storms in 1965-2019 in Figure <ref>. We have included there the corresponding best fit line of linear correlation between the two parameters and the 95% CL lines of the fit. Note that the intercept of this best fit line is close to zero (0.68), indicating that sunspots are, statistically, prerequisite to the occurrence of large CME storms. At intermediate sunspot activity with yearly mean of about 100, the number of large CME storms is about 5 and at higher sunspot activity of 200, twice larger. However, the 95% CL band is quite wide, and allows the storm numbers to vary roughly between 1 and 9 for intermediate sunspot activity and between 6 and 13 for high sunspot activity. We have marked in Figure <ref> the years of each of the five solar cycles (SC20–SC24) with a specific color and marker. We used the sunspot minimum months given at the Solar Influences Data analysis Center (SIDC) (see http://sidc.oma.be/silso/cyclesminmax) to separate the cycles. (The minimum years were allocated to the cycle which included a larger fraction of the respective year.) One can see that there are "outliers" (points beyond the 95% CL lines; stricly speaking they are not outliers but we use this word for lack of a more suitable term) in Fig. <ref> from each of the five solar cycles. Cycle 21 has two outliers, all others one, including cycle 24, which does not deviate from the other cycles in this relation. In order to study if there was a change in the relation between sunspots and large CME storms during SC24 or, in fact, during any of the five cycles 20–24, we have correlated yearly sunspots and large CME storm numbers separately for each of the five solar cycles. We have plotted the corresponding five best fit lines also in Fig. <ref> using the same cycle-specific color scheme as for the yearly dots. The slopes, intercepts, correlation coefficients and p-values of the best fit lines for the five cycles are given in Table <ref>. Figure <ref> and Table <ref> show that the intercepts and even the slopes of the five best fit lines vary considerably from cycle to cycle. The best fit lines of cycle 20 and cycle 23 have almost the same slope as the overall best fit line of all years. However, the intercepts of SC20 and SC23 best fit lines are quite different, leading to SC23 producing typically one large CME storm more for the same amount of sunspot activity than SC20. Still, both of these lines remain well within the 95% CL lines. Correlation coefficients for SC20 and SC23 are also quite close to the correlation coefficient of the overall fit, and correlation is significant for both of these two cycles. Cycles 21 and 24 have lower slopes than the slope of the overall best fit line. The intercept of SC21 is larger than the intercept of SC24, in fact larger than any other intercept. Correlation coefficients for SC21 and SC24 are smaller than for other cycles, explaining only some 30% of variation. In fact, correlation between yearly sunspots and large CME storms is not even significant for cycles 21 and 24 at 95% confidence level, but the fact that the respective p-values are below 0.10 indicates marginal significance. Figure <ref> shows that SC21 has two outlier points, one of them farthest below the lower 95% CL line. Curiously, for both SC21 and SC24, the outlier points correspond to the highest sunspot activity years of the respective cycles. As seen in Figures <ref> and <ref>, there are two peaks in the number of large CME storms during these cycles, separated by a dramatic dropout exactly at the respective sunspot maxima (years 1979 and 1980 in SC21 and year 2014 in SC24). This evolution, which essentially deteriorates the studied correlation in these cycles, closely reminds the structure of the Gnevyshev gap <cit.>. A simultaneous Gnevyshev gap in sunspots is clearly visible (although far less strong than in storm number) only in SC24, but not in SC21 where sunspots even reach the highest yearly value among the five cycles studied. Cycle 22 deviates from the other cycles in all aspects. As seen in Table <ref>, the slope of SC22 is the largest, more than twice the slopes of SC21 and SC24, and some 50% larger than the slope in SC20 and SC23. Most dramatically, there is an almost perfect correlation between the yearly sunspots and large CME storms in SC22, with sunspots explaining almost 95% of the annual variation of large CME storms. The close relation between sunspots and large CME storms in SC22 can also be seen (see Figs. <ref> and <ref>) in that both sunspots and large CME storms have SC22 maximum in 1989 and another, slightly lower peak in 1991. This is another demonstration of a Gnevyshev gap that is simultaneous in both parameters. §.§ Moderate CME storms in five separate solar cycles Same as Figure <ref> but for moderate CME storms. Slopes, intercepts, correlation coefficients and p-values for the correlation between yearly sunspots and moderate CME storm numbers for five solar cycles (years in second column) and for all years in 1965–2019. Cycle Years slope intercept cc p SC20 1965–1975 0.040 2.96 0.72 0.012 SC21 1976–1986 0.052 3.20 0.78 0.005 SC22 1987–1996 0.067 0.27 0.88 0.0007 SC23 1997–2008 0.045 2.43 0.64 0.025 SC24 2009–2019 0.072 0.91 0.76 0.006 All 1965–2019 0.055 1.84 0.78 1.8$*10^{-12}$ Figure <ref> depicts similar correlations for moderate CME storms as Figure <ref> for large CME storms. The corresponding parameters of correlations are given in Table <ref>. The slope of the correlation (0.055) between sunspots and all moderate CME storms (bottom row of Table <ref>) is larger than in the case of all large CME storms (0.044), reflecting the overall larger number of moderate CME storms (see Table <ref> and Figs. <ref> and <ref>). The correlation coefficient for all moderate CME storms (0.78) is almost the same as for large CME storms (0.76), but the p-value is three orders of magnitude smaller, indicating improved significance due to larger statistics. Comparing the best fit correlation lines of the five cycles in Figure <ref> to the corresponding lines in Figure <ref>, one can see that they are more coherently aligned between themselves and also with the overall best fit line in Figure <ref>. As Tables <ref> and <ref> show, the slopes vary from 0.040 to 0.072 (by 0.032) for moderate storms and from 0.024 to 0.067 (by 0.043) for large storms. Similarly, the spread of the five intercepts is smaller for moderate than large CME storms. Accordingly, the improved significance of the correlation between all moderate CME storms and sunspots is due to the five cycles having more similar correlations for moderate CME storms than for large CME storms. This is most likely due to the fact that the yearly number of moderate CME storms is larger than the number of large CME storms, which reduces statistical fluctuations between the five cycles for moderate CME storms. Note also that the correlation between sunspots and moderate CME storms is significant for all five cycles. Curiously, for some cycles, the correlation parameters remain surprisingly similar for moderate and large CME storms. This is most clearly valid for SC20 for which the slopes and correlation coefficients are exactly the same. Even the p-values are almost the same. The slopes of moderate CME storms remain almost the same as for large CME storms also in SC22 and SC23, but the significance of correlation is higher for large CME storms in these two cycles. Still, correlation is highly significant in SC22, where the correlation coefficient for moderate storms (as for large storms) is highest among all cycles, with sunspots explaining almost 80% of the annual variation of moderate CME storms. SC23 has the weakest correlation and the largest number of outlier points for moderate CME storms. The slopes and correlation coefficients in SC21 and SC24 are larger for moderate than large CME storms. Correlations are also 95% CL significant for moderate CME storms (but only marginally significant for large CME storms) in these two cycles. The slope for moderate CME storms is largest in SC24, even slightly larger than the corresponding slope in SC22. §.§ CME storms at solar minima Scatter plot of three-year means of sunspots and CME storms calculated for the minimum years of the four solar cycles SC20–SC22 and SC24 (blue circles), together with the corresponding best fit line (black thick line) and the two 95% CL lines black dashed lines). Red square denotes SC23 which is not included in the fit. Blue solid line denotes the best fit line between yearly sunspots and all CME storm numbers. A notable feature in the yearly number of all CME storms (see the second panel of Figs. <ref> and <ref>) is the low number of CME storms around the sunspot minimum between cycles 23 and 24 (to be called minimum of SC23). In fact, there were no CME storms at all in two successive years of 2007–2008. There was only one other year (1964) with no CME storms within the whole 58-year time interval (1964–2021) with solar wind flow data. Moreover, since the solar wind flow data was very partial in 1964 (see Fig. <ref>), the lack of CME storms in 1964 is not as solidly founded as in 2007–2008 when the coverage was almost 100%. Even the latest solar minimum after the weak cycle SC24 had at least two CME storms in each year. Moreover, during the four-year time interval in 2006–2009, there were altogether only 5 CME storms, out of which only one was large. In comparison, there were 12 (17) CME storms in the second-weakest (third-weakest) 4-year minimum of 2018–2021 (1993–1996, respectively) in 1964–2021. Note also that these results on the minimum-time storm numbers are similarly reproduced in both the Dxt and Dst indices, which gives strong support on their validity. In order to study the relation between yearly sunspots and yearly CME storm numbers during solar minima, we calculated the mean of yearly CME storm numbers in three successive years around each solar minimum when the number of CME storms was smallest. Figure <ref> shows these three-year mean CME storm numbers for the five cycle minima as a function of the corresponding three-year mean sunspot numbers. (Cycles 20–22 and 24 are depicted with blue circles, cycle 23 is depicted as a separate red square). We have also plotted there the best fit line (and the 95% CL lines) for the correlation between the sunspots and the three-year CME storm numbers for the four minima (cycles 20–22 and 24), leaving the minimum of SC23 out of the fit. The correlation coefficient is fairly high (cc = 0.95) and correlation is significant but, because of the small number of points, the p-value is only 0.047. Interestingly, Figure <ref> shows that the three-year mean CME activity at the minimum of SC23 is below the best fit correlation line of the four other cycles. This gives additional evidence for the view that CME activity in the minimum after cycle 23 was exceptionally weak, and underlines the uniqueness of this minimum among all other solar minima during the space age. We have also included in Figure <ref> the best fit line for the correlation between yearly sunspots and yearly numbers of all CME storms. This correlation is better (cc = 0.855; p $< 10^{-16}$) than between yearly sunspots and large CME storms (see Fig. <ref> and Table <ref>) or sunspots and moderate CME storms (Fig. <ref> and Table <ref>). The best fit line of the all-CME correlation is above the 95% confidence lines of the minimum time correlation. However, since the slope (0.123) of the all-CME best fit line is slightly smaller than the slope for the minimum time correlation, the all-CME line will fall within the 95% CL limit at sunspot value of about 55–60 and will cross the three-year minimum best fit line at about 95. This indicates a small nonlinearity in the relation between sunspots and CME storms at small sunspot numbers. § HSS/CIR STORMS AND HCS EVOLUTION WSO coronal source surface synoptic maps (radial model with source surface distance at 3.25 solar radii) for one (June–July) month in (left column) 1993 (CR1870), 1994 (CR1884), 1995 (CR1897), 1996 (CR1910) and in (right column) 2015 (CR2165), 2016 (CR2178), 2017 (CR2192), 2018 (CR2205). Maps are redrawn using WSO data and depict the coronal source surface field for one Carrington rotation in longitude (x-axis from 0$^\circ$ to 360$^\circ$) - latitude (y-axis; from -70$^\circ$ to +70$^\circ$) grid. Light gray background color denotes the positive polarity region, dark gray color the negative polarity region. (Both minima are positive polarity times). Neutral line is marked in black thick line and corresponds to the center of the heliospheric current sheet. Curves on both sides of the neutral line (above NL: blue line; below NL: red line) mark isolines of the source surface field intensity at $\pm0.5, \pm1, \pm2.5$, and $\pm5$ $\mu$T. As noted in Section <ref>, about 43% of all large HSS/CIR storms occurred in the six years from 1991–1996 (see bottom panel of Figs. <ref> and <ref>). Such an exceptional occurrence of intense HSS storms at this time is most likely due to the fact that the HCS was exceptionally thin during the declining phase of SC22 and the subsequent minimum <cit.>, allowing the Earth to reach higher heliomagnetic latitudes and more of high-speed streams than during the corresponding time of other solar cycles. Large HSS/CIR storms had their all-time maximum of four storms per year in two consecutive years in 1993–1994. However, interestingly, the moderate HSS/CIR storms (see the fourth panel of Figs. <ref> and <ref>) had their cycle maximum one year later, in 1995, and the weak HSS/CIR storms (the third panel of Figs. <ref> and <ref>) in 1996. (The latter was the all-time maximum for Dst and second highest peak for Dxt). A similar temporal ordering can be found in the peak number of large, moderate and weak HSS/CIR storms during all cycles of the space age. Moderate HSS/CIR storms tend to peak not earlier than large HSS/CIR storms, and weak HSS/CIR storms not earlier than moderate HSS/CIR storms. During the declining phase of SC23, the peak number of large HSS/CIR storms (3/4 in Dxt/Dst) was found in 2002, while the peak of moderate and weak HSS/CIR storms was one year later in 2003. During SC24, one large HSS/CIR storm occurred in 2013 and also in 2015, according to both indices. Moderate HSS/CIR storms had their maximum in 2015, and weak storms in 2015 and 2017 in Dxt and only in 2017 in Dst. During the earlier cycles, the peak numbers and timings are less certain because of the larger number of data gaps in solar wind flow classification, especially in the 1980s (see bottom panel of Fig. <ref>). Still, the same ordering is followed during the declining phase of SC20, with the peak of large and moderate HSS/CIR storms occurring in 1973, and weak HSS/CIR storms in 1974. In cycle 21, according to Dxt, large and moderate HSS/CIR storms had their peak in 1984 and weak storms in 1985. According to Dst, they occurred in 1984, 1983 and 1987, respectively. Note that the results on the temporal ordering of HSS/CIR storm numbers do not only apply to storm peaks but also to the temporal ordering of the bulk of corresponding HSS/CIR storms. For all cycles and for both storm indices, the mean of the distribution of moderate HSS/CIR storms occurs earlier in the declining phase than the mean of weak HSS/CIR storms. This temporal ordering in the occurrence of HSS/CIR storms of different intensity is related to a systematic evolution of the heliospheric magnetic field, in particular of the structure of the heliospheric current sheet, in the declining phase of the solar cycle. Figure <ref> depicts coronal (source surface) synoptic maps of the Wilcox Solar Observatory (WSO) for one solar rotation in four years in the mid-1990s (1993–1996), in the declining to minimum phase of SC22 (left column), as well as in four years in the late 2010s (2015–2018), in the declining phase of SC24 (right column). We have selected a rotation including the June month of each year, because the Earth is then close to the solar equator, viewing both hemispheres roughly equally. The coronal (source surface) magnetic field is calculated from photospheric field observations at WSO with a potential field source surface (PFSS) model <cit.>. We have selected here the radial model, which assumes that the photospheric field is radial, and the distance of the source surface is at 3.25 solar radii. This model needs no polar correction and depicts the structure of the HCS very clearly. Neutral line, the solar magnetic equator and the center of the heliospheric current sheet, is denoted as a thick black line in each map of Figure <ref>. Neutral line divides the coronal (source surface) magnetic field into two opposite polarity regions, which for both SC22 and SC24 are ordered in the same sense of positive (negative) polarity field dominating in the northern (southern, respectively) hemisphere. (Both are minima of positive solar polarity). Positive polarity field is depicted in the synoptic maps with light gray color and negative polarity field with dark gray. The white region between these two opposite-polarity regions is the heliospheric current sheet (streamer belt), whose center is the neutral line. Synoptic maps also include other curves which tend to roughly follow the neutral line structure. These curves are colored in blue above the neutral line and in red below it, and mark the isolines of the coronal (source surface) field intensity at $\pm0.5, \pm1, \pm2.5$, and $\pm5\,\mu$T. The $\pm1\,\mu$T lines (second blue and second red lines) define the HCS region and their latitudinal separation is used as an approximate width of the HCS. Accordingly, the closer the isolines are to each other, the larger the magnetic field gradients around the neutral line are, and the closer to the NL the polar coronal holes and the related high-speed streams are. Figure <ref> shows that the estimated HCS is rather narrow, about 20–25$^\circ$ wide in heliographic latitude during all the years of SC22 (1993–1996), while it is considerably wider, about 35–40$^\circ$ in SC24 (2015–2018). This is in agreement with the above noted exceptionally thin HCS during the declining phase of SC22 and the subsequent minimum <cit.>. The thin current sheet allows a better access for the fast solar wind from polar coronal holes to low heliolatitudes. Therefore, the Earth is, on an average, more exposed to the high-speed solar wind streams in the 1990s than, e.g., in the 2010s. The occurrence and temporal duration of high-speed solar wind streams at the Earth depends not only on the thickness of the HCS, but also on its curviness, which is mainly determined by the solar dipole tilt angle. The dipole tilt angle has a systematic variation along the solar cycle. It has its maximum of 90$^\circ$ at the time of polarity reversal, which happens close to the sunspot maximum. The tilt angle decreases slowly but unsteadily during the declining phase to its minimum (typically close to 0$^\circ$) at or soon after the sunspot minimum. One can see in Figure <ref> the typical decrease of the HCS tilt (curviness) from 1993 to 1996 (left maps) and from 2015 to 2018 (right maps). Note that these maps are from slightly different parts of the declining phase of the two cycles, reflecting the different evolution of polar fields. This also leads to the different timing of HSS/CIR storms during these two cycles, as discussed above. In SC24, the first map in 2015 is from the early declining phase, while the first year of SC22 (1993) is already in the mid-declining phase. Accordingly, the tilt is slightly larger in 2015 than in 1993 (see top row of Fig. <ref>). Anyway, in both years, the HCS is quite curved and regions of positive magnetic polarity from the northern hemisphere intrude into the southern hemisphere, and vice versa. This is clear, e.g, in 1993 when the northern field intrudes into the southern hemisphere in longitude range of about 120–210$^\circ$ and about 270–360$^\circ$, or in 2015 when the southern field intrudes strongly into the northern hemisphere in the longitude range of about 110–220$^\circ$. During such intrusions the Earth is subjected to the effect of HSS streams from coronal holes (and related CIRs developing during the passage from the Sun to the Earth). Since the solar wind speed increases with heliomagnetic latitude, i.e., with the distance from the neutral line, the further the intrusion takes the neutral line from the solar equator (or, more exactly, from the ecliptic), the faster the solar wind measured at the Earth is and the longer the HSS stream lasts. Thus, since the tilt is decreasing during the declining phase of the solar cycle, the intrusions most effectively produce large HSS/CIR storms in the early to mid-declining phase. Moreover, since the HMF intensity has its cycle maximum in the early declining phase, the southward HMF component, which controls the reconnection in the dayside magnetosphere, is, on an average, also slightly stronger at this time. These two effects lead to the fact that the cycle maxima of large HSS/CIR storms are typically seen in the early to mid-declining phase, before the corresponding maxima of moderate or weak HSS/CIR storms. In 1994 and in 2016 (second row of Fig. <ref>), the HCS structure is quite dipolar, but the tilt is already smaller than one year earlier. Although the Earth's largest distance from the NL is now smaller (Earth reaches slightly lower heliomagnetic latitudes) than one year earlier, the Earth is still during a large fraction of time outside of the HCS. The tilt is further reduced in 1995 and 2017 (third row of Fig. <ref>), and again in 1996 and 2018 when the Earth was mostly within the HCS region, at least in June. As seen in Figures <ref> and <ref>, the number of moderate HSS/CIR storms in SC24 systematically declines from the maximum in 2015 onwards, as the tilt is decreasing. During SC22, the maximum of large HSS/CIR storms was found in 1993–1994 when the tilt was still fairly large, while the moderate HSS/CIR storms peaked in 1995. Note that Figure <ref> depicts the HCS in June when the Earth is at low heliographic latitudes. However, during the high-latitude periods in Spring and Fall, when the Earth is below or, respectively, above the solar equator, the HCS thickness plays a crucial role. Then, if the HCS is narrow, as in the declining phase of SC22, the Earth can be exposed to fast solar wind streams around equinoxes. For example, in 1996, at the maximum year of weak HSS/CIR storms of SC22, some 75% of weak storms occurred in Spring or Fall, and only 25% in Summer and Winter. Accordingly, the evolution of the properties (in particular thickness and tilt) of the HCS and solar magnetic fields over the solar cycle and over the whole space age can explain all the observed changes in the occurrence of HSS/CIR storms noted in earlier Sections. First of all, the temporal ordering of the cycle maxima of large, moderate and weak HSS/CIR storms (in this order) in the declining phase of each solar cycle reflects the systematic decrease of the HCS tilt (curviness) during the respective phase of each solar cycle. This decrease, again, follows the regular dynamics of solar magnetic fields that also control the solar dipole tilt. As explained in the Introduction, surges of new magnetic flux create extensions of polar coronal holes to lower latitudes and, thereby, affect the dipole tilt. When activity subsides in the declining phase, less surges appear and the dipole tilt decreases. Since large (weak) HSS/CIR storms require a larger (smaller, respectively) tilt, their occurrence maxima follow the decreasing tilt during the declining phase. We note that the HCS affects the HSS/CIR storm occurrence even at other times of the solar cycle, but the effect is far less clear and systematic because the solar wind speed does not have the same latitudinal ordering at other times (especially around solar maxima) as in the declining to minimum phase of the cycle. Secondly, the change in overall solar activity during the space age and its effect upon the structure of solar magnetic fields and the HCS can explain the shift in the location of cycle maxima of HSS/CIR storms. As noted in Section <ref>, during SC20–SC22 the cycle maxima of large and moderate HSS/CIR storms are located in the late declining phase, but during the last two cycles SC23–SC24 they have shifted to the early to mid-declining phase of the cycle (see two lowest panels of Figs. <ref> and <ref>). The same shift is also seen in the timing of cycle maxima of all weak storms (see second panel of Fig. <ref>). After the reversal of the solar dipole, surges of new-polarity magnetic flux strengthen the polar fields. In the case of weak cycles, there are fewer surges, whence the polar fields remain weaker, polar coronal holes smaller and the HCS region wider. Also, extensions of polar coronal holes only occur in the early to mid-declining phase of the cycle, leading to early maxima of HSS/CIR storms in weak cycles. Moreover, because of the thick HCS, the Earth stays within the slow wind of the HCS region from quite early on in the declining phase, which reduces the occurrence even of weak HSS/CIR storms. On the other hand, during strong solar cycles, there is more of new magnetic flux emerging on the solar surface in the form of sunspots and other active regions. Accordingly, there are more surges that strengthen the polar fields and form coronal hole extensions, which can occur even in the late declining phase of the cycle. These changes lead to a later maximum of HSS/CIR storms in these cycles. Strong polar fields also push the HCS region thinner, which allows the occurrence of weak HSS/CIR storms even around sunspot minimum. This evolution culminated during the extreme cycle of SC22, when the solar polar fields and the HMF attained their maximum value during the space age <cit.>. Note that the field intensity isolines reach $\pm5 \mu$T in 1990s (left plots of Fig. <ref>) but only $\pm2 \mu$T in 2010s (right plots of Fig. <ref>). The exceptionally strong polar fields in SC22 produced an exceptionally narrow HCS during the subsequent declining phase and minimum <cit.>, which allowed the HSS streams to occasionally reach the Earth even during the sunspot minimum. Quite appropriately, the only cycle when the maximum of weak HSS/CIR storms was on a minimum year, was cycle 22. After SC22, the weakening sunspot activity produced weaker solar polar fields <cit.> and smaller polar coronal holes <cit.> in SC23 and SC24, which led to a thicker HCS during the respective declining phases and minima <cit.>. This is also seen in the fact that, during this millennium, the Earth has spent an increasing fraction, roughly half of the time within the HCS/streamer belt region. Also, the slow wind of the streamer belt has become a more important source of weak magnetic storms than ever during the space age. § DISCUSSION AND CONCLUSIONS We have studied in this paper the occurrence of magnetic storms of different intensities during the whole space age from 1957 until 2021. We have used both the original storm index, the Dst index, and its recalculated and corrected version, the Dxt index. These two indices have differences, e.g., in overall normalization and quiet-time levels. Therefore the yearly storm numbers in the four intensity categories extracted from these two indices slightly differ between each other. However, even despite a small systematic long-term trend in the ratio of the two indices, there is no significant difference in the relative occurrence of storms of different intensities according to the two indices. Rather, we find that all of the main results on the long-term occurrence of magnetic storms and their implications about the evolution of solar magnetic fields and the solar wind are the same using either of the two indices. Moreover, the changes in the Sun during the space age leading to a varying number of magnetic storms are much larger than the differences in the storm numbers between the two indices. There were in total 2526 magnetic storms during the space age according to the Dxt index and 2743 storms, i.e., some 8.6% more, according to the Dst index. This implies that there were, on an average, 39/42 storms per year, i.e., roughly three storms per solar rotation. About 45% of all storms were weak storms (-50 nT$<$Dxt/Dst$\leq$-30 nT), 40% moderate storms (-100 nT$<$Dxt/Dst$\leq$-50 nT), 12% intense storms (-200 nT$<$Dxt/Dst$\leq$-100 nT) and 3% major storms (Dxt/Dst$\leq$-200 nT). So, roughly speaking, almost a half of storms were weak storms and three quarters of the rest were moderate storms. Almost exactly the same percentages are found for both indices. The two indices gave also very closely similar storm peak mean values of about -38 nT, -68 nT, -131 nT and -277/-276 nT for weak, moderate, intense and major storms. We also used solar wind flow type information <cit.> in order to assign magnetic storms occurring in 1964-2021 to their three main solar wind drivers, the coronal mass ejections (CME), the high-speed solar wind streams (HSS/CIR) and the slow wind region related to the streamer belt and the heliospheric current sheet (HCS). The HSS/CIR streams produced a bit more than one thousand (1012/1129) solar wind-classified Dxt/Dst storms, almost exactly a half of all solar wind-classified storms in 1964–2021. There were 785/800 CME storms, making a good third (38.3%$/$35.7%) of all solar wind-classified storms, and roughly one CME storm per solar rotation, on an average. Slow solar wind produced some 300 (262/315) storms, making 12.7%$/$14.0% of all solar wind-classified storms, roughly one slow wind storm in a couple of rotations. The three solar wind streams contributed very differently to the different storm intensity categories. CME streams were the cause of all the 48/51 major storms (Dxt/Dst$\leq$ -200 nT) occurring in 1964-2021. Although CME storms may also include an effect of a HSS, this result proves that the generation of a major storm without a CME was extremely unlikely during the space age. CMEs produced 84.3%/84.8% of intense (-200 nT$<$Dxt/Dst$\leq$-100 nT) storms, HSS/CIR streams produced 14.5%/14.4% of them and slow wind streams only 1.2%/0.8%. Moderate storms were caused slightly more often by HSS/CIR streams (47.5%/49.9%) than by CME (45.6%/41.7%) streams, or by slow wind (6.9%/8.4%). HSS/CIR streams (61.8%/62.1%) were clearly the dominant source of weak storms, but even the slow wind streams (21.3%/23.0%) produced more weak storms than CMEs (16.9%/14.9%). The whole solar-terrestrial environment during the space age is characterized by a slow, but unsteadily evolving decrease of solar magnetic activity after the maximum of solar cycle 19, the peak of the Modern Grand Maximum (MGM). This long-term evolution also affects the sources, occurrence frequencies, intensities and other properties of magnetic storms. Note also that, since the three main solar wind drivers of magnetic storms depend on different solar parameters and vary at different time scales, the long-term decrease in solar activity affects differently on storms of different origin or different intensity. Coronal mass ejections are mainly produced by fairly new magnetic flux emerging on the solar surface. This emergence is evidenced and traditionally even quantified by sunspots. It has been shown that the occurrence of CMEs follows the sunspot cycle <cit.>see, e.g.,>[]Webb_and_Howard_1994, Gopalswamy_2004, Cremades_2007, Robbrecht_ASR_2006, Webb_LRSP_2012. Here we verified that CMEs producing moderate and, separately, large (intense or major) magnetic storms vary closely with the changing sunspot activity, not only over the solar cycle but even at longer time scales. Large storms, which are almost exclusively produced by CMEs, occurred most frequently in the four years (1957–1960) of the maximum and early declining phase of SC19, the peak of the MGM. CME storms mainly follow sunspots within a year, thus being produced by fairly newly emerged flux, rather than distributed flux. Sunspots explain typically 60% of the variation of the yearly number of large and moderate CME storms. On an average, the yearly number of large CME storms is zero if no sunspots exist, suggesting that active regions without sunspots are not effective in producing large CME storms. The yearly number of large CME storms increases to 4–5 storms for an intermediate sunspot number of 100 (version 2.0, <cit.>). On the other hand, about two moderate CME storms in a year can occur even if no sunspots exist. This suggests that, e.g., active regions without sunspots, filaments and other forms of distributed (not newly-emerged) solar magnetic fields can produce CMEs that are sufficiently strong for moderate storms. Increasing yearly sunspot number to 100 (200) increases the yearly number of moderate CME storms, on an average, to about 7 (13) storms. We studied the correlation between sunspots and yearly number of CME storms separately for each of the five full solar cycles (SC20–SC24) included within the space age. Most cycles had significant correlations between sunspots and the yearly number of large and, separately, of moderate CME storms. For large CME storms in SC21 and SC24 correlation was only marginal, which is due to the fact that large CME storms have two separate peaks around the respective sunspot maxima, with a deep minimum in between. This evolution reminds of the so-called Gnevyshev gap which is quite a common feature in the solar cycle evolution of several solar and heliospheric parameters <cit.>. Since sunspots typically depict only a small decrease during a possible Gnevyshev gap (and hardly any in SC21), this leads to a couple of years where storm numbers are small but sunspot numbers high, which deteriorates the overall correlation. Because such a two-peak structure does not seem to be common to all cycles, it likely results from random fluctuations of rather small numbers of yearly large CME storms. This will be studied later in more detail. For moderate CME storms the storm numbers are larger, which reduces the effect of random fluctuations. Moreover, as discussed above, moderate CME storms can be produced even without sunspots. These facts may explain the larger number of moderate storms even at the corresponding Gnevyshev gap minimum, which makes the correlation between sunspots and moderate CME storms more significant even for SC21 and SC24, although a (less deep) Gnevyshev gap is seen also in the number of moderate storms for both cycles. The best correlation between sunspots and large and, separately, moderate CME storms was found during cycle 22. During SC22, sunspots explain 94% and 77%, i.e., far more than the average 60%, of the variation of yearly number of large and moderate CME storms, respectively. Cycle 22 also marks the highest peak in the number of large CME storms during the solar wind flow covered period (SC20–SC24). Sunspots depict a clear (but not very deep) Gnevyshev gap during this cycle, and large CME storms follow this evolution closely, without an excessive decrease between the two peaks. Thus, the cycle evolution of sunspots and large CME storms is very similar during SC22, leading to the extremely high correlation. On the other hand, the Gnevyshev gap of moderate CME storms in SC22 is relatively deeper than in large CME storms. This reduces the respective correlation with sunspots in SC22 below that for large storms, but is still the best for moderate storms among all cycles. The parameters of the linear correlation between sunspots and large and moderate CME storms also vary considerably from cycle to cycle. For large CME storms, SC20 and SC23 have almost the same slope as the overall best fit line for all years. SC21 and SC24 have considerably lower slopes, while SC22 has a slope more than twice higher than SC21 and SC24. For moderate CME storms, the five slopes deviate less from each other and from the common mean, probably because of better statistics due to a larger number of storms. However, despite some difference in correlation parameters, the best fit lines of all cycles remained within the 95% CL boundaries of the overall fit. Thus, interestingly, no cycle clearly deviated from the other cycles in the relation between sunspots and large or moderate CME storms. This applies also to the low cycle 24. However, when studying the relation between sunspots and all CME storms around sunspot minima (using 3-year means in order to increase statistics), we found that the minimum between cycles 23 and 24 deviates from the other four minima by remaining below the 95% CL correlation boundary. This minimum was very weak in sunspot activity, although still slightly more active than the following minimum. However, it was exceptionally weak in CME storms, with no CME storms at all in two years 2007–2008. Although this minimum breaks the common correlation between sunspot activity and CME storm occurrence at sunspot minima, the remaining years of the two annexing cycles, SC23 and SC24, recover this relation for full cycles and make it agree with all other cycles. We have shown in this paper that there is an intimate connection between the occurrence of HSS/CIR storms and the structure of the heliospheric current sheet, in particular its latitudinal width (thickness) and the tilt (curviness). A thin HCS makes large gradients of solar wind properties around the neutral line (center of the HCS region), whence the Earth is more often affected by high-speed solar wind streams. On the other hand, the tilt of the HCS determines how high (northern or southern) heliomagnetic latitudes the Earth can reach. A large tilt in the HCS can produce an intrusion of high-speed solar wind to low heliographic latitudes and to the ecliptic. The tilt angle of the HCS decreases fairly systematically from very high (about $90^\circ$) to very low tilt angles during the declining phase of the solar cycle. Three-year running mean numbers of CME (red line) and HSS/CIR (blue line) Dxt storms. Panels depict (from top to bottom) all storms, weak storms, moderate storms and large storms. Vertical lines denote the locations of cycle maxima of HSS/CIR storms of respective intensity. Annual sunspots are depicted as gray background in each panel. This decrease of the HCS tilt angle controls the occurrence of HSS/CIR storms of different intensity in the declining phase of all solar cycles. Due to this decrease, intense HSS/CIR storms tend to occur not later than moderate HSS/CIR which, again, tend to occur not later than weak HSS/CIR storms. We could see this ordering already in the yearly number of HSS/CIR storms depicted in Figures <ref> and <ref>. However, since random fluctuations may have some effect on the ordering of peaks and, thereby, cast doubt on our results, we have calculated three-year running means of the yearly CME and HSS/CIR storm numbers, separately for weak, moderate, large and all Dxt storms, and depicted them in Figure <ref>. Three-year running means weight the distribution of storms more widely and, therefore, are less vulnerable to random fluctuations of yearly peaks. We have included in Figure <ref> also vertical lines to indicate the cycle maxima of HSS/CIR storms of each intensity class. (In case two years had the same mean number, the first was marked; this only applies to large HSS/CIR storms where numbers are small). The vertical lines of Figure <ref> clearly demonstrate the above discussed ordering of the cycle maxima of HSS/CIR storms. During all the five cycles SC20–SC24 of space age, the cycle peak of large HSS/CIR storms did not occur later than the corresponding peak of moderate HSS/CIR storms which, again, was not later than the peak of weak HSS/CIR storms in the same cycle. This ordering is most clear in the declining phase of SC22, where the three-year mean maxima of large, moderate and weak HSS/CIR storms follow each other in steps of one year from 1993 to 1994 and 1995. As noted earlier, cycle 22 is also exceptional in the number of large HSS/CIR storms, which is dramatically clearly seen in the bottom panel of Figure <ref>. More than 40% of all large HSS/CIR storms in 1964–2021 occurred in the six years of 1991–1996. Although neither moderate nor weak HSS/CIR storms attained their space age peak during SC22, the three-year running mean number of all HSS/CIR storms maximized in 1994, in the late declining phase of SC22. These results are due to the exceptionally thin HCS in the declining phase of SC22 <cit.>, which resulted from exceptionally strong solar polar and heliospheric fields in the declining phase of this cycle <cit.>. Note also that the solar magnetic polarity is positive around the minimum between cycles 22 and 23, which enhances geomagnetic activity at high heliolatitudes (in Spring and Fall) by the Russell-McPherron (RMP) mechanism <cit.>. While this mechanism alone does not cause the exceptional storminess of cycle 22, the thin current sheet also increases the occurrence of dominant HMF sectors at high heliolatitudes, (the so-called Rosenberg-Coleman effect <cit.>), which enhances the effectivity of the RMP mechanism during cycle 22. Thus, the increased HSS activity and the enhanced RMP effectivity, both due to the exceptionally thin HCS, lead to the exceptional occurrence of large HSS/CIR storms during the declining phase of cycle 22. Similarly as for CME storms, the long-term decrease of sunspot activity during the space age also affects the occurrence of HSS/CIR storms. However, the decrease during the space age was not steady or continuous, neither in sunspot activity nor in magnetic storms. The total number of storms experienced two dramatic dropouts, one in the declining phase of SC19, from the record storm level in 1957–1960 until the second-lowest minimum in 1965, and the other in the declining phase of SC23, which continued to the low cycle 24 (see Fig. <ref>). Between these two dropouts, both sunspot and storm activity first increased from SC20 on until cycle 22, and then slightly decreased in SC23. The increase in sunspot activity from SC20 on strengthened the intensity of polar magnetic fields, which culminated in SC22 <cit.>. Strong polar fields pressed the HCS region exceptionally thin in the declining phase of SC22, producing a record of large HSS/CIR storms during the space age. Since SC22, polar fields have continued to decline, which has widened the HCS region. During a wide HCS, the Earth can reach the fast solar wind only at times when the HCS tilt angle is large. Therefore, the maxima of HSS/CIR storms occur earlier in the declining phase of weak cycles of SC23 and SC24 than during the earlier, stronger cycles. This can be seen in the yearly storm numbers of Figures <ref> and <ref>, and it can even more clearly be seen in the three-year running mean storm numbers of Figure <ref>. The shift of cycle peaks of HSS/CIR storms from the late declining phase in the more active cycles SC20–SC22 to the early declining phase in the recent, more quiescent cycles SC23 and SC24 is, very appropriately, most clearly visible for the large HSS/CIR storms. In SC20–SC22, the maxima of large HSS/CIR storms are found some 3–4 years after the respective sunspot maximum, but in SC23 only 1–2 years later and in SC24 the two maxima occur in the same year. Since the cycle maxima of moderate and weak HSS/CIR storms occur, as discussed above, later (or not at least earlier) than large HSS/CIR storm peaks, this long-term shift is less clearly seen for the weaker HSS/CIR storms. For moderate storms, this shift is still seen even in the smoothed numbers of Figure <ref> (peaks in SC20–SC22 are 4–6 years, in SC23 2–3 years and in SC24 2 years after the respective sunspot maximum). Note, first, that the 3-year smoothing moves the weak storm peak in SC23 to 2007, while for yearly storm numbers it was in 2003 (see Figs. <ref> and <ref>). Year 2003 was the peak storm year for both weak and moderate (and, naturally, all) HSS/CIR storms, and only one year after the peak in large HSS/CIR storms. This year is known to be the record year of geomagnetic activity in space age <cit.> and one of the record years of HSS occurrence (see Fig. <ref>). In this year, the HSSs mainly originated from low-latitude coronal holes <cit.>. Accordingly, the occurrence of HSS/CIR storms at this time was less strongly controlled by solar polar coronal holes than in the other cycles. Secondly, the occurrence of weak HSS/CIR storms in SC24 (and partly even in the declining phase of SC23) was increasingly affected by the widening of the HCS. As noted earlier, the yearly number of weak HSS/CIR storms in the minimum of SC24, and later in the early ascending phase of SC25, is not much less than in the prior declining phase (see Figs. <ref> and <ref>). This leads to the fairly constant (smoothed) number of weak HSS/CIR storms in Figure <ref>) since the maximum of SC24. Thirdly, since the HCS has been exceptionally wide for about 15–20 years, the Earth stays, for the same tilt angle, longer inside the HCS region. Thus, the decrease of solar activity and solar polar fields since SC22 have led to the fact that the Earth stays, since the declining phase of SC23, more than 50% of the time within the HCS and the related slow solar wind (streamer belt). This has greatly reduced the number of intense and moderate HSS/CIR storms, and significantly raised the relative fraction of slow solar wind as the source of weak magnetic storms. Concluding, we have studied here the occurrence of magnetic storms of different intensity over the space age (1957–2021). Space age is characterized by a slow, unsteady decrease of solar activity, which directly affects the CME storms and, indirectly, via its effects to the structure of the heliospheric current sheet, also the HSS/CIR storms. We find that the variations in sunspot activity closely explain the variation in the yearly number of CME storms over the whole space age. All cycles separately comply to this common, rather tight connection between sunspots and CME storms. However, during the weak solar minimum between cycles 23 and 24 the number of CME storms is smaller than predicted by the common relation. During four years from 2006–2009, there were only five CME storms, out of which only one was large. In comparison, there were 12 CME storms in the second-weakest four-year time interval of 2018–2021. Increasing sunspot activity from SC20 to SC22 increased solar polar fields, which led to an exceptionally thin HCS during the declining phase of SC22, producing a record of large HSS/CIR storms. Subsequent decrease in sunspot activity has weakened solar polar fields and widened the HCS region. These long-term changes in HCS width have affected the occurrence of HSS/CIR storms during the solar cycle so that HSS/CIR storms occur, since SC23, in the early to mid-declining phase of the solar cycle, while during the earlier, more active cycles 20–22, they maximized in the late declining phase. Widening HCS has also increased the significance of the HCS and the related slow solar wind (streamer belt) for the Earth. The Earth spends now, since the declining phase of SC23, i.e., roughly from the start of the new millennium, about 50% of the time, more than ever before during the space age, in the slow solar wind. This has reduced the number of large and moderate HSS/CIR storms and increased the fraction of slow wind as the source of weak magnetic storms. The authors acknowledge financial support by the Academy of Finland to the Postdoctoral Researcher project of L. Holappa (No. 322459) and to the PROSPECT project (No. 321440) of T. Asikainen. T. Qvick was supported by University of Oulu Graduate School. We thank I. Richardson for the updated version of solar wind flow data <cit.>. The Dxt index is maintained by the University of Oulu at http://dcx.oulu.fi/ <cit.>. The Dst index is provided by the WDC for Geomagnetism, Kyoto, Japan at http://wdc.kugi.kyoto-u.ac.jp/wdc/Sec3.html <cit.>. Sunspot data are from the World Data Center SILSO, Royal Observatory of Belgium, Brussels, on-line catalogue at https://www.sidc.be/silso/datafiles. The WSO source surface maps are available at http://wso.stanford.edu/synsourcel.html. Altschuler, MD. Newkirk, G. Magnetic Fields and the Structure of the Solar Corona. I: Methods of Calculating Coronal Fields Magnetic Fields and the Structure of the Solar Corona. I: Methods of Calculating Coronal Sol. Phys.9131-149. , Maliniemi Asikainen, T. , Maliniemi, V. Mursula, K. Modeling the contributions of ring, tail, and magnetopause currents to the corrected Dst index Modeling the contributions of ring, tail, and magnetopause currents to the corrected Dst J. Geophys. Res.115A1412203. Balogh, A. , Bothmer, V. , Crooker, NU. , Forsyth, RJ. , Gloeckler, G. , Hewish, A. Mikić, Z. The Solar Origin of Corotating Interaction Regions and Their Formation in the Inner Heliosphere The solar origin of corotating interaction regions and their formation in the inner heliosphere. Space Sci. Rev.89141-178. , Sittler , Mariani Burlaga, L. , Sittler, E. , Mariani, F. Schwenn, R. Magnetic loop behind an interplanetary shock: Voyager, Helios, and IMP 8 observations Magnetic loop behind an interplanetary shock: Voyager, Helios, and IMP 8 observations. J. Geophys. Res. (Space)86A86673–6684. , McPherron Burton, RK. , McPherron, RL. Russell, CT. An Empirical Relationship Between Interplanetary Conditions and Dst An empirical relationship between interplanetary conditions and Dst. J. Geophys. Res.804204–4214. St. Cyr St. Cyr Cremades, H. St. Cyr, OC. Coronal mass ejections: Solar cycle aspects Coronal mass ejections: Solar cycle aspects. Adv. Space Res.4071042-1048. Crooker, NU. Cliver, EW. Postmodern View of M-regions Postmodern view of J. Geophys. Res.99A1223383–23390. Daglis, IA. Ring Current Dynamics Ring Current Space Sci. Rev.1241-4183-202. Daglis, IA. , Kozyra, JU. , Kamide, Y. , Vassiliadis, D. , Sharma, AS. , Liemohn, MW. Lu, G. Intense space storms: Critical issues and open disputes Intense space storms: Critical issues and open J. Geophys. Res. (Space)108A51208. , Thorne , Baumjohann Daglis, IA. , Thorne, RM. , Baumjohann, W. Orsini, S. The terrestrial ring current: Origin, formation, and decay The terrestrial ring current: Origin, formation, and decay. Rev. Geophys.37407-438. Dessler, AJ. Parker, EN. Hydromagnetic theory of geomagnetic storms Hydromagnetic theory of geomagnetic storms. J. Geophys. Res.64122239-2252. , Tokumaru , Hayashi , Satonaka Fujiki, K. , Tokumaru, M. , Hayashi, K. , Satonaka, D. Hakamada, K. Long-term Trend of Solar Coronal Hole Distribution from 1975 to 2014 Long-term trend of solar coronal hole distribution from 1975 to 2014. Astrophys. J. Lett.8272L41. Gannon, JL. Love, JJ. USGS 1-min Dst index USGS 1-min Dst J. Atm. Sol.-Terr. Phys.73323-334. Gibson, SE. , Kozyra, JU. , de Toma, G. , Emery, BA. , Onsager, T. Thompson, BJ. If the Sun is so quiet, why is the Earth ringing? A comparison of two solar minimum intervals If the Sun is so quiet, why is the Earth ringing? A comparison of two solar minimum intervals. J. Geophys. Res. (Space)114A13A09105. Gnevyshev, MN. Essential features of the 11-year solar cycle Essential features of the 11-year solar cycle. Sol. Phys.51175-183. Gonzalez, WD. , Joselyn, JA. , Kamide, Y. , Kroehl, HW. , Rostoker, G. , Tsurutani, BT. Vasyliunas, VM. What is a geomagnetic storm? What is a geomagnetic J. Geophys. Res. (Space)995771-5792. , Nunes , Yashiro Gopalswamy, N. , Nunes, S. , Yashiro, S. Howard, R. Variability of solar eruptions during cycle 23 Variability of solar eruptions during cycle 23. Adv. Space Res.342391-396. Solar Variability and Climate Change Gosling, JT. The Solar Flare Myth The solar flare myth. J. Geophys. Res. (Space)9818937. , Asbridge , Bame Gosling, JT. , Asbridge, JR. , Bame, SJ. Feldman, WC. Solar Wind Speed Variations: 1962–1974 Solar wind speed variations: 1962–1974. J. Geophys. Res.815061–5070. Gosling, JT. Pizzo, VJ. Formation and Evolution of Corotating Interaction Regions and their Three Dimensional Structure Formation and Evolution of Corotating Interaction Regions and their Three Dimensional Space Sci. Rev.8921-52. Gosling, JT. Riley, P. The acceleration of slow coronal mass ejections in the high-speed solar wind The acceleration of slow coronal mass ejections in the high-speed solar wind. Geophys. Res. Lett.23212867-2870. Häkkinen, LVT. , Pulkkinen, TI. , Pirjola, RJ. , Nevanlinna, H. , Tanskanen, EI. Turner, NE. Seasonal and diurnal variation of geomagnetic activity: Revised Dst versus external drivers Seasonal and diurnal variation of geomagnetic activity: Revised Dst versus external drivers. J. Geophys. Res. (Space)108A21060. , Asikainen Hamada, A. , Asikainen, T. Mursula, K. A Uniform Series of Low-Latitude Coronal Holes in 1973–2018 A Uniform Series of Low-Latitude Coronal Holes in Sol. Phys.296240. Harvey, JW. Sheeley, NR. Coronal Holes and Solar Magnetic Fields Coronal holes and solar magnetic fields. Space Sci. Rev.23139–158. Harvey, KL. Recely, F. Polar Coronal Holes During Cycles 22 and 23 Polar Coronal Holes During Cycles 22 and 23. Sol. Phys.21131-52. , Wilcox Hoeksema, JT. , Wilcox, JM. Scherrer, PH. The Structure of the Heliospheric Current Sheet The structure of the heliospheric current sheet. J. Geophys. Res.889910. Karinen, A. Mursula, K. A new reconstruction of the Dst index for 1932–2002 A new reconstruction of the Dst index for 1932–2002. Ann. Geophys.23475-485. , Mursula , Takalo Karinen, A. , Mursula, K. , Takalo, J. Ulich, T. An erroneous Dst Index in 1971 An erroneous Dst Index in 1971. H. Sawaya-Lacoste (), Solspa 2001, Proceedings of the Second Solar Cycle and Space Weather Euroconference Solspa 2001, proceedings of the second solar cycle and space weather euroconference ( 477,  443-446). , Koskinen Kilpua, E. , Koskinen, HEJ. Pulkkinen, TI. Coronal mass ejections and their sheath regions in interplanetary space Coronal mass ejections and their sheath regions in interplanetary space. Liv. Rev. Solar Phys.1415. , Pesnell , Young Hess Webber Kirk, MS. , Pesnell, WD. , Young, CA. Hess Webber, SA. Automated detection of EUV Polar Coronal Holes during Solar Cycle 23 Automated detection of EUV Polar Coronal Holes during Solar Cycle 23. Sol. Phys.25799-112. Klein, LW. Burlaga, LF. Interplanetary magnetic clouds at 1 AU Interplanetary magnetic clouds at 1 AU. J. Geophys. Res. (Space)87A2613-624. , Timothy Krieger, AS. , Timothy, AF. Roelof, EC. A Coronal Hole and Its Identification as the Source of a High Velocity Solar Wind Stream A coronal hole and its identification as the source of a high velocity solar wind stream. Sol. Phys.29505–525. Love, JJ. Gannon, JL. Revised Dst and the epicycles of magnetic disturbance: 1958–2007 Revised Dst and the epicycles of magnetic disturbance: Ann. Geophys.273101-3131. , Holappa Mursula, K. , Holappa, L. Karinen, A. Correct normalization of the Dst index Correct normalization of the Dst index. Astrophys. Space Sci. Transact.441-45. , Holappa Mursula, K. , Holappa, L. Karinen, A. Uneven weighting of stations in the Dst index Uneven weighting of stations in the Dst index. J. Atm. Sol.-Terr. Phys.73316-322. , Holappa Mursula, K. , Holappa, L. Lukianova, R. Seasonal solar wind speeds for the last 100 years: Unique coronal hole structures during the peak and demise of the Grand Modern Maximum Seasonal solar wind speeds for the last 100 years: Unique coronal hole structures during the peak and demise of the Grand Modern Geophys. Res. Lett.44130-36. , Lukianova Mursula, K. , Lukianova, R. Holappa, L. Occurrence of High-speed Solar Wind Streams over the Grand Modern Maximum Occurrence of High-speed Solar Wind Streams over the Grand Modern Maximum. Astrophys. J.80130. Newell, PT. Gjerloev, JW. SuperMAG-based partial ring current indices SuperMAG-based partial ring current indices. J. Geophys. Res. (Space)117A5A05215. , Iyemori , Sugiura Nose, M. , Iyemori, T. , Sugiura, M. Kamei, T. Geomagnetic Dst index. Geomagnetic Dst index. Web page () of the World Data Center for Geomagnetism (WDC-C2), Kyoto, Japan, accessed April 01, O'Brien, T. McPherron, R. An empirical phase space analysis of ring current dynamics: Solar wind control of injection and decay An empirical phase space analysis of ring current dynamics: Solar wind control of injection and J. Geophys. Res.105A47707-7719. Petrie, GJD. Solar Magnetism in the Polar Regions Solar Magnetism in the Polar Regions. Liv. Rev. Solar Phys.1215. IG. Richardson IG. Richardson Richardson, IG. Solar wind stream interaction regions throughout the heliosphere Solar wind stream interaction regions throughout the Liv. Rev. Solar Phys.1511. IG. Richardson IG. Richardson Richardson, IG. Cane, HV. Near-earth solar wind flows and related geomagnetic activity during more than four solar cycles (1963–2011) Near-earth solar wind flows and related geomagnetic activity during more than four solar cycles (1963–2011). J. Space Weather Space Clim.2A02. IG. Richardson , Cane IG. Richardson Richardson, IG. , Cane, HV. Cliver, EW. Sources of geomagnetic activity during nearly three solar cycles (1972–2000) Sources of geomagnetic activity during nearly three solar cycles (1972–2000). J. Geophys. Res.1071187. IG. Richardson , Cliver IG. Richardson Richardson, IG. , Cliver, EW. Cane, HV. Sources of geomagnetic activity over the solar cycle: Relative importance of coronal mass ejections, high-speed streams, and slow solar wind Sources of geomagnetic activity over the solar cycle: Relative importance of coronal mass ejections, high-speed streams, and slow solar J. Geophys. Res.105A818,203-18,213. JD. Richardson JD. Richardson Richardson, JD. Paularena, KI. Streamer Belt Structure at Solar Minima Streamer belt structure at solar minima. Geophys. Res. Lett.241435–1438. , Berghmans van der Linden Robbrecht, E. , Berghmans, D. van der Linden, RAM. Objective CME detection over the solar cycle: A first attempt Objective CME detection over the solar cycle: A first Adv. Space Res.383475-479. Rosenberg, RL. Coleman, PJ., Jr. Heliographic latitude dependence of the dominant polarity of the interplanetary magnetic field. Heliographic latitude dependence of the dominant polarity of the interplanetary magnetic J. Geophys. Res.745611-5622. Russell, CT. McPherron, RL. Semiannual variation of geomagnetic activity Semiannual variation of geomagnetic activity. J. Geophys. Res.7892–108. Savitzky, A. Golay, MJE. Smoothing and differentiation of data by simplified least squares procedures Smoothing and differentiation of data by simplified least squares procedures. Anal. Chem.361627-1639. , Wilcox Schatten, KH. , Wilcox, JM. Ness, NF. A Model of Interplanetary and Coronal Magnetic Fields A model of interplanetary and coronal magnetic fields. Sol. Phys.6442–455. Sckopke, N. A general relation between the energy of trapped particles and the disturbance field near the Earth A general relation between the energy of trapped particles and the disturbance field near the J. Geophys. Res.71133125-3130. SILSO World Data Center SILSO World Data Center SILSO World Data Center. The International Sunspot Number The International Sunspot Number. International Sunspot Number Monthly Bulletin and online catalogue. Smith, EJ. The global heliospheric magnetic field The global heliospheric magnetic field. A. Balogh, LJ. Lanzerotti ST. Suess (), The Heliosphere through the Solar Activity Cycle The heliosphere through the solar activity cycle ( 79). Smith, EJ. Balogh, A. Decrease in heliospheric magnetic flux in this solar minimum: Recent Ulysses magnetic field observations Decrease in heliospheric magnetic flux in this solar minimum: Recent Ulysses magnetic field observations. Geophys. Res. Lett.35L22103. , Usoskin , Kromer , Schüssler Solanki, SK. , Usoskin, IG. , Kromer, B. , Schüssler, M. Beer, J. Unusual activity of the Sun during recent decades compared to the previous 11,000 years Unusual activity of the Sun during recent decades compared to the previous 11,000 years. Storini, M. , Bazilevskaya, GA. , Fluckiger, EO. , Krainev, MB. , Makhmutov, VS. Sladkova, AI. The GNEVYSHEV gap: A review for space weather The gnevyshev gap: A review for space weather. Adv. Space Res.31895-900. Subramanian, P. Dere, KP. Source Regions of Coronal Mass Ejections Source Regions of Coronal Mass Ejections. Astrophys. J.5611372-395. Sugiura, M. Hourly values of equatorial Dst for the IGY Hourly values of equatorial Dst for the IGY. Ann. Int. Geophys. Year359. Sugiura, M. IAGA Resolution 2 IAGA Resolution 2. IAGA Bulletin 27 IAGA Bulletin 27 ( 123). Madrid, Spain. Sugiura, M. Kamei, T. Equatorial Dst index 1957–1986 Equatorial Dst index 1957–1986. A. Berthelier M. Menvielle (), IAGA Bulletin 40. IAGA Bulletin 40. Saint-Maur-des-Fossés, FranceISGI Publ. Office. Takalo, J. Mursula, K. A model for the diurnal universal time variation of the Dst index A model for the diurnal universal time variation of the Dst J. Geophys. Res. (Space)106A610905-10914. Takalo, J. Mursula, K. On the diurnal UT variation of the Dst index On the diurnal UT variation of the Dst index. H. Sawaya-Lacoste (), Solspa 2001, Proceedings of the Second Solar Cycle and Space Weather Euroconference Solspa 2001, proceedings of the second solar cycle and space weather euroconference ( 477,  483-486). Tsurutani, BT. , Gonzalez, WD. , Gonzalez, ALC. , Guarnieri, FL. , Gopalswamy, N. , Grande, M. Vasyliunas, V. Corotating solar wind streams and recurrent geomagnetic activity: A review Corotating solar wind streams and recurrent geomagnetic activity: A review. J. Geophys. Res. (Space)111A7. Virtanen, II. Mursula, K. Asymmetry of solar polar fields and the southward shift of HCS observed by Ulysses Asymmetry of solar polar fields and the southward shift of HCS observed by Ulysses. J. Geophys. Res.115A14A09110. Virtanen, II. Mursula, K. North-South Asymmetric Solar Cycle Evolution: Signatures in the Photosphere and Consequences in the Corona North-South Asymmetric Solar Cycle Evolution: Signatures in the Photosphere and Consequences in the Corona. Astrophys. J.781299. , Nash Wang, YM. , Nash, AG. Sheeley, NR., Jr. Evolution of the Sun's Polar Fields During Sunspot Cycle 21 - Poleward Surges and Long-Term Behavior Evolution of the Sun's polar fields during sunspot cycle 21 - poleward surges and long-term Astrophys. J.347529–539. , Robbrecht Wang, YM. , Robbrecht, E. Sheeley, NR., Jr. On the Weakening of the Polar Magnetic Fields during Solar Cycle 23 On the Weakening of the Polar Magnetic Fields during Solar Cycle 23. Astrophys. J.7071372-1386. Wang, YM. Sheeley, J., N. R. Solar Wind Speed and Coronal Flux-Tube Expansion Solar Wind Speed and Coronal Flux-Tube Expansion. Astrophys. J.355726. On Dst index (description in the IAGA Bulletin No 40). On Dst index (description in the IAGA Bulletin no 40). Web page () of the World Data Center for Geomagnetism (WDC-C2), Kyoto, Japan, accessed April 01, Webb, DF. Howard, RA. The solar cycle variation of coronal mass ejections and the solar wind mass flux The solar cycle variation of coronal mass ejections and the solar wind mass flux. J. Geophys. Res. (Space)994201-4220. Webb, DF. Howard, TA. Coronal Mass Ejections: Observations Coronal mass ejections: Observations. Liv. Rev. Solar Phys.913. , Mursula , Holappa , Veselovsky Yakovchouk, OS. , Mursula, K. , Holappa, L. , Veselovsky, IS. Karinen, A. Average properties of geomagnetic storms in 1932–2009 Average properties of geomagnetic storms in 1932–2009. J. Geophys. Res. (Space)117A03201. Zhou, X. Smith, EJ. Solar cycle variations of heliospheric magnetic flux Solar cycle variations of heliospheric magnetic flux. J. Geophys. Res.114A133106. Altschuler, MD. Newkirk, G. Magnetic Fields and the Structure of the Solar Corona. I: Methods of Calculating Coronal Fields Magnetic Fields and the Structure of the Solar Corona. I: Methods of Calculating Coronal Sol. Phys.9131-149. , Maliniemi Asikainen, T. , Maliniemi, V. Mursula, K. Modeling the contributions of ring, tail, and magnetopause currents to the corrected Dst index Modeling the contributions of ring, tail, and magnetopause currents to the corrected Dst J. Geophys. Res.115A1412203. Balogh, A. , Bothmer, V. , Crooker, NU. , Forsyth, RJ. , Gloeckler, G. , Hewish, A. Mikić, Z. The Solar Origin of Corotating Interaction Regions and Their Formation in the Inner Heliosphere The solar origin of corotating interaction regions and their formation in the inner heliosphere. Space Sci. Rev.89141-178. , Sittler , Mariani Burlaga, L. , Sittler, E. , Mariani, F. Schwenn, R. Magnetic loop behind an interplanetary shock: Voyager, Helios, and IMP 8 observations Magnetic loop behind an interplanetary shock: Voyager, Helios, and IMP 8 observations. J. Geophys. Res. (Space)86A86673–6684. , McPherron Burton, RK. , McPherron, RL. Russell, CT. An Empirical Relationship Between Interplanetary Conditions and Dst An empirical relationship between interplanetary conditions and Dst. J. Geophys. Res.804204–4214. St. Cyr St. Cyr Cremades, H. St. Cyr, OC. Coronal mass ejections: Solar cycle aspects Coronal mass ejections: Solar cycle aspects. Adv. Space Res.4071042-1048. Crooker, NU. Cliver, EW. Postmodern View of M-regions Postmodern view of J. Geophys. Res.99A1223383–23390. Daglis, IA. Ring Current Dynamics Ring Current Space Sci. Rev.1241-4183-202. Daglis, IA. , Kozyra, JU. , Kamide, Y. , Vassiliadis, D. , Sharma, AS. , Liemohn, MW. Lu, G. Intense space storms: Critical issues and open disputes Intense space storms: Critical issues and open J. Geophys. Res. (Space)108A51208. , Thorne , Baumjohann Daglis, IA. , Thorne, RM. , Baumjohann, W. Orsini, S. The terrestrial ring current: Origin, formation, and decay The terrestrial ring current: Origin, formation, and decay. Rev. Geophys.37407-438. Dessler, AJ. Parker, EN. Hydromagnetic theory of geomagnetic storms Hydromagnetic theory of geomagnetic storms. J. Geophys. Res.64122239-2252. , Tokumaru , Hayashi , Satonaka Fujiki, K. , Tokumaru, M. , Hayashi, K. , Satonaka, D. Hakamada, K. Long-term Trend of Solar Coronal Hole Distribution from 1975 to 2014 Long-term trend of solar coronal hole distribution from 1975 to 2014. Astrophys. J. Lett.8272L41. Gannon, JL. Love, JJ. USGS 1-min Dst index USGS 1-min Dst J. Atm. Sol.-Terr. Phys.73323-334. Gibson, SE. , Kozyra, JU. , de Toma, G. , Emery, BA. , Onsager, T. Thompson, BJ. If the Sun is so quiet, why is the Earth ringing? A comparison of two solar minimum intervals If the Sun is so quiet, why is the Earth ringing? A comparison of two solar minimum intervals. J. Geophys. Res. (Space)114A13A09105. Gnevyshev, MN. Essential features of the 11-year solar cycle Essential features of the 11-year solar cycle. Sol. Phys.51175-183. Gonzalez, WD. , Joselyn, JA. , Kamide, Y. , Kroehl, HW. , Rostoker, G. , Tsurutani, BT. Vasyliunas, VM. What is a geomagnetic storm? What is a geomagnetic J. Geophys. Res. (Space)995771-5792. , Nunes , Yashiro Gopalswamy, N. , Nunes, S. , Yashiro, S. Howard, R. Variability of solar eruptions during cycle 23 Variability of solar eruptions during cycle 23. Adv. Space Res.342391-396. Solar Variability and Climate Change Gosling, JT. The Solar Flare Myth The solar flare myth. J. Geophys. Res. (Space)9818937. , Asbridge , Bame Gosling, JT. , Asbridge, JR. , Bame, SJ. Feldman, WC. Solar Wind Speed Variations: 1962–1974 Solar wind speed variations: 1962–1974. J. Geophys. Res.815061–5070. Gosling, JT. Pizzo, VJ. Formation and Evolution of Corotating Interaction Regions and their Three Dimensional Structure Formation and Evolution of Corotating Interaction Regions and their Three Dimensional Space Sci. Rev.8921-52. Gosling, JT. Riley, P. The acceleration of slow coronal mass ejections in the high-speed solar wind The acceleration of slow coronal mass ejections in the high-speed solar wind. Geophys. Res. Lett.23212867-2870. Häkkinen, LVT. , Pulkkinen, TI. , Pirjola, RJ. , Nevanlinna, H. , Tanskanen, EI. Turner, NE. Seasonal and diurnal variation of geomagnetic activity: Revised Dst versus external drivers Seasonal and diurnal variation of geomagnetic activity: Revised Dst versus external drivers. J. Geophys. Res. (Space)108A21060. , Asikainen Hamada, A. , Asikainen, T. Mursula, K. A Uniform Series of Low-Latitude Coronal Holes in 1973–2018 A Uniform Series of Low-Latitude Coronal Holes in Sol. Phys.296240. Harvey, JW. Sheeley, NR. Coronal Holes and Solar Magnetic Fields Coronal holes and solar magnetic fields. Space Sci. Rev.23139–158. Harvey, KL. Recely, F. Polar Coronal Holes During Cycles 22 and 23 Polar Coronal Holes During Cycles 22 and 23. Sol. Phys.21131-52. , Wilcox Hoeksema, JT. , Wilcox, JM. Scherrer, PH. The Structure of the Heliospheric Current Sheet The structure of the heliospheric current sheet. J. Geophys. Res.889910. Karinen, A. Mursula, K. A new reconstruction of the Dst index for 1932–2002 A new reconstruction of the Dst index for 1932–2002. Ann. Geophys.23475-485. , Mursula , Takalo Karinen, A. , Mursula, K. , Takalo, J. Ulich, T. An erroneous Dst Index in 1971 An erroneous Dst Index in 1971. H. Sawaya-Lacoste (), Solspa 2001, Proceedings of the Second Solar Cycle and Space Weather Euroconference Solspa 2001, proceedings of the second solar cycle and space weather euroconference ( 477,  443-446). , Koskinen Kilpua, E. , Koskinen, HEJ. Pulkkinen, TI. Coronal mass ejections and their sheath regions in interplanetary space Coronal mass ejections and their sheath regions in interplanetary space. Liv. Rev. Solar Phys.1415. , Pesnell , Young Hess Webber Kirk, MS. , Pesnell, WD. , Young, CA. Hess Webber, SA. Automated detection of EUV Polar Coronal Holes during Solar Cycle 23 Automated detection of EUV Polar Coronal Holes during Solar Cycle 23. Sol. Phys.25799-112. Klein, LW. Burlaga, LF. Interplanetary magnetic clouds at 1 AU Interplanetary magnetic clouds at 1 AU. J. Geophys. Res. (Space)87A2613-624. , Timothy Krieger, AS. , Timothy, AF. Roelof, EC. A Coronal Hole and Its Identification as the Source of a High Velocity Solar Wind Stream A coronal hole and its identification as the source of a high velocity solar wind stream. Sol. Phys.29505–525. Love, JJ. Gannon, JL. Revised Dst and the epicycles of magnetic disturbance: 1958–2007 Revised Dst and the epicycles of magnetic disturbance: Ann. Geophys.273101-3131. , Holappa Mursula, K. , Holappa, L. Karinen, A. Correct normalization of the Dst index Correct normalization of the Dst index. Astrophys. Space Sci. Transact.441-45. , Holappa Mursula, K. , Holappa, L. Karinen, A. Uneven weighting of stations in the Dst index Uneven weighting of stations in the Dst index. J. Atm. Sol.-Terr. Phys.73316-322. , Holappa Mursula, K. , Holappa, L. Lukianova, R. Seasonal solar wind speeds for the last 100 years: Unique coronal hole structures during the peak and demise of the Grand Modern Maximum Seasonal solar wind speeds for the last 100 years: Unique coronal hole structures during the peak and demise of the Grand Modern Geophys. Res. Lett.44130-36. , Lukianova Mursula, K. , Lukianova, R. Holappa, L. Occurrence of High-speed Solar Wind Streams over the Grand Modern Maximum Occurrence of High-speed Solar Wind Streams over the Grand Modern Maximum. Astrophys. J.80130. Newell, PT. Gjerloev, JW. SuperMAG-based partial ring current indices SuperMAG-based partial ring current indices. J. Geophys. Res. (Space)117A5A05215. , Iyemori , Sugiura Nose, M. , Iyemori, T. , Sugiura, M. Kamei, T. Geomagnetic Dst index. Geomagnetic Dst index. Web page () of the World Data Center for Geomagnetism (WDC-C2), Kyoto, Japan, accessed April 01, O'Brien, T. McPherron, R. An empirical phase space analysis of ring current dynamics: Solar wind control of injection and decay An empirical phase space analysis of ring current dynamics: Solar wind control of injection and J. Geophys. Res.105A47707-7719. Petrie, GJD. Solar Magnetism in the Polar Regions Solar Magnetism in the Polar Regions. Liv. Rev. Solar Phys.1215. IG. Richardson IG. Richardson Richardson, IG. Solar wind stream interaction regions throughout the heliosphere Solar wind stream interaction regions throughout the Liv. Rev. Solar Phys.1511. IG. Richardson IG. Richardson Richardson, IG. Cane, HV. Near-earth solar wind flows and related geomagnetic activity during more than four solar cycles (1963–2011) Near-earth solar wind flows and related geomagnetic activity during more than four solar cycles (1963–2011). J. Space Weather Space Clim.2A02. IG. Richardson , Cane IG. Richardson Richardson, IG. , Cane, HV. Cliver, EW. Sources of geomagnetic activity during nearly three solar cycles (1972–2000) Sources of geomagnetic activity during nearly three solar cycles (1972–2000). J. Geophys. Res.1071187. IG. Richardson , Cliver IG. Richardson Richardson, IG. , Cliver, EW. Cane, HV. Sources of geomagnetic activity over the solar cycle: Relative importance of coronal mass ejections, high-speed streams, and slow solar wind Sources of geomagnetic activity over the solar cycle: Relative importance of coronal mass ejections, high-speed streams, and slow solar J. Geophys. Res.105A818,203-18,213. JD. Richardson JD. Richardson Richardson, JD. Paularena, KI. Streamer Belt Structure at Solar Minima Streamer belt structure at solar minima. Geophys. Res. Lett.241435–1438. , Berghmans van der Linden Robbrecht, E. , Berghmans, D. van der Linden, RAM. Objective CME detection over the solar cycle: A first attempt Objective CME detection over the solar cycle: A first Adv. Space Res.383475-479. Rosenberg, RL. Coleman, PJ., Jr. Heliographic latitude dependence of the dominant polarity of the interplanetary magnetic field. Heliographic latitude dependence of the dominant polarity of the interplanetary magnetic J. Geophys. Res.745611-5622. Russell, CT. McPherron, RL. Semiannual variation of geomagnetic activity Semiannual variation of geomagnetic activity. J. Geophys. Res.7892–108. Savitzky, A. Golay, MJE. Smoothing and differentiation of data by simplified least squares procedures Smoothing and differentiation of data by simplified least squares procedures. Anal. Chem.361627-1639. , Wilcox Schatten, KH. , Wilcox, JM. Ness, NF. A Model of Interplanetary and Coronal Magnetic Fields A model of interplanetary and coronal magnetic fields. Sol. Phys.6442–455. Sckopke, N. A general relation between the energy of trapped particles and the disturbance field near the Earth A general relation between the energy of trapped particles and the disturbance field near the J. Geophys. Res.71133125-3130. SILSO World Data Center SILSO World Data Center SILSO World Data Center. The International Sunspot Number The International Sunspot Number. International Sunspot Number Monthly Bulletin and online catalogue. Smith, EJ. The global heliospheric magnetic field The global heliospheric magnetic field. A. Balogh, LJ. Lanzerotti ST. Suess (), The Heliosphere through the Solar Activity Cycle The heliosphere through the solar activity cycle ( 79). Smith, EJ. Balogh, A. Decrease in heliospheric magnetic flux in this solar minimum: Recent Ulysses magnetic field observations Decrease in heliospheric magnetic flux in this solar minimum: Recent Ulysses magnetic field observations. Geophys. Res. Lett.35L22103. , Usoskin , Kromer , Schüssler Solanki, SK. , Usoskin, IG. , Kromer, B. , Schüssler, M. Beer, J. Unusual activity of the Sun during recent decades compared to the previous 11,000 years Unusual activity of the Sun during recent decades compared to the previous 11,000 years. Storini, M. , Bazilevskaya, GA. , Fluckiger, EO. , Krainev, MB. , Makhmutov, VS. Sladkova, AI. The GNEVYSHEV gap: A review for space weather The gnevyshev gap: A review for space weather. Adv. Space Res.31895-900. Subramanian, P. Dere, KP. Source Regions of Coronal Mass Ejections Source Regions of Coronal Mass Ejections. Astrophys. J.5611372-395. Sugiura, M. Hourly values of equatorial Dst for the IGY Hourly values of equatorial Dst for the IGY. Ann. Int. Geophys. Year359. Sugiura, M. IAGA Resolution 2 IAGA Resolution 2. IAGA Bulletin 27 IAGA Bulletin 27 ( 123). Madrid, Spain. Sugiura, M. Kamei, T. Equatorial Dst index 1957–1986 Equatorial Dst index 1957–1986. A. Berthelier M. Menvielle (), IAGA Bulletin 40. IAGA Bulletin 40. Saint-Maur-des-Fossés, FranceISGI Publ. Office. Takalo, J. Mursula, K. A model for the diurnal universal time variation of the Dst index A model for the diurnal universal time variation of the Dst J. Geophys. Res. (Space)106A610905-10914. Takalo, J. Mursula, K. On the diurnal UT variation of the Dst index On the diurnal UT variation of the Dst index. H. Sawaya-Lacoste (), Solspa 2001, Proceedings of the Second Solar Cycle and Space Weather Euroconference Solspa 2001, proceedings of the second solar cycle and space weather euroconference ( 477,  483-486). Tsurutani, BT. , Gonzalez, WD. , Gonzalez, ALC. , Guarnieri, FL. , Gopalswamy, N. , Grande, M. Vasyliunas, V. Corotating solar wind streams and recurrent geomagnetic activity: A review Corotating solar wind streams and recurrent geomagnetic activity: A review. J. Geophys. Res. (Space)111A7. Virtanen, II. Mursula, K. Asymmetry of solar polar fields and the southward shift of HCS observed by Ulysses Asymmetry of solar polar fields and the southward shift of HCS observed by Ulysses. J. Geophys. Res.115A14A09110. Virtanen, II. Mursula, K. North-South Asymmetric Solar Cycle Evolution: Signatures in the Photosphere and Consequences in the Corona North-South Asymmetric Solar Cycle Evolution: Signatures in the Photosphere and Consequences in the Corona. Astrophys. J.781299. , Nash Wang, YM. , Nash, AG. Sheeley, NR., Jr. Evolution of the Sun's Polar Fields During Sunspot Cycle 21 - Poleward Surges and Long-Term Behavior Evolution of the Sun's polar fields during sunspot cycle 21 - poleward surges and long-term Astrophys. J.347529–539. , Robbrecht Wang, YM. , Robbrecht, E. Sheeley, NR., Jr. On the Weakening of the Polar Magnetic Fields during Solar Cycle 23 On the Weakening of the Polar Magnetic Fields during Solar Cycle 23. Astrophys. J.7071372-1386. Wang, YM. Sheeley, J., N. R. Solar Wind Speed and Coronal Flux-Tube Expansion Solar Wind Speed and Coronal Flux-Tube Expansion. Astrophys. J.355726. On Dst index (description in the IAGA Bulletin No 40). On Dst index (description in the IAGA Bulletin no 40). Web page () of the World Data Center for Geomagnetism (WDC-C2), Kyoto, Japan, accessed April 01, Webb, DF. Howard, RA. The solar cycle variation of coronal mass ejections and the solar wind mass flux The solar cycle variation of coronal mass ejections and the solar wind mass flux. J. Geophys. Res. (Space)994201-4220. Webb, DF. Howard, TA. Coronal Mass Ejections: Observations Coronal mass ejections: Observations. Liv. Rev. Solar Phys.913. , Mursula , Holappa , Veselovsky Yakovchouk, OS. , Mursula, K. , Holappa, L. , Veselovsky, IS. Karinen, A. Average properties of geomagnetic storms in 1932–2009 Average properties of geomagnetic storms in 1932–2009. J. Geophys. Res. (Space)117A03201. Zhou, X. Smith, EJ. Solar cycle variations of heliospheric magnetic flux Solar cycle variations of heliospheric magnetic flux. J. Geophys. Res.114A133106. More Information and Advice: Math coded inside display math mode \[ ...\] will not be numbered, e.g.,: \[ x^2=y^2 + z^2\] Math coded inside \begin{equation} and \end{equation} be automatically numbered, e.g.,: \begin{equation} x^2=y^2 + z^2 \end{equation} \begin{eqnarray} x_{1} & = & (x - x_{0}) \cos \Theta \nonumber \\ && + (y - y_{0}) \sin \Theta \nonumber \\ y_{1} & = & -(x - x_{0}) \sin \Theta \nonumber \\ && + (y - y_{0}) \cos \Theta. \end{eqnarray}
2022 XXX 1st December Hot & Cold QCD plans for 2023 Long Range Plan for Nuclear Science The ALICE-USA Collaboration††thanks: See Appendix A for the list of collaboration members ALICE-USA Collaboration ## 1 Executive Summary The ALICE experiment was built to study many-body Quantum Chromo-Dynamics (QCD) at high temperature and effectively zero baryon density, using relativistic heavy-ion collisions at the Large Hadron Collider (LHC). These collisions form the Quark Gluon Plasma (QGP), a state of matter where quarks and gluons are no longer confined inside hadrons. The ALICE program centers around the key questions related to QGP phenomena. These include the macroscopic and microscopic properties of the QGP, and the details of the QGP phase transition to hadrons, that is believed to have taken place in the early Universe. At the same time, ALICE’s versatile setup allows for the study of pp collisions, p–Pb collisions, and ultra-peripheral collisions. The associated studies can provide some of the most stringent tests of QCD and Beyond Standard Model searches. They serve as deep probes of the properties of cold nuclear matter, and allow for investigations of stellar and interstellar phenomena. The ALICE-USA collaboration consists of 13 institutes, representing about 6% of the total ALICE authorship. It has, and continues to play, an essential role in all areas of ALICE leadership: physics, instrumentation, and management. ALICE-USA has been involved in $\sim$25% of ALICE’s $\sim$400 ALICE papers since 2009, including many of ALICE’s most cited and high profile publications. We, the ALICE-USA collaboration, will provide two recommendations for the U.S. 2023 Long Range Plan for Nuclear Science period, and beyond. These recommendations are essential for maintaining the continued success and development of the U.S. Nuclear Science community for both Hot and Cold QCD. The recommendations will be accompanied by details of the vital ALICE-USA scientific priorities. We will also describe the broader impact and support needed for U.S. involvement in ALICE. ### Recommendation 1: Continue and broaden the contribution from U.S. institutes for the ALICE program in Runs 3&4. ALICE has just completed a number of major upgrades. ALICE-USA has made vital contributions to the new Inner Tracking System (ITS) and Time Projection Chamber (TPC) readout. ALICE-USA will now utilize these upgrades for a comprehensive physics program. This ‘ALICE 2’ phase also provides a unique opportunity for Hot and Cold QCD studies between the expected times when RHIC discontinues taking data in 2025, and when the EIC begins taking data around 2035. The new detector setup began taking data in Run 3 (2022-2025), and will continue to operate in Run 4 (2029-2032). During these periods, ALICE will collect data from collisions of: Pb–Pb at $\sqrt{s_{\mathrm{NN}}}=5.36$ TeV, pp at $\sqrt{s}=13.6$ TeV, pp at the same energy as Pb–Pb, p–Pb at $\sqrt{s_{\mathrm{NN}}}=8$ TeV, p–O at $\sqrt{s_{\mathrm{NN}}}=9.9$ TeV, and O–O at $\sqrt{s_{\mathrm{NN}}}=7$ TeV. The new data set of Pb–Pb collisions ($\sim$13 nb-1) represents an increase of two orders of magnitude over Runs 1&2\. These increased capabilities in Runs 3&4 will allow for: * • The elucidation of the microscopic parton dynamics underlying QGP properties using hard processes of lower cross section, such as heavy-flavor jets, over a large range of transverse momenta; * • The characterization of the macroscopic QGP properties with extended precision, including the exploration of unknown dynamical QGP transport parameters at LHC energies, such as the the baryon diffusion coefficient; * • Deeper studies of the hadronization of heavy-flavor baryons and mesons produced in high temperature QCD matter; * • Multiple observables with an increased precision to contribute to the development of a unified picture of particle production and QCD dynamics from small (pp) to large (nucleus–nucleus); * • Unique explorations of parton densities in protons and nuclei in a broad ($x$, $Q^{2}$) kinematic range, reaching to $x\sim 10^{-6}$. In relation to the last point, ALICE-USA is one of the key proponents of the Forward Calorimeter (FoCal). This a new detector that will collect data in Run 4. The development, installation, and operation of the FoCal will occur during the LRP duration, and prior to the start of the EIC data taking period. The FoCal is designed to provide unique capabilities to probe the structure of nucleons and nuclei in an unexplored regime of $x$ and $Q^{2}$. This kinematic region, of $x\sim 10^{-6}$, will probe the regime where gluon saturated matter is expected to be dominant. Such a small-$x$ region is not accessible at the EIC. Global fits of this lower $x$ data, with new high precision EIC data, will make an unprecedented step to understanding the full evolution of Parton Distributions Functions (PDFs) in protons and nuclei. The R&D and design phase of the EIC detectors have well defined synergies with the FoCal instrumentation work. We also stress the additional benefits of U.S. involvement in ALICE’s instrumentation R&D for the ITS3 sensors that will be part of the Run 4 setup. These have various applications for the EIC detector projects, as recognized in multiple forums. Support for both of these endeavors is of vital importance. Beyond the scientific and technological implications discussed, the active participation of U.S. institutions in the ALICE collaboration in the coming years would greatly benefit the training of the next generation of scientists. It is vital that PhD students, postdoctoral researchers, and early career faculty, have the opportunity to develop their data analysis, modeling, and interpretation expertise using the largest and most complex datasets of QCD phenomena ever garnered. These younger members of the community will also have the opportunity to participate in the full cycle of detector operation and development in the 2020s. Such experience is essential for maintaining the U.S. as a leader in nuclear science. Therefore, continued support for ALICE research is crucial. We urge a modest expansion of the ALICE efforts to allow groups currently focused at RHIC to also participate prior to EIC operations. ### Recommendation 2: Begin participation in ALICE-3 R&D and construction. ALICE-USA fully supports the ALICE 3 detector, a next generation detector that is designed to operate in Runs 5&6 (2035 and beyond). It is currently in the R&D phase, and construction of the ALICE 3 detector is expected to begin 2028. It represents a unique direction to pursue Hot QCD studies at the highest temperatures possible into and beyond the 2030s. ALICE 3 will have the largest acceptance and most precise tracking ever achieved for a heavy-ion experiment. It is expected to record $\sim$20 times more heavy-ion data compared to Runs 3&4\. The general purpose detector design of ALICE 3 will provide many physics opportunities, which are expected to lead to new discoveries. It will also provide a level of data quality required to transform our understanding of the QGP, in terms of first principles QCD calculations. The key topics unique to ALICE 3, and under the umbrella of current ALICE-USA interests, include: * • Jet hadrochemistry over an unprecedented kinematic range for transverse momentum and rapidity, to provide new ways to investigate QGP jet energy loss and the microscopic structure of the QGP; * • Measurements of multi-heavy-flavor and exotic charm states to provide unique information on hadronization and hadronic interactions at high temperatures; * • Beyond Standard Model physics with searches for axion-like particles in ultra- peripheral collisions, which will leverage the high collision rates and large acceptance. The design of ALICE 3 will also enable measurements of high-precision, multi- differential of electromagnetic radiation from the QGP to probe its early evolution and the restoration of chiral symmetry through the coupling of vector and axial-vector mesons. Measurements of net-quantum number fluctuations over a wide rapidity range will constrain the susceptibilities of the QGP, and test the realization of the cross-over phase transition expected at LHC energies. Finally, for the first time, measurements of the production of ultra-soft photons will be possible. These can quantitatively test the infrared limits of quantum field theories such as QED and QCD. In order to meet these physics goals, support for both the R&D and construction efforts is critically required from the U.S. community. This needs to begin before the end of the current 2023 Long Range Plan for Nuclear Science. Based on its expertise, ALICE-USA is currently exploring contributions to the silicon and calorimeter detectors. These parallel and independent efforts will also benefit the EIC achieving its construction milestones. ## 2 Overview of ALICE upgrades and U.S. impact ALICE is a large acceptance experiment, with world leading hadron identification capabilities. Unique to the LHC experiments, ALICE’s primary focus, design, and proposed upgrades, enable extremely accurate characterizations of the QGP using a multitude of observables. It has extensive capabilities in exploring few-body hadronic and nuclear interactions. The first major upgrade for LHC Runs 3&4 enables an increase of ALICE’s recorded luminosities by two orders of magnitude via the introduction of a continuous readout system in the TPC. The second involves a greatly improved tracking performance using new inner trackers - ITS2 in Run 3, and ITS3 in Run 4. The ITS2 system has a reduction of a factor three in radiation lengths, and factor two improvement in pointing resolution compared to the ITS used in Runs 1&2\. Further factor of three improvements of the pointing resolution for ITS3 (compared to ITS2) will occur via the introduction of wafer-scale ultra-thin silicon strips, which are also being pursued by the EPIC detector at the EIC. Details of these upgrades and others can be found elsewhere [1]. For Run 4, the Forward Calorimeter detector (FoCal) upgrade, with both electromagnetic and hadronic calorimeters, will provide unique capabilities for both Cold and Hot QCD studies at forward rapidities of $3.4<\eta<5.8$ [2]. For Runs 5&6, a completely new detector has been proposed, named ALICE 3 [3]. Designed by heavy-ion physicists for heavy-ion physics, it is a next generation detector with a main tracking system that covers a large pseudorapidity range ($-4<\eta<4$), and can reconstruct charged tracks down to $p_{\rm T}\sim 100$ MeV/c. The light weight and small radiation length design will greatly reduce the background and improve tracking resolution for electromagnetic and heavy-flavor probes, compared to ALICE 2. The pointing resolution at midrapidity is projected to be about three times better than that of the ITS3, which is in part achieved by placing a highly novel vertex detector 5mm from the beamline. The tracking system is placed in a superconducting solenoid with a field of up to $B=2$ T, to obtain a momentum resolution of 1–2% over a broad pseudorapidity and momentum range. This tracking is complemented by multiple sub-detector systems for particle identification; two TOF detectors and a RICH detector. These systems have the ability to identify leptons and photons in the entire thermal emission range of $p_{\rm T}\lesssim 3$ GeV/c, which is inaccessible for other LHC experiments. Charged hadron identification on the $3\sigma$ level is possible up to $p_{\rm T}\sim 14$ GeV/c, and decay hadrons can be reconstructed much more efficiently and cleanly at low-$p_{\rm T}$ compared to ALICE 2. The fast readout systems will be able to record all of the expected heavy-ion luminosity provided by the LHC. The ALICE 3 program aims to collect an integrated luminosity of about 35 nb-1 with Pb–Pb collisions and 18 fb-1 with pp collisions at top LHC energy. The potential to further increase the luminosity for ion running in the LHC by using smaller ions, e.g. 84Kr or 128Xe, as well as further runs with small collision systems, is being explored. Figure 1: Event display of one of the first LHC Run 3 Pb–Pb $\sqrt{s_{\mathrm{NN}}}\leavevmode\nobreak\ =\leavevmode\nobreak\ 5.36$ TeV collisions in November 18th 2022, using many detectors ALICE-USA has contributed toward the construction and upgrade. As of September 2022, the ALICE-USA groups consist of teams from two national laboratories: Lawrence Berkeley (LBNL) and Oak Ridge (ORNL), and nine universities: California (Berkeley), CalPoly, Chicago State, Creighton, Houston, Kansas, Ohio State, Tennessee (Knoxville), Texas (Austin), Wayne State, and Yale. While the majority of the program is supported by DOE, there are two NSF supported university teams. ALICE-USA has been responsible for two completed DOE funded projects - the EMCal/DCal in Runs 1&2 [4], and the most recently completed Barrel Tracking Upgrade (BTU) for Runs 3&4 111See the BTU website at https://sites.google.com/site/alicebtusite for further details. The BTU project involved critical U.S. contributions to both the TPC and ITS upgrades. A key example of the success of all these projects can be observed in Fig. 1, which shows one of the first ALICE Run 3 event-displays of Pb–Pb collisions using many detectors where ALICE-USA has made essential contributions. Thanks to the long and successful experience building and operating calorimeters, ALICE-USA is leading the FoCal project [5]. It also oversees a major computing project providing the necessary U.S. contribution to ALICE’s data processing infrastructure, under the umbrella of the Worldwide LHC Computing GRID (WLCG). In terms of the scientific output, the contributions from ALICE-USA span a wide range of physics topics within the collaboration, and have been an essential part of the ALICE physics program. Although ALICE-USA members are present at an about 6% level in terms of authorship within the collaboration, they have provided leadership in terms of paper committee contributions well beyond the average contributions. ALICE-USA members were involved in 1 of 2 Nature articles published by ALICE [6], and 1 of 2 Nature Physics articles [7]. ALICE-USA members were also involved in 5 of the top 7 most cited physics publications from ALICE [8, 9, 10, 11, 7], and numerous other high-profile publications. These papers span both the hard and soft physics using heavy-ion and pp collisions. ALICE has published just over 400 papers since 2009, with the overall U.S. involvement on the paper committee level being about 25%. In addition to publications, members of the ALICE-USA have held numerous leadership roles across all seniority levels of the collaboration regarding physics output. These include: Deputy Physics Coordinator (two times), multiple Physics Working Group (PWG) Conveners, and Physics Analysis Group Coordinators. Of particular note, out of the 8 ALICE PWGs, currently ALICE-USA members co-convene 4 (50%) of the those PWGs. Over the years members of ALICE- USA were copiously represented in the governance of ALICE – currently with members of the collaboration holding functions as members of the Conference Committee and the Management Board, and as one of the two Deputy Spokespersons. Moreover, in the past ALICE-USA members served as Chair of the Collaboration Board and co-chair of the Editorial Board. Finally, for the $\sim 100$ ALICE-USA members, roughly half are PhD students. At this rate, ALICE-USA institutions have graduated $\sim 50$ PhD students over a 5 year period. Many of our students have achieved leading roles in research positions in our field and beyond, as well as in industry and teaching jobs. ## 3 A brief summary of ALICE accomplishments from Runs 1&2 The observations ALICE has made during Runs 1&2 (2009-2019) have profoundly changed the landscape of QCD at high temperatures and collision energies [12]. They have been carried out in conjunction with other LHC experiments, and with the continuation of the RHIC program - where huge advances in luminosities and centre-of-mass energy coverage have been achieved. ALICE explorations of high temperature QCD have continued to reveal emergent behavior in many-body interactions at the highest possible temperatures in the laboratory. In heavy- ion collisions, ALICE has measured significant yield suppression for a wide range of hadrons and reconstructed jets in both inclusive and coincidence channels, showing that in-medium energy loss occurs at the partonic level, and quantifying its magnitude. ALICE measurements of heavy-flavor yield suppression provide insight into details of this process, via the QCD “dead- cone” effect. This manifests itself in a larger energy loss observed for charm compared to bottom quarks. It has also, for the time, provided direct evidence of the dead-cone effect for charm quarks in pp collisions, using state of the art jet re-clustering algorithms. ALICE jet substructure measurements indicate preferential suppression of wide-angle radiation in the jet shower within the QGP, which is sensitive to color coherence and the space-time structure of jets. In the soft sector, ALICE measurements of identified hadron spectra and anisotropic flow demonstrate that a QGP formed at LHC energies undergoes the most rapid expansion ever observed for a many-body system in the laboratory. The hydrodynamic description of a huge variety of such data has been tested in heavy-ion collisions at the LHC, offering an environment far beyond the usual application in fluid dynamics. It has shown a QGP is strongly coupled at scales of the QGP’s temperature, on the order of a few hundred MeV. ALICE measurements also imply thermalisation effects for charm quarks in a QGP, and when coupled with transport model calculations, the ALICE data demonstrate microscopically how equilibrium can occur on extremely small time scales. In addition, ALICE results have demonstrated that a QGP transitions into chemically equilibrated hadrons. ALICE has the most extensive set of measurements ever achieved regarding identified particle production, including for the first time a statistical description in the charm sector. ALICE investigations into the hadron-gas phase, via resonance and femtoscopic measurements, indicate this phase is prolonged, and that the decoupling of particles from the expanding hadron gas is likely to be a continuous process. Using RHIC and LHC Run 1&2 data, comprehensive efforts at global fitting for precise determination of QGP properties and dynamics are currently underway, utilizing the powerful approach of Bayesian Inference for rigorous comparisons between theory and experiments. Such analyses are more advanced in the soft sector, and are in their infancy in the hard probes sector. ALICE has played a critical role in this endeavor by providing the experimental data, and developing theory-experiment synergies. Along with other RHIC and LHC experiments, ALICE has discovered QGP-like signatures in high multiplicity pp and p-Pb collisions, which probe the thresholds of QGP formation. Such findings have ignited a debate of whether pp and p-Pb collisions at LHC and RHIC energies could create small QGP droplets. ALICE has investigated few-body hadronic interactions on the soft and hard scales to an unprecedented precision. In pp collisions, it has demonstrated charmed hadron fragmentation functions are not universal with respect to e+e- collisions, revealing unique hadronisation mechanisms that enhance charmed baryons compared to mesons. It also has provided world leading constraints regarding how rarely produced hadrons interact with stable nuclear matter, whose behavior have broad implications for understanding various features in the Universe, such as the equation of state of neutron stars and the interstellar composition of dark matter. Finally, ALICE has been a leader in the study of ultra-peripheral heavy-ion collisions since Run 1, exploring cold QCD matter at the highest possible photon-induced energies. ALICE has observed photoproduction of vector mesons on proton and ion targets, namely, the $J/\psi$, $\psi(2S)$, and $\rho$ mesons. $J/\psi$ photoproduction is moderately suppressed on Pb targets compared to free protons, in line with nuclear shadowing effects for gluon parton distribution functions at $x\approx 10^{-3}$. Coherent $\rho$ photoproduction measurements also demonstrate nuclear shadowing for both the Xe and Pb nuclei. ## 4 Key physics questions addressed for LHC Runs 3&4 and beyond from ALICE- USA Despite the many extraordinary scientific findings by the ALICE Collaboration, many open questions remain, and will be addressed using data to be collected in Runs 3&4 at the LHC (2022-2032). This section details such questions, with a description of the leadership ALICE-USA groups intend to provide addressing them. Opportunities using the ALICE 3 detector in the Run 5&6 period (2035 and beyond) will also be discussed. 1. 1. What is the gluonic structure in protons & nucleons at small-$x$? We propose to maximize the physics potential for small-$x$ physics in ALICE with the installation a high-granularity forward calorimeter (FoCal) in the LHC Run 4 (2029–2032). The FoCal will explore the small-$x$ parton structure of nucleons and protons in a kinematic domain only accessible at the LHC [5]. It is designed to provide unique capabilities to investigate Parton Distribution Functions (PDFs) in the unexplored regime of $x\sim 10^{-6}$ and four momentum transfer $Q$ of a few GeV. In this regime, it is expected that the hadronic structure requires non-linear dynamics due to the high gluon densities, leading to gluon saturation. Gluon saturation is a necessary consequence of the non-Abelian nature of QCD. Its discovery would represent a landmark in our understanding of the strong interaction. The primary objectives of the FoCal include: high-precision inclusive direct photon and jets measurements in pp and p–Pb collisions, photon-jet and jet-jet measurements in pp and p–Pb collisions, and measurements of vector mesons photoproduction in ultra-peripheral p–Pb and Pb–Pb collisions. These measurements by the FoCal would constitute an essential part of a comprehensive small-$x$ program at the LHC, with a broad array of complementary probes. While the FoCal has some complementarity to future measurements by the EIC [13] and the LHCb experiment, it can reach considerably lower $x$ values than the EIC. Vector meson photoproduction and electroweak boson studies have provided unique constraints on the gluon parton distribution function for nuclei using Runs 1& 2 data [14, 15, 16]. They provide clear evidence that the partonic structure of nuclei is different compared to free protons, with nuclear shadowing effects increasing with decreasing longitudinal momentum fractions $x$. In addition, no evidence of saturation of the gluon PDF in the proton is observed between HERA and LHC energies, down to $x\sim 10^{-5}$ [17, 18]. For a Pb target, the $J/\psi$ photoproduction cross-section is compatible with moderate gluon shadowing, in line with the leading twist approximation and the central values of the EPS09 and EPPS16 nuclear parameterizations [19, 20, 21, 22]. The ALICE data has smaller uncertainties than these parameterizations, pointing to the utility of UPC data for constraining parton distributions [23]. The comparison of the energy dependence of photoproduced vector mesons measured in p–Pb and Pb–Pb collisions is expected to significantly reduce the theoretical uncertainties [24], thus a systematic program of UPC vector meson measurements is needed. The FoCal will enable measurements of photoproduced J/$\rm\psi$ and $\psi$(2S) mesons to reach a $x\sim 10^{-6}$ and $Q^{2}\approx 3-4$ GeV2. This is a region that cannot be explored by the EIC. The energy dependence of the ratio between $\psi(2S)$ and J/$\rm\psi$ is particularly sensitive to the difference between linear and non-linear gluon evolution [25], hence saturation effects. Photoproduction data in Runs 3&4 will allow ALICE to go beyond measurements of vector meson cross-sections. Studies of open charm [26] and/or dijets [27] are sensitive to gluon distributions with fewer theoretical assumptions than for vector mesons. The study of angular correlations of diffractive dijets [28] is also sensitive to the gluon nuclear density, particularly for low-momentum track-based jets where ALICE would have a competitive advantage. Measurements of $d\sigma/dt$ for coherent and incoherent photoproduction of vector mesons are respectively sensitive to the transverse distribution of gluons in a nucleus (similar to a GPD) [29, 30], and the event-by-event fluctuations in the nuclear configuration, including gluonic hotspots [31, 32, 33]. ALICE will collect large enough data samples to make these measurements for a number of different mesons [24], including the $J/\psi$ and the $\rho$. The new ALICE streaming DAQ system will be important for increasing the vector meson sample sizes, and reducing systematic uncertainties due to the trigger requirements. The FoCal detector will allow ALICE to extend these measurements to even lower $x$ values, providing enhanced sensitivity for gluon saturation. The future UPC measurements using the FoCal will also benefit from having other ALICE detectors. It is anticipated that the Zero Degree Calorimeters can be utilized to distinguish the photon direction in ultra-peripheral Pb–Pb collisions [34]. This will allow for the exploration of the lowest possible $x$ values in Pb. ALICE measurements can also be carried out combining the FoCal together with the central barrel or the forward muon detector. In this way, a larger kinematic range can be explored, which is vital for gluon saturation searches. More details of the physics prospects in ultra-peripheral heavy-ion collisions using the FoCal detector can be found elsewhere [35]. Gluon saturation at low-$x$ has also been postulated to be the source of quantum entanglement effects. These could lead to instant thermalization based on the potential equivalence of entanglement and thermodynamic entropy, and a determination of final state particle production in the context of minimal decoherence effects. In order to study this type of parton-hadron duality, ALICE can complement the EIC measurements in ep and eA systems via measurements in pp, pA and AA. The survival of the coherent state as a function of system size can be measured through particle multiplicities and quantum tomographic correlation functions of particles from low-$x$ processes. Quantum entanglement can also be studied by reconstructing the density state operator using quantum tomography [36]. In this direction, the study of exclusive four $\pi$ photoproduction will provide a search for exited $\rho$ states [37]. 2. 2. What are the large scale nuclear structures and QGP transport parameters? Large scale nuclear structures, such as the nuclear deformation and radius, can be precisely investigated using measurements of anisotropic flow. This is achieved courtesy of the QGP hydrodynamic response, where initial state features are imprinted on final state anisotropic flow observables. These nuclear structures are often accessible via low energy nuclear experiments, but are limited to the electrically charged proton profiles. Measurements in heavy-ion collisions can in principle access the entire nuclear matter profile. These profiles are also inaccessible for DIS nuclear collisions by the virtue of their design to probe sub nucleon scales. Such measurements have already been used to determine 129Xe deformation, with values of the quadrapole deformation $\beta_{2}=0.18\pm 0.02$ obtained in Run 2 from ALICE data [38]. These have never been measured previously. Similar studies have proved equally successful constraining the 238U and 197Au deformations from RHIC data [39]. Firstly, we propose to carry out a suite of anisotropic flow measurements in 16O–16O and p-16O collisions. These will occur in Run 3. Hydrodynamic calculations have shown such measurements are sensitive to $\alpha$ clustering in the 16O nucleus [40], a longstanding pursuit in the low energy nuclear community. The clustering effects are calculable in Lattice QCD. Clustering affects anisotropic flow measurements more strongly at LHC energies than at RHIC. Anisotropic flow and flow fluctuations measurements in central large nuclei A-A collisions are particularly sensitive to the nuclear profile. In addition to Xe–Xe collisions at the LHC, this has also been demonstrated with data from the RHIC Isobar run coupled with hydrodynamic calculations, where these profile parameters have been highly constrained [41]. Anisotropic flow measurements of $v_{2}$ and $v_{3}$ in very central Pb–Pb collisions at the LHC have also revealed the “ultra-central problem” - that is a deviation from the hydrodynamic description where it is expected to be most applicable (for a recent review see here [42]). Proposals to explore this further, such as the introduction of an octupole deformation ($\beta_{3}$) for the Pb nucleus [43], which has been predicted for the doubly magic Pb208 ground state, require measurements of $v_{3}\\{4\\}/v_{3}\\{2\\}$ well beyond the accuracy of those achieved in Runs 1&2 by ALICE [11]. There are also predictions for an quadrapole deformation of $\beta_{2}=0.05$ for Pb208 [44], which would lead to finite value of $v_{2}\\{4\\}$. This has yet to be observed in within statistic precision for previous ALICE data in very central Pb-Pb collisions [45]. Therefore, the greater statistics provided by Runs 3-6 will be crucial in unraveling these nuclear properties. Their constraints are also critical for the determination of QGP transport parameters obtained by the hydrodynamic framework, in the regime where hydrodynamics is most applicable. These transport parameters include the famous shear viscosity over entropy ratio $\eta/s$. Such nuclear profile constraints could be particularly important for ALICE 3, as non double magic nuclei such as 84Kr or 128Xe (with larger deformations compared to the double magic 208Pb nucleus) are being explored for heavy-ion running. The hydrodynamic framework used to describe collective flow involves numerous transport parameters, which characterize the coupling of the QGP. To date, only two of those, the shear and bulk viscosity over entropy ratios ($\eta/s$ and $\zeta/s$), have benefited from extensive constraints using heavy-ion data. Those constraints provide evidence the QGP is the most strongly coupled system ever studied in the lab. The corresponding relaxation times, of the order of 1 fm/c or less, demonstrate how thermalization can be achieved on the most rapid timescales ever observed for a many body system. Comparisons of these extracted transport parameters to fundamental descriptions offer a unique test to Holographic models [46]. These include AdS/CFT, which predicts $\eta/s=1/4\pi$ in the infinite coupling limit. Such approaches assume a correspondence of 5 dimensional strong gravitational fields in a black hole with a 4 dimensional high temperature QCD system. These transport parameters are also predictable in Lattice QCD, but often have much larger uncertainties [47]. Two-particle correlations with net baryons can be used to explore additional transport parameters of the hydrodynamic evolution. The baryon diffusion constant $D_{B}$ is an example of such a parameter beyond $\eta/s$ and $\zeta/s$. It characterizes the mobility of baryon number, and is predicted to be finite at the LHC, despite the fact that $\mu_{B}\sim 0$. A two-particle correlation function has been proposed to constrain $D_{B}$ [48]. It explores correlations of net-baryon fluctuations as a function of azimuthal and rapidity separations. Such an analysis has yet to be carried out from Run 1 or 2 data, since it is statistically challenging. It will be greatly aided by the increase of two orders of magnitude in the Pb–Pb integrated luminosity foreseen for Runs 3&4\. The ALICE detector is particularly suited to this task, given its world leading particle identification capabilities. Finally, the recently published measurements of balance functions (BFs) of identified charged hadrons $(\pi,\rm K,\rm p)$ in Pb–Pb collisions from Runs 1&2 play an important role in constraining the charge diffusion coefficient $D_{e}$ for quarks [49]. However, the central tracking acceptance of $-1<\eta<1$ for Runs 1&2 leads to large uncertainties in $D_{e}$ from ALICE BF data [50]. In turn, this leads to values between $0.5D_{Latt}<D_{e}<4D_{Latt}$ being permissible, where $D_{Latt}$ represents the Lattice QCD prediction. These uncertainties will be significantly reduced when the same measurements are performed using the ALICE 3 setup, as the acceptance where identified charged hadron measurements can be made will increase to $-4<\eta<4$. The reduction in $D_{e}$ uncertainties due to increases in the $\Delta\eta$ coverage are expected to be at least a factor 4. 3. 3. How does the QGP affect hard probes? The energetic partons produced in heavy-ion collisions from hard scatterings in the initial collision undergo successive branching. This results in a parton shower that can be modified in the presence of the QGP. While the produced particles are highly collimated about the direction of the initial parton, they also cover a range of different momentum scales. The properties of these collimated sprays of particles, known as jets, and how they emerge from QCD calculations, have been extensively studied in high-energy physics. Jets are a primary tool for uncovering the details of interactions of partons with the QGP medium. This is because jets lose energy and are modified as they traverse the medium, a phenomenon called jet quenching. ALICE is pursuing a multi-messenger approach that systematically studies different features of jet quenching to explore the microscopic structure of the QGP and its governing degrees of freedom. The inclusive jet and hadron production, as well as their correlations from low- to high-$p_{\rm T}$, allow for an assessment of one of the hallmarks of jet quenching - jet energy loss to the medium. In practice, although energy is conserved, the yields of jets with finite jet resolution parameter $R$ are suppressed in heavy-ion collisions due to jet-medium interactions. The suppression of the leading hadrons and fully reconstructed jets in Pb–Pb collision at the LHC broadened the observations found at RHIC. That is the QGP is opaque to jets over a large energy range [51, 52, 53]. Recent measurements of hadron-jet correlations [54] show a potential explanation for the fate of the lost energy in the jet quenching process - in both the momentum and angular scale. The energy is distributed over large angles, and recovered in low-$p_{\rm T}$ (below 20 GeV/c) jets. Additionally, new results on inclusive jet suppression [55], using Machine Learning techniques [56], show that larger $R$ jets (up to $R=0.6$) are suppressed relative to smaller $R$ jets at $p_{\rm T}$ = 40 GeV/c, and the energy is not yet recovered. In Runs 3&4, ALICE will continue measurements of jet energy loss with inclusive jet suppression and semi-inclusive coincidences of hadron-jet [57], photon-hadron, and photon-jet. The substantial increase in statistics and new experimental techniques, such as mixed events and Machine Learning [56], will provide increased precision and an extended kinematic range for jets, and high-$p_{\rm T}$ hadrons and photons. At the same time, the hadron-jet coincidences are sensitive to the _jet- correlated_ response of the medium to the presence of energetic probes [57]. This flux of medium energy induced by the propagating jet is one of the most sought after consequences of jet-medium interactions [58]. A medium composed of scattering centers - quasi-particles - ought to, with some finite probability, induce large angle Moliere scatterings of the propagating partons [59, 60]. Such a jet deflection should manifest itself in the modified acoplanarity of the di-jet pair in Pb–Pb compared to pp collisions. Utilizing hadron-jet coincidences, ALICE has investigated this effect at the lowest jet $p_{\rm T}$ possible and with a large jet $R$ [54]. Additionally, jet deflection should manifest itself in large momentum kicks to the hard core of the jet [61, 59]. ALICE also uses the groomed hardest $k_{T}$ to probe this by looking at differences between pp and Pb–Pb collisions at larger groomed $k_{\rm T}$ values [55]. Likewise, in this area, Runs 3&4 will provide further precision to put improved quantitative constraints on the probability of this process within the QGP. Jet fragmentation patterns and modification of the jet structure are also key to understanding of the interactions of the medium with the jet. ALICE has for the first time measured fully corrected angular and momentum sub-jet structure of jets [62, 63], and has made advances towards measurements of other quantities such as jet angularities, jet-axes, the Lund Plane, and intrajet hadron correlations [64, 54, 65, 66]. These measurements are compared to both Monte Carlo models and analytical calculations. They offer a new and more stringent view of the jet-medium interactions. In particular, new measurements show the largest angle splittings within a groomed jet are suppressed as compared to small-angle splittings, which can be connected to a characteristic aperture scale at which the jet fragments/prongs interact with the medium incoherently. Moreover, the new measurements of leading subjets suggest sensitivity of the predicted flavor dependence of jet quenching that should be quantified with more precise data. Concurrently, the new measurements of jet- axes disfavor the intra-jet $p_{\rm T}$ broadening, as prescribed by the BDMPS formalism as the main mechanism of energy loss in the QGP [67]. The Run 3&4 data will be instrumental in providing the necessary precision to further investigate such observations. However, most importantly, the new high statistics data will enable new differential measurements [24]. These will study different regions of the parton shower phase-space related to rare/hard vs. multiple-soft medium-induced radiation. This can be achieved via simultaneous measurements of angular and momentum distributions of jet structure such as the Jet Lund Plane [68]. The modification of jets in the medium is expected to depend on the path the jet traveled in the medium. ALICE traditionally measures this in two ways. They involve using correlations and measuring the jet yield suppression with respect to the event-plane angle. Measurements by ALICE of hadron-jet correlations show no dependence on the angle of the jet with respect to the event plane angle within the uncertainties [69]. On the other hand, measurements of the jet $v_{2}$ demonstrate a significant azimuthal anisotropy [70]. ALICE has continued to explore these effects using higher statistics data from Run 2. For example, a technique that utilizes Event Shape Engineering [71] to select events based on their anisotropy within a centrality class, demonstrates an increased suppression of out-of-plane yields for more anisotropic events [72]. This technique can be further pursued using the higher statistics Run 3&4 data to investigate the path length dependence of jet substructure modification. Finally, a first look at photon-jet correlation measurements with ALICE was investigated [73]. These studies demonstrated the ability to measure photon-tagged jets at lower $p_{\rm T}^{\gamma}$ (20 GeV/$c$). This measurement is statistically limited, thus future ALICE photon-jet measurements will be crucial for measuring the absolute energy loss, since the photon provides an estimate of the initial $p_{\rm T}$ of the jet. ALICE has also performed measurements of the heavy-flavor hadrons as an additional probe of parton-medium interactions, with examples found in [74, 75, 76]. Heavy flavor hadrons offer a unique opportunity to study the medium effects because of the well established probe - heavy quarks. These are created early in the hard scatterings, and maintain their flavor identity throughout the evolution of the medium. A new set of measurements focusing on the leading heavy-flavor particles will take advantage of this theoretical control. It will also profit from improved experimental control, such as better discrimination of ”combinatorial jets” - the uncorrelated background to hard scattering flux of energy, over the light flavor measurements. This applies to both measurements of the heavy-flavor jet $p_{\rm T}$ spectra and jet substructure observables. One of the striking findings by ALICE is the observed mass dependence of energy loss. This manifests itself in smaller nuclear modification factors for prompt $D$ mesons (formed from charm) compared to non-prompt $D$ mesons (formed from bottom quarks) at $p_{\rm T}$ $>8$ GeV/$c$. This onset is interpreted as the dead cone effect [6]. In future runs, ALICE will take advantage of its precision tracking to study this effect in more detail, and contrast it with the impact of gluon splitting processes within the parton shower [77]. In general, with the newly upgraded ALICE detector, the data from Runs3&4 are expected to improve the quantification of how light and heavy quarks propagate within the medium. The new high-statistics measurements will bring additional constraints to the extraction of the transverse and longitudinal transport coefficients. These include the jet transport parameter $\hat{q}$ [78, 79], which quantifies the transverse diffusion. The longitudinal drag $\hat{e}$ and diffusion $\hat{e_{2}}$ coefficients [80], as well as the heavy-flavor diffusion coefficient $D_{s}$ [81], are also expected to be better constrained. These coefficients characterize the coupling of such processes to the medium, and the ultimate task is to perform a dedicated set of measurements that will allow further connections to the emergence of the QGP on a microscopic level. Runs 3&4 will also allow for a revisiting of the jet substructure measurements for the heavy-flavor induced jets e.g. angular structure of groomed and ungroomed jets. This can be achieved over a broad kinematic regime, that is largely unique to ALICE (from $p_{\rm T}$ $\approx\leavevmode\nobreak\ 0$ of heavy-flavor hadrons to hundreds of GeV/$c$ jets). The new data will enable ALICE to fully explore the momentum and the angular correlations within the jet substructure with a systematic study of the Lund Plane (see considerations in [24]). With this new data, and new techniques, the flavor dependence of the Lund Plane can be extended beyond the charm sector [6] to the beauty sector [82]. For Runs 5&6, ALICE 3 will provide an environment for precision jet measurements over much larger kinematic ranges than previously accessible at the LHC. Specifically, jet hadrochemistry will become possible over such a range with the excellent PID capabilities for identifying hadrons inside jets. This is valuable to study in vacuum to provide inputs to hadronization models [83], and also in the QGP since the interactions with the medium could change the chemistry make-up of a jet [84]. Additionally, ALICE 3 will provide increased resolution, purity, and statistics for heavy-flavor jets. The increased psuedorapidity coverage will allow for measurements of heavy-flavor jet correlations or photon-heavy-flavor jet correlations with unprecedented purity at low transverse momentum scales, providing insight into the microscopic properties of in-medium energy loss, including its path length dependence. Finally, ALICE 3 allows for the possibility to resolve heavy- flavor hadron pairs inside of jets to probe gluon splittings, which can be used to study the space-time picture of jet quenching [85], and access a pure gluon sample of jets. Finally, while we are in the midst of fully understanding the measurements that need to be done, the continued development of analysis techniques is essential for jets in a complex environment. This is related to both the careful selection of the most useful measurements [86], but also to the development of the appropriate correction strategies allowing for meaningful theoretical comparisons. These include fully corrected substructure measurements [87, 88], and background mitigated hadron-jet correlations down to lowest jet momenta with large jet resolution parameters. At the same time, it is important to encourage the divulgence of uncertainty correlations for the new generation of measurements. Such an activity will take advantage of the newly developed Bayesian inference methods [78, 89]. This will also aid Machine Learning projects that will drive the development of the modeling of jet-medium interactions in the coming years [90, 91, 86, 92]. 4. 4. How do rarely produced hadrons form and interact? A detailed program by ALICE for Runs 3-6 will shed further light on more exotic heavier flavor states in order to unambiguously determine the hadronization mechanisms, and thus the creation of matter, as a function of flavor. In this regard, the measurements of multi-heavy-flavour hadrons and exotic states offer unprecedented sensitivity. To achieve high statistics for multi-heavy-flavour particles, a novel experimental approach is needed to track all their decay products, typically including hyperons, before they decay. This calls very high tracking/vertexing precision very close to the interaction point, particle identification capabilities over a wide transverse momentum range, and high readout rates. Moreover, large acceptance is required, not only for reasons of statistics, but also in order to investigate the dependence of the production of multi-heavy-flavour hadrons on the variation of the heavy quark density with rapidity. Recent flavor dependent hadronization measurements have become a unifying theme for the bulk and jet sector of QCD. The hadronization of a final state parton cannot be described adopting a perturbative approach, and is usually modelled through a phase of string breaking and/or cluster formation, which is considered to be independent of the surrounding parton density [93]. However, as also confirmed by measurements of e.g. baryon yields in pp collisions at the LHC [7, 94], additional hadronization mechanisms may exist, whereby quarks that are close in phase space can combine into colourless hadrons. Dynamic modeling of these processes, in particular in the strange and charm sector, has led to many novel approaches that were implemented in established event generator models, such as PYTHIA. Color reconnection (CR) and multiple parton interactions (MPI) were attempts to parameterize features of the initial state and final state into the actual formation process of hadrons. Furthermore, it was shown that initial state gluon saturation might lead to quantum entanglement effects, which impact the final state hadron multiplicities [95]. Experimental advances in the strange sector include the detailed mapping of strangeness enhancement in proton-proton collisions as a function of centrality [7], and the potential flavor hierarchy in the light/strange hadron formation during the QCD crossover [96]. These led to refinements of the transport codes that were then followed up by more detailed measurements of baryon to meson formation in the strange sector. Such advancements have been studied in the charm sector, as a function of system size [94]. In heavy-ion collisions, where partons may travel freely over distances much larger than the typical hadron sizes, and a dense system of partons close to thermal equilibrium is formed, recombination mechanisms become more dominant. These make the production of baryons and other heavy hadrons more favourable than in pp collisions [97]. First measurements of baryon/meson ratios in the charm sector also indicate a low-$p_{\rm T}$ enhancement consistent with such a picture that has dominated similar measurements in the light and strange sector associated with the soft underlying event [98]. A comprehensive campaign of precision measurements of charm baryon production yields across collision systems is planned for Runs 3&4\. This starts with baryons containing only one charm quark, but several light and strange quark combinations. The evolution of particle production as a function of increasing strangeness or charm content is one of the main open questions regarding hadron formation, which is a key feature of the non-perturbative sector of QCD. For example, LHCb has found a series of tetra- and penta-quark states in the charm sector. A detailed study of strange penta- and hexa-quark states in ALICE found no evidence in the strange sector. The measured upper limits are several orders of magnitude below estimates from statistical hadronization models (SHM), which have done extremely well on postulating the yields of composite objects in the light sector [99]. SHM predictions assume full equilibration of all flavors up to charm. The comparison of the measured yields to these predictions would provide a very sensitive measure of the degree of equilibration of charm quarks in the medium, including its system- size dependence. Ultimately, it must be possible to relate the equilibration level to the transport properties of the QGP via kinetic transport modeling in order to provide a consistent description of the entire phenomenology. Complementing these opportunities, relativistic heavy-ion collisions provide the only available tool to study the formation and interaction of exotic hadronic states inside a medium filled with deconfined color charges. This puts new constraints on the properties of the states, the characterization of their binding potential, and the details of their hadronization mechanisms. Nuclear collisions at the LHC energies, in particular, offer a unique opportunity to test these mechanisms in the presence of large charm-quark densities. In this environment, while the screening of color charges may reduce the yield of certain bound states, medium-induced or medium enhanced mechanisms may enhance their production in the low-$p_{\rm T}$ region, as predicted by statistical hadronization models. The interest of these observations is twofold. On one hand, the rates of formation and dissociation of bound states depend on their binding energy and size, and the measurement of the production rates and anisotropic flow will allow one to distinguish compact multi-quark configurations from molecular states. On the other hand, states whose nature can be determined by other means, for example by measuring interaction potentials through final state interactions, could be exploited as a calibrated probe in nuclear collisions. This would shed light on the underlying mechanisms and timescales driving the hadronization in a deconfined medium, and test the properties of the deconfined medium itself. Predictions for light flavor tetraquark states have existed for over 40 years [100], but the existence of these states has up to now not been confirmed. To begin to address this, ALICE has published three papers on tetraquark studies of the $a_{0}(980)$ resonance, a tetraquark candidate, using kaon-kaon correlations from Runs 1&2 [101, 102, 103] for pp and Pb–Pb collisions. In Runs 3&4, ALICE will continue these two-meson correlation studies for other low lying tetraquark candidates, the $K^{*}_{0}(700)$ and $f_{0}(500)$. The high rate capabilities in Run 3&4, will also allow access to rare phenomena in the hadronic phase formed after the phase transition. These include the production of nuclei, hypernuclei and, combined with excellent heavy-flavour capabilities, possibly the as-yet undiscovered supernuclei (nuclei in which a nucleon is replaced by a charmed baryon). Wide acceptance, again combined with outstanding heavy-flavour performance at low transverse momenta, will allow to extend the studies of hadron-hadron potentials via two-particle correlations to the charm sector. This will provide a powerful tool to investigate the structure of the newly discovered charmed exotic states. In Runs 5&6, the measurement of the production yields of multi-charm baryons, which can only be produced by combination of uncorrelated charm quarks, would provide a qualitatively new handle on the production of heavy-flavour hadrons. Measurements of multi-charm hadrons, such as the $\Xi^{+}_{cc}$ (ccd), $\Xi^{++}_{cc}$ (ccu), $\Omega^{+}_{cc}$ (ccs), $\Omega^{++}_{ccc}$ (ccc), and exotic states such as the newly discovered T${}^{+}_{cc}$ (ccud) [104, 105], would provide a direct window on hadron formation from the QGP. In fact, the yields of multi-charm baryons relative to the number of produced charm quarks are predicted to be significantly enhanced in AA relative to pp collisions [106]. Enhancements are expected by as much as a factor 100 for the recently observed $\Xi_{cc}$ baryon [107], and even by as much as a factor 1000 for the as yet undiscovered $\Omega_{ccc}$ baryon. The observation and precise quantification of such effects, would represent a major discovery for the study of the properties of deconfined matter. In addition, LHC heavy-ion collisions also have a discovery potential for more exotic hadronic states. This is because of the long-lived deconfined medium and the large cross sections for heavy quarks available at the highest energies. These provide an ideal playground for the production of exotic states with e.g. multiple b quarks. The most notable case is the T${}^{-}_{bb}$ bb, whose experimental detection would profit from the long predicted lifetime (c$\tau$ $\sim$ 2.3 mm) due to the stability of the state with respect to strong decays [108] (contrary to the recently discovered, shortly-lived T${}^{+}_{cc}$ state) . Discovery opportunities also exist for compact bound hidden-charm hexaquarks, also predicted to be stable with respect to strong interaction decays [109], and molecular states composed of three D mesons. Studying exotic QCD states in nuclear collisions is therefore of central importance for QCD physics. The feasibility of such studies has been demonstrated by a recent measurement of $\chi_{c1}$(3872) production in Pb–-Pb collisions by the CMS collaboration in the range $p_{\rm T}$ $>$ 10 GeV/c [110]. An immediate goal for ALICE 3 is to measure the production of $\chi_{c1}$(3872) down to the $p_{\rm T}$ region $<$5-6 GeV/c, where a significant enhancement of the yield was predicted [111]. This is not accessible by other LHC experiments. For hadrons containing charm and beauty quarks, scattering experiments are not feasible to determine the interaction strength between hadrons, therefore the only way to access the information is the femtoscopy technique. This technique consists of the measurement of correlations in momentum space for hadron-hadron pairs, and can be used to extract the corresponding scattering parameters [112]. The ALICE 3 upgrade will allow the measurement of several hadron combinations including DD∗, $\Lambda_{c}^{+}\Sigma_{c}^{0;+;++}$ and BB∗ in pp, p–Pb and Pb–Pb collisions and thereby shed light on the nature of many exotic hadrons. In particular, using the same method as mentioned above in studying tetraquark states with meson-meson correlations, the ALICE 3 upgrade will allow the study of the $f_{2}(2010)$ and $\psi(3770)$ tetraquark candidates with $\phi\phi$ and $D^{+}D^{-}$ correlations, respectively. 5. 5. What are the mechanisms that lead to QGP-like signals in small systems? The formation of the QGP in heavy-ions has been confirmed via numerous measurements. A key example is the simultaneous observation of collective flow phenomena traced to a strongly coupled liquid, and jet quenching. Measurements in small collision systems, such as pp and p–Pb at the LHC, have shown particle correlations and strangeness yields in high-multiplicity collisions resemble observations associated with the creation of the QGP [113, 9, 114, 7]. On the other hand, jet quenching in small systems has not been observed. The resolution of this enigma remains one of the most important questions for our field. ALICE has contributed with a stringent limit on energy loss within small collision systems [115], and plans to improve on those measurements with the high-statistics data in the upcoming LHC runs [24]. ALICE can also use machine learning methods to design and measure new jet observables that are maximally modified by traversing small systems [86, 116]. Anisotropic flow measurements in high-multiplicity small system collisions at RHIC and the LHC are consistent with hydrodynamic predictions [117, 118]. This indicates these collisions may create the smallest possible QGP droplets. However, these measurements from Runs 1&2 suffer from large uncertainties due to ambiguities in the non-flow subtraction methods pursued at the LHC [119]. For Run 4, the introduction of the FoCal will enable $\Delta\eta$ separations of up to 9 units for two-particle correlation anisotropic flow measurements of $v_{n}$, with such separations being critical in reducing non-flow. The broad increase from the ALICE 3 acceptance will also reduce non-flow contributions in two-particle correlation $v_{n}$ measurements by a factor of 4 for Runs 5&6 [120]. Four particle $v_{n}$ measurements enjoy a reduction of a factor 64, and this increases exponentially with the number of particles used to determine $v_{n}$. Such reductions are vital in the transition of anisotropic flow measurements in small systems from being qualitative to truly quantitative. Measurements in Run 4 and ALICE 3 are therefore essential for precise quantitative constraints of the studies of the applicability of hydrodynamics at its limits in small systems [121]. The improvements in non- flow suppression and increased statistics are also essential for identified hadron $v_{n}$ measurements vs. $p_{\rm T}$ in small systems. These test hydrodynamic mass ordering at low-$p_{\rm T}$, with arises due to a common radial flow profile. Such measurements, using ALICE’s highly competitive particle identification capacities, will be key for all future precision tests of hydrodynamics in small systems. These tests can also be expanded for the unmeasured $\Xi$ and $\Omega$ $v_{n}$ vs. $p_{\rm T}$, which have proved statistically challenging in Runs 1&2. Finally, the increased statistics for extremely high multiplicity pp collisions in Runs 3&4 will allow for measurements of the $\Omega/\pi$ ratio at $\rm{d}N_{ch}/\rm{d}\eta\sim 96$. This is six times larger than corresponding measurements from Runs 1&2, and fully overlap with ALICE’s Pb–Pb measurements [7]. This will in turn allow for a critical test regarding particle production mechanisms and the interpretation of strangeness enhancment in pp collisions [24]. If this ratio saturates and reaches the thermal limit in Pb–Pb collisions, this would support the idea the QGP can be created in pp collisions. On the other hand, if this ratio continues to increase beyond the Pb–Pb values, this perhaps would favor non-QGP mechanisms, such as color ropes implemented in the DIPSY model [122], that predict a continuing increase. 6. 6. What are the connections and broad impacts of ALICE measurements to other fields of physics? One of the recent major discoveries are gravitational wave signals from neutron star mergers. The tidal deformation during approach and the ring-down of the frequency spectrum after the merger hold potentials clues on the core composition of neutron stars. In particular, this applies for stars with a large mass to radius ratio. Model predictions range from simple neutron matter, to hyperon matter, and deconfined quark matter [123]. Theoretically, the main required ingredient is an equation of state for the system as a function of density and temperature. Experimentally, relativistic heavy-ion experiments, and in particular ALICE with its high luminosity and data rate beyond Run 3, will be able to map out the production of hyperons, hypermatter and quark matter. Since deconfined matter and hyperon production have been discussed in the previously, the focus here is on hyper nuclei production and the measurement of the balance of attractive and repulsive forces in the interaction between hyperons and protons/neutrons through femtoscopic measurements. During Runs 1&2 these studies have shown great promise as documented in a recent Nature article by ALICE [112]. Certain di-baryon correlations in the light and strange sector show attractive potentials, which should lead to bound hexa-quark states. These results should also be viewed in the context of hyper nuclei production. In previous campaigns, ALICE has successfully reconstructed the lightest hypernucleus, the hypertriton and its anti-particle [124]. In Run 4, we hope to extend the statistical sample in Pb–Pb collisions to a level that might make the reconstruction of A=4 hypernuclei, in particular the ${}^{4}_{\Lambda}$H. Their measurement is interesting for precision tests of particle production models [125] and to constrain hyperon-nucleon potentials [126], and thus the formation probability of hyper-matter in dense stellar objects. Regarding Runs 5&6, anti-nuclei and anti-hyper-nuclei with A$>$4 such as anti-${}^{5}_{\Lambda}$He or anti-6Li have yet to be discovered, and may well be in reach of ALICE 3. The ALICE 3 apparatus is ideally suited for the observation of the $A=4$ or $A=5$ hyper-nuclei like ${}^{4}_{\Lambda}$H or ${}^{5}_{\Lambda}$He. The measurement of $A=6$ nuclei would provide precision tests for the formation of bound clusters thanks to the special nature of 6He and 6Li. 6He is the lightest known (anti-)halo-nucleus. Its production is therefore expected to be suppressed in coalescence models with respect to thermal-statistical models, due to its much larger size. 6 Li is a stable isotope with a spin of $J=1$. With respect to the helium isotopes 4He and 6He with $J=0$, 6Li production is therefore expected to be enhanced by the degeneracy factor g of its spin-substates; $g=2J+1=3$. The expected production yields $d$N/$dy$ for A=5 hyper-nuclei and A=6 nuclei are in the 10-9 to 10-11 range in Pb–Pb collisions. 7. 7. What can ALICE achieve regarding Beyond Standard Model (BSM) Physics? Although studying BSM physics is not ALICE’s primary purpose, there are areas where it can make extremely useful contributions, with the benefit of the upgrades and increased statistics in future runs. One such area is the study of light-by-light scattering, $\gamma\gamma\rightarrow\gamma\gamma$, using UPCs. Light-by-light scattering occurs via a box diagram, which includes contributions from all (standard model or BSM) electrically charged particles. In addition, it can be mediated by axion-like particles, $\gamma\gamma\rightarrow A\rightarrow\gamma\gamma$, which leads to a diphoton invariant mass spectrum that is peaked at the axion mass. By virtue of its sensitivity to lower $p_{\rm T}$ particles, ALICE (especially ALICE 3) can probe lower axion masses than ATLAS or CMS. ALICE is also sensitive to two- photon production of $\tau^{+}\tau^{-}$. Such studies of this reaction can be used to put limits on BSM phenomena such as a $\tau$ anomalous magnetic moment. The discovery of stable massive particles (SMPs) beyond the Standard Model (BSM), such as monopoles, gluinos, heavy leptons, etc., would address a number of important questions in modern physics, including the nature of dark matter, and the unification of fundamental forces. The majority of predictions suggest that SMPs are far too massive to be produced in a foreseeable accelerator [127]. However, there are suggestions some of them, e.g. magnetic monopoles, could appear in a mass range accessible at the LHC [128], in particular in Pb–Pb collisions. Detection of such particles is a very challenging due to their highly ionizing nature, which leads to saturation and even malfunction of electronics of e.g. silicon-based detectors [129]. The newly upgraded GEM (Gas Electron Multiplier) based Time Projection Chamber (TPC) of ALICE is free from this challenge, and is capable of measuring very large energy deposits anticipated from monopoles. An exciting avenue to study dark matter signals lies with cosmic-ray antinuclei, such as antihelium. This is considered another promising signature of the existence of weakly-interactive mass particles (WIMP), which represent an important candidate for dark matter [130]. Since the background from hadronic interactions of primary rays are negligible [131], the preliminary evidence of a handful of anti-3He events collected by the AMS collaboration, could be due to a previously neglected process. Namely, these could be the production of $\Lambda_{b}^{0}$ baryons in dark-matter annihilation, and their subsequent decay into anti-3He [132]. However, the decay rates of $\Lambda_{b}^{0}$ to anti-nuclei are not experimentally measured, and serve as crucial inputs to understand the AMS data. The large beauty-quark production cross section at LHC energies, and the identification capabilities for nuclei over practically the full phase space, make ALICE 3 the ideal experiment for such studies. ## References * [1] ALICE Collaboration, “ALICE upgrade physics performance studies for 2018 Report on HL/HE-LHC physics”,. https://cds.cern.ch/record/2661798. * [2] ALICE Collaboration, “A Forward Calorimeter (FoCal) in the ALICE experiment”, Tech. Rep. ALICE-PUBLIC-2019-005, Oct, 2019. https://cds.cern.ch/record/2696471. * [3] ALICE Collaboration, “Letter of intent for ALICE 3: A next-generation heavy-ion experiment at the LHC”, arXiv:2211.02491 [physics.ins-det]. * [4] ALICE Collaboration, “Letter of intent for ALICE 3: A next generation heavy-ion experiment at the LHC”, tech. rep., CERN, Geneva, 2022\. https://cds.cern.ch/record/2803563. * [5] ALICE Collaboration, “Performance of the ALICE Electromagnetic Calorimeter”, arXiv:2209.04216 [physics.ins-det]. * [6] ALICE Collaboration, S. Acharya et al., “Direct observation of the dead-cone effect in quantum chromodynamics”, Nature 605 no. 7910, (2022) , arXiv:2106.05713 [nucl-ex]. [Erratum: Nature 607, E22 (2022)]. * [7] ALICE Collaboration, J. Adam et al., “Enhanced production of multi-strange hadrons in high-multiplicity proton-proton collisions”, Nature Phys. 13 (2017) , arXiv:1606.07424 [nucl-ex]. * [8] ALICE Collaboration, K. Aamodt et al., “Elliptic flow of charged particles in Pb-Pb collisions at 2.76 TeV”, Phys. Rev. Lett. 105 (2010) , arXiv:1011.3914 [nucl-ex]. * [9] ALICE Collaboration, B. B. Abelev et al., “Long-range angular correlations of $\rm\pi$, K and p in p-Pb collisions at $\sqrt{s_{\rm NN}}$ = 5.02 TeV”, Phys. Lett. B 726 (2013) , arXiv:1307.3237 [nucl-ex]. * [10] ALICE Collaboration, K. Aamodt et al., “Centrality dependence of the charged-particle multiplicity density at mid-rapidity in Pb-Pb collisions at $\sqrt{s_{NN}}=2.76$ TeV”, Phys. Rev. Lett. 106 (2011) , arXiv:1012.1657 [nucl-ex]. * [11] ALICE Collaboration, K. Aamodt et al., “Higher harmonic anisotropic flow measurements of charged particles in Pb-Pb collisions at $\sqrt{s_{NN}}$=2.76 TeV”, Phys. Rev. Lett. 107 (2011) , arXiv:1105.3865 [nucl-ex]. * [12] ALICE Collaboration, “The ALICE experiment - A journey through QCD”, arXiv:2211.04384 [nucl-ex]. * [13] R. Abdul Khalek et al., “Science Requirements and Detector Concepts for the Electron-Ion Collider: EIC Yellow Report”, Nucl. Phys. A 1026 (2022) , arXiv:2103.05419 [physics.ins-det]. * [14] J. G. Contreras and J. D. Tapia Takaki, “Ultra-peripheral heavy-ion collisions at the LHC”, Int. J. Mod. Phys. A 30 (2015) . * [15] S. Klein and P. Steinberg, “Photonuclear and Two-photon Interactions at High-Energy Nuclear Colliders”, Ann. Rev. Nucl. Part. Sci. 70 (2020) , arXiv:2005.01872 [nucl-ex]. * [16] S. Klein et al., “New opportunities at the photon energy frontier”, arXiv:2009.03838 [hep-ph]. * [17] ALICE Collaboration, B. B. Abelev et al., “Exclusive $\mathrm{J/}\psi$ photoproduction off protons in ultra-peripheral p-Pb collisions at $\sqrt{s_{\rm NN}}=5.02$ TeV”, Phys. Rev. Lett. 113 no. 23, (2014) , arXiv:1406.7819 [nucl-ex]. * [18] ALICE Collaboration, S. Acharya et al., “Energy dependence of exclusive $\mathrm{J}/\psi$ photoproduction off protons in ultra-peripheral p–Pb collisions at $\sqrt{s_{\mathrm{\scriptscriptstyle NN}}}=5.02$ TeV”, Eur. Phys. J. C 79 no. 5, (2019) , arXiv:1809.03235 [nucl-ex]. * [19] ALICE Collaboration, B. Abelev et al., “Coherent $J/\psi$ photoproduction in ultra-peripheral Pb-Pb collisions at $\sqrt{s_{NN}}=2.76$ TeV”, Phys. Lett. B 718 (2013) , arXiv:1209.3715 [nucl-ex]. * [20] ALICE Collaboration, E. Abbas et al., “Charmonium and $e^{+}e^{-}$ pair photoproduction at mid-rapidity in ultra-peripheral Pb-Pb collisions at $\sqrt{s_{\rm NN}}$=2.76 TeV”, Eur. Phys. J. C 73 no. 11, (2013) , arXiv:1305.1467 [nucl-ex]. * [21] ALICE Collaboration, S. Acharya et al., “Coherent $J/\psi$ and $\psi^{\prime}$ photoproduction at midrapidity in ultra-peripheral Pb-Pb collisions at $\sqrt{s_{\mathrm{NN}}}\leavevmode\nobreak\ =\leavevmode\nobreak\ 5.02$ TeV”, Eur. Phys. J. C 81 no. 8, (2021) , arXiv:2101.04577 [nucl-ex]. * [22] CMS Collaboration, V. Khachatryan et al., “Coherent $J/\psi$ photoproduction in ultra-peripheral PbPb collisions at $\sqrt{s_{NN}}=$ 2.76 TeV with the CMS experiment”, Phys. Lett. B 772 (2017) , arXiv:1605.06966 [nucl-ex]. * [23] K. J. Eskola, C. A. Flett, V. Guzey, T. Löytäinen, and H. Paukkunen, “Exclusive $J/\psi$ photoproduction in ultraperipheral Pb+Pb collisions at the LHC to next-to-leading order perturbative QCD”, arXiv:2203.11613 [hep-ph]. * [24] Z. Citron et al., “Report from Working Group 5: Future physics opportunities for high-density QCD at the LHC with heavy-ion and proton beams”, CERN Yellow Rep. Monogr. 7 (2019) , arXiv:1812.06772 [hep-ph]. * [25] M. Hentschinski et al., “White Paper on Forward Physics, BFKL, Saturation Physics and Diffraction”, arXiv:2203.08129 [hep-ph]. * [26] S. R. Klein, J. Nystrand, and R. Vogt, “Heavy quark photoproduction in ultraperipheral heavy ion collisions”, Phys. Rev. C 66 (2002) , arXiv:hep-ph/0206220. * [27] ATLAS Collaboration, ATLAS Collaboration, “Photo-nuclear dijet production in ultra-peripheral Pb+Pb collisions.” ATLAS-CONF-2017-011, 2017\. * [28] CMS Collaboration, “Azimuthal correlations within exclusive dijets with large momentum transfer in photon-lead collisions”, arXiv:2205.00045 [nucl-ex]. * [29] STAR Collaboration, L. Adamczyk et al., “Coherent diffractive photoproduction of $\rho^{0}$mesons on gold nuclei at 200 GeV/nucleon-pair at the Relativistic Heavy Ion Collider”, Phys. Rev. C 96 no. 5, (2017) , arXiv:1702.07705 [nucl-ex]. * [30] STAR Collaboration, S. R. Klein, “Using coherent dipion photoproduction to image gold nuclei”, SciPost Phys. Proc. 8 (2022) , arXiv:2107.10447 [nucl-ex]. * [31] J. Cepila, J. G. Contreras, and J. D. Tapia Takaki, “Energy dependence of dissociative $\mathrm{J/}\psi$ photoproduction as a signature of gluon saturation at the LHC”, Phys. Lett. B 766 (2017) , arXiv:1608.07559 [hep-ph]. * [32] J. Cepila, J. G. Contreras, M. Krelina, and J. D. Tapia Takaki, “Mass dependence of vector meson photoproduction off protons and nuclei within the energy-dependent hot-spot model”, Nucl. Phys. B 934 (2018) , arXiv:1804.05508 [hep-ph]. * [33] S. R. Klein and H. Mäntysaari, “Imaging the nucleus with high-energy photons”, Nature Rev. Phys. 1 no. 11, (2019) , arXiv:1910.10858 [hep-ex]. * [34] M. Broz, J. G. Contreras, and J. D. Tapia Takaki, “A generator of forward neutrons for ultra-peripheral collisions: $\textbf{n}^{\textbf{O}}_{\textbf{O}}\textbf{n}$”, Comput. Phys. Commun. 253 (2020) , arXiv:1908.08263 [nucl-th]. * [35] A. Bylinkin, J. Nystrand, and D. Tapia Takaki, “Vector meson photoproduction in UPCs with FoCal”, arXiv:2211.16107 [nucl-ex]. * [36] J. C. Martens, J. P. Ralston, and J. D. Tapia Takaki, “Quantum tomography for collider physics: Illustrations with lepton pair production”, Eur. Phys. J. C 78 no. 1, (2018) , arXiv:1707.01638 [hep-ph]. * [37] M. Klusek-Gawenda and J. D. Tapia Takaki, “Exclusive Four-pion Photoproduction in Ultra-peripheral Heavy-ion Collisions at RHIC and LHC Energies”, Acta Phys. Polon. B 51 no. 6, (2020) , arXiv:2005.13624 [hep-ph]. * [38] ALICE Collaboration, S. Acharya et al., “Anisotropic flow in Xe-Xe collisions at $\mathbf{\sqrt{s_{\rm{NN}}}=5.44}$ TeV”, Phys. Lett. B 784 (2018) , arXiv:1805.01832 [nucl-ex]. * [39] G. Giacalone, J. Jia, and C. Zhang, “Impact of Nuclear Deformation on Relativistic Heavy-Ion Collisions: Assessing Consistency in Nuclear Physics across Energy Scales”, Phys. Rev. Lett. 127 no. 24, (2021) , arXiv:2105.01638 [nucl-th]. * [40] N. Summerfield, B.-N. Lu, C. Plumberg, D. Lee, J. Noronha-Hostler, and A. Timmins, “16O 16O collisions at energies available at the BNL Relativistic Heavy Ion Collider and at the CERN Large Hadron Collider comparing $\alpha$ clustering versus substructure”, Phys. Rev. C 104 no. 4, (2021) , arXiv:2103.03345 [nucl-th]. * [41] J. Jia, G. Giacalone, and C. Zhang, “Separating the impact of nuclear skin and nuclear deformation on elliptic flow and its fluctuations in high-energy isobar collisions”, arXiv:2206.10449 [nucl-th]. * [42] A. V. Giannini, M. N. Ferreira, M. Hippert, D. D. Chinellato, G. S. Denicol, M. Luzum, J. Noronha, T. N. da Silva, and J. Takahashi, “Assessing the ultra-central flow puzzle in the Bayesian era”, arXiv:2203.17011 [nucl-th]. * [43] P. Carzon, S. Rao, M. Luzum, M. Sievert, and J. Noronha-Hostler, “Possible octupole deformation of 208Pb and the ultracentral $v_{2}$ to $v_{3}$ puzzle”, Phys. Rev. C 102 no. 5, (2020) , arXiv:2007.00780 [nucl-th]. * [44] S. Raman, C. W. G. Nestor, Jr, and P. Tikkanen, “Transition probability from the ground to the first-excited 2+ state of even-even nuclides”, Atom. Data Nucl. Data Tabl. 78 (2001) . * [45] ALICE Collaboration, B. B. Abelev et al., “Multiparticle azimuthal correlations in p -Pb and Pb-Pb collisions at the CERN Large Hadron Collider”, Phys. Rev. C 90 no. 5, (2014) , arXiv:1406.2474 [nucl-ex]. * [46] S. I. Finazzo, R. Rougemont, H. Marrochio, and J. Noronha, “Hydrodynamic transport coefficients for the non-conformal quark-gluon plasma from holography”, JHEP 02 (2015) , arXiv:1412.2968 [hep-ph]. * [47] C. Ratti, “Lattice QCD and heavy ion collisions: a review of recent progress”, Rept. Prog. Phys. 81 no. 8, (2018) , arXiv:1804.07810 [hep-lat]. * [48] S. Floerchinger and M. Martinez, “Fluid dynamic propagation of initial baryon number perturbations on a Bjorken flow background”, Phys. Rev. C 92 no. 6, (2015) , arXiv:1507.05569 [nucl-th]. * [49] ALICE Collaboration, S. Acharya et al., “General balance functions of identified charged hadron pairs of ($pi$,K,p) in Pb-Pb collisions at $\sqrt{s_{\rm NN}}=2.76$ TeV”, Phys. Lett. B 833 (2022) , arXiv:2110.06566 [nucl-ex]. * [50] S. Pratt and C. Plumberg, “Charge balance functions for heavy-ion collisions at energies available at the CERN Large Hadron Collider”, Phys. Rev. C 104 no. 1, (2021) , arXiv:2104.00628 [nucl-th]. * [51] ALICE Collaboration, K. Aamodt et al., “Suppression of Charged Particle Production at Large Transverse Momentum in Central Pb-Pb Collisions at $\sqrt{s_{NN}}=$ 2.76 TeV”, Phys. Lett. B 696 (2011) , arXiv:1012.1004 [nucl-ex]. * [52] ALICE Collaboration, S. Acharya et al., “Transverse momentum spectra and nuclear modification factors of charged particles in pp, p-Pb and Pb-Pb collisions at the LHC”, JHEP 11 (2018) , arXiv:1802.09145 [nucl-ex]. * [53] ALICE Collaboration, S. Acharya et al., “Measurements of inclusive jet spectra in pp and central Pb-Pb collisions at $\sqrt{s_{\rm{NN}}}$ = 5.02 TeV”, Phys. Rev. C 101 no. 3, (2020) , arXiv:1909.09718 [nucl-ex]. * [54] Reynier Cruz Torres for the ALICE Collaboration, “Jet acoplanarity and energy flow within jets in pb-pb and pp collisions with alice.” url: https://indico.cern.ch/event/895086/contributions/4742231/, 4, 2022. Quark Matter 2022 Conference. * [55] Hannah Bossi for the ALICE Collaboration, “Measurements of the $r$-dependence of inclusive jet suppression and groomed jet splittings in heavy-ion collisions with alice.” url: https://indico.cern.ch/event/895086/contributions/4736731/, 4, 2022. Quark Matter 2022 Conference. * [56] R. Haake and C. Loizides, “Machine Learning based jet momentum reconstruction in heavy-ion collisions”, Phys. Rev. C 99 no. 6, (2019) , arXiv:1810.06324 [nucl-ex]. * [57] ALICE Collaboration, J. Adam et al., “Measurement of jet quenching with semi-inclusive hadron-jet distributions in central Pb-Pb collisions at $\sqrt{s_{\mathrm{NN}}}=2.76$ TeV”, JHEP 09 (2015) , arXiv:1506.03984 [nucl-ex]. * [58] S. Cao and X.-N. Wang, “Jet quenching and medium response in high-energy heavy-ion collisions: a review”, Rept. Prog. Phys. 84 no. 2, (2021) , arXiv:2002.04028 [hep-ph]. * [59] F. D’Eramo, K. Rajagopal, and Y. Yin, “Molière scattering in quark-gluon plasma: finding point-like scatterers in a liquid”, JHEP 01 (2019) , arXiv:1808.03250 [hep-ph]. * [60] J. a. Barata, Y. Mehtar-Tani, A. Soto-Ontoso, and K. Tywoniuk, “Revisiting transverse momentum broadening in dense QCD media”, Phys. Rev. D 104 no. 5, (2021) , arXiv:2009.13667 [hep-ph]. * [61] J. a. Barata, Y. Mehtar-Tani, A. Soto-Ontoso, and K. Tywoniuk, “Medium-induced radiative kernel with the Improved Opacity Expansion”, JHEP 09 (2021) , arXiv:2106.07402 [hep-ph]. * [62] A Large Ion Collider Experiment, ALICE Collaboration, S. Acharya et al., “Measurement of the groomed jet radius and momentum splitting fraction in pp and Pb$-$Pb collisions at $\sqrt{s_{NN}}=5.02$ TeV”, Phys. Rev. Lett. 128 no. 10, (2022) , arXiv:2107.12984 [nucl-ex]. * [63] ALICE Collaboration, “Measurement of inclusive and leading subjet fragmentation in pp and Pb-Pb collisions at $\sqrt{s_{\rm NN}}$ = 5.02 TeV”, arXiv:2204.10270 [nucl-ex]. * [64] James Mulligan for the ALICE Collaboration, “Jet angularity and fragmentation measurements in heavy-ion collisions with alice.” url: https://indico.cern.ch/event/895086/contributions/4715856/, 4, 2022. Quark Matter 2022 Conference. * [65] ALICE Collaboration, “Measurements of the groomed jet radius and momentum splitting fraction with the soft drop and dynamical grooming algorithms in pp collisions at $\sqrt{s}=5.02$ TeV”, arXiv:2204.10246 [nucl-ex]. * [66] ALICE Collaboration, “Physics Preliminary Summary: Measurement of the primary Lund plane density in pp collisions at $\sqrt{s}=\rm{13}$ TeV with ALICE”,. * [67] F. Ringer, B.-W. Xiao, and F. Yuan, “Can we observe jet $P_{T}$-broadening in heavy-ion collisions at the LHC?”, Phys. Lett. B 808 (2020) , arXiv:1907.12541 [hep-ph]. * [68] A. Lifson, G. P. Salam, and G. Soyez, “Calculating the primary Lund Jet Plane density”, JHEP 10 (2020) , arXiv:2007.06578 [hep-ph]. * [69] ALICE Collaboration, S. Acharya et al., “Jet-hadron correlations measured relative to the second order event plane in Pb-Pb collisions at $\sqrt{s_{\rm{NN}}}$ = 2.76 TeV”, Phys. Rev. C 101 no. 6, (2020) , arXiv:1910.14398 [nucl-ex]. * [70] ALICE Collaboration, J. Adam et al., “Azimuthal anisotropy of charged jet production in $\sqrt{s_{\rm NN}}$ = 2.76 TeV Pb-Pb collisions”, Phys. Lett. B 753 (2016) , arXiv:1509.07334 [nucl-ex]. * [71] J. Schukraft, A. Timmins, and S. A. Voloshin, “Ultra-relativistic nuclear collisions: event shape engineering”, Phys. Lett. B 719 (2013) , arXiv:1208.4563 [nucl-ex]. * [72] Caitlin Beattie for the ALICE Collaboration, “Study of path-length dependent energy loss of jets in p-pb and pb-pb collisions with alice.” url: https://indico.cern.ch/event/895086/contributions/4715857/, 4, 2022. Quark Matter 2022 Conference. * [73] Alwina Liu for the ALICE Collaboration, “Isolated-photon production and photon-jet correlations in pp and pb-pb collisions at $\sqrt{s_{\rm NN}}=$ 5.02 tev in alice.” url: https://indico.cern.ch/event/895086/contributions/4715853/, 4, 2022. Quark Matter 2022 Conference. * [74] ALICE Collaboration, S. Acharya et al., “Prompt D0, D+, and D∗+ production in Pb–Pb collisions at $\sqrt{s_{\mathrm{NN}}}$ = 5.02 TeV”, JHEP 01 (2022) , arXiv:2110.09420 [nucl-ex]. * [75] ALICE Collaboration, S. Acharya et al., “Measurement of beauty production via non-prompt ${\rm D}^{0}$ mesons in Pb-Pb collisions at $\sqrt{s_{\rm NN}}$ = 5.02 TeV”, arXiv:2202.00815 [nucl-ex]. * [76] ALICE Collaboration, S. Acharya et al., “Measurement of electrons from semileptonic heavy-flavour hadron decays at midrapidity in pp and Pb-Pb collisions at $\sqrt{s_{\rm{NN}}}$ = 5.02 TeV”, Phys. Lett. B 804 (2020) , arXiv:1910.09110 [nucl-ex]. * [77] M. Voutilainen, “Heavy quark jets at the LHC”, Int. J. Mod. Phys. A 30 no. 31, (2015) , arXiv:1509.05026 [hep-ex]. * [78] JETSCAPE Collaboration, S. Cao et al., “Determining the jet transport coefficient from inclusive hadron suppression measurements using Bayesian parameter estimation”, Phys. Rev. C 104 no. 2, (2021) , arXiv:2102.11337 [nucl-th]. * [79] JETSCAPE Collaboration, R. Ehlers et al., “Bayesian analysis of QGP jet transport using multi-scale modeling applied to inclusive hadron and reconstructed jet data”, in 29th International Conference on Ultra-relativistic Nucleus-Nucleus Collisions. 8, 2022. arXiv:2208.07950 [hep-ph]. * [80] S. Cao, A. Majumder, G.-Y. Qin, and C. Shen, “Drag Induced Radiation and Multi-Stage Effects in Heavy-Flavor Energy Loss”, Phys. Lett. B 793 (2019) , arXiv:1711.09053 [nucl-th]. * [81] G. D. Moore and D. Teaney, “How much do heavy quarks thermalize in a heavy ion collision?”, Phys. Rev. C 71 (2005) , arXiv:hep-ph/0412346. * [82] F. A. Dreyer, G. Soyez, and A. Takacs, “Quarks and gluons in the Lund plane”, JHEP 08 (2022) , arXiv:2112.09140 [hep-ph]. * [83] ATLAS Collaboration, “Dependence of the Jet Energy Scale on the Particle Content of Hadronic Jets in the ATLAS Detector Simulation”,. * [84] S. Sapeta and U. A. Wiedemann, “Jet hadrochemistry as a characteristics of jet quenching”, Eur. Phys. J. C 55 (2008) , arXiv:0707.3494 [hep-ph]. * [85] M. Attems, J. Brewer, G. M. Innocenti, A. Mazeliauskas, S. Park, W. van der Schee, and U. A. Wiedemann, “The medium-modified $g\to c\bar{c}$ splitting function in the BDMPS-Z formalism”, arXiv:2203.11241 [hep-ph]. * [86] Y. S. Lai, J. Mulligan, M. Płoskoń, and F. Ringer, “The information content of jet quenching and machine learning assisted observable design”, JHEP 10 (2022) , arXiv:2111.14589 [hep-ph]. * [87] A. Andreassen, P. T. Komiske, E. M. Metodiev, B. Nachman, and J. Thaler, “OmniFold: A Method to Simultaneously Unfold All Observables”, Phys. Rev. Lett. 124 no. 18, (2020) , arXiv:1911.09107 [hep-ph]. * [88] J. Mulligan and M. Ploskon, “Identifying groomed jet splittings in heavy-ion collisions”, Phys. Rev. C 102 no. 4, (2020) , arXiv:2006.01812 [hep-ph]. * [89] G. Nijs, W. van der Schee, U. Gürsoy, and R. Snellings, “Bayesian analysis of heavy ion collisions with the heavy ion computational framework Trajectum”, Phys. Rev. C 103 no. 5, (2021) , arXiv:2010.15134 [nucl-th]. * [90] Y.-T. Chien and R. Kunnawalkam Elayavalli, “Probing heavy ion collisions using quark and gluon jet substructure”, arXiv:1803.03589 [hep-ph]. * [91] L. Apolinário, N. F. Castro, M. Crispim Romão, J. G. Milhano, R. Pedro, and F. C. R. Peres, “Deep Learning for the classification of quenched jets”, JHEP 11 (2021) , arXiv:2106.08869 [hep-ph]. * [92] L. Liu, J. Velkovska, and M. Verweij, “Identifying quenched jets in heavy ion collisions with machine learning”, arXiv:2206.01628 [hep-ph]. * [93] B. Andersson, G. Gustafson, G. Ingelman, and T. Sjostrand, “Parton Fragmentation and String Dynamics”, Phys. Rept. 97 (1983) . * [94] ALICE Collaboration, S. Acharya et al., “$\Lambda^{+}_{c}$ Production and Baryon-to-Meson Ratios in pp and p-Pb Collisions at $\sqrt{s_{NN}}$=5.02 TeV at the LHC”, Phys. Rev. Lett. 127 no. 20, (2021) , arXiv:2011.06078 [nucl-ex]. * [95] Z. Tu, D. E. Kharzeev, and T. Ullrich, “Einstein-Podolsky-Rosen Paradox and Quantum Entanglement at Subnucleonic Scales”, Phys. Rev. Lett. 124 no. 6, (2020) , arXiv:1904.11974 [hep-ph]. * [96] F. A. Flor, G. Olinger, and R. Bellwied, “Flavour and Energy Dependence of Chemical Freeze-out Temperatures in Relativistic Heavy Ion Collisions from RHIC-BES to LHC Energies”, Phys. Lett. B 814 (2021) , arXiv:2009.14781 [nucl-ex]. * [97] ALICE Collaboration, S. Acharya et al., “Constraining hadronization mechanisms with $\rm\Lambda_{\rm c}^{+}$/D0 production ratios in Pb-Pb collisions at $\sqrt{s_{\rm NN}}=5.02$ TeV”, arXiv:2112.08156 [nucl-ex]. * [98] ALICE Collaboration, S. Acharya et al., “Production of $\Lambda$ and $K^{0}_{s}$ in jets in p–Pb collisions at $\sqrt{s_{NN}}$=5.02 TeV and pp collisions at $\sqrt{s}$=7 TeV”, Phys. Lett. B 827 (2022) , arXiv:2105.04890 [nucl-ex]. * [99] A. Andronic, P. Braun-Munzinger, K. Redlich, and J. Stachel, “Decoding the phase structure of QCD via particle production at high energy”, Nature 561 no. 7723, (2018) , arXiv:1710.09425 [nucl-th]. * [100] R. L. Jaffe, “Multi-Quark Hadrons. 1. The Phenomenology of (2 Quark 2 anti-Quark) Mesons”, Phys. Rev. D 15 (1977) . * [101] ALICE Collaboration, S. Acharya et al., “Measuring K${}^{0}_{\rm S}$K± interactions using Pb-Pb collisions at ${\sqrt{s_{\rm NN}}=2.76}$ TeV”, Phys. Lett. B 774 (2017) , arXiv:1705.04929 [nucl-ex]. * [102] ALICE Collaboration, S. Acharya et al., “Measuring K${}^{0}_{\rm S}$K± interactions using pp collisions at $\sqrt{s}=7$ TeV”, Phys. Lett. B 790 (2019) , arXiv:1809.07899 [nucl-ex]. * [103] ALICE Collaboration, S. Acharya et al., “Measuring K${}^{0}_{\rm S}$K± interactions using pp collisions at $\sqrt{s}=5$ and $\sqrt{s}=13$ TeV”, Phys. Lett. B 833 (2022) , arXiv:2111.06611 [nucl-ex]. * [104] LHCb Collaboration, R. Aaij et al., “Observation of an exotic narrow doubly charmed tetraquark”, Nature Phys. 18 no. 7, (2022) , arXiv:2109.01038 [hep-ex]. * [105] LHCb Collaboration, R. Aaij et al., “Study of the doubly charmed tetraquark $T_{cc}^{+}$”, Nature Commun. 13 no. 1, (2022) , arXiv:2109.01056 [hep-ex]. * [106] A. Andronic, P. Braun-Munzinger, M. K. Köhler, A. Mazeliauskas, K. Redlich, J. Stachel, and V. Vislavicius, “The multiple-charm hierarchy in the statistical hadronization model”, JHEP 07 (2021) , arXiv:2104.12754 [hep-ph]. * [107] LHCb Collaboration, R. Aaij et al., “Observation of the doubly charmed baryon $\Xi_{cc}^{++}$”, Phys. Rev. Lett. 119 no. 11, (2017) , arXiv:1707.01621 [hep-ex]. * [108] E. Hernández, J. Vijande, A. Valcarce, and J.-M. Richard, “Spectroscopy, lifetime and decay modes of the $T^{-}_{bb}$ tetraquark”, Phys. Lett. B 800 (2020) , arXiv:1910.13394 [hep-ph]. * [109] Z. Liu, H.-T. An, Z.-W. Liu, and X. Liu, “Where are the hidden-charm hexaquarks?”, Phys. Rev. D 105 no. 3, (2022) , arXiv:2112.02510 [hep-ph]. * [110] CMS Collaboration, A. M. Sirunyan et al., “Evidence for X(3872) in Pb-Pb Collisions and Studies of its Prompt Production at $\sqrt{s_{NN}}$=5.02 TeV”, Phys. Rev. Lett. 128 no. 3, (2022) , arXiv:2102.13048 [hep-ex]. * [111] A. Andronic, P. Braun-Munzinger, M. K. Köhler, K. Redlich, and J. Stachel, “Transverse momentum distributions of charmonium states with the statistical hadronization model”, Phys. Lett. B 797 (2019) , arXiv:1901.09200 [nucl-th]. * [112] ALICE Collaboration, A. Collaboration et al., “Unveiling the strong interaction among hadrons at the LHC”, Nature 588 (2020) , arXiv:2005.11495 [nucl-ex]. [Erratum: Nature 590, E13 (2021)]. * [113] ALICE Collaboration, B. Abelev et al., “Long-range angular correlations on the near and away side in $p$-Pb collisions at $\sqrt{s_{NN}}=5.02$ TeV”, Phys. Lett. B 719 (2013) , arXiv:1212.2001 [nucl-ex]. * [114] ALICE Collaboration, J. Adam et al., “Forward-central two-particle correlations in p-Pb collisions at $\sqrt{s_{\rm NN}}$ = 5.02 TeV”, Phys. Lett. B 753 (2016) , arXiv:1506.08032 [nucl-ex]. * [115] ALICE Collaboration, S. Acharya et al., “Constraints on jet quenching in p-Pb collisions at $\mathbf{\sqrt{s_{NN}}}$ = 5.02 TeV measured by the event-activity dependence of semi-inclusive hadron-jet distributions”, Phys. Lett. B 783 (2018) , arXiv:1712.05603 [nucl-ex]. * [116] K. Lee, J. Mulligan, M. Płoskoń, F. Ringer, and F. Yuan, “Machine learning-based jet and event classification at the Electron-Ion Collider with applications to hadron structure and spin physics”, arXiv:2210.06450 [hep-ph]. * [117] PHENIX Collaboration, C. Aidala et al., “Creation of quark–gluon plasma droplets with three distinct geometries”, Nature Phys. 15 no. 3, (2019) , arXiv:1805.02973 [nucl-ex]. * [118] B. Schenke, C. Shen, and P. Tribedy, “Running the gamut of high energy nuclear collisions”, Phys. Rev. C 102 no. 4, (2020) , arXiv:2005.14682 [nucl-th]. * [119] S. H. Lim, Q. Hu, R. Belmont, K. K. Hill, J. L. Nagle, and D. V. Perepelitsa, “Examination of flow and nonflow factorization methods in small collision systems”, Phys. Rev. C 100 no. 2, (2019) , arXiv:1902.11290 [nucl-th]. * [120] S. A. Voloshin, A. M. Poskanzer, and R. Snellings, “Collective phenomena in non-central nuclear collisions”, Landolt-Bornstein 23 (2010) , arXiv:0809.2949 [nucl-ex]. * [121] J. L. Nagle and W. A. Zajc, “Small System Collectivity in Relativistic Hadronic and Nuclear Collisions”, Ann. Rev. Nucl. Part. Sci. 68 (2018) , arXiv:1801.03477 [nucl-ex]. * [122] C. Bierlich and J. R. Christiansen, “Effects of color reconnection on hadron flavor observables”, Phys. Rev. D 92 no. 9, (2015) , arXiv:1507.02091 [hep-ph]. * [123] E. Annala, T. Gorda, A. Kurkela, J. Nättilä, and A. Vuorinen, “Evidence for quark-matter cores in massive neutron stars”, Nature Phys. 16 no. 9, (2020) , arXiv:1903.09121 [astro-ph.HE]. * [124] ALICE Collaboration, S. Acharya et al., “Hypertriton Production in p-Pb Collisions at $\sqrt{s_{NN}}$=5.02 TeV”, Phys. Rev. Lett. 128 no. 25, (2022) , arXiv:2107.10627 [nucl-ex]. * [125] F. Bellini and A. P. Kalweit, “Testing production scenarios for (anti-)(hyper-)nuclei with multiplicity-dependent measurements at the LHC”, Acta Phys. Polon. B 50 (2019) , arXiv:1907.06868 [hep-ph]. * [126] J. Haidenbauer and I. Vidana, “Structure of single-$\Lambda$ hypernuclei with chiral hyperon-nucleon potentials”, Eur. Phys. J. A 56 no. 2, (2020) , arXiv:1910.02695 [nucl-th]. * [127] M. Fairbairn, A. C. Kraan, D. A. Milstead, T. Sjostrand, P. Z. Skands, and T. Sloan, “Stable Massive Particles at Colliders”, Phys. Rept. 438 (2007) , arXiv:hep-ph/0611040. * [128] R. Bruce et al., “New physics searches with heavy-ion collisions at the CERN Large Hadron Collider”, J. Phys. G 47 no. 6, (2020) , arXiv:1812.07688 [hep-ph]. * [129] CMS Tracker Collaboration, W. Adam et al., “The effect of highly ionising particles on the CMS silicon strip tracker”, Nucl. Instrum. Meth. A 543 no. 2-3, (2005) . * [130] F. Donato, N. Fornengo, and P. Salati, “Anti-deuterons as a signature of supersymmetric dark matter”, Phys. Rev. D 62 (2000) , arXiv:hep-ph/9904481. * [131] R. Duperray, B. Baret, D. Maurin, G. Boudoul, A. Barrau, L. Derome, K. Protasov, and M. Buenerd, “Flux of light antimatter nuclei near Earth, induced by cosmic rays in the Galaxy and in the atmosphere”, Phys. Rev. D 71 (2005) , arXiv:astro-ph/0503544. * [132] M. W. Winkler and T. Linden, “Dark Matter Annihilation Can Produce a Detectable Antihelium Flux through $\bar{\Lambda}_{b}$ Decays”, Phys. Rev. Lett. 126 no. 10, (2021) , arXiv:2006.16251 [hep-ph]. ## Appendix A The ALICE-USA Collaboration N. Alizadehvandchali9, N. Apadula5, M. Arslandok13, C. Beattie13, R. Bellwied9, J.T. Blair8, F. Bock6, H. Bossi13, A. Bylinkin10, H. Caines13, I. Chakaberia5, M. Cherney3, T.M. CormierI,6, R. Cruz-Torres5, P. Dhankher4, D.U. Dixit4, R.J. Ehlers5,6, W. Fan5, M. Fasel6, F. Flor9, A.N. Flores8, D.R. Gangadharan9, E. Garcia-Solis2, A. Gautam10, E. Glimos11, V. Gonzalez12, A. Hamdi5, R. Hannigan8, J.W. Harris13, A. Harton2, H. Hassan6, L.B. Havener13, C. Hughes11, T.J. Humanic7, A. Hutson9, T. Isidori10, B. Jacak4,5, P.M. Jacobs5, F. Jonas6, A. Khatun10, M. Kim4, J.L. Klay1, S. Klein5, A.G. Knospe9, Y.S. Lai5, E.D. Lesser4, I. Likmeta9, A. Liu4, C. Loizides6, C. Markert8, J.L. Martinez9, A.S. Menon9, J.D. Mulligan5, A.I. Nambrath4, C. Nattrass11, N. Novitzky6, A.C. Oliveira Da Silva11, M.H. Oliver13, L. Pinsky9, M. Płoskoń5, M.G. Poghosyan6, C.A. Pruneau12, R.E. Quishpe9, S. Ragoni3, K.F. Read6,11, O.V. Rueda9, D. Sarkar12, M.H.P. Sas13, J. Schambach6, N.V. Schmidt6, A.R. Schmier11, J.E. Seger3, O. Sheibani9, N. Smirnov13, J. Song9, P.J. Steffanic11, J.D. Tapia Takaki10, C. Terrevoli9, D. Thomas8, A.R. Timmins9, S.A. Voloshin12, S.L. Weyhmiller13, J.R. Wright8 ## Affiliation Notes I Deceased ## Collaboration Institutes 1 California Polytechnic State University, San Luis Obispo, California 2 Chicago State University, Chicago, Illinois 3 Creighton University, Omaha, Nebraska 4 Department of Physics, University of California, Berkeley, California 5 Lawrence Berkeley National Laboratory, Berkeley, California 6 Oak Ridge National Laboratory, Oak Ridge, Tennessee 7 Ohio State University, Columbus, Ohio 8 The University of Texas at Austin, Austin, Texas 9 University of Houston, Houston, Texas 10 The University of Kansas, Lawrence, Kansas 11 University of Tennessee, Knoxville 12 Wayne State University, Detroit, Michigan 13 Yale University, New Haven, Connecticut
S-Dual of Maxwell Chern-Simons Theory Adi Armoni Department of Physics, Faculty of Science and Engineering Swansea University, SA2 8PP, UK <EMAIL_ADDRESS> Abstract We discuss the dynamics of three dimensional Maxwell theory coupled to a level $k$ Chern-Simons term. Motivated by S-duality in string theory we argue that the theory admits an S-dual description. The S-dual theory contains a non- gauge 1-form field, previously proposed by Deser and Jackiw [1] and a level $k$ $U(1)$ Chern-Simons term, ${\cal Z}_{\rm MCS}={\cal Z}_{\rm DJ}{\cal Z}_{\rm CS}$. The couplings to external electric and magnetic currents and their string theory realisations are also discussed. ## 1 Introduction Chern-Simons theory is vastly used in mathematical physics, in condensed matter physics and in string theory [2]. It was studied intensively in the past three decades, yet the dynamics of Yang-Mills Chern-Simons theory is not fully understood at the strong coupling regime. Four dimensional S-duality is an exact duality between two ${\cal N}=4$ super Yang-Mills theories, enabling us to calculate quantities in the strong coupling regime using a dual weakly coupled theory. In the Abelian case it reduces to the old electric-magnetic duality which swaps electric and magnetic fields $F\longleftrightarrow*F\,.$ (1.1) In 3d Abelian S-duality relates the electric field to a dual scalar $f\longleftrightarrow*d\phi\,.$ (1.2) The purpose of this note is to extend S-duality to 3d Maxwell Chern-Simons (MCS) theory, with either a compact or non-compact $U(1)$ gauge group. It should hold on any spin manifold. The Lagrangian of the theory is given by $L=-\frac{1}{2g^{2}}da_{e}\wedge*da_{e}+\frac{k}{4\pi}a_{e}\wedge da_{e}\,.$ (1.3) MCS theory contains a vector boson of mass $M=\frac{g^{2}k}{2\pi}$. At low energies the kinetic term is irrelevant and the theory flows to a pure level $k$ Chern-Simons theory. As explained in section (3) the theory admits a global $\mathbb{Z}_{k}$ 1-form symmetry generated by $G\equiv\exp\left(i\oint(a_{e}-\frac{1}{M}*da_{e})\right)\,.$ (1.4) When the theory is compactified on a torus, the global ${\mathbb{Z}_{k}}$ 1-form symmetry is spontaneously broken, resulting in $k$ degenerate vacua. Several attempts were made to find the S-dual of (1.3). In [1] Deser and Jackiw proposed a ’self-dual model’ (SDM) which describes a massive vector. While SDM describes a massive vector it does not admit a ${\mathbb{Z}_{k}}$ 1-form symmetry neither does it flow to a pure Chern-Simons theory at low energies, hence it cannot be an exact dual of MCS theory. A closely related problem concerns the open string dynamics on a certain Hanany-Witten brane configuration. It is well known [3, 4] that MCS theory lives on the left brane configuration of fig.(1). Type IIB S-duality maps the left configuration into the right configuration. Thus, knowing the field theory that lives on the right configuration will solve the problem of finding the S-dual. In early attempts [3, 4, 5] the authors found gauge theories with a fractional level Chern-Simons term. While the theories they found are classically equivalent to MCS, it cannot be the full answer, as it does not admit the symmetries nor the same dynamics as the electric theory. Fig.​ 1: The electric theory on the left brane configuration is Maxwell Chern- Simons. The magnetic theory, obtained by type IIB S-duality, lives on the right brane configuration. ## 2 Derivation of the duality We may use 4d S-duality between Maxwell theories to derive the 3d duality. In 4d a pure Maxwell theory with a coupling $g$ is dual to a pure Maxwell theory with a coupling $1/g$. Consider the following partition function ${\cal Z}=\int DF_{m}DA_{e}\exp i\int(-\frac{g^{2}}{2}F_{m}\wedge*F_{m}+F_{m}\wedge dA_{e})$ (2.1) $A_{e}$ is the ’electric’ gauge field, $F_{m}$ is a ’magnetic’ gauge invariant 2-form. $g$ is the ’electric’ gauge coupling. Upon integrating over $F_{m}$ we obtain the electric theory ${\cal Z}=\int DA_{e}\exp i\int\left(-\frac{1}{2g^{2}}dA_{e}\wedge*dA_{e}\right)\,.$ (2.2) If instead we integrate over $A_{e}$ we obtain ${\cal Z}=\int DF_{m}\delta(dF_{m})\exp i\int\left(-\frac{g^{2}}{2}F_{m}\wedge*F_{m}\right)\,,$ (2.3) hence it can be written in terms of $A_{m}$ such that $F_{m}=dA_{m}$. ${\cal Z}=\int DA_{m}\exp i\left(-\frac{g^{2}}{2}dA_{m}\wedge*dA_{m}\right)\,.$ (2.4) This is the magnetic theory dual to the electric theory. Let us use dimensional reduction of (2.1) in order to derive the 3d duality. Upon reducing to 3d the 4d 2-form $F_{m}$ becomes a 3d 2-form $f_{m}$ and 1-form $a_{m}$. The 4d gauge field $A_{e}$ becomes a 3d gauge field $a_{e}$ and a scalar $\phi_{e}$. The the 2-form $f_{m}$ and the scalar $\phi_{e}$ decouple from the rest of the action and admit ${\cal Z}=\int Df_{m}D\phi_{e}\exp i\int(-\frac{g^{2}}{2}f_{m}\wedge*f_{m}+f_{m}\wedge d\phi_{e})\,,$ (2.5) which leads to the well known S-duality $d\hat{a}_{m}\longleftrightarrow*d\phi_{e}\,,$ (2.6) where $f_{m}=d\hat{a}_{m}$. Let us focus on the duality between $a_{e}$ and $a_{m}$, which is the prime purpose of this note. We add to the action a Chern-Simons term111The Chern- Simons term can be obtained by a dimensional reduction of a space dependent theta term $\int d^{4}x\,\theta(x)F_{e}\wedge F_{e}$, such that $\theta(x)=\frac{k}{4\pi}H(x^{3})$, where $H(x^{3})$ is the Heaviside step function.. Our proposal is the following ’master’ partition function ${\cal Z}=\int Da_{m}Da_{e}\exp i\int(-\frac{g^{2}}{2}a_{m}\wedge*a_{m}+a_{m}\wedge da_{e}+\frac{k}{4\pi}a_{e}\wedge da_{e})$ (2.7) Note that $a_{m}$ is a gauge invariant 1-form. Upon integration over $a_{m}$ we obtain the electric theory ${\cal Z}=\int Da_{e}\exp i\int(-\frac{1}{2g^{2}}da_{e}\wedge*da_{e}+\frac{k}{4\pi}a_{e}\wedge da_{e})\,,$ (2.8) namely Maxwell Chern-Simons theory. In order to derive the magnetic theory we should use (2.7) and integrate over $a_{e}$. This is a subtle point. Instead, let us use a change of variables $a_{e}=b-\frac{2\pi}{k}a_{m}$, to obtain the following partition function ${\cal Z}=\int Da_{m}Db\exp i\int(-\frac{g^{2}}{2}a_{m}\wedge*a_{m}-\frac{\pi}{k}a_{m}\wedge da_{m}+\frac{k}{4\pi}b\wedge db).$ (2.9) Equation (2.9) is our proposal for the S-dual of Maxwell Chern-Simons theory. The partition function of the magnetic theory is a product of the Deser-Jackiw theory and a level $k$ Chern-Simons term ${\cal Z}_{\rm MCS}={\cal Z}_{\rm DJ}{\cal Z}_{\rm CS}\,.$ (2.10) Note that $a_{m}$ is not a gauge field and therefore the term $\frac{\pi}{k}a_{m}\wedge da_{m}$ is not ill-defined. Both the electric and the magnetic theories describe a massive vector of mass $M=\frac{g^{2}k}{2\pi}$ and a decoupled level $k$ Chern-Simons theory. Both theories exhibit a 1-form ${\mathbb{Z}_{k}}$ symmetry. Let us now provide another argument in favour of our proposal (2.9). We begin with the magnetic brane configuration of fig.(1). It was argued by Gaiotto and Witten [6] that the theory which lives on the intersection of the 3-brane and the tilted 5-brane (without a D5 brane) is ${\cal Z}=\int DaDc\exp i\int(\frac{1}{2\pi}a\wedge dc+\frac{k}{4\pi}c\wedge dc)\,.$ (2.11) In order to understand what happens when we add a D5 brane, let us assume that the terms that we need to add to the action are $k$ independent. Indeed, the information about $k$ is encoded in the tilted fivebrane, not in the threebrane. Let us use $k=0$, since in this case the duality is well understood: the electric theory is pure Maxwell and the magnetic (mirror) theory is a massless scalar. The brane realisation of the duality was provided in the seminal work of Hanany and Witten [7]. We may write the theory of a free massless scalar as follows ${\cal Z}=\int DaDc\exp i\int(a\wedge*a+\frac{1}{2\pi}a\wedge dc)\,,$ (2.12) with $a$ being a gauge invariant 1-form. The equation of motion for $c$ is $*da=0$, namely that $a=d\chi$. Thus, for $k=0$ we obtain a theory of a free scalar $(d\chi)^{2}$, as expected. We found that adding a term $a\wedge*a$ to the action yields a theory that describes the correct dual of Maxwell theory. We propose that $a\wedge*a$ is the missing term in (2.11), namely that by adding it to (2.11) we obtain the dual of MCS for any $k$. Note that ${\cal Z}=\int DaDc\exp i\int(a\wedge*a+\frac{1}{2\pi}a\wedge dc+\frac{k}{4\pi}c\wedge dc)\,$ (2.13) is almost identical to (2.7). An important difference is that Gaitto and Witten introduced a gauge field $a$, whereas in (2.13) we added a term that breaks gauge invariance. We may re-introduce gauge invariance in (2.13) by transforming the fixed gauge vector $a$ into a gauge invariant term by adding a scalar $\eta$ as follows ${\cal Z}=\int DaDcD\eta\exp i\int((a-d\eta)\wedge*(a-d\eta)+\frac{1}{2\pi}a\wedge dc+\frac{k}{4\pi}c\wedge dc)\,\,,$ (2.14) such that under a gauge transformation $a\rightarrow a+d\lambda,\eta\rightarrow\eta+\lambda$, with $a$ a $U(1)$ gauge field. Eq.(2.13) may be viewed as the fixed gauge version of eq.(2.14) with $d\eta=0$. Our proposal (2.9) passes all the requirements from a dual theory: it admits a ${\mathbb{Z}_{k}}$ global symmetry, it flows to pure Chern-Simons theory in the IR, it contains a massive vector of mass $M$ and, finally, when $k=0$ it agrees with the results of Hanany and Witten [7]. As we shall see the brane realisations of both electric and magnetic theories predict the existence of $k$ degenerate vacua. We summarize this section by writing the precise map between the electric and magnetic variables using (2.7) $\displaystyle-g^{2}a_{m}=*da_{e}$ (2.15) $\displaystyle b=a_{e}-\frac{1}{M}*da_{e}$ (2.16) or $a_{e}=b-\frac{2\pi}{k}a_{m}$ (2.17) ## 3 Comments on ${\mathbb{Z}_{k}}$ Let us introduce a Wilson loop in MCS theory. We wish to measure the ${\mathbb{Z}_{k}}$ charge of the loop, namely the number of fundamental strings, $n$, that pass through a a certain contour $C$. We will define an operator $G$ such that $GW_{n}=\exp(i\frac{2\pi n}{k})W_{n}\,,$ (3.1) with $W_{n}$ a Wilson loop of charge $n$, $W_{n}=\exp(in\oint a_{e})$. In order to the define $G$, let us consider the equation of motion in MCS $d(\frac{1}{g^{2}}*da_{e}-\frac{k}{2\pi}a_{e})=j_{e}\equiv dJ_{e}\,,$ (3.2) where $J_{e}$ is the integral of the electric current $j_{e}$ over a disc $D$ such that $C=\partial D$. The setup is depicted in fig.(2). By integrating (3.2) we learn that $\frac{1}{g^{2}}*da_{e}-\frac{k}{2\pi}a_{e}=J_{e}\,.$ (3.3) Fig.​ 2: A Wilson loop passing through a domain $D$ (shaded region) whose boundary is $C=\partial D$. We can therefore define a generator of a ${\mathbb{Z}_{k}}$ symmetry as follows $G=\exp\left(i\frac{2\pi}{k}\oint_{C}(\frac{k}{2\pi}a_{e}-\frac{1}{g^{2}}*da_{e})\right)=\exp\left(i\oint_{C}b\right).$ (3.4) Note that the implication of ${\mathbb{Z}_{k}}$ symmetry is a symmetry $n\rightarrow n+k$, namely that a collection of $k$ strings is topologically isomorphic to a singlet, namely to no strings at all. This is supported by string theory: suppose that we attempt to place the endpoints of $k$ coincident strings on the worldvolume of the D3 brane. The collection of $k$ fundamental strings can transform itself into an anti D-string and a $(k,1)$ string. Instead of ending on the worldvolume of the D3 brane, the D-string can end on an NS5 brane and a $(k,1)$ string can end on a $(1,k)$ fivebrane. Thus string theory predicts that a collection of $k$ strings can be removed from the worldvolume of the 3d gauge theory. A similar phenomenon happens in the magnetic dual if we attempt to introduce $k$ coincident D-strings in the worldvolume of the magnetic theory. When the theory is defined on the torus the ${\mathbb{Z}_{k}}$ symmetry is broken, resulting in $k$ vacua [8][9]. An intuitive explanation is as follows: The level $k$ $U(1)$ Chern-Simons theory is equivalent (using level-rank duality) to a level $1$ $SU(k)$ theory, that admits a ${\mathbb{Z}_{k}}$ centre symmetry. When it is defined on the torus the $SU(k)$ theory deconfines, resulting in $k$ degenerate vacua, parameterized by the eigenvalues of the ’t Hooft loop. The $k$ vacua manifests themselves in both the electric and magnetic brane configurations as follows: the D3 brane may end on any of the $k$ constituents of the fivebranes. Each one of the $k$ choices corresponds to a vacuum. ## 4 Coupling to external sources, Wilson and magnetic loops Consider the coupling of the electric gauge field to a source $j_{e}$, namely $a_{e}j_{e}$. It translates to the coupling $(b-\frac{\pi}{k}a_{m})j_{e}$ in the magnetic side. We therefore suggest that the Wilson loop $W_{e}=\exp i\oint a_{e}$ (4.1) in the electric side is mapped into a magnetic loop of the form $M_{m}=\exp i\oint(b-\frac{2\pi}{k}a_{m})$ (4.2) in the magnetic side. We may use the above map between the Wilson loop in the electric side and its magnetic counterpart to study the dynamics of 3d QED-CS. Using the worldline formalism [10] we can write the partition function of MCS theory coupled to $N_{f}$ massless fields as follows ${\cal Z}_{\rm QED-CS}=\int Da_{e}\exp(iS_{\rm MCS})\sum_{n}\frac{(N_{f}\Gamma_{e})^{n}}{n!}$ (4.3) with $\Gamma_{e}=\int\frac{dt}{t^{\frac{5}{2}}}\int Dx\exp(-\int_{0}^{t}d\tau(\dot{x})^{2})\exp i\oint a_{e}$ (4.4) The duality transformation yields the following partition function ${\cal Z}_{\rm magnetic}=\int Da_{m}Db\exp(iS_{\rm DJ- CS})\sum_{n}\frac{(N_{f}\Gamma_{m})^{n}}{n!}$ (4.5) with $\Gamma_{m}=\int\frac{dt}{t^{\frac{5}{2}}}\int Dx\exp(-\int d\tau(\dot{x})^{2})\exp i\oint(b-\frac{2\pi}{k}a_{m})\,.$ (4.6) It suggests that the dynamics of QED with $N_{f}$ massless flavours is captured by a dual DJ-CS theory coupled to $N_{f}$ massless ’monopoles’. The precise coupling of $a_{m}$ and $b$ to the monopoles is given by (4.6). We may write the dual magnetic theory in a more ’standard’ form ${\cal Z}=\int Da_{m}DbD\bar{\psi}_{m}D\psi_{m}\exp iS_{\rm magnetic}\,,$ (4.7) with $S_{magnetic}$ given by $S_{magnetic}=\int\left(-\frac{g^{2}}{2}a_{m}\wedge*a_{m}-\frac{\pi}{k}a_{m}\wedge da_{m}+\frac{k}{4\pi}b\wedge db+\bar{\psi}_{m}\gamma\wedge\star(i\partial+b-\frac{2\pi}{k}a_{m})\psi_{m}\right)\,.$ (4.8) It is interesting to note that the QED-CS theory is mapped to a theory of interacting massless magnetic ’monopoles’, with a coupling $1/gk$. Thus, when the electrons couple strongly to $a_{e}$, the ’monopoles’ couple weakly to $a_{m}$ and we may use perturbation in theory in the magnetic side to study the strongly coupled electric theory. Following Itzhaki[11] let us define a magnetic (“disorder”) loop in the electric theory $M_{e}=\exp\left(i\oint_{C}(ka_{e}-\frac{2\pi}{g^{2}}*da_{e})\right)\,,$ (4.9) which is mapped into the electric loop in the magnetic side $W_{m}=\exp\left(ik\oint_{C}b\right).$ (4.10) The magnetic loop in the electric side and the electric (Wilson) loop in the magnetic side are trivial [11]. Fig.​ 3: Rectangular Wilson loops can be realised in string theory by ending a pair of F-string and anti F-string on the the threebrane. Similarly Rectangular ’t Hooft loops can be realised by ending a pair of D-string and anti D-string on the three brane. The end of the F-string represents a heavy quark, whereas the end of the D-string represents a heavy monopole. We suggest that a rectangular Wilson loop (or magnetic loop) should be identified with the end points of an F-string and anti F-string (or a D-string and anti D-string) that end on the threebranes of fig.(3). A D-string can end on an NS5 brane instead of a threebrane, hence the magnetic loop in the electric theory should be trivial. Similarly, a F-string can end on a D5 brane instead of a threebrane, hence a Wilson loop in the magnetic theory should be trivial. This is consistent with our definitions of the magnetic loop (4.9) and the Wilson loop (4.10). ## 5 Summary The purpose of this note is to find the S-dual of MCS theory. We found that the dual theory (2.9) contains a non-gauge vector of mass $M$ and a decoupled pure TQFT. The magnetic theory nicely captures the dynamics of the electric theory: a theory with a mass gap that flows in the IR to a TQFT. The duality we uncovered in this note is a precise manifestation of the duality between a topological insulator and a topological superconductor outlined in ref.[12]. It will be interesting to find the S-dual of the non-Abelian $U(N)$ theory that lives on a collection of $N$ coincident $D3$ branes, suspended between tilted fivebranes. The master field of that theory may be obtained by replacing the Abelian 1-forms of (2.7) by non-Abelian 1-forms as follows222I thank Shigeki Sugimoto for suggesting that. ${\cal Z}=\int Da_{m}Da_{e}\exp i\,{\rm tr}\int(-\frac{g^{2}}{2}a_{m}\wedge*a_{m}+a_{m}\wedge(da_{e}+a_{e}\wedge a_{e})+\frac{k}{4\pi}(a_{e}\wedge da_{e}+\frac{2}{3}a_{e}\wedge a_{e}\wedge a_{e}))$ (5.1) together with $a_{e}=b-\frac{2\pi}{k}a_{m}$. Other dualities that involve $SO/Sp$ (and an orientifold in string theory) could also be derived. The generalisation to supersymmetric QED/QCD theories with a CS term [5] is also interesting and can be written down using the worldline formalism, as in section (4). The duality found in this letter is useful to study the strong coupling regime of those theories. Finally, it is well known that MCS theory admits Seiberg duality. The manifestation of the duality using an exchange of fivebranes in the magnetic theory, might teach us about fivebranes dynamics. ### Acknowledgements A.A. would like to thank the Yukawa institute for theoretical physics, where part of this work was done, for warm hospitality. I would also like to thank Mohammad Akhond and Shigeki Sugimoto for numerous discussions and collaboration. I am grateful to Anton Kapustin and Zohar Komargodski for reading and commenting on a draft version of this paper. ## References * [1] S. Deser and R. Jackiw, “’Selfduality’ of Topologically Massive Gauge Theories,” Phys. Lett. B 139 (1984), 371-373 doi:10.1016/0370-2693(84)91833-1 * [2] G. V. Dunne, “Aspects of Chern-Simons theory,” [arXiv:hep-th/9902115 [hep-th]]. * [3] T. Kitao and N. Ohta, “Spectrum of Maxwell-Chern-Simons theory realized on type IIB Brane configurations,” Nucl. Phys. B 578 (2000), 215-238 doi:10.1016/S0550-3213(99)00715-4 [arXiv:hep-th/9908006 [hep-th]]. * [4] O. Bergman, A. Hanany, A. Karch and B. Kol, “Branes and supersymmetry breaking in three-dimensional gauge theories,” JHEP 10 (1999), 036 doi:10.1088/1126-6708/1999/10/036 [arXiv:hep-th/9908075 [hep-th]]. * [5] A. Kapustin and M. J. Strassler, “On mirror symmetry in three-dimensional Abelian gauge theories,” JHEP 04 (1999), 021 doi:10.1088/1126-6708/1999/04/021 [arXiv:hep-th/9902033 [hep-th]]. * [6] D. Gaiotto and E. Witten, “S-Duality of Boundary Conditions In N=4 Super Yang-Mills Theory,” Adv. Theor. Math. Phys. 13 (2009) no.3, 721-896 doi:10.4310/ATMP.2009.v13.n3.a5 [arXiv:0807.3720 [hep-th]]. * [7] A. Hanany and E. Witten, “Type IIB superstrings, BPS monopoles, and three-dimensional gauge dynamics,” Nucl. Phys. B 492 (1997), 152-190 doi:10.1016/S0550-3213(97)00157-0 [arXiv:hep-th/9611230 [hep-th]]. * [8] C. L. Ho and Y. Hosotani, “Operator algebra in Chern-Simons theory on a torus,” Phys. Rev. Lett. 70 (1993), 1360-1363 doi:10.1103/PhysRevLett.70.1360 [arXiv:hep-th/9210103 [hep-th]]. * [9] D. Tong, Lectures notes, https://www.damtp.cam.ac.uk/user/tong/gaugetheory/83d.pdf * [10] M. J. Strassler, “Field theory without Feynman diagrams: One loop effective actions,” Nucl. Phys. B 385 (1992), 145-184 doi:10.1016/0550-3213(92)90098-V [arXiv:hep-ph/9205205 [hep-ph]]. * [11] N. Itzhaki, “Anyons, ’t Hooft loops and a generalized connection in three-dimensions,” Phys. Rev. D 67 (2003), 065008 doi:10.1103/PhysRevD.67.065008 [arXiv:hep-th/0211140 [hep-th]]. [12] * [12] J. Murugan and H. Nastase, JHEP 05 (2017), 159 doi:10.1007/JHEP05(2017)159 [arXiv:1606.01912 [hep-th]].
11institutetext: Univ. Grenoble Alpes, CNRS, IPAG, 38000 Grenoble, France 22institutetext: Observatoire de Paris, PSL University, Sorbonne Université, CNRS UMR 8112, LERMA, 61 Avenue de l’Observatoire, 75014 Paris, France 33institutetext: Departamento de Astronomia de Chile, Universidad de Chile, Santiago, Chile 44institutetext: Instituto de Radioastronomia y Astrofisica, Universidad Nacional Autónoma de México, P.O. Box 3-72, 58090, Morelia, Michoacán, México # Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind. A. de Valon Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind.Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind. C. Dougados Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind.Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind. S. Cabrit Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind. and Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind.Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind. and Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind. F. Louvet Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind. and Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind.Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind. and Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind. L. A. Zapata Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind.Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind. D. Mardones Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind.Modeling the CO outflow in DG Tau B: Swept-up shells versus perturbed MHD disk wind. ###### Abstract Context. The origin of outflows and their exact impact on disk evolution and planet formation remain crucial open questions. DG Tau B is a Class I protostar associated with a rotating conical CO outflow and a structured disk. Hence it is an ideal target to study these questions. Aims. We aim to characterize the morphology and kinematics of the DG Tau B outflow in order to elucidate its origin and potential impact on the disk. Methods. Our analysis is based on Atacama Large Millimeter Array (ALMA) 12CO(2-1) observations of DG Tau B at 0.15′′ (20 au) angular resolution. We developed a tomographic method to recover 2D (R,Z) maps of vertical velocity $V_{\rm Z}$ and specific angular momentum $j=R\times V_{\phi}$. We created synthetic data cubes for parametric models of wind-driven shells and disk winds, which we fit to the observed channel maps. Results. Tomographic analysis of the bright inner conical outflow shows that both $V_{\rm Z}$ and $j$ remain roughly constant along conical surfaces, defining a shear-like structure. We characterize three different types of substructures in this outflow (arches, fingers, and cusps) with apparent acceleration. Wind-driven shell models with a Hubble law fail to explain these substructures. In contrast, both the morphology and kinematics of the conical flow can be explained by a steady conical magnetohydrodynamic (MHD) disk wind with foot-point radii $r_{0}\simeq 0.7-3.4$ au, a small magnetic level arm parameter ($\lambda\leq 1.6$), and quasi periodic brightness enhancements. These might be caused by the impact of jet bow shocks, source orbital motion caused by a 25 MJ companion at 50 au, or disk density perturbations accreting through the wind launching region. The large CO wind mass flux (four times the accretion rate onto the central star) can also be explained if the MHD disk wind removes most of the angular momentum required for steady disk accretion. Conclusions. Our results provide the strongest evidence so far for the presence of massive MHD disk winds in Class I sources with residual infall, and they suggest that the initial stages of planet formation take place in a highly dynamic environment. ###### Key Words.: stars: formation – protoplanetary disk – ISM : jets and outflows – stars : individual: DG Tau B ## 1 Introduction Understanding the origin of protostellar flows is a key element to our full comprehension of the star formation process. Protostellar flows come in two components: high speed collimated jets and slower, often less collimated winds and outflows. We focus here on the slow molecular outflows which are traditionally associated with the earlier stages of star formation. However, they have also recently been detected around more evolved Class II systems (e.g., Pety et al. 2006; Louvet et al. 2018; Fernández-López et al. 2020). Despite their ubiquity, the exact origin of molecular outflows, their link to the high-velocity jets, and their impact on the young forming star and disk are still crucial open questions. Two main paradigms are currently considered. The first traditional model describes these slow outflows as swept-up material, tracing the interaction between an inner jet or a wide-angle wind with the infalling envelope or parent core. These models have been mainly used for interpreting outflows from Class 0 and I stars which are still surrounded by massive envelopes (Zhang et al. 2019; Shang et al. 2020; Lee et al. 2000). However, on small scales (less than a few 1000 au), recent observations have revealed rotating molecular outflows to originate from well within the disk at all evolutionary stages from Class 0 to Class II (e.g., Launhardt et al. 2009; Zapata et al. 2015; Bjerkeli et al. 2016; Tabone et al. 2017; Hirota et al. 2017; Louvet et al. 2018; Zhang et al. 2018; Lee et al. 2018; de Valon et al. 2020; Lee et al. 2021). These observations suggest an alternative paradigm by which these slow molecular outflows, at least at their base, would trace matter directly ejected from the disk, by thermal or magnetic processes. In support of this interpretation, the flow rotation signatures are consistent with an origin from disk radii $r_{0}\simeq 1-50$ au (see references above), where Panoglou et al. (2012) have shown that magnetic disk winds could remain molecular. These two paradigms imply different evolutions for the disk. Jet- and wind- driven shell models predict that an important mass is swept up from the envelope, impacting the reservoir of matter infalling onto the disk. Disk wind models predict an extraction of mass from the disk and, in the case of magnetohydrodynamic (MHD) models, an extraction of angular momentum which could drive disk accretion (Bai et al. 2016). Evaluating the contributions of each of these mechanisms to the slow molecular outflow emission requires high- angular resolution studies of the molecular outflow base. The flux of angular momentum extracted by the rotating molecular wind is estimated in only two sources so far, HH30 and HH212, and it is found sufficient to drive disk accretion at the current observed rate in both cases (Louvet et al. 2018; Tabone et al. 2020). Recent high-angular resolution observations also reveal striking signatures of multiple CO shells in a few sources (Zhang et al. 2019; Fernández-López et al. 2020). Under the classical paradigm where CO outflows trace swept-up shells, they would require short episodic wind and jet outbursts every few 100 yrs. Characterizing and understanding the origin of these variabilities could bring critical insights into the star formation dynamics. We present here an analysis of the DG Tau B CO outflow, based on recent Atacama Large Millimeter Array (ALMA) observations at 0.15′′ resolution by de Valon et al. (2020) (hereafter DV20). DG Tau B is a Class I 1.1 M⊙ protostar located in the Taurus cloud ($\approx$ 140 pc) and is associated with a bipolar atomic jet (Mundt & Fried 1983) and a strongly asymmetric CO outflow first mapped by Mitchell et al. (1997). The bright redshifted CO outflow lobe displays a striking bright and narrow conical shape at its base. Zapata et al. (2015) detect rotation signatures in the same sense as the disk. The ALMA observations by DV20 clearly confirm rotation in the bright inner conical redshifted lobe and show that it is surrounded by a wider and slower outflow. Residual infall signatures are detected at opening angles $\geq 70^{\circ}$, almost tangent to the disk surface. In addition, DV20 report striking substructures in the CO channel maps at different line-of-sight velocities, reminiscent of the nested layers recently identified in HH46/47 by Zhang et al. (2019) and suggesting variability or interaction processes. The exquisite levels of detail provided by these new ALMA observations provide a prime opportunity to distinguish between swept-up and disk wind origins. In Sect. 2, we recall the main properties of the DG Tau B outflow and characterize the three types of substructures visible in the channel maps. In Sect. 3 we present a model-independent analysis of the inner conical outflow component, which allowed us to retrieve 2D maps of the expansion velocity $V_{\rm Z}$ and specific angular momentum $j=RV_{\phi}$. We compare these overall properties with parametric models of wind-driven shells (in Sect. 4) and disk winds (in Sect. 5). We discuss our results and their implications for the origin of the CO outflows in DG Tau B in Sect. 6. Section 7 summarizes our conclusions. ## 2 Summary of outflow structure Figure 1: Substructures in the DG Tau B redshifted CO outflow. Left panels: 12CO channel maps at selected line-of-sight velocities. The white dashed line traces the $\theta=17^{\circ}$ outer limiting cone of the inner outflow as defined by DV20. Right panels: Transverse PV diagrams across the flow averaged over a slice of $\Delta Z=0.2^{\prime\prime}$. Black dotted lines indicate the outer limits of the conical outflow at the specified height. The triangle (a) and the two circle symbols (c and d) highlight the on-axis height of one arch and two different cusps, respectively. The square symbol (b) is located in the extended outer flow. The symbols are represented both on the channel maps and PV diagrams. Figure 1 summarizes the main properties of the DG Tau B redshifted CO outflow as identified in DV20. A narrow, limb brightened conical outflow is visible in the channel maps at $(V-V_{\rm sys})>2.0$ km s-1. Its opening angle decreases from 17∘ at ($V-V_{\rm sys})=2.8$ km s-1 until 12∘ at $(V-V_{\rm sys})\geq 5.0$ km s-1. The sheer-like velocity gradient across the conical layer is best seen in transverse position-velocity (hereafter PV) cuts (Fig. 1), where the flow width clearly narrows down at higher velocity, up $(V-V_{\rm sys})>5.0$ km s-1. The conical outflow is surrounded by a slower and wider outflowing component visible at $(V-V_{\rm sys})<2$ km s-1. This outer flow is visible in PV diagrams as an extended pedestal with a shallower velocity gradient (see Fig. 1). Another striking property of the DG Tau B redshifted outflow, revealed by the ALMA observations in DV20, are brightness enhancements visible in channel maps. These various substructures, illustrated in Fig. 1, can be classified in three types: Bow-shaped intensity enhancements are visible at low velocities $(V-V_{\rm sys})=0.88-1.51$ km s-1 (see Fig. 1, left panels). We refer to these substructures as arches. The radial extent of the biggest arch is larger than the inner conical outflow, implying that this arch is at least partially formed outside of the conical outflow. The arch seems to increase in height with increasing velocity although this phenomenon is difficult to quantify due to the limited spectral sampling. At intermediate velocities, from $(V-V_{\rm sys})=2.46-3.42$ km s-1, thin quasi vertical lines (see white arrow in Fig. 1) are visible inside the conical outflow, close to the edge. They are almost vertical at $(V-V_{\rm sys})=2.46$ km s-1 and more open at higher projected velocity until they become almost tangent to the edge of the conical outflow at $(V-V_{\rm sys})\leavevmode\nobreak\ =\leavevmode\nobreak\ 3.42$ km s-1. We refer to these substructures as fingers. At high velocities, from $(V-V_{\rm sys})=3.1$ to almost 7 km s-1, multiple U-shaped structures are visible inside the conical outflow. We refer to these substructures as cusps. The contrast of these cusps is maximal at $(V-V_{\rm sys})=4.37$ km s-1 and decreases with increasing velocity. The cusps show a signature of apparent acceleration: their projected distance from the source increases with increasing projected velocity. Table 1: Characteristics of the observed arches and cusps Arches --- Name | position at | Radial extension at | Aspect Ratio at | $V_{\rm A}$aa$a$$(V-V_{\rm sys})=1.19$ km s-1(”) | $V_{\rm A}$aa$a$$(V-V_{\rm sys})=1.19$ km s-1(”) | $V_{\rm A}$aa$a$$(V-V_{\rm sys})=1.19$ km s-1 A0 | 14.9$\pm$ 0.2 | 9.7$\pm$ 0.3 | 1.4 A1 | 9.4$\pm$ 0.2 | ? | ? A2 | 6.2$\pm$ 0.2 | 1.5$\pm$ 0.3 | 1.2 A3 | 3.7$\pm$ 0.2 | 1$\pm$ 0.2 | 1.4 Cusps Name | position at | $N_{\rm chan}$ | derivative | $V_{\rm U}$bb$b$$(V-V_{\rm sys})=4.37$ km s-1(”) | | (”/km s-1) U0 | 11.9 $\pm$ 0.2 | (5) | 3.2 $\pm$ 0.5 U1 | 8.2 $\pm$ 0.2 | (6) | 2.2 $\pm$ 0.3 U2 | 6.5 $\pm$ 0.1 | (4) | 1.4 $\pm$ 0.5 U3 | 5.0 $\pm$ 0.1 | (6) | 0.7 $\pm$ 0.4 U4 | 3.6 $\pm$ 0.1 | (3) | 0.5 $\pm$ 0.2 U5 | 2.2 $\pm$ 0.1 | (4) | 0.4 $\pm$ 0.2 111 Table 1 lists the characteristics of the main arches and cusps. We identify four arches (A0 to A3) and six cusps (U0 to U5). On the channel map at $(V-V_{\rm sys})=1.19$ km s-1 we derived the maximal height of each arch on axis (at $\delta x=0$) and the maximal radial extension. We divided these two values to derive the arch aspect ratio. The cusps are also characterized from the channel maps at $3.73\boldsymbol{\leq}(V-V_{\rm sys})\leq 5.32$ km s-1 (See Fig. 22). At higher velocities, the cusps could not be characterized because the outflow signal-to-noise ratio (S/N) decreases drastically. Moreover, the region at $\delta z<2.2^{\prime\prime}$ was not studied because the cusps locations are complex to identify due to overlapping structures. We derived the cusp reference height on-axis on the channel map at ($V-V_{\rm sys}$) = 4.37 km s-1. The apparent acceleration of each cusp, in (′′)/km s-1 listed in Table 1, was obtained by measuring the average spatial shift of the cusp between two consecutive channel maps (taking as error bar the rms dispersion between measurements in different channels). Internal discrete structures are also visible in transverse PV-diagrams as pseudo-ellipses (see Fig. 1). The top and bottom of the ellipses seem to match with respectively the top of some arches and bottom of some cusps (see Fig. 1). This potentially indicates that cusps and arches are linked to the same phenomenon. We present a model-dependent study of these ellipses in Sect. 5. ## 3 Tomography of the inner conical outflow In this section, we develop a model-independent method that allows us to recover the dynamics and the morphology of the inner conical outflow component. This method assumes that the outflow is axisymmetric. We later discuss possible departures from axisymmetry and their implications on the analysis conducted here. ### 3.1 Method We followed Louvet et al. (2018) who modeled the outflow of HH30 at a given vertical offset by an emitting ring with radius R and extended their method to take into account the inclination of the outflow. For this purpose, we defined the outflow and the observer reference systems (see Fig. 2). On the outflow reference system, $\@vec{Z}$ is defined by the outflow axis and $\@vec{X}$ is tangent to the plane of sky. The observer reference system is defined by $\@vec{\delta z}$ the projection of the outflow axis onto the plane of the sky, $\@vec{\delta y}$, the line-of-sight direction, and $\@vec{\delta x}=\@vec{X}$, in the plane of the sky. The inclination of the outflow $i$ is then defined by the angle between $\@vec{\delta y}$ and $\@vec{Z}$. In the case of edge-on disks such as HH30, the two reference systems are identical. We modeled one layer of the outflow at a specified height Z by an emitting ring of radius R and azimuthal angle $\phi$ (see Fig. 2) with $X=R\cos{\phi}$ and $Y=R\sin{\phi}$. For each ring the velocity components are defined in cylindrical coordinates with: $V_{\rm R}(R,Z)$, $V_{\rm Z}(R,Z)$ and $V_{\phi}(R,Z)$ (see Fig. 2). Hence, an emitting ring is defined by 5 parameters: $Z$, $R$, $V_{\rm R}(R,Z)$, $V_{\rm Z}(R,Z)$ and $V_{\phi}(R,Z)$. Figure 2: Principle of tomographic reconstruction method. Top panel: 3D representation of the two reference systems used: the outflow $(X,Y,Z)$ in black and the observer $(\delta x,\delta y,\delta z)$ in yellow. The colored circles illustrate the emitting rings defined by five parameters (see text). The colored dots trace the locations at $\phi=0$ and $\phi=\pi$ along the emitting rings. Bottom panels: Transverse PV diagrams at $\delta z=2.5^{\prime\prime}$ across the flow axis and averaged over a slice of $\Delta z=0.2^{\prime\prime}$. In the left panel is shown the schematic projection of the colored rings and corresponding colored dots in the PV diagram. Because of the flow inclination, the rings are in fact projected at slightly different heights. Their real projection is studied in Sect. D.1. The white dots in the right panel illustrate the outer limits of the PV diagram. The red curve shows the polynomial fitting of the two edges. The observational coordinates on a position-position-velocity (PPV) data cube are defined by the projection of the outflow on the plane of sky ($\@vec{\delta x},\@vec{\delta z}$) and the projected velocities on the line of sight $V_{los}=-\@vec{V}\cdot\@vec{e_{\rm y}}$ with redshifted velocities considered as positive. This depends on $R$,$Z$ and $\phi$ as: $\displaystyle\delta x$ $\displaystyle=$ $\displaystyle R\cos{\phi}$ (1) $\displaystyle\delta z$ $\displaystyle=$ $\displaystyle Z\sin{i}-R\sin{\phi}\cos{i}$ (2) $\displaystyle V_{\rm los}$ $\displaystyle=$ $\displaystyle-V_{\rm z}\cos{i}-V_{\phi}\cos{\phi}\sin{i}-V_{\rm R}\sin{\phi}\sin{i}.$ (3) A transverse PV diagram corresponds to a pseudo-slit of the data cube perpendicular to the flow axis. This corresponds to a solution of Eqs. 1,2,3 with $\delta z=cst$. In the case of edge-on flows, a ring traces a perfect ellipse in the PV diagram. A fit of these ellipses give complete information about the morphology and dynamics of the outflow as shown by Louvet et al. (2018). In the case of an inclined outflow such as DG Tau B, different rings overlap on the transversal PV diagrams (see Fig. 2 left). Hence it was not possible to fit them individually. However constraints on some of the ring parameters could be derived from characterizing the outer limits of the PV diagram. The radius of the ring corresponds on the first order to $\delta x_{max}=\delta x(\phi\approx 0,\pi)$. In addition, the projected velocities at the edge of the ellipses $V_{\rm los}(\delta x_{\rm max})$ allow one to recover both $V_{\rm Z}(R,Z)$ and $V_{\phi}(R,Z)$ from the following equations: $\displaystyle Z$ $\displaystyle=$ $\displaystyle\frac{\delta z}{\sin{i}}$ (4) $\displaystyle R$ $\displaystyle=$ $\displaystyle\delta x_{\rm max}$ (5) $\displaystyle V_{\rm Z}$ $\displaystyle\simeq$ $\displaystyle\frac{V_{\rm los}(\delta x_{\rm max})+V_{\rm los}(-\delta x_{\rm max})}{-\leavevmode\nobreak\ 2\cos{i}}$ (6) $\displaystyle V_{\phi}$ $\displaystyle\simeq$ $\displaystyle\frac{V_{\rm los}(\delta x_{\rm max})-V_{\rm los}(-\delta x_{\rm max})}{2\sin{i}}.$ (7) By consequence, characterizing the outer limits of the PV diagram along multiple heights allowed us to recover a 2D map of the expansion velocity $V_{\rm Z}$ and specific angular momentum $j=RV_{\phi}$. In the following, we used the inclination derived from DV20 at $i=117^{\circ}\pm 2$. To characterize the outer shape of the transverse PV diagram at a given $\delta z$, we derived the maximal projected velocity for each value of $\pm\delta x$. Numerically, we computed the gradient of the emission profile at a fixed $\delta x$ and localized its maximum. We also derived an uncertainty on $V_{\rm los}$ which is found to vary in the range 0.05 to 0.2 km s-1. We used in the following a mean value of 0.1 km s-1. Figure 2 illustrates our method to determine $V_{\rm los}$ on the edges of the PV diagram.This procedure failed at low radii (or high velocities) because of the low S/N and the almost vertical profile that generated high uncertainties on the velocity estimate. The determination of the velocity was also limited by our spectral resolution of 0.3 km s-1. We fit the two edges for each PV diagram with a polynomial curve (see Fig. 2). We also applied a Gaussian filter with $\delta z$ using a standard deviation of 0.16′′. ### 3.2 Results Figure 3: Tomographic maps of $V_{\rm Z}$ (left panel) and specific angular momentum $j$ (right panel) in the outflow referential. The black dashed line traces the conical fit of the region $V_{\rm Z}=10-11$ km s-1. The white (resp. black) contours in the left (resp. right) panel show $V_{\rm Z}$ contours. The red dashed lines indicate the height of the two extrema in specific angular momentum. Using Eqs. 5, 4, 6, and 7, a tomographic map of $V_{\rm Z}(R,Z)$ and $V_{\rm\phi}(R,Z)$ could be recovered. We show the specific angular momentum $RV_{\rm\phi}(R,Z)$ instead of the rotation alone as this is more meaningful in the understanding of the dynamics. Figure 3 shows the resulting tomographic map of $V_{\rm Z}$ and $RV_{\phi}$ in the outflow referential. The tomography efficiently traces the conical shape visible on the channel maps. Curves of constant $V_{\rm Z}$ trace conical surfaces with semi-opening angles varying from 12∘ for the highest velocities to 17∘ for the lowest velocities. $V_{\rm Z}$ radially decreases from $\approx$ 14 km s-1 to 5 km s-1. This range of velocities is conserved until at least $Z=1200$ au. Figure 4: Specific angular momentum $j$ along curves of constant $V_{\rm Z}$. The corresponding range of $V_{\rm Z}$ is shown in the box. The uncertainty of specific angular momentum was obtained by propagating the $V_{\rm proj}$ uncertainty. The red dashed line corresponds to the median value of each curve. The specific angular momentum derived from the tomographic study varies from 0 to 140 au km s-1 and is consistently in the same sense as the disk rotation. At $Z<500$ au the specific angular momentum increases with radius from $\approx 30$ au km s-1 in the inner radius to $\approx 100$ au km s-1 on the outer radius. The specific angular momentum is also roughly constant on conical lines of constant $V_{\rm Z}$ until $Z\approx 500$ au (see Fig. 4). Our average value around 70 au km s-1 is consistent with the previous estimate of DV20 of $\approx$ 65 au km s-1. Two extrema in the specific angular momentum map can be observed at $Z=550-800$ au and $900-1100$ (see Fig. 4). At the lowest altitude, the specific angular momentum reaches zero while at the highest altitude the specific angular momentum increases up to $j>140$ au km s-1. These irregularities are also visible on channel maps. They correspond to regions where bumps in the cones are observed: toward $\delta x<0$ at $\approx 4.5^{\prime\prime}$, $\delta x>0$ at $\approx 6.5^{\prime\prime}$ (see the channel map $(V-V_{\rm sys})=4.37$ km s-1 on Fig. 1). These bumps may be due to local radial displacements of the outflow axis, due for example to wiggling. We study the impact of small amplitude wiggling in 5.2. ### 3.3 Limitations and biases In this section, we discuss the different biases and limitations of this tomographic study. Firstly, this study could not recover the radial velocity component $V_{\rm R}$ as it impacts mostly the size of the ellipse at $x=0$ where all the ellipses are stacked. A model-dependent study to characterize this radial velocity will be achieved in Sect. 5. Furthermore, in order to apply this method, it is critical that the centers of the rings are not significantly displaced from $\delta x=0$. Such displacements can be induced by a poorly estimated outflow position axis (PA) or by outflow axis wiggling. We derived the PA of the redshifted outflow in Sect. A at $\rm PA=295^{\circ}\pm 1^{\circ}$. This value is in very good agreement with the disk rotation axis $\rm PA=115.7^{\circ}\pm 0.3^{\circ}$ determined by DV20. We also determined an upper limit of 0.5∘ for the wiggling of the CO outflow axis. We discuss in Appendix 5.2 the impact of possible low-amplitude wiggling on our results. The different biases were also computed. We show in Appendix D.1 that assuming that the maximal radial extent corresponded to $\phi=0,\pi$ was partially inaccurate, and could introduce a bias in the estimate of $R$, $Z$, $V_{\rm Z}(R,Z)$ and $V_{\phi}(R,Z)$. The effect of ellipse stacking and its effect on the estimate of the velocities was studied in Appendix D.2. We evaluate at $\lesssim$ 20% the potential bias in our estimate of the conical outflow dynamics. Our estimates of $V_{\rm Z}$ and specific angular momentum are overestimated and underestimated respectively (see Appendix D.2). We estimate the bias on $R$ and $Z$ to be respectively $<1.5\%$ and $<3\%$, resulting in an error $<3\%$ on our estimate of the opening angle $\theta$. The highly asymmetric pedestal emission visible on the two sides of the PV diagram at $(V-V_{\rm sys})<2$ km s-1 traces the outer region (see Fig. 1). DV20 show that this region is outflowing and surrounds the conical outflow. We did not attempt to apply the tomographic method to this region. From its morphology in the channel maps, we derived for this component an opening angle $>30^{\circ}$. Such a large opening angle produces a bias of $\approx 60\%$ in the estimation of $V_{\rm Z}$ using our reconstruction method (see Appendix D). Moreover, this pedestal may potentially be explained by the top of one large ellipse, with the extremal region located at velocities $<\leavevmode\nobreak\ 1$ km s-1, absorbed by the medium. In that case, our reconstruction method is not applicable. By consequence, we did not apply our method for line-of-sight velocities $(V-V_{\rm sys})<2$ km s-1. In the following, we investigate to which extent wind-driven shells and disk winds can account for both the conical velocity stratification determined here and the striking substructures (arches, fingers, cusps) identified in Sect. 2. ## 4 Wind-driven shell modeling The traditional interpretation proposed for CO molecular outflow cavities around young stars is that they trace shells of ambient material swept up by a wide-angle wind or by jet bow shocks (see Cabrit et al. 1997; Lee et al. 2000; Arce et al. 2007, for reviews). In this section, we investigate the simplest and most widely used model to interpret CO outflow observations, namely the wind-driven shell (hereafter WDS) solution of Lee et al. (2000) where the shell is a parabola that expands radially in all directions with a velocity proportional to the local distance from the source (hereafter referred to as the Hubble law). Such a shell structure is predicted under a set of specific conditions in the wind and ambient medium222It is obtained when a wide-angle wind with velocity varying with angle as $V_{w}\propto\cos\theta$ and density varying as $\propto 1/(r^{2}\sin^{2}\theta)$ sweeps-up a static, flattened isothermal core with density $\propto\sin^{2}\theta/r^{2}$, and they mix instantly in the shell. We note that the radial shell expansion results from instant mixing, while the Hubble law derives from the identical radial fall- off of wind and ambient density (both $\propto 1/r^{2}$), which yields a shell speed that is constant over time (Shu et al. 1991). Finally, the parabolic shell shape derives from the combined $\theta$-dependencies of the densities and wind speed (Lee et al. 2001).(see Sect. 6 for details). This simple WDS model is recently shown by Zhang et al. (2019) to reproduce several features of the multiple CO shell structures at the base of the HH46/47 molecular outflow. Therefore it is natural to investigate whether the same WDS model can also reproduce the morphology and kinematics of the DG Tau B outflow, on smaller spatial scales. Following Lee et al. (2000), the parabolic morphology and the radial Hubble- law kinematics of the shell can be empirically described by two parameters, $C$ and $\tau$, through: $\centering Z=C\times R^{2}\qquad\qquad V_{\rm Z}=\frac{Z}{\tau}\qquad\qquad V_{\rm R}=\frac{R}{\tau},\@add@centering$ (8) where $\tau$ is the age of the shell, $C$ is the inverse size of the parabola at $\theta=45\degr$ (where $Z=R=1/C$), and the product $\tau C$ defines the shell expansion speed at each polar angle $\theta=\arctan(R/Z)$ through: $\displaystyle V_{\rm Z}(\theta)$ $\displaystyle=\frac{1}{\tau C}\left(\frac{Z}{R}\right)^{2}=\left({\tau C\tan^{2}\theta}\right)^{-1}$ (9) $\displaystyle V_{\rm R}(\theta)$ $\displaystyle=\frac{1}{\tau C}\left(\frac{Z}{R}\right)=\left(\tau C\tan\theta\right)^{-1}.$ (10) The above equations always produce ellipses in both channel maps and transverse PV diagrams (Lee et al. 2000). This is a direct result of the assumed Hubble law, where the shell velocity vector is proportional to the position vector. A channel map at a given line-of-sight velocity is then equivalent to making a cut through the shell at a given depth along the line of sight and this cut is shaped as an ellipse. Similarly, a transverse PV diagram has the same (elliptical) shape as a cut through the shell at the corresponding projected height. On channel maps, the ellipse is projected at increasing distances from the source with increasing velocity, due to the Hubble law. Similarly, on transverse PV diagrams, the mean velocity of the ellipse increases with the distance of the PV cut from the source (see Figs. 24 and 26 in Lee et al. 2000). In Appendix E, we derived analytical formulae for the center of the ellipse in channel maps as well as for its aspect ratio. Interestingly, we find that the ellipse aspect ratio only depends on the inclination $i$ and is equal to $1/|\cos{i}|$ for the classical model of Eq. 8 (see Appendix E). In the following, we attempt to fit with this model first the low-velocity outer flow component, and then the bright conical outflow and its discrete structures. ### 4.1 Low-velocity outer flow Figure 5: Comparison of the low-velocity outer CO outflow with a classical WDS model. Left panels: 12CO individual channel maps at different line-of- sight velocities tracing the low-velocity outer CO outflow. The white contours trace the model of a WDS of parabolic shape defined by $C=10^{-3}$ au-1, dynamical age $\tau=6000$ yr and specific angular momentum $j=250\pm 50$ au km s-1 with an inclination of $i=117^{\circ}$. Right panel: Transverse PV cut at $\delta z=7.5^{\prime\prime}$ averaged over a slice of $\Delta Z=0.2^{\prime\prime}$. The red ellipse traces the WDS model. ($V-V_{\rm sys}$) units are km s-1. The wide and low-velocity outflow at $(V-V_{\rm sys})<2$ km s-1 shows several properties suggestive of a ”classical” parabolic WDS with a Hubble law. Its outer border in channel maps has a parabolic shape, and it exhibits a larger offset from the origin at higher line-of-sight velocities (see Fig.1 left panels). Although such WDS models do not usually consider rotation, we included rotation to properly fit the large left-right asymmetry observed in the channel maps. To reduce the number of free parameters, we considered that the specific angular momentum $j$ is the same at all positions of the swept-up shell. $C$ was then fixed by the global parabolic shape of the cavity, $\tau$ by its spatial shift between the channel maps, and $j$ by its global left- right asymmetry. Figure 5 (left panels) shows that the outer contour of the low-velocity outflow and its increased spatial offset with velocity are well fit by a WDS obeying Eq. 8 with parameters $C=10^{-3}$ au-1, $\tau$ = 6000 years and $j=250\pm 50$ au km s-1. The rightmost panel in Fig. 5 shows that this shell model reproduces well the most extended, lowest velocity emission of the broad pedestal in transverse PV cuts; the predicted blue-shifted emission from the front side of the shell falls very close to systemic velocity, consistent with its nondetection in our data. On the other hand, our assumption of a thin parabolic shell does not match the observed outflow thickness at high altitudes.This discrepancy is visible at $\delta z\approx 15^{\prime\prime}$ on Fig. 5 where the observed width of the emissive outer layer is $\approx 3^{\prime\prime}$, significantly larger than predicted by our model. The shell should actually have a thickness $\simeq 3^{\prime\prime}\simeq 500$ au. ### 4.2 Conical outflow and discrete structures Figure 6: Dynamical times ($\tau=Z/V_{\rm Z}$) derived from the tomographic map of $V_{\rm Z}$ are shown in color. The white lines correspond to a parabolic curve with $1/C$ varying from 30 to 135 au, in steps of 15 au. In this section, we attempt to model the conical outflow and its discrete structures (arches, cusps, fingers, described in Sect. 2) by a stacking of several parabolic wind-driven shells with a Hubble velocity law. Several qualitative features are suggestive of such a model: the loop shapes of the arches at low velocity are reminiscent of the ellipses predicted in channel maps (see Appendix E), the apparent acceleration of the cusps (increased altitude with increasing velocity) is reminiscent of the predicted Hubble law dynamics, and finally, in the conical flow studied by tomography, contours of constant $\tau=Z/V_{\rm Z}$ follow quasi-parabolic curves above $Z\simeq 400$ au (see Fig. 6). Hence we investigate below whether the conical outflow could be made of successive nested parabolic wind-driven shells, where the apparent continuous aspect of the tomography would be an artifact of our limited spatial and spectral sampling, and the discrete structures (arches, cusps, fingers) would trace a few individual shells brighter than average. Contrary to the slow outer flow modeled in Sect. 4.1, the left-right asymmetry in these faster flow regions is small. Therefore, we neglected rotation when fitting WDS models to the channel maps. We set the WDS axis inclination equal to the large-scale disk inclination derived from ALMA studies ($i=63^{\circ}\pm 2^{\circ}$, DV20), leading to $i=180\degr-63\degr=117\degr$ in the redshifted lobe. Here, we find that the ”classical” WDS model of Lee et al. (2000) encounters a major problem, as shown in the top row of Fig. 7: the predicted aspect ratio of ellipses in channel maps, $A=1/|\cos{i}|$ (see Appendix E) is too large ($\simeq 2.2$). In order to match the observed aspect ratio of the arches ($\approx 1.4$), the inclination of the shell axis should be $i\approx 135\degr$ instead of $i=117\degr$. In the WDS model, however, the direction of shell elongation is not arbitrary but must follow the direction of both highest wind density (traced by the axial jet) and lowest ambient density (traced by core flattening). Proper motions of jet knots in DG Tau B imply a jet axis inclination of $i\geq 65^{\circ}$ for the blue-shifted lobe (Eislöffel & Mundt 1998), hence $i\leq 115^{\circ}$ for the redshifted lobe. This limit agrees within 2° with the large-scale disk inclination determined by ALMA ($i=63\degr\pm 2\degr$, DV20), which should follow the core flattening. Therefore, we can exclude a shell axis at $i\approx 135\degr$ as a solution to the ellipse aspect-ratio problem of the WDS model of Lee et al. (2000). Figure 7: Comparison of 12CO channel maps at three different line-of-sight velocities (color maps) with predicted ellipses for parabolic WDS (white contours). The model used on each row is sketched in the left-most panel. Top row: Classical model with radial Hubble expansion (see Eq. 8 and Lee et al. 2000). Middle and bottom rows: Modified model with ”collimated” expansion (see Eq. 11). Green, blue, and yellow contours highlight specific height ranges, indicated in the first column. White dotted lines outline the radial boundary defined by the shell ellipses with increasing velocity. Model parameters are listed in Table 2. $\delta x$ and $\delta z$ units are arcseconds. Since the ellipse aspect ratio in channel maps does not depend on $\tau$ nor $C$ (see Appendix E) the only possibility to reduce it without changing the shell parabolic shape is to modify the shell dynamics. The maximum height of the ellipse is reached on-axis ($\delta x=0$) where the projected velocity is greatly affected by the radial velocity component $V_{\rm R}$ (see Eq. 3 with $\phi=\pm\frac{\pi}{2}$). To keep a small number of model parameters, we thus chose to add an ad hoc free parameter $\eta$ that modified the radial velocity as: $\centering Z=CR^{2}\qquad\qquad V_{\rm Z}=\frac{Z}{\tau}\qquad\qquad V_{\rm R}=\eta\frac{R}{\tau}.\@add@centering$ (11) In Appendix E, we show that the ellipse aspect ratio in this modified model is set at $A=(\eta\leavevmode\nobreak\ \tan^{2}{i}+1)|\cos{i}|$. Since we wanted to reduce the aspect ratio, we needed $\eta<1$. In other words, we needed a velocity vector that is more collimated (forward-directed) than the radial shell expansion in the original WDS model of Lee et al. (2000). We refer to this ad hoc model as ”modified collimated WDS.” As shown in the second row of Fig. 7, the shape of the largest arch A0 at $(V-V_{\rm sys})=1.19$ km s-1 is well fit by a modified collimated WDS with $\eta=0.6$. The smallest two arches A2,A3 and the smallest cusp U5 are also well fit by a collimated WDS with $\eta=0.5$ (Fig. 7, bottom row). The parameters of these best-fit solutions are listed in Table 2. The inferred shell dynamical times have typical intervals of $\Delta\tau\simeq 300-750$ yrs, similar to those inferred by Zhang et al. (2019) in HH46-47. The fit velocity values $V_{\rm Z}$ at $\theta=14\degr$ (in the region of the conical flow) are also listed333We note that $\tan{14\degr}\sim 1/4$ hence $V_{\rm Z}(14\degr)\simeq 16/(\tau C)$, cf. Eq. 9. Not surprisingly, their range of $\simeq 7.5-16$ km s-1 is similar to our ”model-independent” tomographic results for $V_{\rm Z}$ in the conical flow region. Table 2: Parameters of parabolic wind-driven shells with collimated Hubble law fit to arches and cusps in Fig. 7. Name | $C$ | $\tau$ | $\eta$ | $V_{\rm Z}(\theta=14\degr)$aafootnotemark: $a$ ---|---|---|---|--- of model | (au-1) | (yr) | | (km s${}^{-1})$ WDS-A0 | 0.003 | 1600 | 0.6 | 16 WDS-A2 | 0.01 | 850 | 0.5 | 9 WDS-A3 | 0.02 | 500 | 0.5 | 7.5 WDS-U5 | 0.04 | 220 | 0.5bbfootnotemark: $b$ | 8.5 444The collimated parabolic WDS model is described by Eq. 11: $C$ is the inverse characteristic parabola size, $\tau$ the shell age along $Z$, $\eta$ the $V_{\rm R}$ reduction factor ($\eta=1$ for a radial flow). aafootnotemark: $a$ $V_{\rm Z}(\theta)=1/(\tau C\tan^{2}\theta)$ (see Eq. 9). bbfootnotemark: $b$In the case of WDS-U5, we do not have any constraint on the aspect ratio, and set $\eta$ at 0.5. Figure 8: Kinematic evolution along the flow axis. Panels a) Position- velocity diagrams averaged over a slice of width $\Delta Z=0.2^{\prime\prime}$ at three different $\delta z$ positions along the flow. The red contours trace solutions of collimated WDS used to fit A2, A3, and U5 (third line on Fig. 7). Panel b) In black is represented the maximal velocity of emission of the PV diagram. In red is shown the center velocity of the ellipses on the PV diagrams. Panel c) The gray region highlights the conical outflow domain. The black lines represent the parabolic morphology of the three solutions. The red dots show for each solution the region where the projected velocity reach $V_{\rm max}$. The Hubble law in the WDS model predicts an ever-increasing shell speed at higher altitudes, until it reaches the polar wind speed, which is $\simeq 125$ km s-1 according to the redshifted jet speed in DG Tau B (Eislöffel & Mundt 1998). In contrast, the maximum line-of-sight velocity with detectable emission in our transverse PV cuts (averaged between the left and right sides) is found to stay roughly constant with altitude at $V_{\rm max}\simeq 8\pm 1$ km s-1 (see Fig. 8b). Therefore, all successive wind-driven shells should be truncated, or have their CO emission strongly suppressed, above the point where they reach $V_{\rm Z}=V_{\rm max}/\cos{i}\simeq 18$ km s-1. This velocity limit is close to the molecule dissociation limit $\simeq$ 20 km s-1 in dense hydrodynamical shocks (see e.g., Wilgenbus et al. 2000). Therefore, the disappearance of CO emission above a certain speed might be explained by shock-dissociation of ambient CO. Figure 8c shows that the corresponding truncation region for the best-fit WDS models in Table 2 has a rough conical shape with $\theta\simeq 9\degr$. However, our ad hoc modified collimated WDS model meets two serious issues, detailed below. The model predicts a full ellipse in each channel map (white contours in Fig. 7), which is not observed. In contrast, discrete structures highlight only a portion of ellipse, depending on the velocity range (see Fig. 7 and Sect. 2): the ellipse top at low-velocities (arches), ellipse flanks at mid-velocities (fingers), and ellipse bottom at high-velocities (cusps). We find that a transition from arches at low velocity to cusps at high velocity can only be obtained if emission is restricted to a range of heights from $z_{\rm min}$ to $z_{\rm max}$, as illustrated by the colored contours in Fig. 7. Serious discrepancies still remain with observations, however: In the broadest shell, WDS-A0, the extents of Arch A0 and Cusp U1 require inconsistent ranges of emitting heights, and the predicted ”fingers” at intermediate velocity are much wider than observed (see blue and green contours in middle row of Fig. 7). In the smaller inner WDS, full ellipses are still predicted in intermediate velocity channels, which are not observed (see yellow contours in bottom row of Fig. 7). The same problems remain even if we adopt conical shapes for the shells instead of parabolae, hence the above discrepancies appear intrinsically linked to the assumed Hubble-law dynamics. Another serious issue is that the best-fitting value of $\eta$ in our modified collimated WDS models is always close to 0.5 (see Table 2). A ratio $V_{\rm R}/V_{\rm Z}=0.5R/Z$ corresponds to a velocity vector locally tangent to the parabola. Hence the shell is not expanding but stationary. The physical justification for the Hubble law in the WDS model, namely a shell expanding at constant speed over time (Shu et al. 1991), is then no longer applicable. If the emitting material is moving parallel to the shell, a velocity increasing in proportion to distance would require, instead, a constant accelerating force of unknown nature operating out to z=3000 au, which is totally unphysical. In summary, we find that only the outer faint, low-velocity flow in DG Tau B can be reproduced with the parabolic WDS model with radial Hubble law proposed by Lee et al. (2000). In contrast, the bright conical outflow at mid to high velocity, although reminiscent of WDS models because of the apparent acceleration of its discrete structures, cannot be explained by such models, even when ad hoc modifications to the kinematics, emissivity range, and shape are introduced. The model faces important issues which seem intrinsically linked to the Hubble-law dynamics. We discuss in Sect. 6.1 the implications of these results in the context of more general wind or jet-driven shell scenarios. ## 5 Disk-wind modeling In this section, and alternatively to the WDS models considered in Sect. 4, we investigate a simple kinematical disk wind model for the DG Tau B redshifted outflow where the conical morphology visible in Fig. 3 would trace the trajectory of CO molecules ejected from the disk. Although we cannot derive $V_{\rm R}$ in a model independent way, from the external contours of transverse PV diagrams (see Sect. 3), we showed in Sect. 2 the existence of brighter elliptical structures visible on transverse PV diagrams. Assuming that they trace each a specific layer of the flow along which $V_{\rm Z}$ and $V_{\rm R}$ stay roughly constant with height, we fit these ellipses to derive both the $V_{\rm R}$ and $V_{\rm Z}$ components of the velocity (see Appendix C). Figure 23 shows that the derived velocity directions are parallel to the conical contours of constant $V_{Z}$. This comforts our hypothesis that the trajectory of the outflow follow lines of constant $V_{\rm Z}$. From this hypothesis, we derived the collimation and kinematics of the streamlines using the tomography and created a synthetic data cube of the conical disk wind outflow. ### 5.1 Steady disk-wind model We made the assumption that the flow is axisymmetric and that the matter has reached its terminal velocity and has a constant poloidal velocity along its trajectory. We fit this trajectory by a conical surface defined by an angle $\theta$ from the Z-axis and an anchoring radius $r_{0,\rm geo}$. We extracted from the tomography the specific angular momentum, $j=R\times V_{\phi}$, along curves of constant $V_{\rm Z}$ (see Fig. 4) and derived a median value for each streamline. We defined the uncertainty of this value as the standard deviation of specific angular momentum. We also computed the poloidal velocity $V_{\rm P}=V_{\rm Z}/\cos{\theta}$. Figure 9: Disk-wind properties derived from the tomography along lines of constant $V_{\rm Z}$. Red dashed curves show fits as a function of the poloidal velocity of the streamline $V_{\rm p}$: the anchoring radius of the streamline $r_{0,\rm con}$ was fit by a third-order polynomial, the angle of the streamline with the flow axis $\theta$ by a power law and the product $V_{\rm P}\times j$ by a constant value of 570 au km${}^{2}s^{-2}$. Blue dotted lines represent the extrapolation used to model the high-velocity component that could not be mapped by tomography due to its low S/N. Figure 9 represents the derived values of $r_{0,\rm geo}$, $\theta$ and $V_{\rm P}\times j$ for each streamline of constant $V_{\rm P}$. We fit the variation of anchoring radius $r_{0,\rm con}$ with the poloidal velocity by a polynomial law. The variation of ejection angle $\theta$ was fit by a power law. $V_{\rm P}\times j$ was taken constant at 570$\pm$ 50 au km2 s-2 for all the streamlines. The fits were achieved using nonlinear least squares, The equations are as follow: $\displaystyle r_{0}$ $\displaystyle\simeq$ $\displaystyle 0.18-5.33V_{\rm p}+44.29V_{\rm p}^{2}-73.23V_{\rm p}^{3}$ (12) $\displaystyle\theta$ $\displaystyle\simeq$ $\displaystyle 12.63+348.57V_{\rm p}^{-2.55}$ (13) $\displaystyle j\times V_{\rm p}$ $\displaystyle\simeq$ $\displaystyle 570.$ (14) Figure 10: Comparison of observations with steady disk-wind model. Left panels: Individual 12CO channel maps computed from the global disk-wind model at selected line-of-sight velocities (top row) are compared to observations (bottom row). The color scale is the same for all the channel maps. Right panels: Transverse PV diagrams at two positions $\delta z$ along the flow and averaged over a slice of width $\Delta Z=0.2^{\prime\prime}$. The background grayscale image shows the observations and the red contours trace the predictions from the disk-wind model. $(V-V_{\rm sys})$ units are km s-1. Here $r_{0}$ is in au, $V_{\rm P}$ in km s-1, $\theta$ in degrees, and $j$ in au km s-1. We modeled the disk wind with axisymmetric conical streamlines with the dynamics and morphology laws derived in Fig. 9 and created a synthetic data cube of the conical outflow. We set the external and slower layer at $V_{\rm p}=6$ km s-1, corresponding to the smallest value that could be mapped with our tomography (see Sect. 3. We set the internal, faster velocity at 20 km s-1. The parameters in the velocity range $V_{\rm p}=14-20$ km s-1, not covered by the tomography, were determined from an extrapolation of our fits (blue dotted line in Fig. 9). This extrapolation was done in order to describe the almost-vertical high-velocity component not described by the tomography due to insufficient S/N. For each layer, we set the initial value of $V_{\phi}(R)$ assuming $V_{\rm p}\times j=570$ au km2 s-2 (see Fig. 9). We assumed optically thin emission throughout the outflow, which is justified by the observed ratios of 13CO/12CO (see DV20). We did not consider a variation of emissivity with radius of ejection nor with height (see Sect. F). Proper modeling would require CO chemistry and temperature profiles, which is well beyond the scope of this paper. Projection and beam convolution effects were also taken into account. Figure 10 shows synthetic channel maps and PV diagrams for our model compared with observations. The global morphology of the outflow at $(V-V_{\rm sys})>2$ km s-1 as well as its variation with line-of-sight velocity are well recovered as expected since we use the tomographic results to constrain the wind collimation and kinematics. This model does not attempt to describe the extended outflow surrounding the cone at low velocities $(V-V_{\rm sys})=1.15$ km s-1 (see Sect. 2, Fig. 10). To describe completely this extended low- velocity component, we would need to extrapolate the disk-wind model at larger ejection radii. However, as this component falls at absorbed cloud velocities, we are not able to derive model-independent constraints on the dynamics of this component. Proper modeling would require time-consuming and uncertain parameter space exploration. We choose therefore to focus in the following on the conical outflow; nonetheless, the disk wind could be more extended radially than we describe with our current modeling. Although effective to describe the global morphology of the outflow, our simple axisymmetric and steady disk-wind model does not reproduce the different substructures identified in our observations: cusps and arches and the local deviations of specific angular momentum at $Z\approx 600$ and 1000 au. In the following subsections, we discuss two small perturbations of our disk-wind models which could explain the various substructures observed. ### 5.2 Wiggling of the flow axis Although we do not detect a clear signature of wiggling in our data, we cannot exclude a small amplitude wiggling of the CO axis $<0.5^{\circ}$ (see Appendix A). A wiggling of the outflow axis could explain the variations observed on the specific angular momentum tomographic map. Indeed, in order to determine the specific angular momentum using Eq. 7, we assumed that the center of the layer is located at $\delta x=0$ at all heights. If the center is shifted toward $\delta x>0$ or $\delta x<0$, the specific angular momentum computed with our method will be respectively higher or lower than the true value. This effect is more critical if the PV diagram shows a strong velocity gradient, which is the case for DG Tau B. In this section, we investigate this effect, and show that small amplitude wiggling may also create the substructures observed (cusps, fingers and arches). Figure 11: Synthetic data cubes computed from the generic disk-wind model with flow axis precession: with a precession angle of 0.2∘ (left panels) and 0.5∘ (right panels) and assuming a constant precession period $\tau_{\rm p}=400$ yrs (top row) or a constant precession spatial wavelength $\Lambda_{\rm p}=800$ au (bottom row). For each model are shown both a channel map at $(V-V_{\rm sys})=4.37$ km s-1 and the corresponding tomography of $j$ derived with the method described in Sect. 3. The color scale of $j$ ranging from 0 to 140 au km s-1 is identical to the color scale of Fig. 3. We modifed the disk-wind model presented in Sect. 5.1 to add a precession of the outflow axis. Each conical layer of the outflow precesses with an angle $\alpha$ and a precession period $\tau_{\rm p}$. This is an extension of the model developed by Masciadri & Raga (2002) for jets, modified to take into account the conical morphology of the outflow and inclination to the plane of the sky. We modeled both a prograde and retrograde precession. However, due to the small value of $\alpha$ in our models, the two models give very similar results. We present here results for the prograde model only. We first assumed that all disk-wind layers precess with the same $\alpha$ and the same precession period $\tau_{\rm p}$. Due to the velocity shear across the outflow, the spatial period $\Lambda_{\rm p}=\tau_{\rm p}\times V_{\rm p}$ then varies between layers according to the poloidal velocity. We also investigated a precession model where the spatial period $\Lambda_{\rm p}$ is constant across all streamlines. A constant $\Lambda_{\rm p}$ corresponds to a variation of precession period as $\tau_{\rm p}\propto V_{\rm p}^{-1}\propto r_{0}^{0.5}$. We visually fit $\tau_{\rm p}$ and $\Lambda_{\rm p}$ to best reproduce the location of the two extrema variations in the specific angular momentum map separated by $\simeq$ 400 au. For each model, we computed synthetic data cubes and derived the specific angular momentum map using the same method used in Sect. 3 for the observations. Fig. 11 shows the resulting channel maps and the specific angular momentum maps for the two precession models (constant $\tau_{\rm p}$ and constant $\Lambda_{\rm p}$) with two different precession angles (0.2 and 0.5∘) compatible with the upper limit derived for the CO outflow axis wiggling in Annex A. Precession models with constant $\tau_{\rm p}=400$ yrs successfully reproduce the channel maps morphology, in particular the cusps at high- velocity and arches at low velocities (see Fig. 28). A best match to the intensity contrast is obtained for $\alpha=0.5^{\circ}$. However, the resulting map of specific angular momentum $j$ is not fully consistent with our observations. Indeed, as $\Lambda_{\rm p}$ is different for each layer, the perturbations of specific angular momentum are not localized at one specific height, as in the observations. The modified model with constant $\Lambda_{\rm p}=800$ au for all layers better reproduce the positions of the two extrema at $Z=600$ and $Z=1000$ au in the specific angular momentum map. However, the cusps have a lower intensity contrast than observed, even with the maximum allowed wiggling angle of 0.5∘. In addition, this model predicts clear detectable wiggling on the edges of the cone in channel maps, which is not seen in the observations. A model in between these two extremes, that is to say with $\alpha\simeq 0.5\degr$ and a precession period $\tau_{\rm p}$ increasing more slowly than $r_{0}^{0.5}$ may better account for all observational properties. A remaining discrepancy with observations is that none of the wiggling models reproduce the short spacing of $\simeq 1.5^{\prime\prime}=200$ au between the inner cusps, as well as the apparent increase of cusp separation with distance from the source (see Table 1, Fig. 28), although this latter effect is mostly seen in the farthest cusp A0 and may result from a lack of sensitivity. The constant $\Lambda_{\rm p}$ model predicts a projected separation between the cusps of $\Lambda_{\rm p}\sin(i)$, corresponding to $\simeq 5^{\prime\prime}$, while the constant $\tau_{\rm p}$ model a twice smaller separation typically. However, we stress that our wiggling models are probably too simplistic as they do not take into account the (magneto)-hydrodynamical interactions between the layers. Masciadri & Raga (2002) have shown that simulations depart rapidly from analytical solutions in the case of jet wiggling due to precession. This difference is potentially even greater with a shearing outflow. Dedicated numerical simulations are required to fully test this scenario. Nonetheless, this model is a promising candidate to explain the variation of specific angular momentum along the DG Tau B outflow. We discuss in Sect. 6 possible wiggling mechanisms and their implications. ### 5.3 Emissivity enhancements in the disk wind In this section, we investigate an alternative model where cusps and arches are created by localized emissivity enhancements in the conical disk wind. We first derived their location from the tomography, assuming that they are axisymmetric, and then created a synthetic data cube to compare with our observations. Figure 12: Line-of-sight velocity on-axis (at $\delta x=0$ ) of the model disk-wind flow surfaces as a function of their poloidal velocity. Orange circles show the front side ($\phi=\frac{\pi}{2}$), and green circles show the back side ($\phi=-\frac{\pi}{2}$) of the disk-wind flow surfaces. In gray, we show the line-of-sight velocity range where the arches and cusps are observed (see Annex B). In light gray, is represented the low S/N domain where the cusps could not be characterized. Figure 12 shows the projected velocity on-axis at $\delta x=0$ for the front side and the back side ($\phi=\pm\frac{\pi}{2}$) of each conical wind layer in our model, predicted from Eq. 3. We also represent the domain of line-of-sight velocities where arches and cusps positions could be measured in channel maps as described in Sect. 2. The cusps observed at $(V-V_{\rm sys})>5$ km s-1 could not be characterized in Sect. 2 because of low S/N. Figure 12 shows that if the observed substructures are due to axisymmetric emissivity enhancements in the conical outflow, the cusps come from the back side ($\phi\approx-\frac{\pi}{2}$) and the arches from the front side ($\phi\approx+\frac{\pi}{2}$) of the enhanced ring. Figure 12 also shows that wind layers with poloidal velocities $V_{\rm p}>9$ km s-1 are not located on the arch domain, and are located on the low S/N cusp domain, which would require weaker emissivity enhancements in the fastest internal layers. Each cusp was characterized by its projected height at $\delta x=0$ at a specific projected line-of-sight velocity $V_{\rm los}$. The projected height of the cusp corresponds to an equation $Z(R)$ derived from Eq. 2, with $\phi=-\frac{\pi}{2}$. Similarly, the projected velocity of the cusps corresponds to a poloidal velocity as shown in Fig. 12 and assuming $V_{\rm R}$, a conical line of constant $V_{\rm Z}$ in the tomography. As a result, we can associate the cusp observed in a given channel map with a $(R,Z)$ location on the $V_{\rm Z}$ tomography map. This location would be the intersection between the $Z(R)$ equation derived from the projected height, and the conical line derived from the projected velocity. Figure 13: Tomographic map of $V_{\rm Z}$ from Fig. 3 extrapolated from $Z=1200$ au to $Z=3000$ au using the conical wind model from Fig. 9 . Gray hatched regions named SU0-5 represent the locus of emissivity enhancements required to explain the cusps on-axis positions at $V-V_{\rm sys}=3.73-5.32$ km s-1 (see Fig. 22). White dots represent the solutions for the cusps at $V-V_{\rm sys}=4.37$ km s-1. Red dotted lines indicate the height of the two extrema in specific angular momentum (from Fig. 3). The gray points located on the $R=0$ axis represent the positions of jet knots observed by Podio et al. (2011). The red dots indicate the positions with associated errors of the ellipses identified in transverse PV diagrams (see Sect. C) and indicate that emissivity enhancements extend to the inner streamlines. Figure 13 shows the tomography of the conical outflow using equations from Fig. 9 for extrapolation at $Z>1200$ au. The white dots correspond to the solutions (named SU0 to SU5) for the cusps heights on axis identified in the channel map at $(V-V_{\rm sys})=4.37$ km s-1. We were able to follow six different cusps in up to six different channel maps (see Fig. 22), which allowed us to reconstruct the shape for some of these enhancements, shown as hatched areas in Fig. 13. As mentioned in Appendix B, the cusps are also visible at higher line-of-sight velocities, but could not be characterized reliably due to their lower S/N. Consequently, the hatched area shown in Fig. 13 should be extended toward the inner streamlines. Moreover, density enhancements could also be present at $\delta z<2.2^{\prime\prime}$ but were not identified by our procedure.The apparent acceleration of the cusps seen in channel maps can be readily reproduced, in this disk-wind model, by axisymmetric emissivity enhancements that cross obliquely the flow streamlines. The upward shift of each cusp at higher velocity (apparent acceleration) is simply a result of the velocity shear across flow streamlines. It does not require a Hubble-law dynamics in the underlying flow. Interestingly enough, the density enhancements SU3-4 are located close to the two extrema of specific angular momentum, at $Z=600$ and 1000 au, suggesting a possible link. We could not derive the shape of the emissivity enhancements for $V_{\rm p}$ ¿ 9 km s-1, due to the low S/N of the cusps at the corresponding projected velocities $(V-V_{\rm sys})>5.32$ km s-1 (see Fig. 12). We computed the dynamical age of these emissivity enhancements with $\tau_{\rm w}=\frac{Z}{V_{\rm Z}}$. This value would give the true dynamical age of the enhancement if it is created by variability of ejection from the disk surface. The derived values of $\tau_{\rm w}$ for cusps visible on multiple channel maps are shown in Fig. 14. The dynamical age is almost constant within each cusp with only a slight decrease with increasing poloidal velocities, especially for U3 & U4. The difference in dynamical ages between two successive cusps varies from 190 to 490 years and roughly increases linearly with the age. Figure 14: Dynamical evolution of the enhancements.Left panel: Dynamical age $\tau=\frac{Z}{V_{\rm Z}}$ of the enhancements SU0-5 as a function of $V_{\rm P}$. The red dotted line represents the mean value. Right panel: Dynamical age versus timescale between adjacent cusps ($\tau_{i}-\tau_{i+1}$). The uncertainties are determined by propagating the errors on the cusp positions. These solutions were determined using only the properties of the cusps on-axis (at $\delta x=0$), corresponding to $\phi=-\frac{\pi}{2}$. In order to check if the reconstructed enhancements were consistent with the full cusp morphology in all channel maps, we developed a 3D model where we modified the emissivity profile of the synthetic disk-wind model presented in Sect. 5.1. We multiplied the underlying DW emissivity profile by Gaussian components representing the emissivity enhancement. We used the mean $\tau_{i}$ values derived from Fig. 14 to determine the location of the enhancements along each wind streamline. In order to reproduce the observed width and intensities in the channel maps, we set the full width at half maximum of the Gaussian component at 23 years and the maximal emissivity enhancement at $G_{\rm i}=3$. We introduced these enhancements only in the external layers of the outflow ($V_{\rm P}<9$ km s-1), since we did not have constraints on their location at higher velocities. Figure 15 shows computed and observed channel maps at different line-of-sight velocities. The contour of the modeled enhancement SU0 is also represented on top of the observations. This model successfully reproduces the morphology of the cusps as well as their apparent offset from source with increasing velocity. Interestingly enough, the model used to reproduce the cusps also matches the arches at low velocity, with the locations of the apexes consistent with our observations (see Fig. 28). The intensity contrast is also roughly recovered with our $G_{\rm i}$ value, meaning that the outflow brightness is locally multiplied by three. Figure 15: 12CO channel maps at selected line-of-sight velocities for the steady conical disk-wind model (bottom row) and observations (top row). White contours highlight the predicted emission from the modeled density enhancement SU0. The intensity color scale is the same for all channel maps. The channel map at ($V-V_{\rm sys})=1.51$ km s-1 shows that only the top of the arches is reproduced. Indeed, the flanks of the largest arches are wider than the conical flow region modeled by our disk wind in Fig. 9. In order to fit completely the arches, we would need to extend the disk-wind model to larger radii and opening angles and lower poloidal velocities as suggested in Sect. 5.1, and extend the density enhancements to these regions. However, due to cloud absorption at low velocities $<2$ km s-1, we do not have model- independent tomographic constraints on the streamline shape and kinematics in this slow external flow. Therefore we cannot determine a reliable solution for the emissivity enhancements producing these arch flanks. ## 6 Discussion In this section, we use the results of our parametric modeling of the flow (Sect. 4,5) and tomographic study to critically discuss two possible origins for the small-scale redshifted CO outflow in DG Tau B: 1) a stacking of multiple shells swept-up by an inner wind (or jet), without any contribution from an extended disk wind (Sect. 6.1), 2) an extended disk wind (Sect. 6.2), with internal perturbations causing the observed substructures (Sect. 6.3). We find that our measurements of rotation put stringent constraints on each of these scenarios, and we discuss physical implications for the ejection process and its relation to the disk accretion process. ### 6.1 Stacking of wind-driven shells Figure 16: Specific angular momentum map (in color) for the model of ballistic rotating infall (Ulrich 1976) with $R_{\rm d}=700$ au and $M_{\star}=1.1M_{\odot}$. The angular momentum varies from 25 au km s-1 to 900 au km s-1. The green hatched and red filled regions represent respectively the limits of the conical outflow derived from the tomography in Sect. 3 and the parabolic shape of the classical Hubble-law WDS model fit to the external CO outflow.The white and red contours outline the infalling streamlines with a specific angular momentum similar to the conical outflow $j=40-100$ au km s-1 (white), and to the WDS model $j=200-300$ au km s-1 (red). The black hatched region outline the streamlines with initial $\theta_{0}=70\pm 5^{\circ}$ reproducing the infall signatures seen in DG Tau B (DV20). We discuss here a scenario where the redshifted CO outflow in DG Tau B can be accounted for by a stacking of multiple swept-up shells resulting from the interaction of an episodic wide-angle wind (or jet) with the ambient medium. Such a scenario is recently proposed by Zhang et al. (2019) to reproduce the multiple shell structures at the base of the HH46/47 outflow. They find good agreement with the WDS model of Lee et al. (2000, 2001), namely a parabolic layer undergoing radial expansion following a Hubble law $\@vec{V}\propto\@vec{r}$. Below we summarize the key successes and failures encountered in Sect. 4 by the same WDS model when applied to the various flow components in DG Tau B, and we show that our rotation measurements raise additional issues for this model in terms of cavity refilling. #### 6.1.1 Outer flow component In Sect. 4.1, we find that the morphology and kinematics of the low-velocity outflow can be well reproduced by the simple WDS model of Lee et al. (2000). We now examine whether a WDS origin is physically consistent with the large specific angular momentum inferred from our modeling, $j_{\rm outer}\simeq 250\pm 50$ au km s-1, in the same sense as the disk (see Sect. 4.1). Studies of rotation signatures in pre-stellar and proto-stellar cores show that specific angular momentum stops decreasing with radius below scales $\leq 3000$ au and becomes roughly constant (Ohashi et al. 1997; Gaudel et al. 2020). This ”plateau” is interpreted as the region where infall motions start to dominate, and specific angular momentum is roughly conserved along streamlines. Depending on the object, the specific angular momentum in the ”plateau” is $\approx$ 40–400 au km s-1 ($0.2-2\times 10^{-3}$ km s-1 pc, Gaudel et al. 2020). Our estimate for the outer flow in DG Tau B, $j_{\rm outer}\simeq 250\pm 50$ au km s-1 falls well within this range. In addition, infall signatures are identified around DG Tau B at large polar angles $\theta\simeq 70\degr$ (DV20). No such signatures are seen at smaller polar angles, but rotational flattening predicts lower envelope densities there (Ulrich 1976), hence they might be too faint for detection. It thus appears promising to consider that infalling material might dominate the rotation in the outer CO layer. Strictly speaking, the Hubble law assumed in the WDS model of Lee et al. (2000) is only valid for a static ambient medium with a $1/r^{2}$ radial density decrease555A static medium ensures that, after full mixing, the shell expands in the same radial direction as the wind, while the $1/r^{2}$ ambient density decrease ensures that the expansion speed is constant over time (the density ratio between the wind and ambient medium being independent of radius); both properties then together yield the ”Hubble law” $\@vec{V}=\@vec{r}/\tau$ (Shu et al. 1991; Lee et al. 2001). A rotating infalling envelope, in contrast, has a non-radial motion and a flatter radial density law $\propto 1/r^{1.5}$ (Ulrich 1976). However, the calculations of López-Vázquez et al. (2019) for such an ambient medium show that the WDS expansion remains quasi radial, except close to the mid-plane, and with almost constant speed after 200 yrs. Our simple model in Sect. 4.1 thus remains roughly valid if the wind expands into an infalling envelope. Using the shell rotation speeds computed by López-Vázquez et al. (2019), we expect a shell specific angular momentum close to that of the infalling material immediately ahead of it. Therefore, we consider in Fig. 16 the spatial distribution of specific angular momentum in a rotating, free-falling envelope from Ulrich (1976), with a centrifugal radius $R_{d}=700$ au and central mass $M_{\star}=1.1M_{\odot}$ appropriate to DG Tau B (DV20). The predicted $j$ values along the fit parabolic outer flow boundary are very similar to the observed one, $j_{\rm outer}$. This detailed comparison confirms that an infalling envelope in DG Tau B, if present up to polar angles $\theta\simeq 30\degr$, could provide enough specific angular momentum to explain the rotation in the outer CO layer. However, this analysis is highly simplistic. First, we consider here that the specific angular momentum of the envelope is fully transferred to the entrained layer. This assumption gives an upper limit for the shell rotation velocity, as turbulent mixing would decrease its specific angular momentum. Secondly, the spherical and ballistic infall model used here (Ulrich 1976) does not take into account the effects of pressure gradients or the magnetic field. Pressure gradients could potentially increase the specific angular momentum of the infalling envelope at large polar angles, through ”pushing” outer infalling streamlines toward the axis, while the magnetic field would decrease it due to magnetic breaking. Dedicated numerical simulations of the interaction of an infalling material with an inner wind component and taking into account all these effects are needed to fully test the entrainment scenario for the outer CO layer. A serious issue with this interpretation, however, is the young inferred shell age, $\tau=V_{\rm Z}/Z=6000$ yrs (Sect. 4.1), much younger than the true age of DG Tau B. A first way out of this ”short age problem” would be that the interface between wind and envelope in DG Tau B is not expanding, but static. A static shell may form when mixing between the wind and the ambient material is not instantaneous, as assumed in most WDS models, but very gradual. Shocked ambient material is then slowly entrained along the shell surface by the shocked wind in a thin turbulent mixing-layer. The static shell shape and mixing-layer properties were recently computed in the case of a free-falling rotating envelope by Liang et al. (2020). Using again $R_{d}=700$ au and $M_{\star}=1.1M_{\odot}$ for DG Tau B (DV20), the specific angular momentum in the mixing-layer is predicted to be $j\approx 0.15\sqrt{GM_{\star}R_{d}}$ $\simeq 120$ au km s-1, twice lower than estimated in the outer flow. Therefore, a static wind/envelope interface does not seem able to explain the rotating outer flow. A second way out of the short age problem would be that the outer flow component does not trace the first wind encounter with the infalling envelope, but a more recent wind outburst from 6000 yrs ago. To provide its high angular momentum material to the young shell, however, the infalling envelope should somehow manage to penetrate and ”refill” all the older shells created by previous (unobserved) wind outbursts. Whether such an efficient cavity refilling by the envelope is physically possible on $\leq 6000$ yr timescales is a difficult open question, well outside the scope of the present paper. As shown below, the issue of cavity ”refilling” becomes even more acute when the WDS scenario is applied to the inner conical flow. #### 6.1.2 Inner conical outflow In contrast to the outer flow component, we find in Sect. 4.2 that the morphology and kinematics of the inner conical outflow and its bright substructures (arches, fingers and cusps) cannot be reproduced by the model of parabolic WDS with radial Hubble law used by Lee et al. (2000) and Zhang et al. (2019), even after several ad hoc modifications. We identify two serious issues: (1) The observed aspect ratio of arches in channel maps at low velocity is significantly shorter than predicted by the original model (ellipse aspect ratio = $1/\cos{i}$); it can be reproduced by a more collimated WDS model where the flow is parallel to the parabola; but that is no longer physically consistent with a Hubble-law velocity field, which requires an expanding shell (Shu et al. 1991). (2) The shell models fitting the arches at low-velocity and cusps at high velocity do not agree with the observed emission morphology in channel maps at mid-velocities, predicting fingers that are broader than the cone, or full ellipses that are not seen. We find that assuming a conical shell instead of a parabolic one, but keeping a Hubble law, still creates the same problems. They appear intrinsically caused by the Hubble-law velocity field, regardless of the shell detailed morphology. Therefore, any model where the shell is expanding quasi-radially and at nearly constant speed over time will fail to reproduce our observations. This, in particular, discards all models where the wide-angle wind and the ambient medium share a similar power-law in $r$, and where they instantly mix in the shell (e.g., the models of López-Vázquez et al. 2019; Shang et al. 2006). Simulations including magnetic field (Wang et al. 2015; Shang et al. 2020) and stationary solutions (Liang et al. 2020) show the formation of a shear layer along the shell more in line with DG Tau B. However the first model predicts a shell anchoring radius increasing with time, while the second has a shell anchored near the centrifugal radius $R_{d}\simeq 700$ au (DV20). This is inconsistent with the small observed anchoring radius of the DG Tau B conical outflow ($\leq 50$ au, DV20). Alternative models of swept-up shells exist involving an infalling sheet (Cunningham et al. 2005) or a jet instead of a wide angle wind (e.g., in Downes & Cabrit 2007), but they have no analytical solutions. Therefore, dedicated numerical simulations would be required to test them in DG Tau B. In the following, we show that the specific angular momentum measured in the conical outflow by tomography, $j_{\rm cone}\simeq 40-100$ au km s-1 (see Fig. 3), raises additional issues for the swept-up shell scenario. We first note that a wide-angle ”X-wind” cannot explain the observed rotation in the conical outflow; with a launching radius $\simeq 0.05-0.1$ au from the central protostar and a wind magnetic lever arm parameter $\lambda=j_{X}/j_{\rm Kep}\simeq 3$ (Shang et al. 1998), its specific angular momentum is predicted at $j_{X}\approx 20-30$ au km s-1, a factor two to three times less than observed. In addition, the wind cannot dominate the swept-up shell mass unless it is slower than twice the shell speed666ram pressure equilibrium between the reverse shock in the wind and the forward shock in the static ambient medium imposes $\rho_{\rm w}(V_{\rm w}-V_{\rm s})^{2}=\rho_{a}V_{\rm s}^{2}$, where $V_{\rm w}$ and $V_{\rm s}$ are the wind and shell speeds, and $\rho_{\rm w}$ and $\rho_{a}$ are the wind and ambient density. The mass-flux entering the shell from the wind side will then dominate over the swept-up mass if $(V_{\rm w}-V_{\rm s})<V_{\rm s}$., which in the present case would require $V_{\rm w}\leq 15$ km s-1 (see Table 2). This is inconsistent with a wind originating from close to the protostar. The ”X-wind” model, for example, has $V_{\rm w}\simeq 150$ km s-1 (Shang et al. 1998). The low observed expansion speeds $\simeq 6-14$ km s-1 in the conical flow (see Fig.3) is more consistent with jet bow shocks dominating the shell mass. A jet magnetic lever arm parameter $\lambda\simeq 9$ would then provide enough angular momentum. Such a scenario, however, cannot explain why regions of lower speed in the conical flow have inversely higher specific angular momentum (see Fig. 3). In a jet bow shock, lower speeds arise where more ambient mass has been swept-up. Assuming the ambient medium provides no angular momentum, the jet angular momentum would get more diluted, and the shell specific angular momentum would drop, instead of increasing. We conclude that the observed rotation in the conical flow cannot come from an inner wind or jet. To reproduce the observed conical flow rotation in the swept-up shell scenario, we thus need an external medium with an important angular momentum. The infalling rotating envelope is an obvious candidate. However, the observed specific angular momentum in the conical flow is twice larger than predicted, in the same region, by the Ulrich infalling envelope model (see Fig. 16). In addition, it is unclear how infalling matter could penetrate and ”refill” the space between the closely spaced shells producing the cone substructures, especially when the region immediately outside the conical flow is instead in outflow motion (see Fig. 5). An alternative would be that the swept-up material originates from the rotating disk atmosphere, at radii $R_{0}\simeq j^{2}/(GM_{\star})\simeq$ 2-9 au. The problem is then to refill the cleared cavities between shells with a ”new” static disk atmosphere. We note that our three shell models fitting the substructures in the conical flow have remarkably identical expansion speeds within 1 km s-1 (see last column of Table 2 for WDS-A2, WDS-A3 and WDS-U5). Assuming that the corresponding wind and jet outbursts were of similar strength, it implies that they met an identical ambient density ahead of them, hence the atmosphere refilling process should be extremely efficient. This appears difficult to achieve unless a large-scale disk wind is present. Realistic simulations of the interaction between an episodic inner wind or jet with an infalling rotating envelope and the disk atmosphere will be required to definitely exclude that swept-up shells with enough angular momentum and appropriate kinematics could be generated. However, at this stage and taking into account all the above-mentioned difficulties, we do not favor this scenario as the origin of the small-scale rotating CO outflow in DG Tau B. In the next section, we discuss an alternative scenario where this rotating outflow traces a (perturbed) extended disk wind. ### 6.2 The disk-wind scenario: constraints on the driving mechanism We therefore favor the scenario in which the inner CO conical outflow traces matter directly ejected from the disk. We show in Sect. 5 that the stratified kinematical structure derived for the conical outflow is suggestive of a quasi-steady disk-wind. In this section, we discuss constraints on the driving mechanism. Disk winds come in different flavors, depending on the main physical mechanism responsible for driving the flow: pure thermal effects in photo-evaporated disk winds (PDW) (Alexander et al. 2014), cold magneto- centrifugal ejection (Pudritz et al. 2007) or a combination of the two processes in the so-called warm or magneto-thermal disk winds (Casse & Ferreira 2000; Bai et al. 2016). In the following we refer to the two last classes of magnetized disk winds as MHD disk winds (hereafter MHD DW). Figure 17: Launching radius of the streamline as a function of its poloidal velocity assuming: conical extrapolation for the geometrical radius (green circles), steady thermal ejection conserving angular momentum (orange circles) and steady cold magneto-centrifugal ejection (blue circles). See text for more details. We do not derive the launching radius for the largest poloidal velocities due to the large uncertainties. Figure 18: Specific angular momentum ($j=R\times V_{\phi}$) versus poloidal velocity $V_{\rm P}$ for steady and axisymmetric MHD disk winds. The black symbols represent values derived from the tomography and averaged along lines of constant $V_{\rm Z}$, with their uncertainty. Red curves show the expected relation from self-similar cold magneto-centrifugal disk winds with $r_{0}$ varying from 0.5 to 100 au and $\lambda$ between 1.52 and 2.3 from Ferreira et al. (2006). The green dot corresponds to a warm MHD disk wind solution from Casse & Ferreira (2000) with $r_{0}=1$ au and $\lambda=1.9$. The green arrow shows the path followed by the solution with an increase of $r_{0}$. The gray box represents the estimated specific angular momentum for the outer low velocity layer on Sect. 4.1, taking a maximal poloidal velocity corresponding to a height of $Z=3300$ au in the outflow referential. #### 6.2.1 Photo-evaporated disk winds In photo-evaporated disk winds, the high energy radiation (UV and X-rays) of the central accreting protostar heats the surface layers of the disk to high temperatures (103-104 K) up to significant radial distances. Beyond the gravitational radius $r_{\rm g}=(GM_{\star})/c_{\rm s}^{2}$, where $c_{\rm s}$ is the sound speed in the upper disk surface layers, thermal energy exceeds the gravitational binding energy. Numerical simulations show that significant mass loss starts before $r_{\rm g}$, at the critical radius $r_{\rm c}\simeq 0.1-0.2r_{\rm g}$. Consequently, matter is ejected and reaches terminal velocities of typically two to three times $c_{\rm s}$ (Alexander et al. 2014, and references therein). The exact properties of the wind depends on the dominant source of high-energy irradiation. Extreme-UV (EUV) heating creates an isothermal ionized layer on the disk surface with temperature $T\simeq 10^{4}$ K ($c_{\rm s}=10$ km s-1) which drives a fast wind ($\sim 35$ km s-1) with mass-loss rates 10-9-10-10 M⊙ yr-1 (Font et al. 2004; Wang & Goodman 2017). X-ray irradiation results in cooler and slower flows ($c_{\rm s}\simeq 3-5$ km s-1, $v\simeq 15-20$ km s-1) but penetrates at higher densities and therefore can drive mass-loss rates up to a few 10-8 M⊙ yr-1 (Picogna et al. 2019). Non ionizing far-UV (FUV) heating mostly drives slow mass-loss ($v\simeq 18$ km s-1) from the outer disk regions. The conical morphology of the DG Tau B CO outflow matches expectations from both self-similar PDW models by Clarke & Alexander (2016) and hydrodynamical simulations by Owen et al. (2011) and Wang & Goodman (2017). In such disk-wind solutions, the specific angular momentum is conserved along the streamlines and is equal to the Keplerian value at the foot-point radius. The observed values of $j$ in the CO conical outflow imply footpoint radii in the range $r_{0}=1.6$ to 8.2 au. Figure 17 shows that these foot-point radii are lower (by a factor $\simeq$ three) than the ones derived from a straight extrapolation of the conical morphology on large scales. This requires a larger opening angle of the streamlines at the base, consistent with the self- similar models of Clarke & Alexander (2016). The values of $r_{0}$ are in line with expectations from EUV dominated PDW models. Indeed, for a 1 M⊙ star and $c_{\rm s}\simeq 10$ km s-1, expected in EUV heated PDW, the gravitational radius $r_{\rm g}\simeq 10$ au and significant mass loss starts at $r_{\rm c}\simeq 1-2$ au. On the other hand, observed terminal velocities of $V_{\rm p}=4-16$ km s-1 require sound speeds $\leq$ 2-8 km s-1 at radii $r_{0}\leq 10$ au, excluding EUV-driven models. The large derived mass flux of $1.7-2.9\times 10^{-7}$ M⊙ yr-1 for the CO conical flow (DV20) within $r_{0}\leq 10$ au however excludes FUV driven winds, which fail by at least one order of magnitude (Wang & Goodman 2017). We show below that it is also inconsistent with the latest X-ray driven models. Indeed, recent X-ray driven photo-evaporation models of Picogna et al. (2019) predict mass loss rates up to 10-7 M⊙ yr-1 for stellar X-ray luminosities $L_{\rm X}\geq 10^{31}$ erg s-1. However, these high $L_{\rm X}$ models also predict a stronger contribution of the outer disk regions to the total mass flux due to the increased penetration of X-ray photons. Figure 9 in Picogna et al. (2019) shows that only 10 % of the total mass flux originates from disk foot-point radii below 10 au. In summary, current PDW models fail to account for the combination of large mass flux and small foot-point radii of $r_{0}=1.6-8.2$ au, derived for the DG Tau B CO conical flow. Last but not least, the survival of CO molecules in such a wind is problematic. The full thermo-chemical computation of Wang & Goodman (2017) shows that in their fiducial models, CO survives only at the very base of the wind in an intermediate layer on scales $z/r\leq 0.6$. However, the models of Wang & Goodman (2017) are EUV dominated and hence result in warm and fully ionized winds. Similar problems are expected in thermally driven winds launched from the inner disk, which require base temperature greater than 2000 K. Therefore, we conclude that pure thermal processes appear highly unlikely as the main driving mechanism of the DG Tau B CO conical wind. #### 6.2.2 Magnetic disk winds Disk winds driven by magnetic forces require a large scale poloidal magnetic field anchored in the disk. This large-scale field exerts a torque on the rotating disk that both ejects matter and removes angular momentum from the disk (Blandford & Payne 1982; Pudritz et al. 2007). The strength of this torque is characterized by the magnetic lever arm parameter $\lambda\simeq(r_{A}/r_{0})^{2}$, where $r_{A}$ is the poloidal Alfven radius and $r_{0}$ the disk foot-point radius of the streamline. In principle, such disk winds can produce at the same time fast and collimated jets originating from the inner streamlines and much slower and less collimated winds originating from outer disk radii. The full kinematics and morphology of these solutions also depend on whether thermal effects are important in the launching regions. Numerical simulations show that the mass loss can be significantly increased when thermal effects are taken into account (Casse & Ferreira 2000; Bai et al. 2016). Such magneto-thermal winds have low to moderate $\lambda$ values but can extract significant mass and angular momentum from the disk. Terminal velocities depend on both the foot-point radii $r_{0}$, $\lambda$ and thermal effects. Under the assumption of steady magnetically driven ejection, the asymptotic values of $V_{\rm p}$ and $V_{\phi}$ are given by (Eqs. 4&5 in Ferreira et al. 2006): $\displaystyle RV_{\phi}$ $\displaystyle=$ $\displaystyle\lambda R_{\rm 0}V_{\phi}(R_{\rm 0})$ (15) $\displaystyle V_{\rm p}$ $\displaystyle=$ $\displaystyle V_{\phi}(R_{\rm 0})\sqrt{2\lambda-3+\beta},$ (16) where $\beta$ encompasses all pressure effects, including thermal and turbulent Alfvén waves (see Ferreira et al. 2006). The streamline foot-point radius can be estimated from these equations assuming cold MHD ejection, ie. negligible thermal effects ($\beta\simeq 0$) following (Anderson et al. 2003). Figure 18 traces the relationship between the mean values of $V_{\rm P}$ and $j=RV_{\phi}$ for the various conical layers of constant $V_{\rm Z}$ derived from the tomography. In this figure is also represented the parameter space ($r_{0}$,$\lambda$) predicted by cold magneto-centrifugal disk-wind models (Ferreira et al. 2006). The mean poloidal velocities and specific angular momentum coincide with a line of constant $\lambda_{\rm cold}\simeq 1.58$ with foot-point radii $r_{0,\rm cold}=0.7-3.4$ au. If thermal effects play a dynamical role at the base of the wind, Eqs. 15,16 show that the values of $\lambda$ and $r_{0}$ derived under the cold assumption are respectively upper and lower limits. This effect is illustrated in Fig. 18 where we plot $V_{\rm P}$ and $j$ from the warmest solution of Casse & Ferreira (2000) with $\lambda=1.9$, at $r_{0}=1$ au. We see that using the cold MHD curves would lead to overestimate $\lambda\simeq 2.5$ and underestimate $r_{0}\simeq 0.7$ au. The cold assumption is only valid if $\beta\ll 2\lambda-3\simeq 0.2$, which would require a cold disk atmosphere and no substantial wind heating. On the other hand, the low derived upper limit on $\lambda\leq 1.6$ is consistent with warm MHD DW models (Casse & Ferreira 2000; Bai et al. 2016; Wang et al. 2019) or cold MHD DW from weakly magnetized disks (Jacquemin-Ide et al. 2019). The derived minimum foot-point radius of $r_{\rm in}\geq 0.7$ au for the CO wind streamlines is in good agreement with the thermo-chemical predictions of Panoglou et al. (2012), who demonstrate that CO molecules magnetically launched from foot-point radii $\geq 1$ au survive in the case of accretion rates in the disk $\geq 10^{-7}$ M⊙ yr-1. Similar results are obtained for warm magneto-thermal wind solutions (Wang et al. 2019). The streamline foot-point radii derived from the kinematics are significantly smaller than the radii obtained from direct conical extrapolation (Fig. 17), suggesting wider opening angle of the streamlines at their base. This is indeed expected in MHD DW solutions where streamlines originally follow a conical trajectory and recollimate on larger scales due to the hoop stress provided by the azimuthal B-field. The constant opening angle of the streamlines observed out to $Z=3000$ au suggests that the magnetic hoop stress drops rapidly above $Z\simeq 50$ au. Contrary to pure thermal disk winds, MHD DW also account for the observed large mass flux in the DG Tau B conical CO flow. Wind mass loss rates in the range 10-8-10-7 M⊙ yr-1 are predicted by the magneto-thermal wind solutions of Wang et al. (2019), on the same order as the accretion rate in the underlying disk and increasing with disk magnetization. We estimated the local ejection efficiency $\xi$ defined as $\dot{M}_{\rm acc}(r)\propto r^{\xi}$ (Ferreira et al. 2006). We estimated $\xi$ from estimates of the disk accretion rate at the inner launching radius of the CO outflow. Indeed, from the mass conservation across the disk region launching the conical CO outflow ($\dot{M}_{\rm acc}(r_{\rm in})=\dot{M}_{\rm acc}(r_{\rm out})-\dot{M}_{\rm DW}$) and the definition of $\xi$, we derived the following expression: $\frac{\dot{M}_{\rm DW}}{\dot{M}_{\rm acc}(r_{\rm in})}=\bigg{(}\frac{r_{\rm out}}{r_{\rm in}}\bigg{)}^{\xi}-1.$ (17) We estimated the accretion rate onto the central star by taking 10 % of the jet mass flux (Ellerbroek et al. 2013). Podio et al. (2011) estimate the red jet mass flux at $6.4\times 10^{-9}$ M⊙ yr-1, giving a mass accretion rate onto the star of $\simeq 6\times 10^{-8}$ M⊙ yr-1. We took this value as a lower limit to $\dot{M}_{\rm acc}(r_{\rm in})$. From the measured mass flux in the conical CO flow $\dot{M}_{\rm DW}=2.3\times 10^{-7}$ M⊙ yr-1 and the disk- wind launching zone $r_{\rm out}/r_{\rm in}\simeq 5$, we then derived an upper limit on $\xi\leq 1$. The mass flux in the conical wind is $\sim 4$ times the estimated accretion rate onto the star, implying that 80% of the mass accreting at $r_{\rm out}$ is being ejected before reaching the star. If the transport of angular momentum in the disk is entirely provided by the torque exerted by the MHD DW, one expects in steady state the following relationship: $\xi=1/(2(\lambda-1))$. The upper limit derived above on $\xi\leq 1$ translates into $\lambda\geq 1.5$. This condition is compatible with our upper limit on $\lambda\leq 1.58$ derived from the kinematics. Thus the CO mass flux combined with the constraints on launching radii and magnetic lever arm appear compatible with an MHD DW extracting all angular momentum required for the disk to accrete from $r_{\rm out}$ to $r_{\rm in}$. If the inner conical outflow is tracing a disk wind, then the outer parabolic outflow cannot be explained by the interaction between the envelope and an inner jet or X-wind. However, the outer outflow could be tracing the interaction between the envelope and outer disk-wind streamlines located outside of the conical outflow ($r_{0}>4$ au). The arches located at least partially outside of the conical outflow as well as the continuous aspect in the PV diagrams suggests an ”intermediate” outflow located between the conical outflow and the outer parabolic surface. Alternatively, the outer flow could also be tracing directly the outer disk-wind streamlines. The derived $j$ and maximal velocity in the outer layer would indicate launching radii $\geq 30$ au and a similar low $\lambda$ value as for the inner cone (see Fig. 18). Moreover, Bai et al. (2016) show that some MHD DW solutions can accelerate until R $\sim$ 100 $r_{0}$, possibly explaining the apparent acceleration of this component seen in our channel maps. The global outflow would then be a continuous MHD DW originating from 0.7 au to $\geq 30$ au. Unfortunately, the morphology and dynamics of the potential disk wind originating from $r_{0}>4$ au could not be studied in detail due to our limited spectral sampling and the absorption by the surrounding cloud or envelope. However, in that scenario, the origin for the difference of emissivity between the bright conical outflow and the faint outer flow is not clear but could reflect the radial distribution of magnetic field strength or surface density in the underlying disk. ### 6.3 Disk-wind scenario: origin of perturbations We discuss here the merits of different models for the origin of the substructures (arches, fingers, and cusps) in the conical flow, observed in the channel maps. #### 6.3.1 Perturbation by jet bow shocks A first potential explanation for the bright substructures seen in the conical outflow is that the steady disk wind is perturbed by nested bow shock wings created by the propagation of the variable axial jet (see Fig. 20, scenario B). Perturbation of the inner streamlines of a rotating disk wind by a large jet bow shock is recently reported in the much younger system of HH 212 by Lee et al. (2021). In DG Tau B, this interpretation is supported by the similar spatial spacings between the inferred locations of the perturbations producing the substructures in the CO conical wind, and the axial jet knots identified in optical images, as shown in Fig. 13 (for z=500-3000 au along the flow axis spacings range between 200-1300 au for the jet knots, 300-700 au for the over- densities). Indeed, in the jet-wind interaction scenario, the over-densities trace the point of contact between each bow shock and the outer disk wind so they propagate along the interface at the bow shock propagation speed, which is similar to the inner jet knot propagation speed. Although the optical knot observations are not synchronous with our ALMA CO observations, and jet knots move away from the source at the jet speed $\simeq 150$ km s-1, the general pattern of knot spacing as a function of distance is set by the underlying jet variability properties, and thus will tend to remain similar at different epochs in a given jet. If, in addition, the jet undergoes low-amplitude wiggling (as frequently seen in young stars), perturbations to the disk wind caused by jet bow shocks would be slightly nonaxisymmetric, possibly explaining the apparent distortions in specific angular momentum along disk-wind streamlines using tomography (see Fig. 4). Finally, this scenario might also explain the lack of recollimation of the conical disk-wind streamlines, due to the additional internal pressure created by the jet driven bow shocks wings. Hydrodynamical simulations of the interaction of a variable inner jet with a slower outer disk wind have been recently performed by Tabone et al. (2018). These simulations show the formation of a dense stationary conical layer closing down at the source, created by the stacking of jet bow shock wings in the disk wind. This shell exhibits local over-densities at the positions of individual jet bow shocks, illustrated by the red regions in Fig. 10 (right panels) in Tabone et al. (2018) . These over-densities globally reproduce the observed shapes of the perturbations in the DG Tau B conical layer: for the perturbations closer-in, the density map is dominated by the regions close to the bow shock apex which bend inward, while for the perturbations farther out the density map is dominated by the bow shock wings which bend outward. This simulation also shows that the conical dense shell mostly retains the velocity of the surrounding disk wind, because the shock is weak and oblique. Therefore, the simulations in Tabone et al. (2018) does not show the characteristic stratification in $V_{\rm Z}$ observed in the DG Tau B inner conical flow. The numerical simulations of Tabone et al. (2018) also reproduce the observed trend of similar spacings between the axial jet knots and the over-densities in the contact region (shown in red in their Fig. 10, right panels). However, the simulations are made in a simplistic configuration where the outer disk wind is assumed to have a constant vertical velocity of 40% of the jet speed, uniform density, and no rotation motion. More realistic simulations taking into account the velocity and density gradients across the outer disk wind, and including rotation and magnetic fields, are strongly needed to reliably test this scenario. #### 6.3.2 Possible origin of CO outflow axis wiggling As discussed in Sect. 5.2, wiggling of the wind ejection axis is the only model explored here that can reproduce simply the variations of angular momentum observed along the DG Tau B outflow. From this analysis, we derived estimates of the wiggling period $\tau\simeq 400$ yrs and semi-amplitude wiggling angle $\alpha\simeq 0.5^{\circ}$. Wind axis wiggling can originate from the precession of the underlying disk angular momentum vector. Disk axis precession can be induced by a mis-aligned companion. In that scenario, orbital periods are expected to be significantly shorter than the disk axis precession period (Terquem et al. 1998), hence the companion would be located well within the disk. A companion in a mis-aligned orbit can open a gap in the disk, separating the dynamical evolution of the inner and outer disks (Zhu 2019). The inner disk then starts to precess with a period $\tau_{\rm p}$ related to the orbital period of the companion by $\tau_{o}/\tau_{\rm p}\simeq 0.37\mu/\sqrt{1-\mu}$, where $\mu=m_{2}/(m_{1}+m_{2})$ is the ratio of the companion mass to the total mass of the binary system, and assuming a small mis-alignment $i_{\rm p}$ as suggested by the maximal wind precession angle $\alpha=0.5\degr$ (see Eq. 27 in Zhu 2019). An additional constraint can be obtained by requiring that the semi-amplitude wiggling angle $\alpha$ is dominated by the precession motion, which translates into the condition $V_{0}\leq V_{\rm Z}\tan(\alpha)$, that is $V_{0}\leq 0.1$ km s-1where $V_{0}$ is the orbital velocity of the flow source. Combining these two constraints, we derive a companion mass ratio $\mu\leq 2.5\times 10^{-3}$ and binary separation: $a\leq 0.5$ au. This separation is smaller than the launching radius of the CO disk wind ($r_{0}\geq 0.7$ au, see Fig. 17). So the precessing disk launching the CO flow would be outside the orbit of the planetary mass companion, which is inconsistent with the scenario investigated here. Indeed, in such a scenario the outer disk is precessing on much longer timescales than the ones given by the formula above. Nixon et al. (2013) and Facchini et al. (2018) investigate circumbinary disk precession around an inner mis-aligned binary system. In such circumstances the inner rim of the circumbinary disk can break from the outer disk and precess. However, large mis-alignments or massive enough companions are necessary for this situation to occur (Facchini et al. 2018). We show that this is not likely in the DG Tau B case. The inner binary system would truncate the disk at 1.5-1.7 times the separation $a$ of the binary (Facchini et al. 2018). If we take an upper limit of $r_{\rm in}=0.7$ au for the inner circumbinary disk rim, to allow the launching of the CO flow, we get an upper limit of $a\leq 0.5$ au for the binary separation, corresponding to an orbital period $\tau_{0}\leq 0.35$ yrs using Third Kepler’s law and a total mass of 1 M⊙ for the binary system (DV20). With a small misalignment (less than a few degrees) between the outer disk and the binary orbital planes, suggested by the small wiggling angle of the CO outflow, a mass ratio of the companion $\mu\leq 10^{-2}$ would be required to get a precession period of 400 yrs (see Eq. 4 in Facchini et al. 2018). Equation 2 in Facchini et al. (2018) then shows that such a low mass companion combined with a small misalignment will not break the inner circumbinary disk. Therefore this second precession scenario can be ruled out to explain the CO outflow wiggling. So far we have investigated only precession as the origin of the wiggling of the CO flow axis. However, wiggling of the disk-wind ejection axis can be also induced by the orbital motion of the CO outflow source in a binary system. The equations of motion of the ejected gas will be equivalent to the precession solution investigated so far. We follow the formulation developed by (Masciadri & Raga 2002; Anglada et al. 2007) assuming an orbital plane perpendicular to the outflow axis. From the third Kepler’s law of motion, $a^{3}=M_{\rm tot}\tau_{o}^{2}$, with $\tau_{o}=400$ yrs and $M_{\rm tot}=1$ M⊙ we can derive the mean separation of the companion at $a\simeq 50$ au. On the other hand, from the semi-amplitude of the wiggling ($\alpha\simeq 0.5^{\circ}$) the orbital velocity of the CO outflow source is constrained at: $V_{0}=V_{\rm CO}\times\tan(\alpha)=0.1$ km s-1, using an average velocity of $\simeq 10$ km s-1 for the CO outflow. This in turn gives the orbital radius of the outflow source around the center of mass of the binary $r_{0}=1.3$ au and the ratio $\mu$ between the mass of the companion and the total mass of the system: $\mu=r_{0}/a=0.025$. Thus a brown dwarf or massive planetary mass companion located at $\simeq 50$ au ($=0.35^{\prime\prime}$) separation would be required to account for the observed wiggling of the CO outflow in the orbital scenario. Such a low mass companion could have escaped direct detection so far (Rodríguez et al. 2012). Strikingly the predicted companion separation is very close to an emission bump at $r=62$ au detected in the continuum emission profile of the disk at millimetric wavelengths (de Valon et al. 2020; Garufi et al. 2020). However no clear gap is detected in the disk emission at this position which would be expected for such a massive companion. Therefore, the precession scenario is excluded to account for the observed wiggling of the CO flow while the orbital scenario requires a brown dwarf or massive planetary mass companion at 50 au separation, which signature we do not clearly see in the disk yet. We also recall that our wiggling models cannot account for the observed variable separations between cusps in channel maps. So we conclude that although attractive to explain some of the substructures observed in the DG Tau B CO outflow, the interpretation of the wiggling scenario faces some difficulties. We discuss below the alternative model where substructures arise from axisymmetric brightness enhancements in the disk wind. #### 6.3.3 A variable disk wind We show in Sect. 5.3 that the cusps, fingers, and a section of the arches can be explained by brightness or density enhancements in the conical outflow. The timescales between the density enhancements are typically a few hundred years (see Fig. 14). Unfortunately, these timescales cannot be directly compared to the ones observed in the DG Tau B jet due to the larger jet velocity and its fading brightness at large distances. However, such timescales are observed on younger molecular outflows. The cluster W43-MM1 in Nony et al. (2020), CARMA-7 in Plunkett et al. (2015) as well as HH46/47 in Zhang et al. (2019) show signatures of variability in molecular outflows with timescales between episodic events typically of a few hundred years. We may be witnessing similar variability in the DG Tau B CO outflow. In the following, we discuss the possibility that these density enhancements are created by variability at the source in the disk-wind launching regions. For the model presented in Fig. 15, we considered for the sake of simplicity that the density bursts take place simultaneously at all radii in the disk (that is, $\tau_{i}=cst$ for all layers). A more physical assumption would be to assume that the density burst propagates radially across the disk with a velocity $V_{\rm prop}$. If we consider that the burst takes place close to the mid-plane ($Z\approx 0$), the expression of the travel time for the density launched from a fixed radius $r_{0}$ is : $\displaystyle\tau(\rm r_{0})$ $\displaystyle=$ $\displaystyle t-t_{\rm eject}(r_{0})$ (18) $\displaystyle t_{\rm eject}(r_{0})$ $\displaystyle=$ $\displaystyle t_{\rm eject}(0)+\frac{r_{0}}{V_{\rm prop}}$ (19) $\displaystyle\tau(\rm r_{0})$ $\displaystyle=$ $\displaystyle\tau(0)-\frac{\rm r_{0}}{\rm V_{\rm prop}},$ (20) where $t$ is the current time and $t_{\rm eject}(r_{0})$ is the epoch of density ejection. Here, $V_{\rm prop}$ is positive when the density burst moves from the inner to the outer disk regions. Figure 19: Dynamical timescales of the density enhancements responsible for the high-velocity cusps in the channel maps, as a function of the launching radius $r_{0}$ of the streamline under the cold MHD disk-wind hypothesis (assuming $\lambda=1.58$). For each cusp, the red dashed line represents the best linear fit of $\tau(r_{0})$ and its corresponding velocity. The one $\sigma$ domain is shown in gray and the uncertainties of the fits are noted on the right. Figure 14 indicates that $\tau$ decreases with increasing poloidal velocity, which is consistent with a density enhancement moving from the outer to the inner regions of the disk. Figure 19 is a modification of Fig. 14 in which we transformed the poloidal velocity into the corresponding radius of ejection under the MHD disk-wind hypothesis, using Eq. 8 from Ferreira et al. (2006) with $\lambda_{\phi}=1.58$ and $\beta=0$. We did not take into account the uncertainty on the estimation of $r_{0}$ ($\simeq 20\%$, see Fig. 17) as these uncertainties would make impossible the estimation of $V_{\rm prop}$. Nonetheless, this study gave an estimate of the propagation velocity. For each cusp, we traced the profile $\tau(\rm r_{0})$ derived from the observed positions of the cusp apex in the different channel maps, as described in Sect. 5.3. Linear fits to these profiles with associated $\boldsymbol{\tau}$ uncertainties are also shown. The slopes of these profiles is directly linked to $1/V_{\rm prop}$, and hence give the radial propagation velocities of the density enhancement at the origin. The average velocity over all cusps is $V_{\rm prop}\approx-0.2$ km s-1. The same study could be achieved with PDW models. In that case, the radius of ejection were multiplied by $\lambda^{2}$ ($=2.5$) and therefore $V_{\rm prop}$ values were also multiplied by 2.5, giving an average radial velocity $V_{\rm prop}\approx-0.5$ km s-1. From uncertainties on the slopes, we derived uncertainties on $V_{\rm prop}$ values in the case of U3, U4, and U5. In the case of U0,1,2, a horizontal solution ($V_{\rm prop}=\infty$) could not be excluded. For these last three fits, we derived the minimal radial velocities for the two opposite directions of propagation. Fig. 19 seems to indicate that the density bursts propagate from the outer to the inner regions of the disk at $\approx|V_{\rm prop}|=0.2-0.5$ km s-1. The accreting density must also propagate toward the inner regions of the conical flow ($r_{0}<2$ au) where the density enhancements were visible but not characterized. This propagation velocity corresponds to $\approx 0.01-0.04\leavevmode\nobreak\ V_{\rm Kep}(r=2-5\leavevmode\nobreak\ \rm au)$. Such accretion velocities match expectations for radial surface velocities in MHD wind-driven accretion (Riols et al. 2020). From this velocity and our estimate of the burst duration of $\simeq$ 23 years, we can constraint the radial extent of the burst propagating along the disk at $\Delta R\approx 1$ au. Figure 20 illustrates the proposed scenario. Episodic density bursts propagating inward are obviously reminiscent of Fu Ori and Ex Ori type variable accretion events. Some Fu Ori have burst durations $<30$ years (e.g., V1515 Cyg, or V1714 Cyg in Hartmann & Kenyon 1996). But periods are usually assumed to be $\approx$ 10${}^{4}-10^{5}$ years. On the other extreme, Ex Ori have typically bursts of a few months and periods of few years. DG Tau B variability seems to be located between these two extrema. However, the mass accretion rate increase in Fu Ori type events is typically three to four orders of magnitude higher than suggested in the DG Tau B outflow by the moderate factor three emissivity enhancement during the bursts. Moreover, models of Fu Ori events predict a global accretion affecting the whole vertical structure of the disk during the high state. In magnetically accreting disk models, most of the mass is concentrated in the mid-plane, where the radial accretion velocity is $\simeq V_{\rm Kep}/10^{3}$ (Riols et al. 2020). The high propagation velocity combined with the moderate emissivity enhancements derived in DG Tau B suggest that the accretion burst takes place locally on the disk surface and is less extreme than in typical Fu Ori phenomena. Such moderate accretion bursts could be due for example to residual infalling envelope material creating a shock wave when infalling into the disk (Hennebelle et al. 2017). It is important to note, however, that this axisymmetric model does not reproduce the local deviations observed in the specific angular momentum map (Fig. 3). Nonaxisymmetric perturbations would be required. Figure 20: Schematic scenarios describing the different components of the DG Tau B redshifted outflow in the disk-wind paradigm (see text). ## 7 Conclusions We present a detailed analysis and modeling of the ALMA 12CO(2-1) observations of the DG Tau B redshifted outflow published in de Valon et al. (2020), with the aim to constrain its origin. Our main conclusions are as follows: * • We identify three classes of discrete structures visible on the 12CO channel maps: arches at low velocities, fingers at medium velocities, and cusps at high velocities. Both cusps and arches show apparent acceleration in channel maps. * • We derived the 2D kinematics of the inner conical outflow using a tomographic method, assuming only axisymmetry of the outflow. We reconstructed 2D maps for both the expansion velocity $V_{\rm Z}$ and specific angular momentum $j=R\times V_{\phi}$. The inner outflow shows a striking $V_{\rm Z}$ shear with faster material closer to the flow axis. Lines of constant $V_{\rm Z}$ are conical from $Z\simeq 50$ au out to $\approx 1200$ au. Specific angular momentum is roughly constant along those lines (except in two localized regions), and it increases outward inversely with $V_{\rm Z}$ from $\simeq 40$ to 100 au km s-1. * • The lower velocity external CO outflow shows a parabolic morphology, apparent acceleration, and a large specific angular momentum of $j=250\pm 50$ au km s-1. This suggests that it is tracing either a swept-up infalling and rotating envelope, or an extended disk wind launched from $\sim 30$ au. * • The conical outflow and the discrete structures could not be described by wind-driven shells with radial Hubble velocity laws. Such models fail to reproduce at the same time the morphologies of the observed structures (arches, fingers, and cusps) in the channel maps. Numerical simulations of the interaction between an episodic jet- and wide-angle wind with an infalling envelope are, however, required to confirm these conclusions. * • Instead, the conical outflow global morphology and kinematics appear consistent with matter directly ejected from the disk. Constraints on the disk-wind foot-point radii were derived at $r_{0}=1.6-8.2$ au (resp. 0.7 - 3.4 au) in the limiting cases where thermal (resp. magneto-centrifugal) processes dominate. However, none of the current photo-evaporated wind models can reproduce the large observed mass flux ( $2.3\times 10^{-7}$ M⊙ yr-1 ) ejected from $r_{0}<10$ au. In contrast, an MHD disk-wind model with a constant magnetic level arm parameter $\lambda\leq 1.58$ can reproduce – at the same time – the flow velocity and angular momentum, as well as the large mass flux if it extracts most of the angular momentum for disk accretion across the wind launching region. The low lambda value is consistent with recent models of warm or weakly magnetized MHD DW (Bai et al. 2016; Jacquemin-Ide et al. 2019). * • The wiggling of the flow axis may explain both the localized deviations of specific angular momentum and the morphology of the substructures in the conical flow. Orbital motion of the flow source in a binary of separation $\simeq 50$ au with companion mass $2.5\times 10^{-3}$ M⊙ can explain the inferred wiggling period and amplitude. Such a low mass companion could have escaped direct detection so far, but it should produce a gap signature in the continuum dust disk emission, which is not currently detected. In addition, the wiggling scenario fails to account for the variable separation between cusps in channel maps. * • Alternatively, the substructures observed in the CO channel maps can be explained by a series of mild (a factor three) density perturbations in the wind launching region, propagating inward at a radial velocity of $\simeq 0.2-0.5$ km s-1, consistent with the surface accretion flow predicted in MHD wind-driven accretion. We derived a typical perturbation width of $\sim 1$ au and intervals of 200-500 years between perturbations. Alternatively the conical morphology and local density enhancements might be explained by the interaction of inner jet bow shocks with the disk wind (Tabone et al. 2018); although, further numerical simulations are required to fully test this hypothesis. The discrete structures that are increasingly observed in Class 0 and Class I outflows on larger scales have been usually interpreted in terms of nested shells swept up by an episodic inner wind. In contrast, we have shown in this paper that the substructures in the DG Tau B outflow appear best explained by density enhancements at the disk surface and that they propagate in a shear- like MHD disk wind. If confirmed, these results would directly demonstrate the link between accretion and ejection processes in embedded sources. DG Tau B is a Class I protostar with a structured disk, infalling flows, and possible variable disk wind. Structures in Class I disks are assumed to trace early stages of planetary formation processes. Therefore, our results suggest that planetary formation is taking place in a very dynamic environment. The impact of such outflowing and infalling flows on the disk and its evolution remains an open question. Additional models and simulations of variable disk wind are needed in order to fully comprehend its impact on disk evolution and planet formation. ###### Acknowledgements. The authors would like to thank the referee, whose comments helped improve the quality of the paper. This paper makes use of the following ALMA data: ADS/JAO.ALMA#2015.1.01108.S, ADS/JAO.ALMA#2017.1.01605.S. ALMA is a partnership of ESO (representing its member states), NSF (USA) and NINS (Japan), together with NRC (Canada), MOST and ASIAA (Taiwan), and KASI (Republic of Korea), in cooperation with the Republic of Chile.The Joint ALMA Observatory is operated by ESO, AUI/NRAO and NAOJ. This work was supported by the Programme National de Physique Stellaire (PNPS) and the Programme National de Physique et Chimie du Milieu Interstellaire (PCMI) of CNRS/INSU co-funded by CEA and CNES. FL acknowledges the support of the Marie Curie Action of the European Union (project MagiKStar, Grant agreement number 841276). ## References * Alexander et al. (2014) Alexander, R., Pascucci, I., Andrews, S., Armitage, P., & Cieza, L. 2014, in Protostars and Planets VI, ed. H. Beuther, R. S. Klessen, C. P. Dullemond, & T. Henning, 475 * Anderson et al. (2003) Anderson, J. M., Li, Z.-Y., Krasnopolsky, R., & Blandford, R. D. 2003, ApJ, 590, L107 * Anglada et al. (2007) Anglada, G., López, R., Estalella, R., et al. 2007, AJ, 133, 2799 * Arce et al. (2007) Arce, H. G., Shepherd, D., Gueth, F., et al. 2007, in Protostars and Planets V, ed. B. Reipurth, D. Jewitt, & K. Keil, 245 * Bai et al. (2016) Bai, X.-N., Ye, J., Goodman, J., & Yuan, F. 2016, ApJ, 818, 152 * Bjerkeli et al. (2016) Bjerkeli, P., van der Wiel, M. H. D., Harsono, D., Ramsey, J. P., & Jørgensen, J. K. 2016, Nature, 540, 406 * Blandford & Payne (1982) Blandford, R. D. & Payne, D. G. 1982, MNRAS, 199, 883 * Cabrit et al. (1997) Cabrit, S., Raga, A., & Gueth, F. 1997, in Herbig-Haro Flows and the Birth of Stars, ed. B. Reipurth & C. Bertout, Vol. 182, 163–180 * Casse & Ferreira (2000) Casse, F. & Ferreira, J. 2000, A&A, 353, 1115 * Clarke & Alexander (2016) Clarke, C. J. & Alexander, R. D. 2016, MNRAS, 460, 3044 * Cunningham et al. (2005) Cunningham, A., Frank, A., & Hartmann, L. 2005, ApJ, 631, 1010 * de Valon et al. (2020) de Valon, A., Dougados, C., Cabrit, S., et al. 2020, A&A, 634, L12 (DV20) * Downes & Cabrit (2007) Downes, T. P. & Cabrit, S. 2007, A&A, 471, 873 * Eislöffel & Mundt (1998) Eislöffel, J. & Mundt, R. 1998, AJ, 115, 1554 * Ellerbroek et al. (2013) Ellerbroek, L. E., Podio, L., Kaper, L., et al. 2013, A&A, 551, A5 * Facchini et al. (2018) Facchini, S., Juhász, A., & Lodato, G. 2018, MNRAS, 473, 4459 * Fernández-López et al. (2020) Fernández-López, M., Zapata, L. A., Rodríguez, L. F., et al. 2020, AJ, 159, 171 * Ferreira et al. (2006) Ferreira, J., Dougados, C., & Cabrit, S. 2006, A&A, 453, 785 * Font et al. (2004) Font, A. S., McCarthy, I. G., Johnstone, D., & Ballantyne, D. R. 2004, ApJ, 607, 890 * Garufi et al. (2020) Garufi, A., Podio, L., Codella, C., et al. 2020, A&A, 636, A65 * Gaudel et al. (2020) Gaudel, M., Maury, A. J., Belloche, A., et al. 2020, A&A, 637, A92 * Guilloteau et al. (2011) Guilloteau, S., Dutrey, A., Piétu, V., & Boehler, Y. 2011, A&A, 529, A105 * Hartmann & Kenyon (1996) Hartmann, L. & Kenyon, S. J. 1996, ARA&A, 34, 207 * Hennebelle et al. (2017) Hennebelle, P., Lesur, G., & Fromang, S. 2017, A&A, 599, A86 * Hirota et al. (2017) Hirota, T., Machida, M. N., Matsushita, Y., et al. 2017, Nature Astronomy, 1, 0146 * Jacquemin-Ide et al. (2019) Jacquemin-Ide, J., Ferreira, J., & Lesur, G. 2019, MNRAS, 490, 3112 * Launhardt et al. (2009) Launhardt, R., Pavlyuchenkov, Y., Gueth, F., et al. 2009, A&A, 494, 147 * Lee et al. (2018) Lee, C.-F., Li, Z.-Y., Hirano, N., et al. 2018, ApJ, 863, 94 * Lee et al. (2000) Lee, C.-F., Mundy, L. G., Reipurth, B., Ostriker, E. C., & Stone, J. M. 2000, ApJ, 542, 925 * Lee et al. (2001) Lee, C.-F., Stone, J. M., Ostriker, E. C., & Mundy, L. G. 2001, ApJ, 557, 429 * Lee et al. (2021) Lee, C.-F., Tabone, B., Cabrit, S., et al. 2021, ApJ, 907, L41 * Liang et al. (2020) Liang, L., Johnstone, D., Cabrit, S., & Kristensen, L. E. 2020, ApJ, 900, 15 * López-Vázquez et al. (2019) López-Vázquez, J. A., Cantó, J., & Lizano, S. 2019, ApJ, 879, 42 * Louvet et al. (2018) Louvet, F., Dougados, C., Cabrit, S., et al. 2018, A&A, 618, A120 * Masciadri & Raga (2002) Masciadri, E. & Raga, A. C. 2002, ApJ, 568, 733 * Mitchell et al. (1997) Mitchell, G. F., Sargent, A. I., & Mannings, V. 1997, ApJ, 483, L127 * Mundt et al. (1987) Mundt, R., Brugel, E. W., & Buehrke, T. 1987, ApJ, 319, 275 * Mundt & Fried (1983) Mundt, R. & Fried, J. W. 1983, ApJ, 274, L83 * Mundt et al. (1991) Mundt, R., Ray, T. P., & Raga, A. C. 1991, A&A, 252, 740 * Nixon et al. (2013) Nixon, C., King, A., & Price, D. 2013, MNRAS, 434, 1946 * Nony et al. (2020) Nony, T., Motte, F., Louvet, F., et al. 2020, A&A, 636, A38 * Ohashi et al. (1997) Ohashi, N., Hayashi, M., Ho, P. T. P., et al. 1997, ApJ, 488, 317 * Owen et al. (2011) Owen, J. E., Ercolano, B., & Clarke, C. J. 2011, MNRAS, 412, 13 * Panoglou et al. (2012) Panoglou, D., Cabrit, S., Pineau Des Forêts, G., et al. 2012, A&A, 538, A2 * Pety et al. (2006) Pety, J., Gueth, F., Guilloteau, S., & Dutrey, A. 2006, A&A, 458, 841 * Picogna et al. (2019) Picogna, G., Ercolano, B., Owen, J. E., & Weber, M. L. 2019, MNRAS, 487, 691 * Plunkett et al. (2015) Plunkett, A. L., Arce, H. G., Mardones, D., et al. 2015, Nature, 527, 70 * Podio et al. (2011) Podio, L., Eislöffel, J., Melnikov, S., Hodapp, K. W., & Bacciotti, F. 2011, A&A, 527, A13 * Pudritz et al. (2007) Pudritz, R. E., Ouyed, R., Fendt, C., & Brandenburg, A. 2007, in Protostars and Planets V, ed. B. Reipurth, D. Jewitt, & K. Keil, 277 * Riols et al. (2020) Riols, A., Lesur, G., & Menard, F. 2020, A&A, 639, A95 * Rodríguez et al. (2012) Rodríguez, L. F., Dzib, S. A., Loinard, L., et al. 2012, Rev. Mexicana Astron. Astrofis., 48, 243 * Shang et al. (2006) Shang, H., Allen, A., Li, Z.-Y., et al. 2006, ApJ, 649, 845 * Shang et al. (2020) Shang, H., Krasnopolsky, R., Liu, C.-F., & Wang, L.-Y. 2020, ApJ, 905, 116 * Shang et al. (1998) Shang, H., Shu, F. H., & Glassgold, A. E. 1998, ApJ, 493, L91 * Shu et al. (1991) Shu, F. H., Ruden, S. P., Lada, C. J., & Lizano, S. 1991, ApJ, 370, L31 * Tabone et al. (2017) Tabone, B., Cabrit, S., Bianchi, E., et al. 2017, A&A, 607, L6 * Tabone et al. (2020) Tabone, B., Cabrit, S., Pineau des Forêts, G., et al. 2020, A&A, 640, A82 * Tabone et al. (2018) Tabone, B., Raga, A., Cabrit, S., & Pineau des Forêts, G. 2018, A&A, 614, A119 * Terquem et al. (1998) Terquem, C., Papaloizou, J. C. B., Nelson, R. P., & Lin, D. N. C. 1998, ApJ, 502, 788 * Ulrich (1976) Ulrich, R. K. 1976, ApJ, 210, 377 * Wang et al. (2019) Wang, L., Bai, X.-N., & Goodman, J. 2019, ApJ, 874, 90 * Wang & Goodman (2017) Wang, L. & Goodman, J. 2017, ApJ, 847, 11 * Wang et al. (2015) Wang, L.-Y., Shang, H., Krasnopolsky, R., & Chiang, T.-Y. 2015, ApJ, 815, 39 * Wilgenbus et al. (2000) Wilgenbus, D., Cabrit, S., Pineau des Forêts, G., & Flower, D. R. 2000, A&A, 356, 1010 * Zapata et al. (2015) Zapata, L. A., Lizano, S., Rodríguez, L. F., et al. 2015, ApJ, 798, 131 * Zhang et al. (2019) Zhang, Y., Arce, H. G., Mardones, D., et al. 2019, ApJ, 883, 1 * Zhang et al. (2018) Zhang, Y., Higuchi, A. E., Sakai, N., et al. 2018, ApJ, 864, 76 * Zhu (2019) Zhu, Z. 2019, MNRAS, 483, 4221 ## Appendix A Determination of the redshifted outflow PA We show in Sect. 3 that the velocity difference between two radially symmetric positions, at a given altitude z in the flow, is directly linked to the rotation velocity of the outflow. However, this derivation is highly sensitive to the determination of the outflow radial center position. Because of the steep gradient in projected velocity versus radius observed in transverse PV diagrams, a small variation in radial position would cause an important variation in the velocity difference. Therefore, the determination of the position axis is a key concern in the characterization of the specific angular momentum of the outflow. To accurately derive the flow center position as a function of the projected altitude, we adopted the following method, illustrated in Fig. 21. At each projected altitude above the disk $\delta z$, we integrated the emissivity from $(V-V_{\rm sys})=6.64$ km s-1 to $(V-V_{\rm sys})=9.77$ km s-1. This gave the radial emissivity profile of the high-velocity emission tracing the limiting inner cone. Indeed as shown in DV20, at large projected velocities, the emission traces an inner cone with almost constant radius. We derived the radial positions of the two peaks tracing the edges of the high-velocity component (See Fig. 21) using Gaussian fitting. From the median value of these two positions, we derived the radial center position and its associated uncertainty. We applied this method after rotating the data cube with three different values of outflow PA (sampled around the disk PA), and the transverse PV diagram was obtained with a slice perpendicular to this axis. The effect of the different PV cut is completely negligible, as the considered variation of outflow P.A is small (less than 2 degrees). Figure 21 shows the derived radial offset as a function of the projected height for three values of the flow position angle (PA). With a PA of $296^{\circ}$ and $294^{\circ}$, the radial offset increases consistently with a miss-estimation of the outflow PA of $\pm 1^{\circ}$. The global offset is minimized with a PA of $295\pm 1^{\circ}$. This is consistent with the atomic redshifted jet PA of 296∘ derived by Mundt et al. (1987) and the PA of the projected disk in the plane of the sky, determined at $25.7\pm 0.3^{\circ}$ by DV20 and $24\pm 1^{\circ}$ by Guilloteau et al. (2011). The outflow rotation could potentially introduce a bias in this method. Indeed, at a given projected velocity, the two sides of a rotating ring fall at different projected radial offsets from the axis, inducing an artificial shift of the position centroid of the flow axis. However, since a rotating ring produces a tilted ellipse in the transverse PV diagram, centered at $V_{\rm Z}\times\cos(i)$, the radial shift has an opposite sense at velocities above and below $V_{\rm Z}\times\cos(i)$. By integrating emission over a broad range of velocity, we thus averaged out this effect. Moreover, if the variation of radial offset was due to flow rotation, the measured radial offsets should decrease with distance from the source, as the rotation velocity decreases (due to the conservation of angular momentum along conical streamlines). This is not consistent with our results in Fig. 21 where the radial offsets stay constant (at our nominal PA) or increase linearly (at non- nominal PAs) with height. Fig. 21 also shows no clear signature of wiggling in the high-velocity component of DG Tau B. This is consistent with the absence of wiggling in the atomic jet observed by Mundt et al. (1991). From the maximum error bars observed, we derived an upper limit for the wiggling angle of $\theta\leq 0.5^{\circ}$. We consider in the following that the redshifted outflow PA is constant for all layers (velocities) at $\rm PA=295^{\circ}$, corresponding to the PA derived here for the high-velocity emission. Figure 21: Determination of the outflow axis PA. Top panels: Determination of the high-velocity component radial center as a function of projected heights above the disk for different outflow axis PA orientations. The black dotted lines show an angle of $\pm 1^{\circ}$ with respect to the central axis. Middle panel: 12CO PV diagram perpendicular to the outflow axis. $(V-V_{\rm sys})$ unit is km s-1. Bottom panel: The black dashed profile shows the emissivity integrated from the PV diagram between $(V-V_{\rm sys})=6.64$ km s-1 to $(V-V_{\rm sys})=9.77$ km s-1, this domain is indicated by the red dashed lines in the middle panel. The red curves show Gaussian fits used to derive the radial positions of the two edge peaks. The vertical black dotted lines in the bottom and middle panels indicate the positions of these peaks. The average of these two radial positions give the radial center at this height. ## Appendix B Characterization of the cusps Figure 22: Characterization of the cusps. Left panels: Individual 12CO channel maps at different line-of-sight velocities illustrating the spatial evolution of the cusps. Six cusps, labeled U0-U5, are clearly identified (U4 is only visible on three channel maps). We identify the projected heights of the bottom of the cusps (red circles) and their evolution in the channel maps (white dashed lines). Right panel: Longitudinal PV diagram at $\delta x=0^{\prime\prime}$ averaged over a slice of $\Delta Z=0.2^{\prime\prime}$. The velocity and height locations of the cusps derived in the channel maps are shown as red dots. The four orange crosses indicate the maximum on-axis projected velocity (back side) of the top of the fit four bright elliptical structures identified in transverse PV diagrams at the corresponding heights (see Fig. 23). ## Appendix C Ellipse fits in the transverse PV diagrams In this section, we discuss an alternative method to derive the radial component of the velocity, not constrained by our tomographic method. At a few positions along the flow, transverse PV diagrams clearly show elliptical structures nested inside the main emission (see Fig. 23). We assumed that each of these ellipses traces one emitting layer of the outflow. If the outflow velocity does not vary drastically with height, one layer of the outflow of fixed radius will be projected as an inclined elliptical structure in the PV diagram (Louvet et al. 2018). The velocity width of the ellipse at $x=0$ is directly linked to the radial velocity component of the layer. We fit ellipses by the naked eye to the structures observed. From these fits, we recovered the radial velocity $V_{\rm R}$ as well as $V_{\rm Z}$ at a few specific positions along the outflow. Figure 23 shows the $V_{\rm Z}$ profile of the outflow reconstructed with the tomographic technique and extrapolated until Z $\approx$ 2700 au. The red arrows trace the poloidal vectors derived from individual elliptical fits. The angle of the arrow is determined by the ratio between $V_{\rm Z}$ and $V_{\rm R}$. The length and colors of the arrows correspond to $V_{\rm Z}$. The $V_{\rm Z}$ values derived from the elliptical fits appear consistent with the estimates from the tomographic study. The poloidal velocity direction is also consistent with the conical lines of constant $V_{\rm Z}$ in the tomography. This suggests that the flow is indeed aligned with these conical lines. Figure 23: Constraints on the radial velocity component $V_{\rm R}$. Left panels: Transverse PV diagrams averaged over a slice of width $\Delta Z=0.2^{\prime\prime}$ at selected positions $\delta z$ along the flow. In red are shown the ellipse fits. We show side by side the PV diagram with and without the fit for more visibility. Right panel: $V_{\rm Z}$ tomography of the outflow in the outflow referential, extrapolated beyond $Z>1200$ au. The red arrows represent the velocity directions determined from the ellipse fits. The arrow colors trace the $V_{\rm Z}$ values determined from the ellipse fit. $R$ and $Z$ units are in au. We studied the relation between the ellipses visible in the transverse PV diagrams and the cusps visible in the channel maps. The on-axis maximal velocity of the fit ellipses was compared to the cusp location in the on-axis longitudinal PV diagram (see Fig. 22). The (R,Z) positions of the fit ellipses were also compared to the cusp locations in the tomographic $V_{Z}$ map (see Fig. 13). The ellipse located at $\delta z=1.1^{\prime\prime}$ comes from a lower altitude region, not included in our cusp analysis because of crowding. The ellipses at $\delta z=2.1^{\prime\prime}$ and $2.9^{\prime\prime}$ seem to extend the U5 cusp, at higher and lower line-of-sight velocities respectively. The ellipse at $\delta z=7.5^{\prime\prime}$ is not so clearly associated with a cusp extension in Fig. 22. It could be the high velocity extension of the U3 cusp, or associated with a fainter cusp located between U2 and U3, not included in our analysis. Moreover, Fig. 24 indicates that the tops of the ellipses located at $\delta z=2.9^{\prime\prime}$ and $\delta z=7.5^{\prime\prime}$ are possibly consistent with a cusp structure which was not included in our study, due to their low S/N. Similarly, Fig. 24 shows that the cusps U1 and U2 also coincide with internal structures in the transverse PV diagram, but the stacking of ellipses in the PV diagram makes the identification more confusing than in the channel maps. The cusps could be precisely located only at moderate projected velocities $<5.3$ km s-1, where they are sufficiently bright (see Fig. B.1), hence they probe outer streamlines; in contrast elliptical structures in transverse PV diagrams are best distinguished at high projected velocities $>5.3$ km s-1where they have less overlap with each other (see Fig. C.1), hence most of them probe inner faster flow streamlines. However, we see clear correspondences between faint cusps in channel maps and the on-axis high- velocity portion of some ellipses in transverse PV cuts, and vice-versa, which demonstrates that they trace different portions of the same underlying substructures, extending across the whole conical outflow. Therefore, assuming a radial flow to deproject the cusp apparent positions seems fully justified. Figure 24: Illustration of the correspondences between ellipses and cusps. Left panels: transverse PV diagrams at $\delta z=2.9^{\prime\prime}$ and $7.5^{\prime\prime}$ (from Fig 23), with red dots indicating the on-axis projected velocities of cusps U1 and U2 at $\delta z=7.5^{\prime\prime}$. Each of these two cusps corresponds in position and velocity to the maximum velocity (back side) of an ellipse in the PV cut. Right panels: channel maps at the same line-of-sight velocity as the orange crosses in the left panels, showing that elliptical structures seem to correspond to fainter cusps in position-position space, but falling outside of the velocity domain where cusp characterization was possible. ## Appendix D Biases in the tomographic reconstruction We studied possible biases in the method presented in Sect. 3 to reconstruct poloidal maps of $V_{\rm Z}$ and $j=R\times V_{\phi}$. We first studied the bias introduced by projection effects considering one single layer of the outflow using the analytical solutions from Eqs. 1,2,3. To study the impact of beam convolution and multiple layers of the outflow, we also applied our tomographic reconstruction method to the synthetic data cube presented in Sect. 10 and estimate the difference between the reconstructed $V_{\rm Z}$ and $j$ and the initial values of the model. ### D.1 Single shell Figure 25: Biases in the tomographic reconstruction for one single conical shell. Top panels: Black curves show computed transverse PV diagrams for one conical layer with constant $V_{\rm Z}$, $V_{\rm R}$, and $j$ seen at different inclinations $i$ and with $\theta_{V}=\theta=17^{\circ}$. The red dashed curve traces an elliptical fit. Bottom Panel: Computed relative bias in $V_{\rm Z}$ (green curves) and $j$ (orange curves) for one conical layer as a function of the inclination and for different combinations of $\theta$ and $\theta_{v}$ : solid curves: $\theta=\theta_{v}=17^{\circ}$, dotted curves: $\theta=17^{\circ}$ & $\theta_{v}=30^{\circ}$, and dashed curves: $\theta=\theta_{v}=30^{\circ}$. In our tomographic study, we assumed that one shell will be projected as an ellipse in transverse PV diagrams. We also assumed that the projected velocities at the extrema radii allowed us to recover $V_{\rm Z}$ and $V_{\phi}$. We determine in this section biases introduced by these two assumptions. We assumed a conical shell of radius R(Z) in the outflow referential and with local opening angle $\theta$, such as $R(Z)=R_{0}+Z\tan{\theta}$. For the DG Tau B inner conical outflow, the opening angle $\theta$ of the layers vary between 17∘ and 12∘, increasing with decreasing velocities. The estimated opening angle of the lower velocity emission contributing to the pedestal is $\theta\simeq 30^{\circ}$. We defined the parameter $\theta_{v}$ corresponding to the angle of the velocity vector with the Z axis in the poloidal plane ($V_{\rm R}=V_{\rm Z}\tan{\theta_{V}}$). For a flow parallel to the conical surface, $\theta=\theta_{V}$. We also assumed a constant specific angular momentum ($V_{\phi}=j/R(\phi)$) as well as constant $V_{\rm Z}$ and $V_{\rm R}$ over the transverse slit width. Solving Eq. 2 with constant $\delta z$ gave: $R(\phi)=\frac{\delta z\tan{\theta}+R_{0}\sin{i}}{\sin{i}-\sin{\phi}\cos{i}\tan{\theta}}.$ (21) In the case $i=\theta$ and $i<\theta$, the cut of the conical outflow will be respectively a parabola and an hyperbola. Consequently, $R(\phi)$ will tend toward infinite. This is not the case for the DG Tau B conical outflow. We then implemented this solution into Eqs. 3,1 for multiple values of the inclination. The resulting transverse PV diagrams are represented in Fig. 25 for $\theta_{V}=\theta=17^{\circ}$. The PV diagrams were generated using Eqs. 3, 2 with the solution from Eq. 21. The difference with an ellipse increases with the inclination but is expected to be small for the DG Tau B case ($i=117^{\circ}$). We assume in Sect. 3 that the radial edges of the ellipse correspond to $\phi=0-\pi$. However, this is an approximation. The radial extrema of the ellipse correspond to solutions of the equation: $\frac{\partial\delta x}{\partial\phi}=0$. For a conical layer, the solution is obtained for: $\sin{\phi}=\frac{\tan{\theta}}{\tan{i}}.$ (22) This difference is small in our situation. However, this leads to bias in the estimation of $R$, $Z$, $V_{\rm Z}$ and $V_{\phi}$. We determined the estimated value of $V_{\rm est,Z}$ and $V_{\rm est,\phi}$ using Eqs. 6,7. We then computed the relative differences with the real values $R_{\rm real}$, $Z_{\rm real}$, $V_{\rm real,Z}$, and $V_{\rm real,\phi}$: $\frac{R_{\rm est}-R_{\rm real}}{R_{\rm real}}=\sqrt{1-\frac{\tan^{2}{\theta}}{\tan^{2}{i}}}-1$ (23) $\frac{Z_{\rm est}-Z_{\rm real}}{Z_{\rm real}}=-\frac{\tan^{2}\theta}{\tan^{2}i}$ (24) $\frac{V_{\rm est,Z}-V_{\rm real,Z}}{V_{\rm real,Z}}=\tan{\theta}\times\tan{\theta_{V}}$ (25) $\frac{V_{\rm est,\phi}-V_{\rm real,\phi}}{V_{\rm real,\phi}}=\sqrt{1-\frac{\tan^{2}{\theta}}{\tan^{2}{i}}}-1.$ (26) Hence, $R$, $Z$, and $j$ are systematically underestimated, while $V_{\rm Z}$ is overestimated. Figure 25 shows the relative biases due to inclination and projection effects in the estimation of $V_{\rm Z}$ and $V_{\phi}$ for different values of $\theta$ and $\theta_{v}$. In the conical outflow, where $\theta_{V}=\theta\leq 17^{\circ}$, the inclination bias is expected to be $\leq 10\%$ in $V_{\rm Z}$ and $Z$, $\leq$ 1 % in $j$ and $R$. ### D.2 Multiple layers Figure 26: Biases in the tomographic reconstruction for a stacking of conical shells. Left panels: The gray area shows the limits of the transverse PV diagram predicted for a stacking of conical shells with a shearing factor $f_{\rm sh}$ (see text). In black is shown the ellipse corresponding to a single layer of maximal radius $R$. The blue and red dots correspond respectively to the velocity of the ellipse at $\pm R$ and the maximal velocity of the PV diagram at $\pm R$. The red dots are the ones used in our method, while the blue dots correspond to the true measurements. The differences in velocities will lead to a bias in our $j$ and $V_{\rm Z}$ estimation. The red line shows the slope of the PV diagram $\frac{\partial V_{\rm los}(\pm R)}{\partial R}$, see text for more details. Right panel: Plot of the relative bias in $V_{\rm Z}$ (green curves) and $j$ (orange curves) due to the shearing aspect of the PV diagram as a function of the $x$ parameter (see text). The solid, dashed, and dotted lines correspond to different configurations of $\theta$ and $\theta_{v}$. The black histogram, labeled C.O., shows the distribution of the $x$ parameter derived at different heights Z and radii R in the conical outflow. The relative biases corresponding to the a and b ellipses shown in the left panels are represented. The orange (resp. green) colored contours trace the relative biases derived from applying the tomographic method to the synthetic data cube. Contours outline 30 to 90 % of the distributions in step of 15 %. In the previous section, we estimated the bias in the estimation of $V_{\rm Z}$, $V_{\rm\phi}$ due to projection effects for one single conical layer. However, the DG Tau B transverse PV diagrams shows a clear shear-like velocity structure suggesting a stacking of layers. We modeled this effect directly in the transverse PV diagrams by stacking the elliptical projections for conical layers of increasing radii at origin $R_{0}$ with the same opening angle $\theta$. The velocity shear in $V_{\rm Z}$ of the conical layers was defined by a shearing parameter $f_{\rm sh}=\frac{\partial V_{\rm Z}}{\partial R}$. The specific angular momentum of each layer, $j$, was assumed constant with $Z$ and vary between 20 and 90 au km s-1, increasing with increasing radius, such as $j\times V_{\rm P}$ was kept constant to mimic the DG Tau B observations. Due to the stacking, the maximal velocity at $\pm R$ does not perfectly describe the velocity of the ellipse of radius R. This effect is larger when $f_{\rm sh}$ is small. Using a similar procedure as before, we derived the relative difference between the estimation of $V_{Z}$ and $j$ with the tomographic method and the input theoretical values. This was achieved for a range of $f_{\rm sh}$ values between 0.2 to 20 km s-1 au-1 and with the three ($\theta$, $\theta_{V}$) configurations studied in the previous subsection. In order to study efficiently this bias, we defined the a-dimensional parameter $x=\frac{[V_{\rm los}(\pm R)]}{R}$ $\times$ ($\frac{\partial[V_{\rm los}(\pm R)]}{\partial R}$ )-1, where we defined $[V_{\rm los}(\pm R)]=1/2(V_{\rm los}(R)+V_{\rm los}(-R))$. This a-dimensional parameter can be derived from the observations. We show in Fig. 26 the predicted relative biases in $V_{\rm Z}$ and $j$ as a function of this parameter $x$. Biases increase with increasing $x$ values illustrating the effect of the velocity shear. We also show the distribution of observed x values computed at each (R,Z) position in the conical outflow. The x values are concentrated around $\simeq 0.5$, suggesting moderate biases in both $V_{\rm Z}$ and $j$ for the conical outflow where $\theta_{V}=\theta\leq 17^{\circ}$ . However, this modeling did not include the effect of beam smearing nor the impact of our polynomial fitting method to describe the shape of the PV diagram. In order to study these effects, we directly applied our tomographic method to the synthetic data cube presented in Sect. 10 and determined the relative differences between the computed and the input $V_{\rm Z}$ and $j$ values at each (R,Z) position along the conical flow. We show in Fig. 26 the derived relative biases in $V_{\rm Z}$ and $j$ as a function of the $x$ parameter. Their distributions are broader and flatter than predicted, especially in $j$, likely due to the combination of velocity shear and beam smearing effects. From this study, we estimate that in the conical part of the outflow (at $(V-V_{\rm sys})\geq 2$ km s-1), the tomographic method suffers from a relative bias $\leq 15$% in the estimation of $V_{\rm Z}$ and $\leq 20$% in the estimation of $j$. The X values extend up to $\approx 1.3$ in the pedestal region. With an opening angle $\theta$, $\theta_{v}>30^{\circ}$, a tomographic study of the low velocity component would suffer from a relative bias larger than 60% and 30% in the estimation of $V_{\rm Z}$ and $j$ respectively. ## Appendix E Wind-driven shell analytical solutions In this section, using Eqs. 1, 2,3 and 11, we derive an analytical solution for predicted channel maps in the case of the generalized WDS model introduced in Sect. 4. The WDS model is defined by three parameters: $C$, $\tau$ and $\eta$, see Eqs. 11. The projection on the plane of the sky ($\delta x$, $\delta z$) for the emissivity map at $(V-V_{\rm sys})=V_{\rm CM}$ could be recovered by solving Eq. 3 with $V_{\rm los}=V_{\rm CM}$ and using Eqs. 11 for $V_{\rm Z}$ and $V_{\rm R}$: $\displaystyle\zeta$ $\displaystyle=$ $\displaystyle\frac{4V_{\rm CM}\tau C\cos{i}}{\eta^{2}\sin^{2}{i}}$ (27) $\displaystyle R$ $\displaystyle=$ $\displaystyle-\eta\frac{\tan{i}}{2C}(\sin{\phi}\pm\sqrt{\sin^{2}{\phi}-\zeta})$ (28) $\displaystyle\delta x$ $\displaystyle=$ $\displaystyle R\cos{\phi}$ (29) $\displaystyle\delta z$ $\displaystyle=$ $\displaystyle CR^{2}\sin{i}-R\sin{\phi}\cos{i}.$ (30) Equation 28 has no solution in the case $\zeta>1$ ($V_{\rm CM}\cos{i}>\frac{\eta^{2}\sin^{2}{i}}{4\tau C}$). This corresponds to the case where no emission is predicted at $V_{\rm los}=V_{\rm CM}$. Similarly, in the situation $0\geq\zeta\geq 1$, only a fraction of $\phi$ will be projected such as $\sin^{2}{\phi}>\zeta$. Developing Eq. 30, we obtained the following: $\displaystyle\delta z$ $\displaystyle=$ $\displaystyle\eta\frac{\sin^{2}{\phi}\sin{i}}{2C}(\eta\tan^{2}{i}+1)-V_{CM}\tau\tan{i}$ (32) $\displaystyle\pm\eta\frac{\sin{\phi}\sin{i}}{2C}(\eta\tan^{2}{i}+1)\sqrt{\sin^{2}{\phi}-\zeta}.$ We reformulated $\delta z$ as $z_{0}+\Delta Z\sin{\beta}$ with: $\displaystyle z_{0}$ $\displaystyle=$ $\displaystyle\eta\frac{\sin{i}}{2C}(\eta\tan^{2}{i}+1)-V_{CM}\tau\tan{i}$ (33) $\displaystyle\Delta Z$ $\displaystyle=$ $\displaystyle\eta\frac{\sin{i}}{2C}(\eta\tan^{2}{i}+1)\sqrt{1-\zeta}$ (34) $\displaystyle\sin{\beta}$ $\displaystyle=$ $\displaystyle\frac{-1+\sin^{2}{\phi}\pm\sin{\phi}\sqrt{\sin^{2}{\phi}-\zeta}}{\sqrt{1-\zeta}}.$ (35) Similarly, manipulating Eq. 29, $\delta x$ could be reformulated as $\delta x=\Delta X\cos{\beta}$, with the following: $\Delta X=\mp\eta\frac{\tan{i}}{2C}\sqrt{1-\zeta}.$ (36) Hence ($\delta x$, $\delta z$) trace an ellipse of center (0, $z_{0}$) and aspect ratio: $\Bigl{|}\frac{\Delta Z}{\Delta X}\Bigr{|}=(\eta\tan^{2}{i}+1)|\cos{i}|.$ (37) Therefore, in the classical WDS models with radial velocity vectors ($\eta=1$), the aspect ratio of the ellipse on the channel maps only depends on the inclination. We represent on the channel maps in Figs. 5, 7, and 8 the two limiting ellipses computed at $V-\delta V/2$ and $V+\delta V/2$ to take into account the width of the channel map. ## Appendix F Global model Figure 27: Schematic representation of the method used to create synthetic data cube. In order to comprehend the impact of projection or convolution effects on the observations, we developed a locally axisymmetric code that allows us to simulate optically thin observations of simple outflow models. We took advantage of the axisymmetric hypothesis that permits to reduce the complexity of the model by defining at each height the radius $R(Z)$ and the cylindrical velocities $V_{\rm z}$, $V_{\rm r}$ and $V_{\phi}$ (See the schematic view of Fig. 2). The dependency between the height and the radius or the velocities vary with the model used. We then created at each height an emitting ring with azimuth parametrized with $\phi$. As the morphology is axisymmetric, neither the radius nor velocities depend on $\phi$. Under the assumption of optically thin emission, the emissivity at position ($Z$,$\phi$) is proportional to the elementary volume $dV=R(Z)d\phi dRdZ$. We added an additional variation of the emissivity with the height and radius as a power-law with parameters $\alpha$ and $\beta$ respectively. Proper modeling of the emissivity would require the temperature and chemistry to be solved, which is well beyond the scope of this model. The positions of the outflow emission on the data cube ($\delta x$,$\delta y$, $V_{\rm proj}$) were then defined by Eqs. 1,2,3. We then created a data cube with the same spectral and spatial resolution than our observations, and placed on each point ($\delta x$,$\delta y$, $V_{\rm proj}$) the emissivity I. Under the assumption of optically thin emission, we summed each emissivity corresponding to the same positions on the data cube. We set a step size of 1∘ for $\phi$ and a fraction of the spatial pixel for Z. We then convolved the data cube by a 2D Gaussian matching the spatial beam characteristics in order to fully simulate the ALMA observations. The code, written in Python 3, is publicly online777https://github.com/Alois- deValon/Axoproj. ## Appendix G Channel maps of disk-wind models Figure 28: 12CO observed (first column) and synthetic disk-wind channel maps at selected line-of-sight velocities. The second and third columns correspond to the two disk-wind models with axis precession presented in Sect. 5.2. The last column corresponds to a disk-wind model with axisymmetric density enhancements, as presented in Sect. 5.3. $\delta x$ and $\delta z$ units are arcseconds.
ON A FAMILY OF $2$-AUTOMATIC SEQUENCES DERIVED FROM ULTIMATELY PERIODIC SEQUENCES AND GENERATING ALGEBRAIC CONTINUED FRACTIONS IN $\mathbb{F}_{2}((1/t))$ (Suites $2$-automatiques à colone vertébrale ultimement périodique) by A. Lasjaunias Warning:This note is not intended to be officially published. The matter exposed here grew from numerous exchanges during the past two months with Yining Hu (at a very and too large distance !). My aim is to report in a first draft, and in a very private and personal way, on a curious mathematical structure. We consider a large family $\Large{\mathcal{F}}$ of infinite sequences over a finite alphabet $\mathcal{A}=\left\\{a_{1},a_{2},...,a_{k}\right\\}$. This family includes a celebrated example of a 2-automatic sequence on the set $\left\\{a,b\right\\}$, called Period-doubling sequence, which has been studied in a previous article [1]. Each sequence ${\bf{s}}$ in $\large{\mathcal{F}}$ is built in the following way, from another sequence ${\boldsymbol{\varepsilon}}=(\varepsilon_{i})_{i\geq 0}$ over $\left\\{a_{1},a_{2},...,a_{k}\right\\}$, this last one being ultimately periodic. Starting from the empty word $W_{0}$, we consider the sequence of words $(W_{n})_{n\geq 0}$ such that for $n\geq 0$ we have $W_{n+1}=W_{n},\varepsilon_{n},W_{n}$ (note the coma is for concatenation and it will be omitted when it is suitable). Hence, we have $W_{1}=\varepsilon_{0}$, $W_{2}=\varepsilon_{0},\varepsilon_{1},\varepsilon_{0}$ etc…Observe by construction that $W_{n+1}$ starts by $W_{n}$ and therefore we may consider $W_{\infty}$ the inductive limit of these words (i.e. the word begining by $W_{n}$ for all $n\geq 0$). Note that, for all $n\geq 1$, $W_{n}$ is a palindrome, centered in $\varepsilon_{n}$, of length $2^{n}-1$. This $W_{\infty}$ represents the sequence ${\bf{s}}$, which we may denote by $\bf{s}(\boldsymbol{\varepsilon})$. Hence, we have : $\textbf{s}(\boldsymbol{\varepsilon})=\varepsilon_{0}\varepsilon_{1}\varepsilon_{0}\varepsilon_{2}\varepsilon_{0}\varepsilon_{1}\varepsilon_{0}\varepsilon_{3}\varepsilon_{0}\varepsilon_{1}\varepsilon_{0}\varepsilon_{2}\varepsilon_{0}....$ In this family $\Large{\mathcal{F}}$, each sequence $\textbf{s}=(s_{i})_{i\geq 0}$ , over $\left\\{a_{1},a_{2},...,a_{k}\right\\}$, generates an infinite continued fraction $\alpha$ in $\mathbb{F}_{2}((1/t))$, denoted $CF(\bf{s})$, by replacing the letters $a_{i}$ by non-constant polynomials in $\mathbb{F}_{2}[t]$ (this choice is arbitrary, hence there is a $CF(\bf{s})$ for each choice but it is considered as unique in the sequel). For basic information on continued fractions, particularly in power series fields, the reader is refered to [2]. Hence we will write : $\alpha=CF({\bf{s}})=[s_{0},s_{1},....,s_{n},....]$ where the $s_{i}$ are the partial quotients in $\mathbb{F}_{2}[t]$. We recall that the sequence of convergents to $\alpha$ is denoted $(x_{n}/y_{n})_{n\geq 1}$. For $n\geq 1$, we have $x_{n}/y_{n}=[s_{0},...,s_{n-1}]=s_{0}+1/(s_{1}+1/....)$ , hence $x_{1}=s_{0}$, $y_{1}=1$ and $x_{2}=s_{0}s_{1}+1$, $y_{2}=s_{1}$ etc… To be more precise about periodic sequences, we introduce the following definition. Definition. _Let $l\geq 0$ and $d\geq 1$ be two integers. An ultimately periodic sequence $\boldsymbol{\varepsilon}$ is called of type $(l,d)$ if : 1) $l=0$ and $\boldsymbol{\varepsilon}$ is purely periodic, with period of length $d$, this being denoted by $\boldsymbol{\varepsilon}=(\varepsilon_{0},\varepsilon_{1},...,\varepsilon_{d-1})^{\infty}$. 2) $l>0$ and $\boldsymbol{\varepsilon}$ is ultimately periodic, with a prefix of length $l$ and a period of length $d$, this being denoted by $\boldsymbol{\varepsilon}=\varepsilon_{0},\varepsilon_{1},...,\varepsilon_{l-1},(\varepsilon_{l},\varepsilon_{l+1},...,\varepsilon_{l+d-1})^{\infty}$. Here all the $\varepsilon_{i}$’s are in $\mathcal{A}$ (assuming that $k\geq l+d$ to allow different values to the terms of the sequence $\boldsymbol{\varepsilon}$)._ Let us illustrate the construction of $\bf{s}(\boldsymbol{\varepsilon})$ in two basic cases : 1) $\boldsymbol{\varepsilon}=a,b,(c)^{\infty}$ then $\textbf{s}(\boldsymbol{\varepsilon})=a,b,a,c,a,b,a,c,....=(abac)^{\infty}$ Note that here $\boldsymbol{\varepsilon}$ is ultimately constant (of type $(3,1)$) and $\bf{s}(\boldsymbol{\varepsilon})$ is periodic. Consequently a basic property on continued fractions shows that $\alpha=CF(\bf{s}(\boldsymbol{\varepsilon}))$ is quadratic over $\mathbb{F}_{2}(t)$. 2) $\boldsymbol{\varepsilon}=(a,b)^{\infty}$ then $\textbf{s}(\boldsymbol{\varepsilon})=abaaabababaaaba....$ Here $\boldsymbol{\varepsilon}$ is of type $(0,2)$ and $\bf{s}(\boldsymbol{\varepsilon})$ is the celebrated sequence mentionned above and called Period-doubling. It has been proved that $\alpha=CF(\bf{s}(\boldsymbol{\varepsilon}))$ satisfies an algebraic equation of degree 4 with coefficients in $\mathbb{F}_{2}[t]$ (see [1]). As it happens in these two simple cases, we are going to prove in the following theorem that all sequences $\bf{s}(\boldsymbol{\varepsilon})$ in $\large{\mathcal{F}}$ generate a continued fraction $\alpha$ which is algebraic over $\mathbb{F}_{2}(t)$. During the proof, the algebraic equation satisfied by $\alpha$ will appear explicitely. Let $\boldsymbol{\varepsilon}$ be a sequence of type $(l,d)$ then we denote by $\mathbb{F}_{2}(\boldsymbol{\varepsilon})$ the subfield of $\mathbb{F}_{2}(t)$ generated by the vector $(\varepsilon_{0},...,\varepsilon_{l+d-1})$ whose $l+d$ coordinates belong to $\mathbb{F}_{2}[t]$. Theorem. _Let $l\geq 0$ and $d\geq 1$ be integers. Let $\boldsymbol{\varepsilon}$ be an ultimately periodic sequence of type $(l,d)$. Let $\bf{s}(\boldsymbol{\varepsilon})$ the sequence in $\mathbb{F}_{2}[t]$ and $\alpha=CF(\bf{s}(\boldsymbol{\varepsilon}))$ the continued fraction in $\mathbb{F}_{2}((1/t))$, both be defined as above. Then there is a polynomial $P$ in $\mathbb{F}_{2}(\boldsymbol{\varepsilon})[x]$ such that $\deg_{x}(P)=2^{d}$ and $P(\alpha)=0$. To be more precise, setting $\beta=1/\alpha$, there are $d+1$ elements in $\mathbb{F}_{2}(\boldsymbol{\varepsilon})$, $A$ and $B_{k}$ for $0\leq k\leq d-1$, such that_ $\beta^{2^{d}}=A+\sum_{0\leq k\leq d-1}B_{k}\beta^{2^{k}}.$ Note that the case $d=1$ is trivial as we saw in case 1) above. Indeed, in that case $\boldsymbol{\varepsilon}$ is ultimately constant. Hence $\textbf{s}(\boldsymbol{\varepsilon})=W,\varepsilon_{l},W,\varepsilon_{l},...=(W,\varepsilon_{l})^{\infty}$ where $W$ is a finite (or empty) word and therefore $\alpha$ is quadratic. In the sequel, we may assume that $d\geq 2$. The proof of the Theorem lies on the existence of a particular subsequence of convergents to $\alpha$. These particular convergents are linked to the structure of the word $\textbf{s}(\boldsymbol{\varepsilon})$. Indeed, it is natural to consider the truncation of the continued fraction $\alpha$ containing the partial quotients from $s_{0}$ up to $s_{2^{n}-2}$ for $n\geq 1$, thus corresponding to the finite word $W_{n}$, of length $2^{n}-1$, mentionned above. Hence, for $n\geq 1$, we set $u_{n}/v_{n}=[s_{0},...,s_{2^{n}-2}]$. We have $(u_{1},v_{1})=(s_{0},1)\quad\text{and}\quad(u_{2},v_{2})=(s_{0}s_{1}s_{2}+s_{0}+s_{2},s_{1}s_{2}+1).$ The first step of the proof was introduced in our previous work concerning the particular case of the Period-doubling sequence. In [1,p. 4 Lemma 3.2.], using basic properties on continuants, we could prove that the pair $(u_{n},v_{n})$ satisfies a simple recurrence relation. Indeed, for $n\geq 1$, we have $(R)\qquad u_{n+1}=\varepsilon_{n}u_{n}^{2}\quad\text{and}\quad v_{n+1}=\varepsilon_{n}u_{n}v_{n}+1,$ with $(u_{1},v_{1})=(\varepsilon_{0},1)$. From $(R)$ we get immediately $v_{n+1}/u_{n+1}=v_{n}/u_{n}+1/u_{n+1}$ and therefore we obtain $v_{n}/u_{n}=\sum_{1\leq i\leq n}1/u_{i}\quad\text{ for}\quad n\geq 1.$ Let us consider $\beta=1/\alpha$. Then $\beta=lim_{n}(v_{n}/u_{n})$ and consequently we have $\beta=\sum_{n\geq 1}1/u_{n}.\qquad eq(0)$ From $eq(0)$, we will show that $\beta$ is algebraic in the following way. By successive elevation to the power 2, for $0\leq i\leq d$, we can define inductively $d+1$ sequences, $(\varepsilon(n,i))_{n\geq 0}$, as follows: $\varepsilon(n,0)=1\quad\text{and}\quad\varepsilon(n,i+1)=\varepsilon(n,i)^{2}\varepsilon_{n+i}\quad\text{for}\quad 0\leq i\leq d-1.$ Note that, for $n\geq 0$, we have $\varepsilon(n,1)=\varepsilon_{n}$. Then we observe that, by elevating $eq(0)$ to the power 2, we get $\beta^{2}=\sum_{n\geq 1}1/u_{n}^{2}=\sum_{n\geq 1}\varepsilon_{n}/u_{n+1}=\sum_{n\geq 1}\varepsilon(n,1)/u_{n+1}.\qquad eq(1)$ Moreover, by successive elevation to the power 2, starting from $eq(0)$ and introducing the sequences $(\varepsilon(n,i))_{n\geq 0}$, we also get, for $0\leq i\leq d$, $\beta^{2^{i}}=\sum_{n\geq 1}\varepsilon(n,i)/u_{n+i}.\qquad eq(i)$ Remark. _For all $1\leq i\leq d$ the sequence $(\varepsilon(n,i))_{n\geq 0}$ is ultimately periodic of type $(l,d)$. Proof by induction. This is true for $i=1$. If $(\varepsilon(n,i))_{n\geq 0}$ is a periodic sequence of type $(l,d)$, then we have, for $n\geq l$, $\varepsilon(n+d,i)=\varepsilon(n,i)$ and consequently $\varepsilon(n+d,i+1)=\varepsilon(n+d,i)^{2}\varepsilon_{n+d+i}=\varepsilon(n,i)^{2}\varepsilon_{n+i}=\varepsilon(n,i+1)$._ Now we introduce a partition of the set of positive integers into $d+1$ subsets: first the finite set $F=\\{k\quad|\quad 1\leq k\leq l+d-1\\}$ and the $d$ subsets $E_{j}=\\{md+l+j\quad|\quad m\geq 1\\}\quad\text{for}\quad 0\leq j\leq d-1.$ $\mathbb{N}^{*}=\bigcup_{j=0}^{d-1}E_{j}\bigcup F.$ Linked to this partition, we introduce $d$ elements, $\beta_{j}$ for $0\leq j\leq d-1$, in $\mathbb{F}_{2}((1/t))$, defined by $\beta_{j}=\sum_{n\in E_{j}}1/u_{n}.\qquad(B)$ Combining $eq(0)$ and $(B)$, and defining $z_{0}\in\mathbb{F}_{2}(\boldsymbol{\varepsilon})$ by $\sum_{k\in F}1/u_{k}$, we can write, $\beta=z_{0}+\sum_{0\leq j\leq d-1}\sum_{n\in E_{j}}1/u_{n}=z_{0}+\sum_{0\leq j\leq d-1}\beta_{j}.\qquad Eq(0)$ Using the above remark, concerning the periodicity of the sequences $(\varepsilon(n,i))_{n\geq 0}$, for $0\leq j\leq d-1$ and for $1\leq i\leq d$, we can write , $\varepsilon(d+l+j-i,i)\beta_{j}=\sum_{m\geq 1}\varepsilon(md+l+j-i,i)/u_{md+l+j}$ $\varepsilon(d+l+j-i,i)\beta_{j}=\sum_{n+i\in E_{j}}\varepsilon(n,i)/u_{n+i}.$ Consequently, we observe that there exists $z_{i}\in\mathbb{F}_{2}(\boldsymbol{\varepsilon})$, a finite sum of the first terms in the series appearing in $eq(i)$, such that, for $1\leq i\leq d$, $eq(i)$ becomes the following equality $\beta^{2^{i}}=z_{i}+\sum_{0\leq j\leq d-1}\varepsilon(d+l+j-i,i)\beta_{j}.\qquad Eq(i)$ (Note that these quantities $z_{i}$ have a different form depending on the triplet $(l,d,i)$. See the three examples below.) Now, let us introduce the following square matrix of order $d$ : $M(d)=(m_{i,j})_{0\leq i,j\leq d-1}\quad\text{where}\quad m_{i,j}=\varepsilon(d+l+j-i,i).$ Introducing two column vectors $B$ and $C$, we observe that the $d$ equations $Eq(i)$, for $0\leq i\leq d-1$, can be summed up introducing the following linear system $(S):\quad M(d).B=C\quad$, where $B=\begin{bmatrix}\beta_{0}\\\ \beta_{1}\\\ \vdots\\\ \beta_{d-1}\end{bmatrix}\quad\text{and}\quad C=\begin{bmatrix}\beta+z_{0}\\\ \beta^{2}+z_{1}\\\ \vdots\\\ \beta^{2^{d-1}}+z_{d-1}\end{bmatrix}.$ We introduce the determinant, $\Delta(d)$, of the matrix $M(d)$ and also the determinant $\Delta(j,d)$ obtained from $\Delta(d)$ by replacing the column vector of rank $j$ by the column vector $C$. Hence, applying Cramer’s rule for solving the linear system $(S)$, we get $\beta_{j}=\Delta(j,d)/\Delta(d)\quad\text{for}\quad 0\leq j\leq d-1.$ Finally, reporting these values for $\beta_{j}$ in $Eq(d)$, we obtain $\beta^{2^{d}}=z_{d}+(\sum_{0\leq j\leq d-1}\varepsilon(l+j,d)\Delta(j,d))/\Delta(d).\quad Eq(*)$ We observe that $\Delta(d)$ belongs to $\mathbb{F}_{2}(\boldsymbol{\varepsilon})$. While, $\Delta(j,d)$ belongs to $\mathbb{F}_{2}(\boldsymbol{\varepsilon})[\beta]$. Indeed, developping the determinant $\Delta(j,d)$ along the column of rank $j$, we get $\Delta(j,d)=c_{j}+\sum_{0\leq k\leq d-1}b_{k,j}\beta^{2^{k}}$. Consequently $Eq(*)$ can be witten as expected ( with coefficients in $\mathbb{F}_{2}(\boldsymbol{\varepsilon})$) : $\beta^{2^{d}}=A+\sum_{0\leq k\leq d-1}B_{k}\beta^{2^{k}}.\qquad Eq(**)$ So the proof of the theorem is complete. We present, here below, three examples. In order to avoid unnecessary complications with the subscripts, we use $(a,b,c)$ for the letters $(\varepsilon_{0},\varepsilon_{1},\varepsilon_{2})$. Example 1: Type (0,2). Period-Doubling sequence . Let us consider the case 2, mentioned above, where $\boldsymbol{\varepsilon}=(a,b)^{\infty}\quad\text{and}\quad\beta=1/CF(\textbf{s}(\boldsymbol{\varepsilon})).$ We have $(\varepsilon_{0},\varepsilon_{1},\varepsilon_{2})=(a,b,a)$ and $(z_{0},z_{1},z_{2})=(1/a,0,1)$. $\mathrm{\Delta(2)=}\begin{vmatrix}1&1\\\ \varepsilon_{1}&\varepsilon_{2}\end{vmatrix}=a+b,$ $\mathrm{\Delta(0,2)=}\begin{vmatrix}\beta+z_{0}&1\\\ \beta^{2}+z_{1}&\varepsilon_{2}\end{vmatrix}\quad\textrm{ and }\quad\mathrm{\Delta(1,2)=}\begin{vmatrix}1&\beta+z_{0}\\\ \varepsilon_{1}&\beta^{2}+z_{1}\end{vmatrix}.$ Hence we get $\Delta(0,2)=\beta^{2}+a\beta+1\quad\text{and}\quad\Delta(1,2)=\beta^{2}+b\beta+b/a.$ Since $\varepsilon(0,2)=\varepsilon_{0}^{2}\varepsilon_{1}=a^{2}b$ and $\varepsilon(1,2)=\varepsilon_{1}^{2}\varepsilon_{2}=b^{2}a$, $Eq(*)$ becomes $(a+b)\beta^{4}=a+b+(\beta^{2}+a\beta+1)a^{2}b+(\beta^{2}+b\beta+b/a)b^{2}a.$ From this, we get (as expected, see [1, p 2, Th 1.1] ) : $\beta^{4}=1+b(a+b)+ab(a+b)\beta+ab\beta^{2}.\qquad Eq(**)$ Example 2: Type (1,2). Here we have : $\boldsymbol{\varepsilon}=a,(b,c)^{\infty}\quad\text{and}\quad\beta=1/CF(\textbf{s}(\boldsymbol{\varepsilon})).$ We have $(\varepsilon_{0},\varepsilon_{1},\varepsilon_{2},\varepsilon_{3})=(a,b,c,b)$ and $(z_{0},z_{1},z_{2})=(1/a+1/ba^{2},1/a^{2},0)$. $\mathrm{\Delta(2)=}\begin{vmatrix}1&1\\\ \varepsilon_{2}&\varepsilon_{3}\end{vmatrix}=b+c$ $\mathrm{\Delta(0,2)=}\begin{vmatrix}\beta+z_{0}&1\\\ \beta^{2}+z_{1}&\varepsilon_{3}\end{vmatrix}\quad\textrm{ and }\quad\mathrm{\Delta(1,2)=}\begin{vmatrix}1&\beta+z_{0}\\\ \varepsilon_{2}&\beta^{2}+z_{1}\end{vmatrix}.$ Hence we get $\Delta(0,2)=\beta^{2}+b\beta+b/a\quad\text{and}\quad\Delta(1,2)=\beta^{2}+c\beta+1/a^{2}+c/a+c/ba^{2}.$ We have $\varepsilon(1,2)=\varepsilon_{1}^{2}\varepsilon_{2}$ and $\varepsilon(2,2)=\varepsilon_{2}^{2}\varepsilon_{3}$. Consequently, $Eq(*)$ becomes $(b+c)\beta^{4}=(\beta^{2}+b\beta+b/a)b^{2}c+(\beta^{2}+c\beta+1/a^{2}+c/a+c/ba^{2})c^{2}b.$ From this, we get : $\beta^{4}=bc(b+c)/a+c^{2}/a^{2}+bc(b+c)\beta+bc\beta^{2}.\qquad Eq(**)$ (Note that changing $c$ into $a$, we have $\boldsymbol{\varepsilon}=a,(b,c)^{\infty}=a,(b,a)^{\infty}=(a,b)^{\infty}$ and we regain the previous example and the same algebraic equation for $\beta$ as above.) Example 3: Type (0,3). Here we have : $\boldsymbol{\varepsilon}=(a,b,c)^{\infty}\quad\text{and}\quad\beta=1/CF(\textbf{s}(\boldsymbol{\varepsilon})).$ We have $(\varepsilon_{0},\varepsilon_{1},\varepsilon_{2},\varepsilon_{3},\varepsilon_{4})=(a,b,c,a,b)$ and $(z_{0},z_{1},z_{2},z_{3})=(1/a+1/ba^{2},1/a^{2},0,1).$ $\mathrm{\Delta(3)=}\begin{vmatrix}1&1&1\\\ \varepsilon_{2}&\varepsilon_{3}&\varepsilon_{4}\\\ \varepsilon_{1}^{2}\varepsilon_{2}&\varepsilon_{2}^{2}\varepsilon_{3}&\varepsilon_{3}^{2}\varepsilon_{4}\end{vmatrix}=ba^{2}(a+c)+ac^{2}(b+c)+cb^{2}(a+b)$ $\mathrm{\Delta(0,3)=}\begin{vmatrix}\beta+z_{0}&1&1\\\ \beta^{2}+z_{1}&a&b\\\ \beta^{4}+z_{2}&c^{2}a&a^{2}b\end{vmatrix}=(a+b)\beta^{4}+(a^{2}b+ac^{2})\beta^{2}+ab(a^{2}+c^{2})\beta+\delta_{0}$ $\mathrm{\Delta(1,3)=}\begin{vmatrix}1&\beta+z_{0}&1\\\ c&\beta^{2}+z_{1}&b\\\ b^{2}c&\beta^{4}+z_{2}&a^{2}b\end{vmatrix}=(a+b)\beta^{4}+(a^{2}b+ac^{2})\beta^{2}+ab(a^{2}+c^{2})\beta+\delta_{1}$ $\mathrm{\Delta(2,3)=}\begin{vmatrix}1&1&\beta+z_{0}\\\ c&a&\beta^{2}+z_{1}\\\ b^{2}c&c^{2}a&\beta^{4}+z_{2}\end{vmatrix}=(a+c)\beta^{4}+(ac^{2}+b^{2}c)\beta^{2}+ac(c^{2}+b^{2})\beta+\delta_{2}$ together with $\delta_{0}=ab(a^{2}+c^{2})z_{0}+(a^{2}b+c^{2}a)z_{1}+(a+b)z_{2}$ $\delta_{1}=cb(a^{2}+b^{2})+(a^{2}b+b^{2}c)z_{1}+(a+b)z_{2}$ $\delta_{2}=ac(c^{2}+b^{2})+(c^{2}a+b^{2}c)z_{1}+(b+c)z_{2}.$ Here, $Eq(*)$ becomes $\beta^{8}=z_{3}+(\sum_{0\leq j\leq 2}\varepsilon(j,3)\Delta(j,3))/\Delta(3)$ and we also have $\varepsilon(0,3)=a^{4}b^{2}c,\qquad\varepsilon(1,3)=b^{4}c^{2}a\quad\text{and}\quad\varepsilon(2,3)=c^{4}a^{2}b.$ Finally combining these values and the four values for the determinants given above, from $Eq(*)$, we get the desired outcome : $\beta^{8}=A+\sum_{0\leq k\leq 2}B_{k}\beta^{2^{k}}.\qquad Eq(**)$ At last, remarkably enough, we can check that the four coefficients in this last equation do not only belong to $\mathbb{F}_{2}(\boldsymbol{\varepsilon})$ but are indeed elements in $F_{2}[t]$ and we have $A=a^{3}b^{2}c+a^{2}b^{2}c^{2}+ab^{3}c^{2}+b^{4}c^{2}+ab^{2}c^{3}+abc^{4}+a^{2}bc+ab^{2}c+abc^{2}+c^{4}+1$ $B_{0}=a^{4}b2c+a^{3}b^{2}c^{2}+a^{2}b^{3}c^{2}+ab^{4}c^{2}+a^{2}b^{2}c^{3}+a^{2}bc^{4}$ $B_{1}=a^{3}b^{2}c+a^{2}b^{2}c^{2}+ab^{3}c^{2}+a^{2}bc^{3}$ and $B_{2}=a^{2}bc+ab^{2}c+abc^{2}.$ An important and last point need to be discussed. Indeed, the reader will probably ask the following question : are the sequences, belonging to the family $\Large{\mathcal{F}}$, $2$-automatic as it is indicated in the title of this note ? There are different ways to characterize automatic sequences. A direct way is to consider the letters of the infinite world as elements in a finite field $\mathbb{F}_{q}$ of characteristic $p$. If a power series $\gamma$ in $\mathbb{F}_{q}((1/t))$ is algebraic over $\mathbb{F}_{q}(t)$, then the sequence of its coefficients is $p$-automatic (Christol’s theorem). Concerning the sequences $\bf{s}(\boldsymbol{\varepsilon})$ described above, in the general case the automaticity will result from a conjecture. First we assume that the $l+d$ elements defining the sequence are in a finite field $\mathbb{F}_{q}$ of characteristic $2$ with $q=2^{s}\geq l+d$, and consequently we may consider the power series $\gamma\in\mathbb{F}_{q}((1/t))$ associated to this sequence. Beginning by the trivial case $d=1$, we have observed that $\bf{s}(\boldsymbol{\varepsilon})$ is utimately periodic and therefore $p$-automatic for all $p$. Note that the power series $\gamma$, associated to it, is rational and consequently it satisfies a polynomial of degree $1$ over $\mathbb{F}_{q}(t)$. We make the following conjecture : Conjecture. _Let $l\geq 0$ and $d\geq 2$ be integers. Let $\boldsymbol{\varepsilon}$ be an ultimately periodic sequence of type $(l,d)$. Let $\bf{s}(\boldsymbol{\varepsilon})$ $=(s_{n})_{n\geq 0}$ be the sequence defined above. Then there exists a finite field $\mathbb{F}_{q}$ of characteristic $2$, containing $l+d$ elements identified with the terms of this sequence so that we may consider $\gamma=\sum_{n\geq 0}s_{n}t^{-n}$ in $\mathbb{F}_{q}((1/t))$ and there is a polynomial $P$ in $\mathbb{F}_{q}(t)[x]$ such that $\deg_{x}(P)=2^{d-1}$ and $P(\gamma)=0$._ In the simpler case $(l,d)=(0,2)$ and $\boldsymbol{\varepsilon}=(a,b)^{\infty}$, already considered several times above, the resulting sequence $\bf{s}(\boldsymbol{\varepsilon})$ (called Period-doubling) is well known to be $2$-automatic. More precisely, the above conjecture is true. Indeed, if we identify the pair $(a,b)$ with the pair $(0,1)$ in $\mathbb{F}_{2}$ then $\gamma=\sum_{n\geq 0}s_{n}t^{-n}\in\mathbb{F}_{2}((1/t))$ satisfies $\gamma^{2}+t\gamma=t^{2}/(t^{2}+1)$. ## References * [1] Y. Hu and A. Lasjaunias. Period-doubling continued fractions are algebraic in characteristic 2. arXiv:2204.01068, 3 Apr 2022. (to appear in Annales de l’Institut Fourier) * [2] A. Lasjaunias. Continued fractions. arXiv:1711.11276, 30 Nov 2017. September 2022.
# Adaptive Safety Evaluation for Connected and Automated Vehicles with Sparse Control Variates Jingxuan Yang * —, Haowei Sun * —, Honglin He * —, Yi Zhang * —, , Shuo Feng * —, and Henry X. Liu * — This work is supported by National Key Research and Development Program under Grant 2021YFB2501200 and National Natural Science Foundation of China under Grant 62133002. (Corresponding author: Shuo Feng.)Jingxuan Yang and Honglin He are with the Department of Automation, Tsinghua University, Beijing 100084, China (email: {yangjx20, hehl21}@mails.tsinghua.edu.cn).Haowei Sun and Henry X. Liu are with the Department of Civil and Environmental Engineering, University of Michigan, Ann Arbor, MI 48109, USA (e-mail: {haoweis, henryliu}@umich.edu).Yi Zhang is with the Department of Automation, Beijing National Research Center for Information Science and Technology (BNRist), Tsinghua University, Beijing 100084, China (e-mail: [email protected]).Shuo Feng is with the Department of Automation, Tsinghua University, Beijing 100084, China and the University of Michigan Transportation Research Institute, Ann Arbor, MI 48109, USA (e-mail: [email protected]). ###### Abstract Safety performance evaluation is critical for developing and deploying connected and automated vehicles (CAVs). One prevailing way is to design testing scenarios using prior knowledge of CAVs, test CAVs in these scenarios, and then evaluate their safety performances. However, significant differences between CAVs and prior knowledge could severely reduce the evaluation efficiency. Towards addressing this issue, most existing studies focus on the adaptive design of testing scenarios during the CAV testing process, but so far they cannot be applied to high-dimensional scenarios. In this paper, we focus on the adaptive safety performance evaluation by leveraging the testing results, after the CAV testing process. It can significantly improve the evaluation efficiency and be applied to high-dimensional scenarios. Specifically, instead of directly evaluating the unknown quantity (e.g., crash rates) of CAV safety performances, we evaluate the differences between the unknown quantity and known quantity (i.e., control variates). By leveraging the testing results, the control variates could be well designed and optimized such that the differences are close to zero, so the evaluation variance could be dramatically reduced for different CAVs. To handle the high-dimensional scenarios, we propose the sparse control variates method, where the control variates are designed only for the sparse and critical variables of scenarios. According to the number of critical variables in each scenario, the control variates are stratified into strata and optimized within each stratum using multiple linear regression techniques. We justify the proposed method’s effectiveness by rigorous theoretical analysis and empirical study of high- dimensional overtaking scenarios. ###### Index Terms: Adaptive safety evaluation, connected and automated vehicles, sparse control variates, high-dimensional scenarios ## I Introduction Testing and evaluation of safety performance are major challenges for the development and deployment of connected and automated vehicles (CAVs). One proposed way is to test CAVs in the naturalistic driving environments (NDE) through a combination of software simulation, test tracks, and public roads, observe their performances, and make statistical comparisons with human drivers. Due to the rarity of safety-critical events in NDE, however, hundreds of millions of miles and sometimes hundreds of billions of miles would be required to demonstrate CAVs’ safety performance at the human-level [1], which is intolerably inefficient. To improve the efficiency and accelerate the evaluation process, the past few years have witnessed increasingly rapid advances in the field of testing scenario library generation (TSLG) [2, 3, 4, 5, 6, 7, 8, 9, 10], where safety-critical testing scenarios are usually purposely generated utilizing prior knowledge of CAVs such as surrogate models (SMs) of CAVs. However, due to the high complexity and black-box properties of CAVs, there exist significant performance dissimilarities between SMs and CAVs under test, which could severely compromise the effectiveness of the generated testing scenarios and decrease the evaluation efficiency. Towards addressing this problem, several adaptive testing and evaluation methods have been proposed [11, 12, 13, 14]. The basic idea of existing methods is to adaptively generate the testing scenarios during the testing process of CAVs. With more testing results of CAVs, more posteriori knowledge of CAVs can be obtained, and therefore the testing scenarios can be more customized and optimized for the CAVs under test. However, most existing methods can only be applied to relatively simple scenarios, and how to handle high-dimensional scenarios remains an open question. For example, Mullins et al. [11] proposed an adaptive sampling method that uses Gaussian process regression (GPR) and $k$-nearest neighbors to discover performance boundaries of the system under test and then updates the SM with new testing results obtained near the performance boundaries. Koren et al. [12] put forward an adaptive stress testing method that uses deep reinforcement learning to find the most-likely failure scenarios. Feng et al. [13] proposed an adaptive testing scenario library generation method using Bayesian optimization techniques with classification-based GPR and acquisition functions to select subsequent testing scenarios and then update the SMs with new testing results. Sun et al. [14] presented an adaptive design of experiments method to detect safety-critical scenarios, which uses supervised machine learning models as SMs to approximate the testing results and devises acquisition functions for updating the SMs. The challenge for adaptively generating high-dimensional scenarios comes from the compounding effects of the “Curse of Rarity” (CoR) and the “Curse of Dimensionality” (CoD) [15]. The CoR refers to the concept that, due to rarity of safety-critical events, the amount of data needed to obtain sufficient information grow dramatically, while the CoD refers to the dimensionality of variables to represent realistic scenarios, which makes the computation cost increase exponentially with the growth of scenario dimensions. Most existing scenario-based testing approaches can only handle short scenario segments with limited background road users, where the decision variables are low- dimensional, which cannot represent the full complexity and variability of the real-world driving environment [16, 17, 18, 19, 20]. Towards addressing this challenge, the naturalistic and adversarial driving environment (NADE) method has been developed in our previous work [21], which can generate high- dimensional highway driving scenarios. However, the NADE did not consider the performance gap between CAVs and SMs, which could also slow down the testing process. To the best of the authors’ knowledge, there is no existing work that can handle the adaptive testing and evaluation problem in high-dimensional scenarios, and the goal of this paper is to fill this gap. Figure 1: Illustration of the adaptive testing and evaluation framework. The focus of this study is the adaptive evaluation method for high-dimensional scenarios, where the sparse control variates method is proposed. In general, the adaptive testing and evaluation methods can be categorized into two types including adaptive testing scenario generation and adaptive testing result evaluation, which are complementary to each other as shown in Fig. 1. Most existing studies focus on the former one, while in this study, we focus on the latter one and propose an adaptive evaluation framework that can handle high-dimensional scenarios. We note that how to realize the former one in high-dimensional scenarios also remains unsolved, which we leave for future study. In the proposed framework, we apply the NADE method to generate high- dimensional testing scenarios, where combinations of multiple SMs are utilized to improve the robustness of the generated scenarios for different CAVs under test. Then we propose a sparse control variate (SCV) method to adjust the testing results and evaluate CAVs’ performance adaptively. Essentially, the SCV method could reduce the estimation variance for the CAV under test and thus reduce the required number of tests, accelerating the evaluation process adaptively. In the following paragraphs, we further explain the major idea of the proposed SCV method. The control variates (CV) method [22] is a popular variance reduction technique applied in research areas such as deep learning [23] and reinforcement learning [24]. Suppose we want to estimate $\mu\triangleq\mathbb{E}_{p}[f(X)]$ by Monte Carlo sampling [25], where $p$ is the probabilistic distribution of the random variable $X$ and $f$ is the performance index of interest. Instead of directly estimating the unknown quantity $\mu$, the control variates method estimates the differences between the unknown quantity and known quantity as $\mu^{\prime}\triangleq\mathbb{E}_{p}[f(X)-h(X)+\theta]$, where $h(X)$ is the control variate and $\theta\triangleq\mathbb{E}_{p}[h(X)]$ is a known value. Then, if $h(X)$ correlates with the performance index $f(X)$ (hence can provide some information about $f(X)$), the estimation variance of $\mu^{\prime}$ will always be less than directly estimating $\mu$ [26]. For testing and evaluation of CAVs, the control variate $h(X)$ can be designed by utilizing the prior knowledge of CAVs (e.g., different SMs). $h(X)$ usually contains adjustable control parameters, which can be optimized by leveraging the testing results. In such way, the information about the CAV under test could be incorporated, which makes the adaptive evaluation possible. However, due to the CoD, the computation cost of optimal control parameters will increase exponentially with the growth of scenario dimensions, so directly applying the ordinary CV method in high-dimensional scenarios is problematic. Figure 2: Illustration of the sparse control variates method. The SCV are constructed by only considering critical variables (represented as red dots in testing scenarios). The testing results are stratified into strata according to the number of critical variables and then adjusted by SCV within each stratum. Finally, the performance index are obtained by summing up these evaluation results with proportion weights. To address this problem, we propose the sparse control variates (SCV) method, as shown in Fig. 2. The key idea is to construct the SCV by only considering the sparse but critical variables (e.g., behaviors of principal other vehicles at critical moments), following the similar idea from [21] that handles the CoD. However, the number of critical variables varies in different testing scenarios, which cannot be handled by ordinary CV method. To address this issue, in the SCV method, we stratify the testing scenarios into strata according to the number of critical variables. Then the control parameters can be optimized by multiple linear regression (MLR) [27] within each stratum, and the final evaluation results are obtained by summing up those evaluation results in each stratum with the proportion weights. Since the number of critical variables is much less than the dimension of testing scenarios, the computation cost of optimal control parameters for SCV could be greatly reduced, overcoming the CoD challenge. To verify the proposed method, we theoretically analyze its accuracy, efficiency, and optimality. The theorems show that our method is unbiased, and its estimation variance is nearly proportional to the best one that all the SMs used for generating testing scenarios could have. Moreover, under certain assumptions about the SMs, our method can provide a zero-variance estimator. To validate our method, the high-dimensional overtaking scenarios with large- scale naturalistic driving data are investigated. Simulation results show that our method can further accelerate the evaluation process by about one order of magnitude for different types of CAV models, comparing with the estimation efficiency in NADE. Compared with our previously published conference paper about SCV [28], the new contributions of this paper are listed as follows. First, we significantly extend our methodology into high-dimensional scenarios and establish the theoretical analysis for the accuracy, efficiency, and optimality of the proposed method with rigorous proofs. Second, a more realistic overtaking case study with large-scale naturalistic driving data is investigated to systematically validate the performances of our method. The remainder of this paper is organized as follows. Section II provides preliminary knowledge for the generation of NDE and NADE. Section III formulates the adaptive testing and evaluation problem and elaborates the challenges of applying ordinary CV for adaptive safety evaluation. To address these challenges, in Section IV, the SCV method is proposed. Then Section V and VI verify and validate the accuracy and efficiency of the proposed method from the theoretical and experimental perspectives, respectively. Finally, Section VII concludes the paper and discusses future research. ## II Preliminaries ### II-A Naturalistic Driving Environment Testing As discussed above, the prevailing approach for CAV evaluation is to test CAVs in the naturalistic driving environments (NDE) [29], observe their performances, and make statistical comparisons with human drivers. In NDE, one of the vehicles is the automated vehicle (AV) under test and the others are background vehicles (BVs), which can be formulated as Markov games [30]. A Markov game for $N$ agents (i.e., BVs) is defined by a set of states $\mathcal{S}$ describing the positions and velocities of all vehicles and a collection of action (i.e., acceleration) sets $\mathcal{A}_{1},\dots,\mathcal{A}_{N}$, one for each agent in NDE. The total action space is denoted as $\mathcal{A}=\mathcal{A}_{1}\times\cdots\times\mathcal{A}_{N}$. Then a scenario is defined as the time series of the states of all vehicles and the actions of all agents, i.e., $x=(s_{0},a_{0},\dots,s_{T},a_{T})\in\mathcal{X},$ (1) where $x$ represents the scenario, $\mathcal{X}$ is the set of all feasible scenarios, $s_{t}\in\mathcal{S}$ is the state of all vehicles at time $t$, $a_{t}\in\mathcal{A}$ is the action of all agents at time $t$, and $T$ is the time horizon. Let $\Omega=\mathcal{X}$ be the sample space incorporating all feasible scenarios. Consider the probability space $(\Omega,\mathcal{F},\mathbb{P})$, where $\mathcal{F}\triangleq 2^{\Omega}$ is the power set of $\Omega$ and $\mathbb{P}$ is a probability measure on $\mathcal{F}$. Let $X:x\mapsto x$, $\forall x\in\mathcal{X}$ be the random variable of scenarios. For testing and evaluation of CAVs, the crash event is usually of most interest, which can be defined as $A=\\{x\in\mathcal{X}:s_{T}\in\mathcal{S}_{c}\\}$, where $\mathcal{S}_{c}$ is the set of all crash states. Then the crash rate is selected as the performance index, which can be computed as $\mu=\mathbb{P}(A)=\mathbb{E}_{p}[\mathbb{I}_{A}(X)]=\sum_{x\in\mathcal{X}}\mathbb{P}(A|x)p(x),$ (2) where $\mathbb{I}_{A}$ is the indicator function of $A$, and $p$ is the naturalistic joint distribution of $x$. The essence of testing AV in NDE is to estimate the performance index $\mu$ by Monte Carlo simulation, i.e., $\hat{\mu}_{n}=\frac{1}{n}\sum_{i=1}^{n}\mathbb{P}(A|X_{i}),\quad X_{i}\sim p.$ (3) ### II-B Naturalistic and Adversarial Driving Environment Generation The NDE faces the CoR, making its estimation catastrophically inefficient. To improve the estimation efficiency, the importance sampling (IS) technique [19, 17, 18] has been used to sample testing scenarios from the importance function $q$, which puts more weights on crash-prone scenarios. In IS, the performance index can be estimated as $\hat{\mu}_{q}=\frac{1}{n}\sum_{i=1}^{n}\frac{\mathbb{P}(A|X_{i})p(X_{i})}{q(X_{i})},\quad X_{i}\sim q.$ (4) However, the IS method faces the CoD if the testing scenarios are high- dimensional [31]. To address both the CoR and the CoD, the naturalistic and adversarial driving environment (NADE) [21] has been proposed to only sample critical variables of testing scenarios from importance functions, while other variables remain their naturalistic distributions. Denote $x=(x_{c},x_{-c})$, where $x_{c}=\\{x_{c_{1}},\dots,x_{c_{l}}\\}$ is the set of critical variables, $c_{1},\dots,c_{l}$ are called the critical moments, $l=0,1,\dots,L$ is the number of control steps (i.e., the number of critical variables in $x_{c}$), and $x_{-c}$ is the set of other variables. Let $X_{c}:x\mapsto x_{c}$ be the random variable of critical variables and $X_{-c}:x\mapsto x_{-c}$ be the random variable of other variables, then we have $X=(X_{c},X_{-c})$. The importance function can then be formulated as $q(x)=q(x_{c})p(x_{-c})$, and therefore the performance index can be estimated in NADE as $\tilde{\mu}_{q}=\frac{1}{n}\sum_{i=1}^{n}\frac{\mathbb{P}(A|X_{i})p(X_{c,i})}{q(X_{c,i})},\quad X_{i}\sim q,$ (5) where $X_{c,i}$ is the random variable of critical variables of $X_{i}$. ## III Problem Formulation ### III-A Adaptive Testing and Evaluation Due to the black-box property and various types of CAVs, how to adaptively test and evaluate CAVs remains a major challenge. One way of adaptive testing and evaluation is adaptively generating testing scenarios. For example, we can minimize the estimation variance by optimizing the importance function, i.e., $\min_{q\in\mathcal{Q}}~{}\mathrm{Var}_{q}\left(\frac{\mathbb{P}(A|X)p(X)}{q(X)}\right),$ (6) where $\mathcal{Q}$ is the function space of $q$. Better importance functions can be found by leveraging the posteriori knowledge of CAVs obtained from testing results. Then the testing scenarios can be adaptively generated by sampling from updated importance functions. In this paper, we focus on another way of adaptive testing and evaluation, i.e., adaptively evaluating weighted testing results. Specifically, the control variates (CV) method is adopted. This problem can be formulated as $\min_{h\in\mathcal{H}}~{}\mathrm{Var}_{q}\left(\frac{\mathbb{P}(A|X)p(X)}{q(X)}-h(X)\right),$ (7) where $h:\mathcal{X}\to\mathbb{R}$ is the control variate and $\mathcal{H}$ is the function space of $h$. The goal is to further reduce the estimation variance by optimizing $h$ in $\mathcal{H}$, leveraging the testing results. ### III-B Control Variates Control variates are widely used as a basic variance reduction technique in Monte Carlo simulation. They can be usefully combined with the mixture importance sampling, where individual importance functions can serve as CV. In mixture IS, the scenarios $X_{i},i=1,\dots,n$ are sampled from the mixture importance function $q_{\alpha}=\sum_{j=1}^{J}\alpha_{j}q_{j}$, where $\alpha_{j}\geqslant 0$, $\sum_{j=1}^{J}\alpha_{j}=1$ and the $q_{j}$ are importance functions. One commonly used way is to construct CV by using the linear combination of individual importance functions as $h_{\beta}(X)=\sum_{j=1}^{J}\beta_{j}\left[\frac{q_{j}(X)}{q_{\alpha}(X)}-1\right],$ (8) where $\beta=(\beta_{1},\dots,\beta_{J})^{\top}$ is the control vector, $\beta_{j}\in\mathbb{R}$ are control parameters, and $q_{j}/q_{\alpha}-1$ are individual control variate. Combining the control variate $h_{\beta}$ with mixture IS gives the estimation $\hat{\mu}_{q_{\alpha},\beta}=\frac{1}{n}\sum_{i=1}^{n}\left[\frac{\mathbb{P}(A|X_{i})p(X_{i})}{q_{\alpha}(X_{i})}-h_{\beta}(X_{i})\right]$ (9) for $X_{i}\sim q_{\alpha}$. The unbiasedness of $\hat{\mu}_{q_{\alpha},\beta}$ is guaranteed since $\mathbb{E}_{q_{\alpha}}[\hat{\mu}_{q_{\alpha},\beta}]=\mathbb{E}_{q_{\alpha}}\left[\frac{\mathbb{P}(A|X)p(X)}{q_{\alpha}(X)}-h_{\beta}(X)\right]=\mu,$ (10) where the second equality is obtained from the unbiasedness of IS and $\mathbb{E}_{q_{\alpha}}[h_{\beta}(X)]=0$. The variance of $\hat{\mu}_{q_{\alpha},\beta}$ can be compared to that of IS with individual importance functions $q_{j}$. We have the following lemma. ###### Lemma 1 Let $\beta^{*}$ be any minimizer over $\beta$ of $\mathrm{Var}_{q_{\alpha}}(\hat{\mu}_{q_{\alpha},\beta})$, then $\mathrm{Var}_{q_{\alpha}}(\hat{\mu}_{q_{\alpha},\beta^{*}})\leqslant\min_{1\leqslant j\leqslant J}\frac{\sigma_{q_{j}}^{2}}{n\alpha_{j}},$ (11) where $\sigma_{q_{j}}^{2}$ is the asymptotic variance of $\hat{\mu}_{q_{j}}$, i.e., $\sigma_{q_{j}}^{2}=\mathrm{Var}_{q_{j}}\left(\frac{\mathbb{P}(A|X)p(X)}{q_{j}(X)}\right),~{}j=1,\dots,J.$ (12) ###### Proof: This is the Theorem 2 in [32]. ∎ It can be seen from Lemma 1 that the variance of $\hat{\mu}_{q_{\alpha},\beta}$ will be zero if any one of the $q_{j}$ is optimal. This is a significant feature because we can nearly omit the influence of all other worse-performed importance functions. In applications, using only one SM to test CAVs is usually under huge risk, because the performance gap between the SM and various types of CAVs may be too large to give a good estimation efficiency. Therefore, to ensure the robustness, we can combine multiple SMs to test the CAVs. However, there often exist some poor- performed SMs that will compromise the overall estimation efficiency. Using mixture IS with CV provides an effective way to ensure both good estimation efficiency and robustness to various types of CAVs. In practice, the optimal control vector $\beta^{*}$ is usually unknown, and its estimation $\hat{\beta}$ can be obtained by multiple linear regression (MLR). Denote the weighted testing results as $Y_{i}=\mathbb{P}(A|X_{i})p(X_{i})/q_{\alpha}(X_{i})$, $i=1,\dots,n$, and the individual control variate as $Z_{ij}=q_{j}(X_{i})/q_{\alpha}(X_{i})-1$, $i=1,\dots,n$, $j=1,\dots,J-1$. Then the $\hat{\beta}$ is given as the vector of coefficients obtained from MLR of $Y_{i}$ on $Z_{ij}$. In essence, this process is to search for the best control variate defined in Eq. (8) in the function space spanned by individual control variate $q_{j}/q_{\alpha}-1$. However, challenges of estimating optimal control parameters arise when the testing scenarios are high-dimensional. ### III-C CoD of Control Variates Considering the Markov chain structure of scenarios with $T+1$ time steps, the mixture importance function is given by $q_{\alpha}(x)=q_{\alpha}(s_{0})\prod_{t=0}^{T}q_{\alpha}(a_{t}|s_{t}),~{}\forall x\in\mathcal{X},$ (13) where $q_{\alpha}(s)=\sum_{j=1}^{J}{\alpha_{j}q_{j}(s)},~{}\forall s\in\mathcal{S}$, and $q_{\alpha}(a|s)=\sum_{j=1}^{J}{\alpha_{j}q_{j}(a|s)},~{}\forall a\in\mathcal{A},~{}s\in\mathcal{S}$. It can be found that $q_{\alpha}(x)$ is the product of $T+2$ individual importance functions and thus is also the summation of $J^{T+2}$ combinations of different importance functions at each time step. Specifically, these individual importance functions are $q_{j_{0},\dots,j_{T+1}}(x)=q_{j_{0}}(s_{0})q_{j_{1}}(a_{0}|s_{0})\cdots q_{j_{T+1}}(a_{T}|s_{T}),$ (14) where $j_{0},\dots,j_{T+1}=1,\dots,J$. Then the individual control variate are given by $q_{j_{0},\dots,j_{T+1}}/q_{\alpha}-1$. To find the estimation of optimal control parameters, we have to conduct MLR of $n$ weighted testing results on $J^{T+2}$ individual control variate. The number $J^{T+2}$ will increase exponentially with the dimension of scenarios, leading to the CoD of MLR. For example, if we have $J=10$ individual importance functions and the testing scenarios last for 10 seconds at a frequency of 10 Hz, then the number of individual control variate will be 10102. This means that a matrix with dimension 10102 should be inverted in MLR, which is not tractable. Moreover, the situation will get even worse if the duration of scenarios grows to several hours, which are common in daily driving yet far from being tractable. The following section aims to address this challenge. ## IV Adaptive Safety Evaluation with Sparse Control Variates In this section, we will address the CoD discussed above and show how to estimate the optimal control parameters. ### IV-A Sparse Control Variates We propose the sparse control variates (SCV) method to address the CoD of applying CV in high-dimensional scenarios. Specifically, the SCV are constructed by only considering the importance functions of only sparse and critical variables in high-dimensional testing scenarios. The number of critical variables is usually much less than the dimension of scenarios in NADE. Therefore, the number of SCV is also much less than the number of ordinary CV, which could greatly address the CoD. However, as the number of SCV varies in different testing scenarios, we can not directly apply SCV to the weighted testing results. Towards addressing this issue, we propose to stratify the testing scenarios into strata according to the number of critical variables and then apply SCV within each stratum. Let $\mathcal{X}_{l}=\\{x\in\mathcal{X}:|x_{c}|=l\\}$, $l=0,1,\dots,L$ be the stratum of scenarios that are controlled $l$ steps, satisfying $\bigcup_{l=0}^{L}\mathcal{X}_{l}=\mathcal{X}$. Using mixture importance function $q_{\alpha}$, the estimation of the performance index in NADE is $\tilde{\mu}_{q_{\alpha}}=\frac{1}{n}\sum_{i=1}^{n}\frac{\mathbb{P}(A|X_{i})p(X_{c,i})}{q_{\alpha}(X_{c,i})},\quad X_{i}\sim q_{\alpha}.$ (15) The performance index of scenarios in stratum $\mathcal{X}_{l}$ can be written as $\mu_{l}\triangleq\mathbb{E}_{p}[\mathbb{I}_{A}(X)\mathbb{I}_{\mathcal{X}_{l}}(X)]$, $l=0,1,\dots,L$, then we have $\mu=\sum_{l=0}^{L}\mathbb{E}_{p}[\mathbb{I}_{A}(X)\mathbb{I}_{\mathcal{X}_{l}}(X)]=\sum_{l=0}^{L}\mu_{l}.$ (16) Similar to Eq. (15), the estimation of $\mu_{l}$ is given by $\tilde{\mu}_{l,q_{\alpha}}=\frac{1}{n}\sum_{i=1}^{n}\frac{\mathbb{P}(A|X_{i})\mathbb{I}_{\mathcal{X}_{l}}(X_{i})p(X_{c,i})}{q_{\alpha}(X_{c,i})},$ (17) and then we have $\displaystyle\tilde{\mu}_{q_{\alpha}}$ $\displaystyle=\sum_{l=0}^{L}\frac{1}{n}\sum_{i=1}^{n}\frac{\mathbb{P}(A|X_{i})\mathbb{I}_{\mathcal{X}_{l}}(X_{i})p(X_{c,i})}{q_{\alpha}(X_{c,i})}=\sum_{l=0}^{L}\tilde{\mu}_{l,q_{\alpha}}.$ (18) Let $q_{j_{1},\dots,j_{l}}(x)=p(x_{-c})q_{j_{1}}(x_{c_{1}})\cdots q_{j_{l}}(x_{c_{l}})$ be the importance functions that sample $x_{-c}$ from $p$ and sample $x_{c_{1}},\dots,x_{c_{l}}$ from $q_{j_{1}},\dots,q_{j_{l}}$ respectively, where $j_{1},\dots,j_{l}=1,\dots,J$, $l=1,\dots,L$. Then the individual importance functions of critical variables are given by $q_{j_{1},\dots,j_{l}}(x_{c})$. Denote the linear combination of these individual importance functions as $\tilde{h}_{l}(x)\triangleq\sum_{j_{1},\dots,j_{l}}\beta_{l,j_{1},\dots,j_{l}}q_{j_{1},\dots,j_{l}}(x),~{}l=1,\dots,L,$ (19) where $\beta_{l,j_{1},\dots,j_{l}}\in\mathbb{R}$ are associated control parameters. Then the SCV are given by $h_{l}(x_{c})=\frac{\tilde{h}_{l}(x_{c})\mathbb{I}_{\mathcal{X}_{l}}(x_{c})}{q_{\alpha}(x_{c})}-\theta_{l},~{}l=1,\dots,L,$ (20) where $\theta_{l}\triangleq\mathbb{E}_{q_{\alpha}}\big{[}\tilde{h}_{l}(X)\mathbb{I}_{\mathcal{X}_{l}}(X)/q_{\alpha}(X)\big{]}$. Therefore, the estimation $\tilde{\mu}_{l,q_{\alpha}}$ in Eq. (17) can be evaluated with SCV as $\displaystyle\tilde{\mu}_{l,q_{\alpha},\beta_{l}}$ $\displaystyle=\frac{1}{n}\sum_{i=1}^{n}\left[\frac{\mathbb{P}(A|X_{i})\mathbb{I}_{\mathcal{X}_{l}}(X_{i})p(X_{c,i})}{q_{\alpha}(X_{c,i})}-h_{l}(X_{c,i})\right]$ (21) $\displaystyle=\frac{1}{n}\sum_{i=1}^{n}\frac{\mathbb{P}(A|X_{i})p(X_{c,i})-\tilde{h}_{l}(X_{c,i})}{q_{\alpha}(X_{c,i})}\mathbb{I}_{\mathcal{X}_{l}}(X_{i})+\theta_{l}$ for $l=1,\dots,L$, where $\beta_{l}=\mathrm{vec}(\beta_{l,j_{1},\dots,j_{l}})$ is the vector of control parameters, and $\mathrm{vec}(\cdot)$ is the vectorization operator that flattens a tensor into a long vector. Note that there is no critical variable for $l=0$, and thus we set $\beta_{0}\triangleq 0$. In summary, the performance index estimated by the proposed SCV method is given by $\tilde{\mu}_{q_{\alpha},\beta}=\sum_{l=0}^{L}\tilde{\mu}_{l,q_{\alpha},\beta_{l}},$ (22) where $\beta=\\{\beta_{l}\\}_{l=0}^{L}$ is the set of all control vectors. ### IV-B Optimal Control Parameters To estimate the optimal control parameters that minimize the estimation variance, multiple linear regression (MLR) technique is applied in each stratum. Let $\mathbb{X}_{l}\triangleq\\{X_{i}|X_{i}\in\mathcal{X}_{l},i=1,\dots,n\\}$ be the set of sampled scenarios with $l$ controlled steps, $n_{l}\triangleq\sum_{i=1}^{n}\mathbb{I}_{\mathcal{X}_{l}}(X_{i})$ be the number of tests with $l$ controlled steps and $d_{l}\triangleq J^{l}$ be the number of SCV, $l=1,\dots,L$. Denote the vector of testing results as $Y_{l}\triangleq\left[\frac{\mathbb{P}(A|X_{i})p(X_{i})}{q_{\alpha}(X_{i})}~{}\mathrm{for}~{}X_{i}\in\mathbb{X}_{l}\right]\in\mathbb{R}^{n_{l}},$ (23) the individual SCV as $h^{\prime}_{j_{1},\dots,j_{l}}(x_{c})=\frac{q_{j_{1},\dots,j_{l}}(x_{c})}{q_{\alpha}(x_{c})}-\sum_{x_{c}\in\mathcal{X}_{l}}q_{j_{1},\dots,j_{l}}(x_{c}),$ (24) for $l=1,\dots,L$. Then the matrix of individual SCV can be formulated as $H_{l}\triangleq\left[\mathrm{vec}\left(h^{\prime}_{j_{1},\dots,j_{l}}(X_{c,i})\right)~{}\mathrm{for}~{}X_{i}\in\mathbb{X}_{l}\right]\in\mathbb{R}^{n_{l}\times d_{l}},$ (25) for $l=1,\dots,L$. Then the regression formula is given by $Y_{l}\approx\eta_{l}+H_{l}\beta_{l}$. The MLR of $Y_{l}$ on $H_{l}$ is to find the optimal solution of the following optimization problem, i.e., $\min_{\eta_{l},\beta_{l}}~{}f(\eta_{l},\beta_{l})=\|Y_{l}-\eta_{l}-H_{l}\beta_{l}\|_{2}^{2}.$ (26) Letting the partial derivatives of $f$ with respect to $\eta_{l}$ and $\beta_{l}$ both equal zero, we have $\hat{\eta}_{l}=1^{\top}Y_{l}/n_{l}$ and $\hat{\beta}_{l}=(H_{l}^{\top}H_{l})^{-1}H_{l}^{\top}Y_{l}$, assuming that the control matrix $M_{l}\triangleq H_{l}^{\top}H_{l}\in\mathbb{R}^{d_{l}\times d_{l}}$ is invertible. Then the estimated performance index is $\hat{\mu}_{l}=n_{l}\hat{\eta}_{l}/n$. In practice the control matrix may often not be invertible, then we use singular value decomposition (SVD) [33] to compute the regression coefficients $\hat{\beta}_{l}$, and the rank of the control matrix is $\mathrm{rank}(M_{l})=\mathrm{rank}(H_{l})\leqslant\min\\{n_{l},d_{l}\\}.$ (27) If $n_{l}<d_{l}$, then the control matrix $M_{l}$ will be singular and has utmost $n_{l}$ nonzero singular values. As the number of tests $n_{l}$ in $\mathbb{X}_{l}$ will not grow exponentially with the number of control steps $l$, the rank of the control matrix will also not, albeit the dimension $d_{l}=J^{l}$ of the control matrix increases exponentially with $l$. In conclusion, solving the optimal control parameters for SCV is tractable and will not face the CoD challenge. We will further demonstrate this in Subsection VI-E. ## V Theoretical Analysis This section theoretically justifies the accuracy, efficiency and optimality of the proposed SCV method. ### V-A Accuracy Analysis We first prove that the estimation is unbiased. ###### Theorem 1 Let $\tilde{\mu}_{q_{\alpha},\beta}$ be given by Eq. (22) where $q_{\alpha}>0$ whenever $\mathbb{P}(A|x)p(x)>0$, then $\mathbb{E}_{q_{\alpha}}[\tilde{\mu}_{q_{\alpha},\beta}]=\mu$. ###### Proof: To establish unbiasedness, write $\displaystyle\mathbb{E}_{q_{\alpha}}[\tilde{\mu}_{q_{\alpha},\beta}]$ $\displaystyle=\mathbb{E}_{q_{\alpha}}\left[\sum_{l=0}^{L}\tilde{\mu}_{l,q_{\alpha},\beta_{l}}\right]$ (28) $\displaystyle=\sum_{l=0}^{L}\mathbb{E}_{q_{\alpha}}\left[\tilde{\mu}_{l,q_{\alpha}}-\frac{\tilde{h}_{l}(X)}{q_{\alpha}(X)}\mathbb{I}_{\mathcal{X}_{l}}(X)+\theta_{l}\right]$ $\displaystyle=\sum_{l=0}^{L}(\mu_{l}-\theta_{l}+\theta_{l})=\mu.$ ∎ ###### Remark 1 This theorem indicates that the estimation is unbiased if the control parameters $\beta$ are independent of the sample data. It’s worth noting that in practice the control parameters are usually estimated by the sample data, which would bring a bias. However, that bias is ordinarily negligible (please see Section 8.9 in [26] for more discussions). ### V-B Efficiency Analysis Next, we evaluate the efficiency of the SCV method. The variance of the estimation $\tilde{\mu}_{q_{\alpha},\beta}$ is $\mathrm{Var}_{q_{\alpha}}(\tilde{\mu}_{q_{\alpha},\beta})=\sigma_{q_{\alpha},\beta}^{2}/n$, where $\sigma_{q_{\alpha},\beta}^{2}$ is the asymptotic variance of $\tilde{\mu}_{q_{\alpha},\beta}$, i.e., $\sigma_{q_{\alpha},\beta}^{2}=\mathrm{Var}_{q_{\alpha}}\left(\sum_{l=0}^{L}\frac{\mathbb{P}(A|X)p(X)-\tilde{h}_{l}(X)}{q_{\alpha}(X)}\mathbb{I}_{\mathcal{X}_{l}}(X)\right)$ (29) for $X\sim q_{\alpha}$. Denote $Z_{l}\triangleq\frac{\mathbb{P}(A|X)p(X)-\tilde{h}_{l}(X)}{q_{\alpha}(X)}\mathbb{I}_{\mathcal{X}_{l}}(X),~{}l=0,\dots,L,$ (30) then the asymptotic variance $\sigma_{q_{\alpha},\beta}^{2}$ can be expressed as $\sigma_{q_{\alpha},\beta}^{2}=\mathrm{Var}_{q_{\alpha}}\left(\sum_{l=0}^{L}Z_{l}\right)=\mathbb{E}_{q_{\alpha}}\left[\left(\sum_{l=0}^{L}\Big{[}Z_{l}-\mathbb{E}_{q_{\alpha}}[Z_{l}]\Big{]}\right)^{2}\right].$ (31) Let $L^{\prime}=L+1$, then by convexity of quadratic function and Jensen’s inequality, we have $\displaystyle\sigma_{q_{\alpha},\beta}^{2}$ $\displaystyle\leqslant\mathbb{E}_{q_{\alpha}}\left[L^{\prime}\sum_{l=0}^{L}\Big{(}Z_{l}-\mathbb{E}_{q_{\alpha}}[Z_{l}]\Big{)}^{2}\right]$ (32) $\displaystyle=L^{\prime}\sum_{l=0}^{L}\mathrm{Var}_{q_{\alpha}}(Z_{l}).$ Denote $\sigma_{l,q_{\alpha},\beta_{l}}^{2}\triangleq\mathrm{Var}_{q_{\alpha}}(Z_{l})$ and the asymptotic variance of $\tilde{\mu}_{l,q}$ over $\mathcal{X}_{l}$ as $\sigma_{l,q}^{2}$, i.e., $\sigma_{l,q}^{2}\triangleq\sum_{x\in\mathcal{X}_{l}}\left(\frac{\mathbb{P}(A|x)p(x)}{q(x)}-\mu_{l}\right)^{2}q(x),~{}l=1,\dots,L,$ (33) then we have the following theorem. ###### Theorem 2 If $\beta^{*}$ is any minimizer of $\sigma_{q_{\alpha},\beta}^{2}$, then $\displaystyle\sigma_{q_{\alpha},\beta^{*}}^{2}$ $\displaystyle\leqslant L^{\prime}\sigma_{0,p,\beta_{0}}^{2}$ (34) $\displaystyle\quad+L^{\prime}\sum_{l=1}^{L}\min_{j_{1},\dots,j_{l}}\left\\{\frac{\sigma_{l,q_{j_{1},\dots,j_{l}}}^{2}}{\prod_{\ell=1}^{l}\alpha_{j_{\ell}}}+3\left(\frac{\mu_{l}}{\prod_{\ell=1}^{l}\alpha_{j_{\ell}}}\right)^{2}\right\\}.$ ###### Proof: Take $\sigma_{1,q_{\alpha},\beta_{1}}^{2}$ as an example. Following the proof in [32], we consider the particular vector $\beta_{1}$ having $\beta_{1,1}=0$ and $\beta_{1,j}=-\mu_{1}\alpha_{j}/\alpha_{1}$ for $j>1$. Let $r_{1}(x)\triangleq[\mathbb{P}(A|x)p(x)-\mu_{1}q_{1}(x)]\mathbb{I}_{\mathcal{X}_{1}}(x)$, then we have $\sum_{x\in\mathcal{X}}r_{1}(x)=\mu_{1}(1-\xi_{1})$, where $\xi_{1}\triangleq\sum_{x\in\mathcal{X}_{1}}q_{1}(x)$, $\xi_{1}\in[0,1]$. Substituting these values, we find that for this $\beta_{1}$, $\displaystyle Z_{1}$ $\displaystyle=\frac{\mathbb{P}(A|X)p(X)-\tilde{h}_{1}(X)}{q_{\alpha}(X)}\mathbb{I}_{\mathcal{X}_{1}}(X)$ (35) $\displaystyle=\frac{\mathbb{P}(A|X)p(X)-\mu_{1}q_{1}+\mu_{1}q_{1}-\tilde{h}_{1}(X)}{q_{\alpha}(X)}\mathbb{I}_{\mathcal{X}_{1}}(X)$ $\displaystyle=\frac{r_{1}(X)}{q_{\alpha}(X)}+\frac{\mu_{1}}{\alpha_{1}}\mathbb{I}_{\mathcal{X}_{1}}(X),$ and $\mathbb{E}_{q_{\alpha}}[Z_{1}]=\mu_{1}\alpha_{1,1}/\alpha_{1}$, where $\alpha_{1,1}\triangleq\alpha_{1}+\sum_{j=2}^{J}\alpha_{j}\linebreak\sum_{x\in\mathcal{X}_{1}}q_{j}(x)$, $\alpha_{1,1}\in[0,1]$. Therefore, we have $\displaystyle\sigma_{1,q_{\alpha},\beta_{1}}^{2}$ $\displaystyle=\mathbb{E}_{q_{\alpha}}\left[\Big{(}Z_{1}-\mathbb{E}_{q_{\alpha}}[Z_{1}]\Big{)}^{2}\right]$ (36) $\displaystyle=\sum_{x\in\mathcal{X}}\left[\frac{r_{1}(x)}{q_{\alpha}(x)}+\frac{\mu_{1}}{\alpha_{1}}\Big{(}\mathbb{I}_{\mathcal{X}_{1}}(x)-\alpha_{1,1}\Big{)}\right]^{2}q_{\alpha}(x)$ $\displaystyle\triangleq V_{1,1}+V_{1,2}+V_{1,3},$ where $\displaystyle V_{1,1}$ $\displaystyle\triangleq\sum_{x\in\mathcal{X}}\frac{r_{1}^{2}(x)}{q_{\alpha}(x)}=\sum_{x\in\mathcal{X}}\frac{[\mathbb{P}(A|x)p(x)-\mu_{1}q_{1}(x)]^{2}}{q_{\alpha}(x)}\mathbb{I}_{\mathcal{X}_{1}}(x)$ (37) $\displaystyle\leqslant\sum_{x\in\mathcal{X}_{1}}\frac{[\mathbb{P}(A|x)p(x)-\mu_{1}q_{1}(x)]^{2}}{\alpha_{1}q_{1}(x)}=\frac{\sigma_{1,q_{1}}^{2}}{\alpha_{1}},$ $\displaystyle V_{1,2}$ $\displaystyle\triangleq\sum_{x\in\mathcal{X}}\frac{2\mu_{1}r_{1}(x)(\mathbb{I}_{\mathcal{X}_{1}}(x)-\alpha_{1,1})}{\alpha_{1}}$ (38) $\displaystyle=\frac{2\mu_{1}^{2}(1-\xi_{1})(1-\alpha_{1,1})}{\alpha_{1}}\leqslant 2\left(\frac{\mu_{1}}{\alpha_{1}}\right)^{2},$ and $\displaystyle V_{1,3}$ $\displaystyle\triangleq\sum_{x\in\mathcal{X}}\left[\frac{\mu_{1}(\mathbb{I}_{\mathcal{X}_{1}}(x)-\alpha_{1,1})}{\alpha_{1}}\right]^{2}q_{\alpha}(x)$ (39) $\displaystyle\leqslant\sum_{x\in\mathcal{X}}\left(\frac{\mu_{1}}{\alpha_{1}}\right)^{2}q_{\alpha}(x)=\left(\frac{\mu_{1}}{\alpha_{1}}\right)^{2}.$ Therefore, we conclude that $\sigma_{1,q_{\alpha},\beta_{1}^{*}}^{2}\leqslant\sigma_{1,q_{\alpha},\beta_{1}}^{2}\leqslant\frac{\sigma_{1,q_{1}}^{2}}{\alpha_{1}}+3\left(\frac{\mu_{1}}{\alpha_{1}}\right)^{2}.$ (40) By making similar arguments for $j=2,\dots,J$, we have $\sigma_{1,q_{\alpha},\beta_{1}^{*}}^{2}\leqslant\min_{j}\left\\{\frac{\sigma_{1,q_{j}}^{2}}{\alpha_{j}}+3\left(\frac{\mu_{1}}{\alpha_{j}}\right)^{2}\right\\}.$ (41) It’s straightforward to extend the proof for $l=2,\dots,L$, then Eq. (34) is established. ∎ ###### Remark 2 For $l=1$, we expect to get approximately $n_{1}\alpha_{j}$ scenarios in $\mathcal{X}_{1}$ from the importance function $q_{j}$. The quantity $\sigma_{1,q_{j}}^{2}/\alpha_{j}$ in Eq. (41) is the variance we would obtain from $n_{1}\alpha_{j}$ such scenarios alone. It is hard to imagine that we could do better in general, because when $\sigma_{1,q_{j}}^{2}=\infty$ for all but one of the mixture components it is guaranteed that those bad components do not make the estimation worse than what we would have had from the one good importance function. Moreover, if there exists an optimal importance function in $q_{j}$, then the minimum value of $\sigma_{1,q_{j}}^{2}/\alpha_{j}$ will be zero, which will greatly reduce the estimation variance. It should be noted that the upper bound for variance in Eq. (41) contains a residual term $3(\mu_{1}/\alpha_{j})^{2}$, which is the cost for stratifying the scenarios. ### V-C Optimality Analysis Under the following assumptions, the estimation variance of the SCV method can be zero. ###### Assumption 1 The scenarios in $\mathcal{X}_{0}$ will not be sampled by $q_{\alpha}$, i.e., $q_{\alpha}(x)=0$, $\forall x\in\mathcal{X}_{0}$. ###### Assumption 2 The control policy satisfies $|x_{c}|=1$, i.e., the number of critical variable of all sampled scenarios is 1. ###### Assumption 3 There exists an optimal control policy such that $\mathbb{P}(A|x_{c})=\mathbb{P}(A|x)$, which means that the critical variable $x_{c}$ can totally dominate the crash probability. ###### Assumption 4 There exists an optimal importance function among $q_{j}$. Without loss of generality, let $q_{1}$ be the optimal importance function, i.e., $q_{1}(x_{c})\triangleq\mathbb{P}(A|x_{c})p(x_{c})/\mu$. ###### Theorem 3 Under Assumptions 1, 2, 3 and 4, if $\beta^{*}$ is any minimizer of $\sigma_{q_{\alpha},\beta}^{2}$, then $\sigma_{q_{\alpha},\beta^{*}}^{2}=0$. ###### Proof: From Assumptions 1 and 2, we know that all sampled scenarios will only be controlled once, i.e., $\mathcal{X}=\mathcal{X}_{1}$ and $\mu=\mu_{1}$, then $Z_{1}=\frac{r_{1}(X)}{q_{\alpha}(X)}+\frac{\mu_{1}}{\alpha_{1}}\mathbb{I}_{\mathcal{X}_{1}}(X)=\frac{r_{1}(X)}{q_{\alpha}(X)}+\frac{\mu_{1}}{\alpha_{1}},$ (42) and $\mathbb{E}_{q_{\alpha}}[Z_{1}]=\mu_{1}\alpha_{1,1}/\alpha_{1}=\mu_{1}/\alpha_{1}$. Therefore, the asymptotic variance $\sigma_{1,q_{\alpha},\beta_{1}}^{2}$ is $\displaystyle\sigma_{1,q_{\alpha},\beta_{1}}^{2}$ $\displaystyle=\mathbb{E}_{q_{\alpha}}\left[\Big{(}Z_{1}-\mathbb{E}_{q_{\alpha}}[Z_{1}]\Big{)}^{2}\right]$ (43) $\displaystyle=\sum_{x\in\mathcal{X}}\frac{r_{1}^{2}(x)}{q_{\alpha}(x)}\leqslant\frac{\sigma_{1,q_{1}}^{2}}{\alpha_{1}}.$ By Assumptions 3 and 4, we have $\mathbb{P}(A|x_{c})=\mathbb{P}(A|x)$ and $q_{1}(x_{c})=\mathbb{P}(A|x_{c})p(x_{c})/\mu$, then $\displaystyle\sigma_{1,q_{1}}^{2}$ $\displaystyle=\sum_{x\in\mathcal{X}_{1}}\left(\frac{\mathbb{P}(A|x)p(x)}{q_{1}(x)}-\mu_{1}\right)^{2}q_{1}(x)$ (44) $\displaystyle=\sum_{x\in\mathcal{X}}\left(\frac{\mathbb{P}(A|x_{c})p(x_{c})}{q_{1}(x_{c})}-\mu\right)^{2}q_{1}(x)=0.$ Therefore, we conclude that $\sigma_{q_{\alpha},\beta^{*}}^{2}=\sigma_{1,q_{\alpha},\beta_{1}}^{2}=0$. ∎ ###### Remark 3 Assumption 1 suggests that the scenarios in $\mathcal{X}_{0}$ should not be sampled. Since there are no crash in these scenarios, they can not make any contribution to the estimation. Assumption 2 requires that the number of critical variable is 1, because stratifying scenarios into different strata leads to some residual terms (e.g., $3(\mu_{1}/\alpha_{j})^{2}$ in Eq. (41)) in estimation variance that can not be eliminated. Assumption 3 indicates that the critical variables should dominate the crash probability, since otherwise we may lose some critical information about the scenarios and obtain the suboptimal testing results. Assumption 4 requires that one of the importance functions should be optimal, together with Assumption 3 further reducing the asymptotic variances to zero. Although in practice these assumptions may not be fully satisfied, they could provide useful guidance for us to implement the SCV method. ###### Remark 4 The theorems in this section hold regardless of the specifics of SMs, which may be constructed by traditional traffic models or by neural networks. ## VI Overtaking Case Study ### VI-A Overtaking Scenarios Figure 3: Illustration of the overtaking scenarios. The overtaking scenarios are shown in Fig. 3, where the leading vehicle (LV) runs at the left lane, the background vehicle (BV) follows LV and the automated vehicle (AV) runs at the right lane. If BV cuts in to the right lane, then AV will follow BV and may rear-end BV, resulting in a crash. The state of the overtaking scenarios can be formulated as $s\triangleq\big{(}v_{\mathrm{BV}},R_{1},\dot{R}_{1},R_{2},\dot{R}_{2}\big{)},$ (45) where $R_{1}\triangleq x_{\mathrm{LV}}-x_{\mathrm{BV}}$, $\dot{R}_{1}\triangleq v_{\mathrm{LV}}-v_{\mathrm{BV}}$, $R_{2}\triangleq x_{\mathrm{BV}}-x_{\mathrm{AV}}$, and $\dot{R}_{2}\triangleq v_{\mathrm{BV}}-v_{\mathrm{AV}}$. The $x_{\mathrm{BV}}$, $x_{\mathrm{LV}}$, $x_{\mathrm{AV}}$ are the positions and $v_{\mathrm{BV}}$, $v_{\mathrm{LV}}$, $v_{\mathrm{AV}}$ are the velocities of BV, LV and AV, respectively. The action of the overtaking scenario is defined as the actions of LV and BV, i.e., $a\triangleq(a_{\mathrm{LV}},a_{\mathrm{BV}})$. We note that the overtaking scenarios are more stochastic and complicated than simple scenarios such as cut-in scenarios and car-following scenarios, since the BV in overtaking scenarios may have many chances to cut in, resulting in different cut-in scenarios and car-following scenarios between BV and AV. This is the reason why overtaking scenarios are always much more high-dimensional than cut-in scenarios. ### VI-B Generation of NDE The essence of NDE is to provide a driving environment where all BVs travel like humans. To generate NDE, the probability distributions of the behaviors of all BVs should be consistent with the naturalistic driving data (NDD) [29]. In this paper, the probability distributions of free-driving, car-following, and cut-in behaviors are extracted from the NDD of the Safety Pilot Model Deployment (SPMD) [34] program and Integrated Vehicle-Based Safety System (IVBSS) [35] at the University of Michigan, Ann Arbor. The initial state is set as $s_{0}=[v_{\mathrm{BV},0},R_{1,0},\dot{R}_{1,0},R_{2,0},\dot{R}_{2,0}],$ (46) where $v_{\mathrm{BV},0}$, $R_{1,0}$, $\dot{R}_{1,0}$ are sampled from the naturalistic distributions of car-following scenarios, $R_{2,0}\sim\mathcal{U}(20~{}\text{m},100~{}\text{m})$, $\dot{R}_{2,0}\sim\mathcal{U}(-5~{}\text{m/s},-10~{}\text{m/s})$, where $\mathcal{U}$ is the uniform distribution. After sampling the initial state, all vehicles select actions independently and simultaneously for each time step (0.1 s). The cut-in maneuver of BV is set completed within one time step. The car-following maneuver of AV is controlled by the intelligent driver model (IDM)[36]. The simulation continues until AV rear-ends BV or maximum simulation time (20 s) reached. Typically, the dimension of overtaking scenarios will exceed 1400 (201 time steps, each with 5 state variables and 2 action variables), leading to the high-dimensionality challenge. ### VI-C Generation of NADE The goal of NADE is to generate high-dimensional testing scenarios where the behaviors of BVs are adjusted only at critical moments, while keeping naturalistic distributions as in NDE at other time steps [21]. To construct the importance function, the maneuver criticality of BV is evaluated at each time step, which is defined as the multiplication of the exposure frequency and the maneuver challenge. The exposure frequency represents the probability of each action given current state in NDE. The maneuver challenge measures the probability of crash between AV and BV given current state and action. Since the AV models are usually black-boxes, the surrogate models (SMs) are adopted to approximate the maneuver challenge. In this paper, we use IDM and full velocity difference model (FVDM) [36] as SMs with different parameters: (1) IDM, denoted as SM-I; (2) FVDM with $a_{\min}=-1$ m/s2, denoted as SM-II; (3) FVDM with $a_{\min}=-6$ m/s2, denoted as SM-III. Then the importance functions can be obtained from the maneuver criticalities estimated by these SMs. Readers can find more technical details in [21]. ### VI-D Application of SCV Input: $p$, $q_{\alpha}$, $X_{c,i}$, and $\mathbb{P}(A|X_{i})$, $i=1,\dots,n$ Output: $\tilde{\mu}_{q_{\alpha},\hat{\beta}}$, $\mathrm{Var}_{q_{\alpha}}(\tilde{\mu}_{q_{\alpha},\hat{\beta}})$ 1 initialize $Y_{l}$ and $H_{l}$ as empty arrays, $l=0,\dots,L$; 2 initialize $n_{l}=0$, $l=0,\dots,L$; 3 for _$i\leftarrow 1$ to $n$_ do 4 $l\leftarrow$ number of control steps of $X_{c,i}$; 5 $n_{l}\leftarrow n_{l}+1$; 6 if _$l=0$_ then 7 append $Y_{l}$ with $\mathbb{P}(A|X_{i})$; 8 append $H_{l}$ with 0; 9 10 else 11 append $Y_{l}$ with $\mathbb{P}(A|X_{i})p(X_{c,i})/q_{\alpha}(X_{c,i})$; 12 append $H_{l}$ with $\mathrm{vec}(q_{j_{1},\dots,j_{l}}(X_{c,i})/q_{\alpha}(X_{c,i}))$, $j_{1},\dots,j_{l}=1,\dots,J-1$; 13 14 end if 15 16 end for 17for _$l\leftarrow 0$ to $L$_ do 18 $H_{l}\leftarrow H_{l}-\mathrm{average}(H_{l})$; 19 $\mathrm{MLR}\leftarrow$ multiple linear regression of $Y_{l}$ on $H_{l}$; 20 $\hat{\beta}_{l}\leftarrow$ estimated coefficients from $\mathrm{MLR}$; 21 $\hat{\eta}_{l}\leftarrow$ estimated intercept from $\mathrm{MLR}$; 22 $\tilde{\mu}_{l,q_{\alpha},\hat{\beta}_{l}}\leftarrow n_{l}\hat{\eta}_{l}/n$, $Z_{l}\leftarrow Y_{l}-H_{l}\hat{\beta}_{l}$; 23 24 end for 25$Z\leftarrow[Z_{0},\dots,Z_{L}]$; 26 $\tilde{\mu}_{q_{\alpha},\hat{\beta}}\leftarrow\sum_{l=0}^{L}\tilde{\mu}_{l,q_{\alpha},\hat{\beta}_{l}}$, $\mathrm{Var}_{q_{\alpha}}(\tilde{\mu}_{q_{\alpha},\hat{\beta}})\leftarrow\mathrm{var}(Z)$; 27 return $\tilde{\mu}_{q_{\alpha},\hat{\beta}}$, $\mathrm{Var}_{q_{\alpha}}(\tilde{\mu}_{q_{\alpha},\hat{\beta}})$; Algorithm 1 Adaptive safety evaluation with sparse control variates by multiple linear regression As shown in Algorithm 1, the SCV method can be applied to adjust the testing results and reduce estimation variance after testing AV in NADE. The key is to use importance functions of only sparse and critical variables to construct SCV, and then apply MLR of weighted testing results on SCV in each stratum. Finally, the estimated performance index is given by the summation of weighted intercepts obtained from MLR in all strata. ### VI-E Evaluation Results We validate the accuracy and efficiency of AV evaluation in NDE and NADE by the simulation of overtaking scenarios. The simulation is parallel conducted using 100 threads on a computer equipped with AMD® EPYC™ 7742 CPU and 512 GB RAM. Fig. 4 shows the crash rates of AV in NDE and NADE, respectively. The crash rate in NDE is presented as the black line in Fig. 4, with the bottom $x$-axis as its number of tests. The blue line in Fig. 4 represents the crash rate in NADE, and the top $x$-axis is the number of tests. The light shadow gives the 90% confidence interval. It can be seen that the crash rates in NDE and NADE converge to the same value, while NADE requires a much smaller number of tests. To measure the estimation precision of the crash rate, the relative half-width (RHW) [19] is adopted as the metric. The threshold of RHW is set to 0.3. To reach this threshold, NADE requires 6.76 $\times$ 106 number of tests, while NDE requires 1.21 $\times$ 108 number of tests, as shown in Fig. 5. It can be found that NADE can accelerate the evaluation by about 17.90 times compared with NDE. We note that the acceleration ratio is smaller than that in [21], because combinations of multiple various SMs are applied in this paper, which improves the robustness yet decreases the efficiency. The goal of the adaptive evaluation is to improve the efficiency while keeping the robustness. Figure 4: Crash rates of AV in NDE and NADE, where the dashed line is the crash rate estimated by NDE. Figure 5: RHW of AV evaluation in NDE and NADE, where the dashed line represents the RHW threshold (0.3). Figure 6: Crash rate of AV using NADE and SCV for (a) $n=5.92\times 10^{5}$, (b) $n=1.29\times 10^{6}$, (c) $n=4\times 10^{6}$, (d) $n=7\times 10^{6}$ and (e) $n=1\times 10^{7}$, where $n$ is the total number of tests and the dashed line is the crash rate estimated by NDE; (f) RHW of AV evaluation using NADE and SCV, where the dashed line in black represents the RHW threshold (0.3) and 5 dashed lines in orange correspond to (a)-(e). To investigate the performance of the SCV method, the accuracy and efficiency of AV evaluation in NADE with and without SCV are compared. It can be seen in Fig. 6 (a)-(e) that the crash rates of NADE and SCV converge to the same value for different number of tests. Fig. 6 (f) shows that the required numbers of tests of NADE and SCV for reaching the RHW threshold are 6.76 $\times$ 106 and 5.92 $\times$ 105, respectively, resulting in a further acceleration ratio of 11.42. The weighted testing results before and after being adjusted by SCV with different number of control steps are compared in Fig. 7 (a)-(i), and Fig. 7 (j) shows the total 107 adjusted testing results. It can be seen that the SCV method is able to adjust the testing results into a much narrower interval, especially for relatively large number of control steps (e.g., $l\geqslant 4$), resulting in a considerable reduction of the estimation variance. Figure 7: Adjusted testing results by NADE and SCV for (a)-(i) the number of control steps (#CS) from 1 to 9 and (j) total 107 testing results. Figure 8: Number of tests, number of SCV and maximum rank of control matrices for different number of control steps. The detailed regression processes of the SCV method are also investigated. Fig. 8 shows the number of tests, the number of SCV and the maximum rank of the control matrices for the number of control steps $l=1,\dots,9$, respectively. Note that for $l\geqslant 10$, we only use the first 9 control steps to construct the SCV. It can be seen that the maximum number of tests appears at $l=6$ and then the number of tests decreases to a relatively low level. As shown in Eq. (27), the maximum rank of the control matrices is the minimum value between the number of tests and the number of SCVs, and hence will not grow exponentially with the number of control steps, although the number of SCVs will do. Therefore, the SVD of control matrices is always tractable in each stratum and the optimal control parameters can be found to minimize the estimation variance. Since the scenario generation processes are stochastic, the testing and evaluation results are usually not the same in different experiments. Therefore, to find the average performances, we shuffle the testing results 200 times to bootstrap them and obtain the frequency distributions of the required number of tests (RNoT) in NDE and NADE. The average RNoT of NDE and NADE are 1.20 $\times$ 108 and 8.71 $\times$ 106, respectively. Therefore, the average acceleration ratio (AAR) of NADE with respect to NDE is 13.78. The testing results of SCV are also bootstrapped by 200 times. For cases with maximum RHW below 0.3, we use the RNoT when the maximum RHW is reached. The average RNoT of SCV is 1.29 $\times$ 106, resulting in an AAR of 6.76 times compared with NADE. ### VI-F Generalizability Analysis TABLE I: AARs of SCV where AV admits IDMs with different $\alpha$ values, and the rightmost column corresponds to the VT-IDM. $\alpha$ | 0.5 | 1.0 | 1.5 | 2.0 | 2.5 | 3.0 | VT-IDM ---|---|---|---|---|---|---|--- AAR | 11.52 | 9.02 | 7.87 | 6.76 | 7.73 | 10.90 | 7.30 $\alpha$ | 3.5 | 4.0 | 4.5 | 5.0 | 5.5 | 6.0 | AAR | 13.44 | 11.95 | 11.12 | 10.61 | 10.45 | 10.05 | In the above experiments, we have set the AV model the same as SM-I, i.e., they are both IDMs with same parameters. To investigate the generalizability of the SCV method for different AV models, the IDMs with a series of parameters $\alpha=0.5,1.0,\dots,6.0$ are chosen as AV models. The AARs of SCV compared with NADE are shown in Table I. The testing results of all AV models are shuffled 200 times to obtain the AARs. It can be seen that the minimum AAR appears at $\alpha=2.0$, where the AV model is the same as SM-I, while the maximum AAR appears at $\alpha=3.5$. The mean AAR for different AV models is 10.12. Therefore, the SCV method can further accelerate the evaluation process by about one order of magnitude for various types of AV models. Moreover, the AARs of SCV with AV models different from SM-I are always greater than that of AV model the same as SM-I. The reason is that although using AV models different from SM-I will do harm to both the estimation efficiency of NADE and SCV, the damage to NADE is more than to SCV. In addition, we also select the calibrated IDM in [37] (denoted as VT-IDM) as the AV model to further validate the generalization performance of the SCV method. The testing results shuffled 200 times give an AAR of 7.30 for SCV compared with NADE, which is shown at the rightmost column in Table I. Therefore, the SCV method can also increase the evaluation efficiency considerably for AV model with completely different calibrated parameters. This is not a surprising result because the only requirement for the SCV method to work is that the SMs and the AV model have some correlation, and more correlation contributes to more variance reduction. Although the VT-IDM and IDM have totally different parameters, they are still correlated to some extent. ## VII Conclusion In this paper, we propose an adaptive safety evaluation framework for CAVs in high-dimensional scenarios with a newly developed sparse control variates (SCV) method. To address the CoD, the SCV are constructed by only considering the sparse and critical variables of testing scenarios and stratified into strata accordingly. By optimizing the SCV leveraging the testing results within each stratum, the estimation variance is significantly reduced for different CAVs adaptively, accelerating the evaluation process. The accuracy, efficiency and optimality of the proposed method are verified and validated by both theoretical analysis and empirical studies. Comparing with the evaluation efficiency in NDE and NADE, our method is always more efficient particularly for CAVs that are different from SMs. It has been noted that adaptive testing scenario generation and adaptive testing result evaluation are two complementary approaches for adaptive testing and evaluation of CAVs. How to develop the former in high-dimensional scenarios deserves further investigation. ## References * [1] N. Kalra and S. M. Paddock, “Driving to safety: How many miles of driving would it take to demonstrate autonomous vehicle reliability?” _Transportation Research Part A: Policy and Practice_ , vol. 94, pp. 182–193, 2016. * [2] A. Li, S. Chen, L. Sun, N. Zheng, M. Tomizuka, and W. Zhan, “Scegene: Bio-inspired traffic scenario generation for autonomous driving testing,” _IEEE Transactions on Intelligent Transportation Systems_ , 2021. * [3] J. Wang, A. Pun, J. Tu, S. Manivasagam, A. Sadat, S. Casas, M. Ren, and R. Urtasun, “Advsim: Generating safety-critical scenarios for self-driving vehicles,” in _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 2021, pp. 9909–9918. * [4] T. Menzel, G. Bagschik, and M. Maurer, “Scenarios for development, test and validation of automated vehicles,” in _2018 IEEE Intelligent Vehicles Symposium (IV)_. IEEE, 2018, pp. 1821–1827. * [5] Y. Tian, K. Pei, S. Jana, and B. Ray, “Deeptest: Automated testing of deep-neural-network-driven autonomous cars,” in _Proceedings of the 40th International Conference on Software Engineering_ , 2018, pp. 303–314. * [6] D. Rempe, J. Philion, L. J. Guibas, S. Fidler, and O. Litany, “Generating useful accident-prone driving scenarios via a learned traffic prior,” in _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 2022, pp. 17 305–17 315. * [7] L. Li, W.-L. Huang, Y. Liu, N.-N. Zheng, and F.-Y. Wang, “Intelligence testing for autonomous vehicles: A new approach,” _IEEE Transactions on Intelligent Vehicles_ , vol. 1, no. 2, pp. 158–166, 2016. * [8] L. Li, Y.-L. Lin, N.-N. Zheng, F.-Y. Wang, Y. Liu, D. Cao, K. Wang, and W.-L. Huang, “Artificial intelligence test: A case study of intelligent vehicles,” _Artificial Intelligence Review_ , vol. 50, no. 3, pp. 441–465, 2018. * [9] L. Li, X. Wang, K. Wang, Y. Lin, J. Xin, L. Chen, L. Xu, B. Tian, Y. Ai, J. Wang _et al._ , “Parallel testing of vehicle intelligence via virtual-real interaction,” _Science Robotics_ , 2019. * [10] S. Riedmaier, T. Ponn, D. Ludwig, B. Schick, and F. Diermeyer, “Survey on scenario-based safety assessment of automated vehicles,” _IEEE access_ , vol. 8, pp. 87 456–87 477, 2020. * [11] G. E. Mullins, P. G. Stankiewicz, R. C. Hawthorne, and S. K. Gupta, “Adaptive generation of challenging scenarios for testing and evaluation of autonomous vehicles,” _Journal of Systems and Software_ , vol. 137, pp. 197–215, 2018\. * [12] M. Koren, S. Alsaif, R. Lee, and M. J. Kochenderfer, “Adaptive stress testing for autonomous vehicles,” in _2018 IEEE Intelligent Vehicles Symposium (IV)_. IEEE, 2018, pp. 1–7. * [13] S. Feng, Y. Feng, H. Sun, Y. Zhang, and H. X. Liu, “Testing scenario library generation for connected and automated vehicles: an adaptive framework,” _IEEE Transactions on Intelligent Transportation Systems_ , vol. 23, no. 2, pp. 1213–1222, 2022. * [14] J. Sun, H. Zhou, H. Xi, H. Zhang, and Y. Tian, “Adaptive design of experiments for safety evaluation of automated vehicles,” _IEEE Transactions on Intelligent Transportation Systems_ , 2021. * [15] H. X. Liu and S. Feng, ““curse of rarity” for autonomous vehicles,” _arXiv preprint arXiv:2207.02749_ , 2022. * [16] S. Feng, Y. Feng, X. Yan, S. Shen, S. Xu, and H. X. Liu, “Safety assessment of highly automated driving systems in test tracks: A new framework,” _Accident Analysis & Prevention_, vol. 144, p. 105664, 2020. * [17] S. Feng, Y. Feng, C. Yu, Y. Zhang, and H. X. Liu, “Testing scenario library generation for connected and automated vehicles, part i: Methodology,” _IEEE Transactions on Intelligent Transportation Systems_ , vol. 22, no. 3, pp. 1573–1582, 2021. * [18] S. Feng, Y. Feng, H. Sun, S. Bao, Y. Zhang, and H. X. Liu, “Testing scenario library generation for connected and automated vehicles, part ii: Case studies,” _IEEE Transactions on Intelligent Transportation Systems_ , vol. 22, no. 9, pp. 5635–5647, 2021. * [19] D. Zhao, H. Lam, H. Peng, S. Bao, D. J. LeBlanc, K. Nobukawa, and C. S. Pan, “Accelerated evaluation of automated vehicles safety in lane-change scenarios based on importance sampling techniques,” _IEEE Transactions on Intelligent Transportation Systems_ , vol. 18, no. 3, pp. 595–607, 2016. * [20] D. Zhao, X. Huang, H. Peng, H. Lam, and D. J. LeBlanc, “Accelerated evaluation of automated vehicles in car-following maneuvers,” _IEEE Transactions on Intelligent Transportation Systems_ , vol. 19, no. 3, pp. 733–744, 2017. * [21] S. Feng, X. Yan, H. Sun, Y. Feng, and H. X. Liu, “Intelligent driving intelligence test for autonomous vehicles with naturalistic and adversarial environment,” _Nature Communications_ , vol. 12, no. 1, pp. 1–14, 2021. * [22] R. Y. Rubinstein and R. Marcus, “Efficiency of multivariate control variates in monte carlo simulation,” _Operations Research_ , vol. 33, no. 3, pp. 661–677, 1985. * [23] W. Grathwohl, D. Choi, Y. Wu, G. Roeder, and D. Duvenaud, “Backpropagation through the void: Optimizing control variates for black-box gradient estimation,” in _International Conference on Learning Representations_ , 2018\. * [24] C.-A. Cheng, X. Yan, and B. Boots, “Trajectory-wise control variates for variance reduction in policy gradient methods,” in _Conference on Robot Learning_. PMLR, 2020, pp. 1379–1394. * [25] A. Shapiro, “Monte carlo sampling methods,” _Handbooks in operations research and management science_ , vol. 10, pp. 353–425, 2003. * [26] A. B. Owen, _Monte Carlo theory, methods and examples_. Stanford, 2013. * [27] D. J. Olive, “Multiple linear regression,” in _Linear Regression_. Springer, 2017, pp. 17–83. * [28] J. Yang, H. He, Y. Zhang, S. Feng, and H. X. Liu, “Adaptive testing for connected and automated vehicles with sparse control variates in overtaking scenarios,” in _IEEE International Intelligent Transportation Systems Conference (ITSC)_. IEEE, 2022. * [29] X. Yan, S. Feng, H. Sun, and H. X. Liu, “Distributionally consistent simulation of naturalistic driving environment for autonomous vehicle testing,” _arXiv preprint arXiv:2101.02828_ , 2021. * [30] R. Lowe, Y. I. Wu, A. Tamar, J. Harb, O. Pieter Abbeel, and I. Mordatch, “Multi-agent actor-critic for mixed cooperative-competitive environments,” _Advances in neural information processing systems_ , vol. 30, 2017. * [31] S.-K. Au and J. Beck, “Important sampling in high dimensions,” _Structural safety_ , vol. 25, no. 2, pp. 139–163, 2003. * [32] A. Owen and Y. Zhou, “Safe and effective importance sampling,” _Journal of the American Statistical Association_ , vol. 95, no. 449, pp. 135–143, 2000\. * [33] M. E. Wall, A. Rechtsteiner, and L. M. Rocha, “Singular value decomposition and principal component analysis,” in _A practical approach to microarray data analysis_. Springer, 2003, pp. 91–109. * [34] D. Bezzina and J. Sayer, “Safety pilot model deployment: Test conductor team report,” _Report No. DOT HS_ , vol. 812, no. 171, p. 18, 2014. * [35] J. Sayer, D. LeBlanc, S. Bogard, D. Funkhouser, S. Bao, M. L. Buonarosa, A. Blankespoor _et al._ , “Integrated vehicle-based safety systems field operational test: Final program report,” United States. Joint Program Office for Intelligent Transportation Systems, Tech. Rep., 2011. * [36] J. W. Ro, P. S. Roop, A. Malik, and P. Ranjitkar, “A formal approach for modeling and simulation of human car-following behavior,” _IEEE Transactions on Intelligent Transportation Systems_ , vol. 19, no. 2, pp. 639–648, 2017. * [37] J. Sangster, H. Rakha, and J. Du, “Application of naturalistic driving data to modeling of driver car-following behavior,” _Transportation research record_ , vol. 2390, no. 1, pp. 20–33, 2013.
1 # Unexpected Scaling in Path Copying Trees Ilya Kokorin ITMO UniversityRussia<EMAIL_ADDRESS>, Alexander Fedorov IST AustriaAustria<EMAIL_ADDRESS>, Trevor Brown University of Waterloo, CanadaCanada<EMAIL_ADDRESS>and Vitaly Aksenov ITMO University, RussiaRussia<EMAIL_ADDRESS> ###### Abstract. Although a wide variety of handcrafted concurrent data structures have been proposed, there is considerable interest in universal approaches (henceforth called Universal Constructions or UCs) for building concurrent data structures. These approaches (semi-)automatically convert a sequential data structure into a concurrent one. The simplest approach uses locks that protect a sequential data structure and allow only one process to access it at a time. The resulting data structures use locks, and hence are blocking. Most work on UCs instead focuses on obtaining non-blocking progress guarantees such as obstruction-freedom, lock-freedom, or wait-freedom. Many non-blocking UCs have appeared. Key examples include the seminal wait-free UC by Herlihy, a NUMA- aware UC by Yi et al., and an efficient UC for large objects by Fatourou et al. We borrow ideas from persistent data structures and multi-version concurrency control (MVCC), most notably path copying, and use them to implement concurrent versions of sequential persistent data structures. Despite our expectation that our data structures would not scale under write-heavy workloads, they scale in practice. We confirm this scaling analytically in our model with private per-process caches. ††conference: ACM SIGPLAN Conference on Programming Languages; January 01–03, 2018; New York, NY, USA††journalyear: 2018††doi: ††copyright: none††ccs: Software and its engineering General programming languages††ccs: Social and professional topics History of programming languages ## 1\. Introduction Although a wide variety of handcrafted concurrent data structures have been proposed, there is considerable interest in universal approaches (henceforth called _Universal Constructions_ or UCs) for building concurrent data structures. These approaches (semi-)automatically convert a sequential data structure into a concurrent one. The simplest approach uses locks (Herlihy et al., 2020; Lamport, 1987) that protect a sequential data structure and allow only one process to access it at a time. The resulting data structures use locks, and hence are blocking. Most work on UCs instead focuses on obtaining non-blocking progress guarantees such as _obstruction-freedom_ , _lock- freedom_ or _wait-freedom_. Many non-blocking UCs have appeared. Key examples include the seminal wait-free UC (Herlihy, 1991) by Herlihy, a NUMA-aware UC (Yi et al., 2021) by Yi et al., and an efficient UC for large objects (Fatourou et al., 2020) by Fatourou et al. In this work, we consider the simpler problem of implementing persistent (also called functional) data structures, which preserve the old version whenever the data structure is modified (Okasaki, 1999). Usually this entails copying a part of the data structure, for example, the path from the root to a modified node in a tree (Kaplan, 2018), so that none of the existing nodes need to be changed directly. We borrow ideas from persistent data structures and multi version concurrency control (MVCC) (Sun et al., 2019), most notably path copying, and use them to implement concurrent versions of sequential persistent data structures. Data structures implemented this way can be highly efficient for searches, but we expect them to not scale in write-heavy workloads. Surprisingly, we found that a concurrent treap implemented in this way obtained up to 2.4x speedup compared to a sequential treap (Seidel and Aragon, 1996) with 4 processes in a write-heavy workload. We present this effect experimentally, and analyze it in a model with private per-processor caches: informally, as the number of processes grows large, speedup in our treap of size $N$ tends to $\Omega(\log N)$. ## 2\. Straightforward Synchronization for Persistent Data Structures In the following discussion, we focus on rooted data structures, but one could imagine generalizing these ideas by adding a level of indirection in data structures with more than one entry point (e.g., one could add a dummy root node containing all entry points). We store a pointer to the current version of the persistent data structure (e.g., to the root of the current version of a persistent tree) in a Read/CAS register called Root_Ptr. Read-only operations (queries) read the current version and then execute sequentially on the obtained version. Note that no other process can modify this version, so the sequential operation is trivially atomic. Modifying operations are implemented in the following way: 1) read the current version; 2) obtain the new version by applying the sequential modification using path copying (i.e., by copying the root, and copying each visited node); 3) try to atomically replace the current version with the new one using CAS; if the CAS succeeds, return: the modifying operation has been successfully applied; otherwise, the data structure has been modified by some concurrent process: retry the execution from step (1). This approach clearly produces a lock-free linearizable data structure. We expect read-only operations to scale extremely well. Indeed, two processes may concurrently read the current version of the persistent data structure and execute read-only persistent operations in parallel. However, modification operations seemingly afford no opportunity for scaling. When multiple modifications contend, only one can finish successfully, and the others must retry. For example, consider concurrent modification operations on a set: 1) process P calls insert(2) and fetches the current pointer RP; 2) process Q calls remove(5) and fetches the current pointer RP; 3) P constructs a new version $\texttt{RP}_{\texttt{P}}$ with key 2; 4) Q constructs a new version $\texttt{RP}_{\texttt{Q}}$ without key 5; 5) P successfully executes CAS(&Set.Root_Pointer, RP, $\texttt{RP}_{\texttt{P}}$); 6) Q executes CAS from RP to $\texttt{RP}_{\texttt{Q}}$ but fails; thus, Q must retry its operation. Successful modifications are applied sequentially, one after another. Intuitively, this should not scale at all in a workload where all operations must perform successful modifications. As we will see in Section 4, this intuition would be incorrect. ## 3\. Analysis The key insight is that failed attempts to perform updates load data into processor caches that may be useful on future attempts. To better understand, consider the binary search tree modification depicted in Fig. 1. Suppose we want to insert two keys: 5 and 75. We compare how these insertions are performed sequentially and concurrently. At first, we consider the sequential execution. We insert key 5 into the tree. It should be inserted as a left child of 10. Thus, we traverse the tree from the root to the leaf 10. On the way, we fetch nodes {40, 30, 20, 10} into the processor’s cache. Note this operation performs four uncached loads. Now, we insert 75. It should be inserted as the right child of 70. Our traversal loads four nodes: {40, 50, 60, 70}. Node 40 is already cached, while three other nodes must be loaded from memory. Thus, we perform three uncached loads, for a total of seven uncached loads. Now, we consider a concurrent execution with two processes, in which P inserts 5 and Q inserts 75. Initially, both processes read Root_Ptr to load the current version. Then, 1) P traverses from the root to 10, loading nodes {40, 30, 20, 10}, and 2) Q traverses from the root to 70, loading nodes {40, 50, 60, 70}. Each process constructs a new version of the data structure, and tries to replace the root pointer using CAS. Suppose P succeeds and Q fails. Q retries the operation, but on the new version (Fig. 1). Note that the new version shares most nodes with the old one. Figure 1. The new version (green) of the tree shares its nodes with the old version (white) Q inserts 75 into the new version. Again, the key should be inserted as the right child of 70. Q loads four nodes {40, 50, 60, 70} from the new version of the tree. Crucially, nodes {50, 60, 70} are already cached by Q. This retry only incurs one cache miss! Thus, there are only five serialized loads in the concurrent execution, compared to seven in the sequential execution. ### 3.1. High-level analysis We use a simple model that allows us to analyze this effect. (The full proof appears in Appendix A.) In this model, the processes are synchronous, i.e., they perform one primitive operation per tick, and each process has its own cache of size $M$. We show that for a large number of processes $P$, the speedup is $\Omega(\log N)$, where $N$ is the size of the tree. Now, we give the intuition behind the proof. To simplify it, we suppose that the tree is external and balanced, i.e., each operation passes though $\log N$ nodes. We also assume that the workload consists of successful modification operations on keys chosen uniformly at random. We first calculate the cost of an operation for one process: $(\log N-\log M)\cdot R+\log M$ where $M=O(N^{1-\varepsilon})$ is the cache size and $R$ is the cost of an uncached load. This expression captures the expected behaviour under least-recently- used caching. The process should cache the first $\log M$ levels of the tree, and thus, $\log M$ nodes on a path are in the cache and $\log N-\log M$ are not. To calculate the throughput in a system with $P$ processes, we suppose that $P$ is quite large ($\approx\Omega(min(R,\log N))$). Thus, each operation performs several unsuccessful attempts, ending with one successful attempt, and all successful attempts (over all operations) are serialized. Since the system is synchronous, each operation attempt $A$ loads the version of the data structure which is the result of a previous successful attempt $A^{\prime}$. The nodes evicted since the beginning of $A$ are those created by $A^{\prime}$. One can show that in expectation only two nodes on the path to the key are uncached. Finally, the successful attempt of an operation incurs cost $2\cdot R+(\log N-2)$. Since successful attempts are serialized, the expected total speedup is $\frac{(\log N-\log M)\cdot R+\log M}{2\cdot R+(\log N-2)}$ giving $\Omega(\log N)$ with $R=\Omega(\log N)$. ## 4\. Experiments We implemented a lock-free treap and ran experiments comparing it with a sequential treap in Java on a system with an 18 core Intel Xeon 5220. Each data point is an average of 15 trials. We highlight the following two workloads. (More results appear in Appendix B.) ### 4.1. Batch inserts and batch removes Suppose we have $P$ concurrent processes in the system. Initially the set consists of $10^{6}$ random integer keys. Processes operate on mutually disjoint sets of keys. Each process repeatedly: inserts all of its keys, one by one, then removes all of its keys. Since the key sets are disjoint, each operation successfully modifies the treap. We report the speedup for our treap over the sequential treap below. ### 4.2. Random inserts and removes In this workload, we first insert $10^{6}$ random integers in [$-10^{6}$; $10^{6}$], then each process repeatedly generates a random key and tries to insert/remove it with equal probability. Some operations do not modify the data structure (e.g., inserting a key that already exists). Workload | Seq Treap | UC 1p | UC 4p | UC 10p | UC 17p ---|---|---|---|---|--- Batch | $451\,940$ | 0.89x | 1.23x | 1.47x | 1.47x Random | $419\,736$ | 1.48x | 2.38x | 3.07x | 3.19x ## References * (1) * Fatourou et al. (2020) Panagiota Fatourou, Nikolaos D Kallimanis, and Eleni Kanellou. 2020. An efficient universal construction for large objects. _arXiv_ (2020). * Herlihy (1991) Maurice Herlihy. 1991\. Wait-free synchronization. _ACM Transactions on Programming Languages and Systems (TOPLAS)_ 13, 1 (1991), 124–149. * Herlihy et al. (2020) Maurice Herlihy, Nir Shavit, Victor Luchangco, and Michael Spear. 2020. _The art of multiprocessor programming_. Newnes. * Kaplan (2018) Haim Kaplan. 2018\. Persistent data structures. In _Handbook of Data Structures and Applications_. Chapman and Hall/CRC, 511–527. * Lamport (1987) Leslie Lamport. 1987\. A fast mutual exclusion algorithm. _ACM Transactions on Computer Systems (TOCS)_ 5, 1 (1987), 1–11. * Okasaki (1999) Chris Okasaki. 1999\. _Purely functional data structures_. Cambridge University Press. * Seidel and Aragon (1996) Raimund Seidel and Cecilia R Aragon. 1996. Randomized search trees. _Algorithmica_ 16, 4 (1996), 464–497. * Sun et al. (2019) Y Sun, G Blelloch, W Lim, and A Pavlo. 2019\. On supporting efficient snapshot isolation for hybrid workloads with multi-versioned indexes. _VLDB_ 13, 2 (2019). * Yi et al. (2021) Z Yi, Y Yao, and K Chen. 2021. A Universal Construction to implement Concurrent Data Structure for NUMA-muticore. In _50th ICPP_. 1–11. ## Appendix A Mathematical model ### A.1. Sequential execution Let us estimate how much time is spent on executing $T$ operations sequentially on a binary search tree. Suppose our binary search tree is _external_ , i.e., data is contained only in leaves, while internal nodes maintain only routing information. Suppose tree contains $N$ keys and the tree is balanced, therefore the tree height is $O(\log N)$. We suppose _uniform workload_ : all keys from the tree are accessed uniformly at random. Suppose the cache size is $M=O(N^{1-\varepsilon})$, therefore, approximately upper $\log M$ levels of the tree are cached, while $\log N-\log M$ lower levels of the tree are not (Fig. 2). Figure 2. Upper levels of the tree are cached, while lower levels reside in RAM Each operation first loads $\log M$ nodes from the cache, spending $1$ time unit per each cache fetch. After that, it loads $\log N-\log M$ nodes from the RAM, spending $R$ time units per RAM fetch. Thus, the sequential execution will take $T\cdot\left(\log M+R\cdot\left(\log N-\log M\right)\right)$ time units to finish, where $T$ is the number of operations. ### A.2. Concurrent execution Suppose we have $P$ concurrent processes $\\{t_{i}\\}_{i=1}^{P}$ executing operations concurrently, while each process has its own cache of size larger than $\log N$. In our model we assume that each successful try of a modifying operation causes $p-1$ unsuccessful tries of modifying operations on other processes (Fig. 3). Figure 3. Each successful try of an operation causes unsuccessful tries of $p-1$ operations We also assume that operation completion events are distributed among processes in a round-robin pattern: first process $t_{1}$ executes its successful try of the operation, then process $t_{2}$ executes its successful try of the operation, and so on. Finally, $t_{P}$ manages to complete its operation, the next process to get its successful try is yet again $t_{1}$ (Fig. 4). Figure 4. Nearly each successful modifying operation consists of $P$ retries: $P-1$ unsuccessful and one successful As follows from the diagram, almost each successful try of an operation is preceded by $P-1$ unsuccessful retries (except for $P-1$ first successful operation, which are preceded by the lower number of unsuccessful retries). Let us estimate, how long the first retry takes to execute. We must load $\log N$ nodes, none of which might be cached. Thus, we spend $R\cdot\log N$ time units on the first retry. Let us estimate now how much time we spend on subsequent retries. We begin with estimating, how many nodes on the path to the requested leaf have been modified (Fig. 5). Figure 5. The number of modified nodes on the path to the requested node Consider the successful modifying operation $op$, that led to a latest failure of our CAS and made us retry our operation the last time. Remember, that arguments of operations are chosen uniformly at random, therefore: * • There is $\frac{1}{2}$ probability that $op$ modified some leaf from Root->Right subtree, thus, the number of modified nodes on our path is $1$; * • Similarly, there is $\frac{1}{4}$ probability that the number of modified nodes on our path is $2$; $\ldots$ * • Similarly, there is $\frac{1}{2^{k}}$ probability that the number of modified nodes on our path is $k$. Thus, we can calculate the expected number of modified nodes on our path $\sum\limits_{k=1}^{\log N}\frac{k}{2^{k}}\leq\sum\limits_{k=1}^{\infty}\frac{k}{2^{k}}=2$. Thus, the expected number of modified nodes on our path is not greater than $2$. Modified nodes were created by another process, thus they do not exist in our process cache. Therefore, they should be loaded out-of-cache, while all the remaining nodes reside in the local cache and can be loaded directly from it. Therefore, we spend $2\cdot R$ time on average to load all the necessary nodes. In addition, we spend $\log N-2$ time on average to load all the necessary nodes from the the local cache. Therefore, we spend $2\cdot R+\log N-2$ time to fetch all the nodes required for a last operation retry. An operation execution consists of the first retry, executed in $R\cdot\log N$ and $P-1$ subsequent retries executed in $(P-1)\cdot(2\cdot R+\log N-2)$. Thus, a single operation is executed in $R\cdot\log N+(P-1)\cdot(2\cdot R+\log N-2)$. Therefore, we execute $T$ operations in $\frac{T\cdot R\cdot\log N+T\cdot(P-1)\cdot(2\cdot R+\log N-2)}{P}$ time, since we execute these operations in parallel on $P$ processes. To measure the speedup we simply divide the sequential execution time by parallel execution time: $\frac{T\cdot\left(\log M+R\cdot\left(\log N-\log M\right)\right)}{\frac{T\cdot R\cdot\log N+T\cdot(P-1)\cdot(2\cdot R+\log N-2)}{P}}=P\cdot\frac{\log M+R\cdot\left(\log N-\log M\right)}{R\cdot\log N+(P-1)\cdot(2\cdot R+\log N-2)}$. This gives us $\Omega(\log N)$ speedup when $P=\Omega(min(R,\log N))$ and $R=\Omega(\log N)$. ## Appendix B Experiments on other processors We did the same experiments on Intel Xeon Platinum 8160 with 24 cores and AMD EPYC 7662 with 64 cores. Workload | Seq Treap | UC 1p | UC 6p | UC 12p | UC 23p ---|---|---|---|---|--- Batch | $638\,600$ | 0.93x | 1.31x | 1.37x | 1.08x Random | $487\,161$ | 1.24x | 3.23x | 3.55x | 2.8x Table 1. Results for Intel Xeon Platinum 8160. Workload | Seq Treap | UC 1p | UC 8p | UC 16p | UC 32p | UC 63p ---|---|---|---|---|---|--- Batch | $459\,580$ | 0.96x | 1.7x | 1.91x | 1.55x | 1.02x Random | $396\,898$ | 1.36x | 3.63x | 2.41x | 2.81x | 2.3x Table 2. Results for AMD EPYC 7662. Unfortunately, one can see that the results are not so impressive when the number of processes is large enough. We suggest that the bottleneck for our benchmarks occurs in Java memory allocator.
11institutetext: Center of Ubiquitous Computing Faculty of Information Technology and Electrical Engineering University of Oulu, Finland # The Limits of Learning and Planning: Minimal Sufficient Information Transition Systems Basak Sakcak Vadim Weinstein Steven M. LaValle This work was supported by a European Research Council Advanced Grant (ERC AdG, ILLUSIVE: Foundations of Perception Engineering, 101020977), Academy of Finland (projects PERCEPT 322637, CHiMP 342556), and Business Finland (project HUMOR 3656/31/2019). (e-mail: [email protected]). ###### Abstract In this paper, we view a policy or plan as a transition system over a space of information states that reflect a robot’s or other observer’s perspective based on limited sensing, memory, computation, and actuation. Regardless of whether policies are obtained by learning algorithms, planning algorithms, or human insight, we want to know the limits of feasibility for given robot hardware and tasks. Toward the quest to find the best policies, we establish in a general setting that minimal information transition systems (ITSs) exist up to reasonable equivalence assumptions, and are unique under some general conditions. We then apply the theory to generate new insights into several problems, including optimal sensor fusion/filtering, solving basic planning tasks, and finding minimal representations for feasible policies. ###### keywords: planning, sensing uncertainty, information spaces, theoretical foundations ## 1 Introduction Robotics increasingly appears as an application area for other fields. It is a frequent target for designing and testing machine learning algorithms, planning algorithms, sensor fusion methods, control laws, and so on. This may lead many to believe that robotics itself does not have its own, unique theoretical core (on this, we agree with Koditschek [6]). Thus, are we missing something? Surely robotics is not a pure algorithmic problem or pure nonlinear control problem. Could there be a theory that plays a similar role to Turing machines for computer science, or ${\dot{x}}=f(x,u)$ over differentiable manifolds for control theory, and yet is distinct from both? Can we formulate and potentially answer questions such as: Does a solution even exist to the problem? What are the minimal necessary components to solve it? What should the best learning approach imaginable produce as a representation? Such questions would be analogous to existence and uniqueness in control and dynamical systems, or decidability and complexity (especially Kolomogorov) in theoretical computer science. | | ---|---|--- (a) | | (b) Figure 1: (a)The internal robot brain is defined as an ITS that interacts with the external world (robot body and environment). (b)Coupled internal and external systems mathematically capture sensing, actuation, internal computation, and the external world. This paper proposes a robotics theory that is built from the input-output relationships between a programmable mechanical system (robot) and its environment via sensing and actuation; see Figure 1(a). The key is to focus on necessary and sufficient conditions that a robot’s internal processor (i.e., “brain”) must maintain to solve required tasks, such as collision-free navigation or coverage. We assume that the robot hardware and actuation model are fixed, and that for a given set of tasks in a space of environments, we must determine the weakest amount of sensing, actuation, and computation that would be sufficient for solving tasks. We will call such conditions minimal sufficient: If you take away anything from a minimal sufficient system, the tasks will become unsolvable. We introduce the notion of an information transition system (ITS) to formally model the robot’s brain (as well as any other system observers). The “information” part of an ITS is inspired by von Neumann’s definitions in the context of sequential games with hidden information (and not Shannon’s later notion of information theory). This inspired the development of information spaces [8] (Chapter 11) as a foundation of planning with imperfect state information due to sensing uncertainty. The concept of sufficient information mappings appears therein. It is generalized in this paper, and the state space of each ITS will in fact be an information space. In our work, the ITS and its underlying information space serve as the domain over which a plan or policy can be expressed and analyzed. Note that prior work in planning usually assumes that the space one plans over is fixed, as in a configuration space or state (phase) space based on the robot’s mobility. Even the information spaces described in [8] remain fixed in the planning phase. A notable exception is by O’Kane and Shell [11], in which information spaces for passive filtering are reduced algorithmically, and is closely connected to this paper. All such spaces will be considered here as potential information spaces, and we intend to reduce or collapse them as much as possible in the development of an information-feedback plan. This is perhaps closer to the goals of machine learning, in which candidate representations are determined through optimization of discrepancies with respect to input- output data. In this paper, we in fact consider both model-free and model- based ITSs, in alignment with the choices commonly found in machine learning [2]. The robot’s ITS is coupled to the physical world, which is itself modeled as a transition system. Note that the physical world model is not “given” to the robot; we will formalize notions of “who gets what” information in Section 3. The coupled system is inspired by neuroscience models (for example, [4]). Many of the concepts in this paper build upon [12], in which we recently proposed an enactivist-oriented model of cognition based on information spaces. By enactivist [5], it is meant that the necessary brain structures emerge from sensorimotor interaction and do not necessarily have predetermined, meaningful components as in the classical representationalist sense. Section 2 provides a mathematical formulation of robot-environment interaction as transition systems. Section 3 develops notions of sufficiency and minimality over the space of possible ITSs. Section 4 applies the general concepts to address what it means to solve both passive (filtering) and active (planning/control) tasks minimally and Section 5 provides simple examples. ## 2 Mathematical Models of Robot-Environment Systems #### 2.0.1 Internal and external systems In this paper, we will consider a robot embedded in an environment and describe this system as two subsystems, named internal and external, connected through symmetric input-output relations. External refers to the system describing the physical world, and internal is the complement of it. This interaction is shown in Figure 1(b). In this sense, the states of the external and internal systems are similar to the use of the term in control theory and computer science, respectively. External system corresponds to the totality of the environment and the robot body within it. Let $X$ denote the set of states of this system; a state could be for example, the configuration of the robot in a known environment (or within a set of possible environments) or its phase. There are no restrictions on $X$; it may be discrete, an $n$-dimensional manifold, a function space, and so on. Next, let $U$ be the set of control inputs (also referred to as actions) such that when applied at state $x\in X$ causes it to change according to a state transition function $f:X\times U\rightarrow X$. The set $U$ can also be anything: a finite or infinite discrete set, a compact or non- compact manifold, and so on. Similarly, at each state $x$, $y=h(x)$ is the output in which $h:X\rightarrow Y$ is a state-based sensor mapping and $Y$ is the set of all possible observations. The internal system (robot’s brain) observes the external system through a sensor mapping and interacts with it through a selection of actions with respect to a policy (alternatively, we can call it a plan or a strategy). Therefore, the input to the internal system is an observation and its output is an action. The states of this system correspond to the retained information gathered through the outcomes of actions in terms of sensor observations. To this end, the basis of our mathematical formulation of the internal is the notion of __information space_ (I-space)_ [8]. We will use the term __information state_ (I-state)_ to refer to the state of the internal system and denote it with ${\iota}$, and ${\cal I}$ will denote the set of all I-states, that is, the _I-space_. Note that the notions I-space and I-state are not exclusive to the internal system and we will use them in a more general setting in the following sections. Similar to the external system, the internal system evolves with each $y\in Y$ according to the information transition function $\phi:{\cal I}\times Y\rightarrow{\cal I}$. The output then corresponds to the control command given by an information feedback policy $\pi:{\cal I}\rightarrow U$. Finally, we can write the coupled dynamical system composed of these two subsystems defined as external and internal as $\displaystyle\centering x^{\prime}\@add@centering$ $\displaystyle=f(x,\pi({\iota}))$ $\displaystyle y=$ $\displaystyle h(x)$ $\displaystyle{\iota}^{\prime}$ $\displaystyle=\phi({\iota},h(x))$ $\displaystyle u=$ $\displaystyle\pi({\iota}).$ (1) Whereas the equations on the left side describe the evolution of this coupled system, the ones on the right show the respective outputs of each subsystem. Given an initial state $(x_{1},{\iota}_{1})\in X\times{\cal I}$, there exists a unique state-trajectory. Suppose the system evolves in discrete stages. For the external system, starting from an initial state $x_{1}$, each stage $k$ corresponds to applying an action $u_{k}$ which then yields the next stage $k+1$ and the next state $x_{k+1}=f(x_{k},u_{k})$. As the system evolves through stages, ${\tilde{x}}_{k}=(x_{1},x_{2},\dots,x_{k})$, ${\tilde{u}}_{k-1}=(u_{1},u_{2},\dots,u_{k-1})$, ${\tilde{y}}_{k}=(y_{1},y_{2},\dots,y_{k})$, correspond to the state, action and observation histories up to stage $k$, respectively. Note that applying the action $u_{k}$ at stage $k$ would result in a transition to state $x_{k+1}$ and the corresponding sensor reading $y_{k+1}=h(x_{k+1})$. The same applies for the internal system, we can describe its evolution, starting from an initial I-state ${\iota}_{0}$ following the state transition equation ${\iota}_{k}=\phi({\iota}_{k-1},y_{k})$. At stage $k$, $\pi({\iota}_{k})$ would produce the action $u_{k}$. Note that the stage index of the I-state starts from $0$, this corresponds to any prior information the internal system might have regarding the external; ${\iota}_{1}$ is then obtained using ${\iota}_{0}$ and $y_{1}$. Furthermore, $\pi({\iota}_{0})=u_{0}=()$ for all policies $\pi$, meaning that no action is outputted at this stage. #### 2.0.2 Generalizing to transition systems Without loss of generality, we can describe the internal and external subsystems as transition systems of the form $(S,\Lambda,T)$ in which $S$ is the set of states, $\Lambda$ is the set of names for the outgoing transitions, and $T\subset S\times\Lambda\times S$ is a ternary relation describing the transitions. If for each $(s,\lambda)\in S\times\Lambda$ there is a unique $s^{\prime}\in S$ such that $(s,\lambda,s^{\prime})\in T$, then we will write this system as $(S,\Lambda,\tau)$ in which $\tau:S\times\Lambda\rightarrow S$ is a function, and call the system an automaton. This corresponds to a deterministic system. Note that our definition of an automaton differs from the one usually used in computer science in the sense that ours do not necessarily have a start state and a set of accepting states, and it is not necessarily finite. Suppose ${\cal T}:S\times\Lambda\rightarrow{\rm pow}(S)$, in which ${\rm pow}(\cdot)$ denotes the power set. Then, the transition system $(S,\Lambda,{\cal T})$ is a nondeterministic automaton. In [12], we have used the notion of state-relabeled transition systems to model the internal and external systems. A state-relabeled transition system is the quintuple $(S,\Lambda,T,\sigma,L)$ in which $\sigma:S\rightarrow L$ is a labeling function and $(S,\Lambda,T)$ is a transition system. Preimages of a labeling function $\sigma$ induce a partitioning of the state space $S$. Let $S/\sigma$ be the set of equivalence classes $[s]_{\sigma}$ induced by $\sigma$ such that $S/\sigma=\\{[s]_{\sigma}\mid s\in S\\}$ and $[s]_{\sigma}=\\{s^{\prime}\in S\mid\sigma(s^{\prime})=\sigma(s)\\}$. Then, we can define a new transition system $(S/\sigma,\Lambda,T/\sigma)$ called the _quotient_ of $(S,\Lambda,T)$ by $\sigma$, in which $T/\sigma=\\{\left([s]_{\sigma},\lambda,[s^{\prime}]_{\sigma}\right)\mid(s,\lambda,s^{\prime})\in T\\}$. Note that $(S/\sigma,\Lambda,T/\sigma)$ is a reduced version of $(S,\Lambda,T)$. We might be interested in finding a labeling function $\sigma$ such that the corresponding quotient transition system is as simple as possible while ensuring that it is still useful. In the following sections, we will provide motivations for a reduction and discuss in more detail the requirements on $\sigma$ for the quotient system to be useful. Considering the deterministic case and the description of an automaton given above, external and internal systems can be written as state-relabeled automata $(X,U,f,h,Y)$ and $({\cal I},Y,\phi,\pi,U)$, respectively, in which $h$ and $\pi$ are considered as labeling functions. Interpreting the labels as the output of a transition system, coupled internal-external system can be described in terms of the state-relabeled transition systems formulation too such that output of one transition system is an input for another. Described this way, coupling of two transition systems result in unique paths in either automaton initialized at a particular state. ## 3 Sufficient Information Transition Systems ### 3.1 Information transition systems In the general setting, an I-state corresponds to the available (stored) information at a certain stage with respect to the action and observation histories. Consequently, an I-space refers to the collection of all possible I-states. We will use the term __information transition system_ (ITS)_ to refer to a transition system whose state space is an I-space. We have already used the notion of I-space while modeling the internal system representing the robot brain, which makes it an ITS. Here, we extend the notion of an ITS to include different perspectives from which the external and the coupled system is viewed. In particular, we identified three perspectives corresponding to 1) a plan executor which corresponds to the robot brain 2) a planner, and 3) an (independent) observer. With a slight abuse of previously introduced notation and terminology, we will use the term “internal” to refer to any system that is not the external and we will use ${\cal I}$ to denote a generic I-space. We describe an ITS in a robot-centric way such that an observation will refer to a sensor-reading, that is, $y$. However, an independent observer defined over the coupled system can observe, at stage $k$, both the action taken $u_{k-1}$ and the corresponding sensor-reading $y_{k}$. Recall that the information regarding the external is obtained through the sensor-mapping and any potential prior knowledge. Suppose that no policy is fixed over the I-space. Then, the corresponding internal system can be modeled as an ITS of the form $({\cal I},U\times Y,\phi$), in which $\phi:{\cal I}\times(U\times Y)\rightarrow{\cal I}$ is a state (information) transition function, if it is deterministic. We will then use the term _deterministic information transition system_ (DITS) to refer to them. Otherwise, it is called a _nondeterministic information transition system_ (NITS) and described as $({\cal I},U\times Y,\Phi)$, in which $\Phi\subseteq{\cal I}\times(U\times Y)\times{\cal I}$ is the transition relation. This formulation corresponds to the perspectives other than the plan executor such that it is possible to take any action from an I-state as it is not constrained by a policy, hence the outgoing transitions are determined by the elements of $U\times Y$. A plan executor corresponds to the internal system (robot’s brain) described in the previous section. The only information regarding the external is gained through manipulating the state of the external system through actions and obtaining the corresponding sensor readings. Recall the representation used in the previous section, that is, $({\cal I},Y,\phi)$, and $\pi:{\cal I}\rightarrow U$ a labeling function. This can be considered as a constrained version of the DITS described in the previous paragraph such that the transitions are restricted to those that can be realized under $\pi$. To show that, we augment the definition of internal system corresponding to the robot brain such that the transitions now also correspond to labels. Let $({\cal I},U\times Y,\Phi)$ be the augmented transition system describing the internal such that $\Phi=\\{({\iota},(u,y),{\iota}^{\prime})\in{\cal I}\times(U\times Y)\times{\cal I}\mid u=\pi({\iota})\land{\iota}^{\prime}=\phi({\iota},y)\\},$ (2) by construction, this augmented ITS is also deterministic111We could use the same approach for the external system too. In that case, let $(X,Y\times U,F)$ be this augmented transition system corresponding to the external, in which $F$ is the set of transitions such that $F=\\{(x,(y,u),x^{\prime})\in X\times(Y\times U)\times X\mid y=h(x)\land x^{\prime}=f(x,u)\\}$. Further creating bipartite graphs (for either system) such that transitions from a state correspond either to an observation $y\in Y$ or to an action $u\in U$ allows us to describe the coupling as a form of intersecting two automata. However, because it is not central to this paper we will not elaborate on this topic.. Suppose $({\cal I},U\times Y,\Phi^{\prime})$ is the DITS that is not constrained by a policy. Then, $\Phi\subseteq\Phi^{\prime}$. ### 3.2 History information spaces An I-space constitutes the state space of an ITS. Therefore, we describe the basic I-space named _history I-space_ denoted as ${\cal I}_{hist}$. It will be used to derive other I-spaces as well. A _history I-state_ at stage $k$ corresponds to all the information that is gathered through sensing (and potentially also through actions) up to stage $k$ assuming perfect memory. Let ${\eta}_{k}$ denote the history I-state at stage $k$, that is ${\eta}_{k}=({\eta}_{0},{\tilde{u}}_{k-1},{\tilde{y}}_{k})$, in which ${\eta}_{0}$ is the initial condition. Recall that ${\tilde{u}}_{0}$ is assumed to be the null-tuple, hence, ${\tilde{u}}_{k}$ starts with $u_{1}$ for any $k>1$. Let ${\cal I}_{0}$ be the set of initial conditions whose description varies with the available prior information. We defer the descriptions of possible ${\cal I}_{0}$ to the following paragraph. The history information space at stage $k$ is expressed as ${\cal I}_{k}={\cal I}_{0}\times{\tilde{U}}_{k-1}\times{\tilde{Y}}_{k}$. In general, the number of stages that the system will go through is not fixed. Therefore, we can define history I-space as the union over all $k\in\mathbb{N}$, that is, ${\cal I}_{hist}=\bigcup_{k\in\mathbb{N}}{\cal I}_{k}$. The DITS corresponding to ${\cal I}_{hist}$ becomes $({\cal I}_{hist},U\times Y,{\phi_{hist}})$, in which ${\eta}_{k}={\phi_{hist}}({\eta}_{k-1},u_{k-1},y_{k})={\eta}_{k-1}{}^{\frown}u_{k-1}{}^{\frown}y_{k}$ and ⌢ is the concatenation operation that adds an element at the end of a sequence. We consider two categories of initial conditions depending on whether information regarding the state space $X$ of the external system is available or not. Suppose $X$ or any information regarding $X$ is not given. Then, an I-state at stage $k$ simply is ${\eta}_{k}=({\tilde{u}}_{k-1},{\tilde{y}}_{k})$, that is, the concatenation of action and observation histories up till stage $k$. We call this type of history I-space, the _model-free history I-space_, and respectively call the corresponding ITS, _model-free history ITS_. In this case, we can treat ${\eta}_{0}$ as ${\eta}_{0}=()$. Thus, ${\cal I}_{0}=\\{()\\}$. For the second category of initial conditions, full or partial information regarding $X$, against which the actions and observations can be interpreted, is given. We will then use the terms _model-based history I-space_ and _model-based history ITS_ to refer to the respective I-space and ITS. The initial condition ${\eta}_{0}$ could be (i) a known state $x_{1}\in X$ such that ${\cal I}_{0}=X$, (ii) a set of possible initial states $X_{1}\subset X$ such that ${\cal I}_{0}={\rm pow}(X)$ or (iii) a probability distribution $P(x_{1})$ over $X$ such that ${\cal I}_{0}\subseteq\mathcal{P}(X)$, in which $\mathcal{P}(X)$ is the set of all probability distributions over $X$. ### 3.3 Sufficient state-relabeling In [12] we have introduced a notion of _sufficiency_ that substantially generalizes the definition in Chapter 11 of [8] and is presented here for completeness. ###### Definition 3.1 (Sufficient state-relabeling). Let $(S,\Lambda,T)$ be a transition system. A labeling function $\sigma:S\rightarrow L$ defined over the states of a transition system is sufficient if and only if for all $s,q,s^{\prime},q^{\prime}\in S$ and all $\lambda\in\Lambda$, the following implication holds: $\sigma(s)=\sigma(q)\land(s,\lambda,s^{\prime})\in T\land(q,\lambda,q^{\prime})\in T\implies\sigma(s^{\prime})=\sigma(q^{\prime}).$ If $\sigma$ is defined over the states of an automaton $(S,\Lambda,\tau)$, then $\sigma$ is sufficient iff for all $s,q\in S$ and all $\lambda\in\Lambda$, $\sigma(s)=\sigma(q)$ implies that $\sigma(\tau(s,\lambda))=\sigma(\tau(q,\lambda))$. Consider the stage-based evolution of external system $(X,U,f,h,Y)$ with respect to the action (control input) sequence ${\tilde{u}}_{k-1}=(u_{1},\dots,u_{k-1})$. This corresponds to the state and observation histories till stage $k$, that are ${\tilde{x}}_{k}=(x_{1},\dots,x_{k})$ and ${\tilde{y}}_{k}=(y_{1},\dots,y_{k})$. Recall that applying $u_{k}$ at stage $k$ would result in a transition to $x_{k+1}$ and the corresponding observation $y_{k+1}=h(x_{k+1})$. Hence, in this context, sufficiency of $h$ implies that given the label $y_{k}=h(x_{k})$ and the action $u_{k}$, it is possible to determine the label $y_{k+1}=h(x_{k+1})$. One interpretation of sufficiency of $h$ is that the respective quotient system sufficiently represents the underlying system up to the induced equivalence classes. This notion is similar to minimal realization of a system, that is, the minimal state space description that models the given input-output measurements (see for example [7]). Second interpretation is in a predictive sense. Suppose the quotient system is known. Then, the label $y_{k+1}=h(x_{k+1})$ can be determined before the system gets to $x_{k+1}$, using the current label $y_{k}$ and the action to be applied $u_{k}$. Furthermore, under a fixed policy, complete observation-trajectory can be determined from the initial observation by induction. Now, consider an internal system with a labeling function ${\kappa}:{\cal I}\rightarrow{\cal I}^{\prime}$, that is, $({\cal I},U\times Y,\phi,\kappa,{\cal I}^{\prime})$, and its evolution with respect to the observation history $\tilde{y}=(y_{1},\dots,y_{k})$. At stage $k$, the state of the automaton is ${\iota}_{k}$ and with $(u_{k},y_{k+1})$ the system transitions to ${\iota}_{k+1}=\phi({\iota}_{k},u_{k},y_{k+1})$. Sufficiency of ${\kappa}$ implies that given ${\kappa}({\iota}_{k})$, $u_{k}$, and $y_{k+1}$, we can determine ${\kappa}({\iota}_{k+1})$. This is equivalent to the definition introduced in Chapter 11 of [8] and makes it a special case for Definition 3.1. ### 3.4 Derived information transition systems Even though it seems natural to rely on a history ITS, dimension of a history I-space increases linearly with each stage, making it impractical in most cases. Thus, we are interested in defining a reduced ITS that is more manageable.Furthermore, this would largely simplify the description of a policy for a planner or a plan executor. Recall the quotient of a transition system by a a labeling function. We rewrite $({\cal I}_{hist},U\times Y,\phi_{hist})$ as $({\cal I}_{hist},U\times Y,\Phi_{hist})$, in which $\Phi_{hist}=\\{({\eta},(u,y),\phi_{hist}({\eta},u,y))\in{\cal I}_{hist}\times(U\times Y)\times{\cal I}_{hist}\\}.$ (3) We can introduce an _information mapping_ (I-map) ${\kappa}:{\cal I}_{hist}\rightarrow{\cal I}_{der}$ that categorizes the states of ${\cal I}_{hist}$ into equivalence classes through its preimages. In this case, ${\kappa}$ serves as a labeling function and a reduction can be obtained in terms of the quotient of $({\cal I}_{hist},U\times Y,\Phi)$ by ${\kappa}$, that is, $({\cal I}_{hist}/{\kappa},U\times Y,\Phi/{\kappa})$. It is crucial that the derived ITS is a DITS so that the labels can be determined using only the derived ITS without making reference to the history ITS.Considering the quotient system derived from $({\cal I}_{hist},U\times Y,\phi)$, which is a DITS by definition, by ${\kappa}$, we can not always guarantee that the resulting ITS is deterministic. This depends on the I-map used for state-relabeling as stated in the following proposition. ###### Proposition 3.2. For all non-empty $U$ and $Y$, and for the corresponding ${\cal I}_{hist}$, there exists a labeling function ${\kappa}$ such that the quotient of $({\cal I}_{hist},U\times Y,\phi)$ by $\kappa$, that is, $({\cal I}_{hist}/{\kappa},U\times Y,\Phi/{\kappa})$, in which $\Phi$ is defined as in (3), is not a DITS. ###### Proof 3.3. Let ${\kappa}:{\cal I}_{hist}\rightarrow\\{l_{1},l_{2}\\}$ such that ${\kappa}^{-1}(l_{1})=\\{{\eta}_{k}=({\tilde{u}}_{k-1},{\tilde{y}}_{k})\in{\cal I}_{hist}\mid{\tilde{u}}_{k-1}=(u_{i})_{i=1,\dots,k-1},u_{i}=u,\forall i=1,\dots,k-1\\}$ is the set of histories that correspond to applying the same action for $k-1$ times and ${\kappa}^{-1}(l_{2})$ is its complement, that is, ${\kappa}^{-1}(l_{2})={\cal I}_{hist}\setminus{\kappa}^{-1}(l_{1})$. Then, there exist sequences ${\eta}_{k-2}=({\tilde{u}}_{k-3},{\tilde{y}}_{k-2})$ and ${\eta}_{k-1}=({\tilde{u}}_{k-2},{\tilde{y}}_{k-1})$ such that ${\eta}_{k-2}={\eta}_{k-1}{}^{\frown}(u,y)$ and ${\eta}_{k}={\eta}_{k-1}{}^{\frown}(u,y)$ for which ${\kappa}({\eta}_{k-2})={\kappa}({\eta}_{k-1})=l_{2}$ and ${\kappa}({\eta}_{k})=l_{1}$. Thus, $\\{([{\eta}_{k-2}]_{\kappa},(u,y),[{\eta}_{k-1}]_{\kappa}),([{\eta}_{k-1}]_{\kappa},(u,y),[{\eta}_{k}]_{\kappa})\\}\in\Phi/{\kappa}.$ Since $[{\eta}_{k-2}]_{\kappa}=[{\eta}_{k-1}]_{\kappa}$ and $[{\eta}_{k-1}]_{\kappa}\neq[{\eta}_{k}]_{\kappa}$, the transition corresponding to $([{\eta}_{k-1}]_{\kappa},(u,y))$ is not unique; thus, $({\cal I}_{hist}/{\kappa},U\times Y,\Phi/{\kappa})$ is not deterministic.∎ Note that Proposition 3.2 holds also in the case of a generic ITS $({\cal I},U\times Y,\phi)$, with non-history I-states, if $\exists s,s^{\prime},q,q^{\prime}\in{\cal I}$ such that $\\{(s,(u,y),s^{\prime}),(q,(u,y),q^{\prime})\\}\in\Phi$, in which $\Phi$ is defined using $\phi$ as in (3). Then, any I-map ${\kappa}$ such that ${\kappa}(s)={\kappa}(q)$ and ${\kappa}(s^{\prime})\neq{\kappa}(q^{\prime})$ results in a quotient system that is not a DITS. For the quotient system derived from $({\cal I}_{hist},U\times Y,\phi)$ to be a DITS depends on the sufficiency of ${\kappa}$. In [12] it is shown that the quotient of a transition system $(S,\Lambda,T)$ by a labeling function $\sigma$ is an automaton (recall our definition) if and only if $(S,\Lambda,T)$ is full222A transition system $(S,\Lambda,T)$ is full, if $\forall s\in S,\lambda\in\Lambda$ there exists at least one $s^{\prime}\in S$ with $(s,\lambda,s^{\prime})\in T$. and $\sigma$ is sufficient. As $\phi_{hist}$ is a function with domain ${\cal I}_{hist}\times(U\times Y)$, it is full, then, the following follows from [12] as a special case. ###### Proposition 3.4. Let $({\cal I}_{hist}/{\kappa},U\times Y,\Phi_{hist}/{\kappa})$ be the quotient of $({\cal I}_{hist},U\times Y,\phi_{hist})$ by ${\kappa}$, in which $\Phi$ is defined as in (3), then $({\cal I}_{hist}/{\kappa},U\times Y,\Phi/{\kappa})$ is a DITS if and only if ${\kappa}$ is sufficient. For an I-map ${\kappa}:{\cal I}_{hist}\rightarrow{\cal I}_{der}$, $({\cal I}_{hist}/{\kappa},U\times Y,\Phi_{hist}/{\kappa})$ is isomorphic to $({\cal I}_{der},U\times Y,\Phi_{der})$, in which $\Phi_{der}=\\{({\kappa}({\eta}),(u,y),{\kappa}({\eta}^{\prime}))\mid({\eta},(u,y),{\eta}^{\prime})\in\Phi_{hist}\\}$ [12]. Thus, we can use the labels introduced by ${\kappa}$ as the new (derived) I-space and the corresponding quotient system as the derived ITS. Suppose, ${\kappa}$ is sufficient. Then, the derived ITS is a DITS, meaning that given the I-state ${\iota}_{k-1}\in{\cal I}_{der}$, and ($u_{k-1}$, $y_{k}$), ${\iota}_{k+1}\in{\cal I}_{der}$ can be uniquely determined. Consequently, we can write the derived ITS as $({\cal I}_{der},U\times Y,\phi_{der})$ in which $\phi_{der}:{\cal I}_{der}\times(U\times Y)\rightarrow{\cal I}_{der}$ is the new information transition function. Therefore, we no longer need to rely on the full histories and the history ITS and can rely solely on the derived ITS. This is shown in the first two rows of the following diagram: ${{\cal I}_{hist}}$${{\cal I}_{hist}}$${{\cal I}_{hist}}$${{\cal I}_{hist}}$${{\cal I}_{hist}}$ ${{\cal I}_{der}}$${{\cal I}_{der}}$${{\cal I}_{der}}$${{\cal I}_{der}}$${{\cal I}_{der}}$ ${{\cal I}_{min}}$${{\cal I}_{min}}$${{\cal I}_{min}}$${{\cal I}_{min}}$${{\cal I}_{min}}$ ${{\cal I}_{task}}$${{\cal I}_{task}}$${{\cal I}_{task}}$${{\cal I}_{task}}$${{\cal I}_{task}.}$$\scriptstyle{u_{1},y_{2}}$$\scriptstyle{{\kappa}}$$\scriptstyle{u_{2},y_{3}}$$\scriptstyle{{\kappa}}$$\scriptstyle{u_{3},y_{4}}$$\scriptstyle{{\kappa}}$$\scriptstyle{u_{4},y_{5}}$$\scriptstyle{{\kappa}}$$\scriptstyle{{\kappa}}$$\scriptstyle{u_{1},y_{2}}$$\scriptstyle{{\kappa}^{\prime}}$$\scriptstyle{u_{2},y_{3}}$$\scriptstyle{{\kappa}^{\prime}}$$\scriptstyle{u_{3},y_{4}}$$\scriptstyle{{\kappa}^{\prime}}$$\scriptstyle{u_{4},y_{5}}$$\scriptstyle{{\kappa}^{\prime}}$$\scriptstyle{{\kappa}^{\prime}}$$\scriptstyle{u_{1},y_{2}}$$\scriptstyle{{\kappa}^{\prime\prime}}$$\scriptstyle{u_{2},y_{3}}$$\scriptstyle{{\kappa}^{\prime\prime}}$$\scriptstyle{u_{3},y_{4}}$$\scriptstyle{{\kappa}^{\prime\prime}}$$\scriptstyle{u_{4},y_{5}}$$\scriptstyle{{\kappa}^{\prime\prime}}$$\scriptstyle{{\kappa}^{\prime\prime}}$ (4) Note that we can similarly define an I-map that maps any derived I-space to another. An example is given in (4) as the mappings ${\kappa}^{\prime}:{\cal I}_{der}\rightarrow{\cal I}_{min}$ and ${\kappa}^{\prime\prime}:{\cal I}_{min}\rightarrow{\cal I}_{task}$. The corresponding quotient system is deterministic for ${\kappa}^{\prime}$, indicating that it is sufficient. However, the quotient system by ${\kappa}^{\prime\prime}$ derived from ${\cal I}_{min}$ is not deterministic, hence, ${\kappa}^{\prime\prime}$ is not sufficient, as the next I-state can not be uniquely determined. Note that an I-map whose domain is ${\cal I}_{hist}$ can also be defined as composition of the mappings along the column of the diagram. For instance, ${\kappa}_{min}:{\cal I}_{hist}\rightarrow{\cal I}_{min}$ is the composition of ${\kappa}$ and ${\kappa}^{\prime}$, that is, ${\kappa}_{min}={\kappa}^{\prime}\circ{\kappa}$ (same for ${\kappa_{task}}:{\cal I}_{hist}\rightarrow{\cal I}_{task}$). ### 3.5 Lattice of information transition systems We fix ${\cal I}_{hist}$, which corresponds to fixing the set of initial states ${\cal I}_{0}$. Then, each I-map ${\kappa}$ defined over ${\cal I}_{hist}$ induces a partition of ${\cal I}_{hist}$ through its preimages, denoted as ${\cal I}_{hist}/{\kappa}$. An I-map ${\kappa}^{\prime}$ is a refinement of ${\kappa}$, denoted as ${\kappa}^{\prime}\succeq{\kappa}$, if $\forall A\in{\cal I}_{hist}/{\kappa}^{\prime}$ there exists a $B\in{\cal I}_{hist}/{\kappa}$ such that $A\subseteq B$. Let $K({\cal I}_{hist})$ denote the set of all partitions over ${\cal I}_{hist}$. Refinement induces a partial ordering since not all partitions of ${\cal I}_{hist}$ are comparable. The partial ordering given by refinements form a lattice of partitions over ${\cal I}_{hist}$, denoted as $(K({\cal I}_{hist}),\succeq)$. At the top of the lattice, there is the partition induced by an identity I-map (or equivalently, by a bijection), ${\kappa}_{id}:{\cal I}_{hist}\rightarrow{\cal I}_{hist}$, since all of its elements are singletons (all equivalence classes contain exactly one element), making it the maximally distinguishable case. Conversely, we can define a constant mapping ${\kappa}_{const}:{\cal I}_{hist}\rightarrow{\cal I}_{const}$ for which ${\cal I}_{hist}/{\kappa}_{const}$ is a singleton, that is, ${\cal I}_{const}=\\{{\iota}_{const}\\}$, which then will be at the bottom of the lattice. In turn, ${\kappa}_{const}$ yields the minimally distinguishable case as all histories now belong to a single equivalence class. This idea is similar to the notion of the _sensor lattice_ defined over the partitions of $X$ [9, 13]. Indeed, if we take ${\cal I}_{0}=X$ and consider ${\kappa}_{est}:{\cal I}_{hist}\rightarrow X$, the ordering of partitions of ${\cal I}_{hist}$ such that ${\cal I}_{hist}/{\kappa}_{est}$ is the least upper bound gives out the sensor lattice. As motivated in previous sections, we are interested in finding a sufficient I-map such that the quotient ITS derived from the history ITS is still deterministic. Notice that the constant I-map ${\kappa}_{const}$ is sufficient by definition since for all $(u,y)\in U\times Y$, and all ${\eta},{\eta}^{\prime}\in{\cal I}_{hist}$, we have that ${\kappa}_{const}({\eta})={\kappa}_{const}({\eta}^{\prime})$ and ${\kappa}_{const}(\phi_{hist}({\eta},(u,y))={\kappa}_{const}(\phi_{hist}({\eta}^{\prime},(u,y)).$ On the other hand, in certain cases it is crucial to differentiate certain histories from the others. This will become clear in the next section when we describe the notion of a task. Suppose ${\kappa}$ is a labeling that partitions ${\cal I}_{hist}$ into equivalence classes that are of importance and suppose that ${\kappa}$ is not sufficient. Then, we want to find a refinement of ${\kappa}$ that is sufficient. This will serve as a lower bound on the lattice of partitions over ${\cal I}_{hist}$ since for any partition such that ${\cal I}_{hist}/{\kappa}$ is a refinement of it, the classes of histories that are deemed crucial will not be distinguished. The following defines the refinement of ${\kappa}$ that ensures sufficiency and a minimal number of equivalence classes. ###### Definition 3.5. Let $({\cal I}_{hist},U\times Y,{\phi_{hist}})$ be a history ITS and ${\kappa}$ an I-map. A _minimal sufficient refinement_ of ${\kappa}$ is a sufficient I-map ${\kappa}^{\prime}$ such that there does not exist a sufficient I-map ${\kappa}^{\prime\prime}$ that satisfy ${\kappa}^{\prime}\succ{\kappa}^{\prime\prime}\succeq{\kappa}$. ###### Remark 3.6. It is shown in [12] that the minimal sufficient refinement of ${\kappa}$ defined over the states of an automaton $(S,\Lambda,\tau)$ is unique. ## 4 Solving Tasks Minimally #### 4.0.1 Definition of a task We now connect the general ITS concepts to the accomplishment of particular tasks. We have two categories: 1) active, which corresponds to planning and executing an information-feedback policy that forces a desirable outcome in the environment, and 2) passive, which means only to observe the environment without being able to effect changes. Recall from Section 3.4 that there may be model-free or model-based formulations. In the model-free case, tasks are specified using a logical language over ${\cal I}_{hist}$ which will result in a labeling and derived I-space ${\cal I}_{task}$ and associated I-map ${\kappa_{task}}$ that corresponds to the “resolution” at which the tasks are specified. Various logics are allowable, such as propositional or a temporal logic. The resulting sentences of the language involve combinations of predicates that may assign true or false values to subsets of ${\cal I}_{hist}$. Solving an active task (or tasks) requires that a sentence of interest becomes true during execution of the policy. This is called satisfiability. For example, the task may be to simply reach some goal set $G\subset{\cal I}_{hist}$, causing a predicate in-goal$({\cal I}_{hist})$ to become satisfied (in other words, be true). Using linear temporal logic, more complex requirements, such as cycling through a finite sequence of subsets forever while avoiding others, can be specified [3]. Solving a passive task only requires maintaining whether a sentence is satisfied, rather than forcing an outcome; this corresponds to filtering. Whether the task is active or passive, if satisfiability is concerned with a single, fixed sentence, then a task-induced labeling (or task labeling for short), that is, ${\kappa_{task}}$, over ${\cal I}_{hist}$ assigns two labels: Those I-states that result in true and those that result in false. A task labeling may also be assigned for a set of possible sentences by assigning a label to each set of the common refinement of the partition of ${\cal I}_{hist}$ induced by each possible sentence. In the model-based case, tasks are instead specified using a language over $X$, and sentence satisfiability must be determined by an I-map that converts history I-states into expressions over $X$. #### 4.0.2 Problem families It is assumed that the state-relabeled transition system $(X,U,f,h,Y)$ describing the external system is fixed, but it is unknown or partially known to the observer (a robot or other observer). Filtering (passive case) requires maintaining the label of an I-state attributed by ${\kappa_{task}}$. Since ${\kappa_{task}}$ is not necessarily sufficient, we can not guarantee that the quotient system by ${\kappa_{task}}$ is a DITS (Propositions 3.2 and 3.4). Thus, relying solely on the quotient system by ${\kappa_{task}}$, we can not determine the class that the current history belongs to (see the last row in (4)) and, hence can not determine whether a sentence describing the task is satisfied (or which sentences are satisfied). Suppose the sets $U$ and $Y$ are specified, and at each stage $k$, $u_{k-1}$ is known and $y_{k}$ is observed. The following describes the problem for a passive task given a state-relabeled (history) ITS $({\cal I}_{hist},U\times Y,\phi_{hist},{\kappa_{task}},{\cal I}_{task})$, in which ${\kappa_{task}}:{\cal I}_{hist}\rightarrow{\cal I}_{task}$ is a task labeling that is not sufficient, and ${\cal I}_{task}$ is the corresponding I-space. ###### Problem 4.1 (Find a sufficient I-space filter). Find a sufficient refinement of ${\kappa_{task}}$. Note that ${\cal I}_{hist}/{\kappa_{task}}$ determines a lower bound on the partitioning of ${\cal I}_{hist}$ which is interpreted as the crucial information that can not be lost. Consequently, histories belonging to different equivalence classes with respect to ${\kappa_{task}}$ must always be distinguished from each other. However, Problem 4.1 does not an upper bound. At the limit, a bijection from ${\cal I}_{hist}$ is always a sufficient refinement of ${\kappa_{task}}$. As stated previously, using history ITS can create computational obstructions in solving problems. This motivates the following problem. ###### Problem 4.2 (Filter minimization). Find a minimal sufficient refinement of ${\kappa_{task}}$. We now consider a basic planning problem, for which ${\cal I}_{task}=\\{0,1\\}$, such that ${\kappa_{task}}^{-1}(1)\subset{\cal I}_{hist}$ is the set of histories that achieve the goal, and ${\kappa_{task}}^{-1}(0)\subset{\cal I}_{hist}$ is its complement. Most planning problems refer to finding a labeling function $\pi$ such that, when used to label the states of the internal system, guarantees task accomplishment. Then $\pi$ is called a feasible policy, which is defined in the following. Consider an external system $(X,f,U,h,Y)$. Let $\mathcal{R}_{X}({\cal I}_{task})\subseteq X$ be the set of initial states for which there exist a $k$ and histories ${\tilde{x}}_{k}$, ${\tilde{u}}_{k-1}$, and ${\tilde{y}}_{k}$, such that $x_{i+1}=f(x_{i},u_{i})$ and $y_{i}=h(x_{i})$ for all $0<i<k$, and ${\eta}_{k}\in{\kappa_{task}}^{-1}(1)$, in which ${\eta}_{k}$ is the history I-state corresponding to ${\tilde{u}}_{k-1}$ and ${\tilde{y}}_{k}$. ###### Definition 4.3 (Feasible policy for ${\cal I}_{task}$). Let $({\cal I},Y,\phi,\pi,U)$ and $(X,f,U,h,Y)$ be the state-relabeled transition systems corresponding to internal and external systems, respectively. Labeling function $\pi$ defined over ${\cal I}$ is a feasible policy for ${\cal I}_{task}$ if for all $x\in\mathcal{R}_{X}({\cal I}_{task})$, the history corresponding to the coupled internal-external system initialized at $({\iota}_{0},x)$ belongs to ${\kappa_{task}}^{-1}(1)$. Most problems in the planning literature consider a fixed DITS and look for a feasible policy for ${\cal I}_{task}$. This yields the following problem. Typically, the I-space considered is $X$ which corresponds to state estimation. Note that a DITS, in other words, the robot brain, can be seen as an I-space filter itself. ###### Problem 4.4 (Find a feasible policy). Given $({\cal I},Y,\phi)$, an internal system (robot brain), find a labeling function $\pi:{\cal I}\rightarrow U$ that is a feasible policy for ${\cal I}_{task}$. We can further extend the planning problem to consider an unspecified internal system which refers to finding a DITS, and a policy such that the resulting histories satisfy the task description, that is, he problem of jointly finding an I-space-filter and a policy defined over its states. Recall from (2) that a policy constrains the transitions of an ITS to the ones that are realizable under that policy. ###### Problem 4.5 (Find a DITS and a feasible policy). Given $({\cal I}_{hist},U\times Y,\phi_{hist})$, for which ${\kappa_{task}}:{\cal I}_{hist}\rightarrow{\cal I}_{task}$ is a task labeling and ${\cal I}_{task}$ is the corresponding I-space, find a sufficient I-map ${\kappa}:{\cal I}_{hist}\rightarrow{\cal I}$ and a feasible policy $\pi$ for ${\cal I}_{task}$ as a labeling function for the resulting quotient system by ${\kappa}$. Note that ${\kappa_{task}}$ can already be sufficient, so that it is the minimal sufficient refinement of itself, however, this does not necessarily imply the existence of a feasible policy defined over ${\cal I}_{task}$. Therefore, we are not looking for a refinement of ${\kappa_{task}}$ while describing the DITS over which the policy is defined. On the other hand, we can still talk about a notion of minimality. Let $({\cal I},Y,\phi,\pi,U)$ be a state-relabeled DITS that solves Problem 4.5 (or similarly Problem 4.4) which is the quotient system of history ITS by ${\kappa}$, then, $({\cal I},Y,\phi,\pi,U)$ is _minimal for $\pi$_ if there does not exist a sufficient I-map ${\kappa}^{\prime}$ with ${\kappa}\succ{\kappa}^{\prime}$ for which there exists a $\pi^{\prime}$ for the quotient system by ${\kappa}^{\prime}$ that satisfies $\pi({\iota})=\pi^{\prime}({\kappa}^{\prime}({\iota}))$. #### 4.0.3 Learning a sufficient ITS Although learning and planning overlap significantly, some unique issues arise in pure learning (see also [12]). This corresponds to the case when ${\cal I}_{task}$ is not initially given but needs to be revealed through interactions with the external system, that is, respective action and observation histories. It is assumed that whether the sentence (or sentences) describing the task is satisfied or not can be assessed at a particular history I-state. We can address both filtering and planning problems defined previously within this context, considering model-free and model-based cases. In the model-free case, the task is to compute a minimal sufficient ITS that is consistent with the actions and observations. Variations include lifelong learning, in which there is a single, ‘long’ history I-state, or more standard learning in which the system can be restarted, resulting in multiple trials, each yielding a different history I-state. In the model-based case, partial specifications of $X$, $f$, and $h$ may be given, and unknown parameters are estimated using the history I-state(s). Different results are generally obtained depending on what assumptions are allowed. For example, do identical history I-states imply identical state trajectories? If not, then set-based, nondeterministic models may be assumed, or even probabilistic models based on behavior observed over many trials and assumptions on probability measure, priors, statistical independence, and so on. ## 5 Illustrative Examples In this section we provide some simple examples to show how the ideas presented in this paper apply to filtering and planning problems. All problems can be posed as well in a machine learning context for which ${\cal I}_{task}$ is not given but it is revealed through interactions between the internal and external as the input-output data. Figure 2: (a) State-relabeled history ITS described in Example 5.1, and labeling function ${\kappa_{task}}$ yellow colored states correspond to states that satisfy the task description. (b) Equivalence classes induced by ${\kappa}^{\prime}$; the minimal sufficient refinement of ${\kappa_{task}}$. (c) Quotient of the history ITS by ${\kappa}^{\prime}$. (d)DITS describing the internal system solving the planning problem described in Example 5.2. (e) Environment used in Examples 5.1,5.2, the obstacle (an open disk) is shown in black. (f) L-shaped corridor; $l_{1},l_{2}\leq l$. Let $E\subseteq{\mathbb{R}}^{2}$ be a bounded planar environment (see Figure 2(e)) that is partitioned into regions separated by gates Each gate is either green or red whose color can be detected by the robot’s color sensor and follows the rule that each region shares a boundary with exactly two gates; one green and one red. The set of possible observations are $Y=\\{r,g\\}$. ###### Example 5.1. This example considers a filtering problem from the perspective of an independent observer. Suppose the actions taken by the robot are not observable and the only information about the system is the history of readings coming from the robot’s color sensor; for example, $(r,r,r,g,r,g)$. Then, the history I-space is the set of all finite length sequences of elements of $Y$, that is, ${\cal I}_{hist}=Y^{*}$, which refers to the free monoid generated by the elements of $Y$ (or the Kleene star of $Y$). Hence, the history ITS can be represented as an infinite binary tree. The task is to determine whether the robot crosses the gates consistently (in a clockwise or counterclockwise manner) or not. Hence, the preimages of ${\kappa_{task}}:{\cal I}_{hist}\rightarrow{\cal I}_{task}$ partition ${\cal I}_{hist}$ into two subsets: one which the condition is satisfied and the others. The labeling induced by ${\kappa_{task}}$ is shown in Figure 2(a). Clearly, ${\kappa_{task}}$ is not sufficient since there exist I-states ${\eta},{\eta}^{\prime}$ such that ${\kappa_{task}}({\eta})={\kappa_{task}}({\eta}^{\prime})$ and there exists a $y$ for which ${\kappa_{task}}({\phi_{hist}}({\eta},y))\neq{\kappa_{task}}({\phi_{hist}}({\eta}^{\prime},y))$; for example consider ${\eta}=(r,g)$, ${\eta}^{\prime}=(r,g,r)$ and $y=g$. A sufficient refinement of ${\kappa_{task}}$ can be obtained (equivalence classes shown in 2(b)), denote as ${\kappa}^{\prime}$, for which the quotient DITS is shown in Figure 2(d). Furthermore, ${\kappa}^{\prime}$ is a minimal sufficient refinement of ${\kappa_{task}}$ since it follows from Proposition 4.2 that if a labeling is not minimal then there is a minimal one that is strictly coarser and is still sufficient. However, neither of the subsets that belong to ${\cal I}_{hist}/{\kappa}^{\prime}$ can be merged, since merging ${\iota}_{nt}$ (colored gray) with anything else violates the condition that ${\kappa}^{\prime}$ is a refinement of ${\kappa_{task}}$ and any pairwise merge of the others violate sufficiency. Suppose the robot has a boundary detector, and executes a bouncing motion using the two motion primitives $U=\\{u_{1},u_{2}\\}$, in which $u_{1}$ is move forward and $u_{2}$ is rotate in place. Let a basic motion be move forward and bounce off the walls, which can be implemented using the elements of $U$. We assume that the boundary detector and color sensor readings do not arrive simultaneously. ###### Example 5.2. We now consider a planning problem (that belongs to the class described in Problem 4.5) for which the goal is to ensure that the robot crosses the gates consistently. The history I-space of the planner is ${\cal I}_{hist}=(U\times Y)^{*}$ and the preimages of ${\kappa_{task}}$ partition ${\cal I}_{hist}$ into two sets; the histories that satisfy the predicate and the ones that do not. Then, a DITS with only three states (see Figure 2(d)) can be derived using the mapping ${\kappa}:{\cal I}_{hist}\rightarrow{\cal I}$, in which ${\cal I}=\\{i_{0},i_{1},i_{2}\\}$ such that for $\pi({\iota}_{1})$ boundary with the red gate is set as a wall, $\pi({\iota}_{2})$ the boundary with the green gate is set as a wall, and for $\pi({\iota}_{0})$ no boundary with gates are considered as a wall. We assume that a bouncing motion can be determined using the motion primitives so that the resulting trajectory will strike every open interval in the boundary of every region infinitely often, with non-zero, non-tangential velocities [1]. Consider a robot in an L-shaped planar corridor (Figure 2(f)). Let $\mathcal{E}_{l}$ be the set of all such environments such that $l_{1},l_{2}\leq l$, in which $l_{1}$ and $l_{2}$ are the dimensions of the corridor bounded by $l$. We assume that the minimum length/width is larger than the robot radius, that is, $1$. The state space $X$ is defined as the set of all pairs $(q,E_{i})$, in which $(q_{1},q_{2})\in E_{i}$, and $E_{i}\in\mathcal{E}_{l}$. The action set is $U=\\{0,1\\}\times\\{0,1\\}$ which corresponds to moving one step in one of the 4 cardinal directions; if the boundary is reached, the state does not change. The robot has a sensor that reports $1$ if the motion is blocked. ###### Example 5.3. Consider a model-based history ITS with ${\eta}_{0}\subset X$ that specifies the initial position as $q_{0}=(0,0)$ but does not specify the environment so that it can be any $E_{i}\in{\cal{E}}_{l}$. Let ${\cal I}_{hist}$ be its set of states and ${\kappa}_{ndet}:{\cal I}_{hist}\rightarrow{\rm pow}(X)$ be an I-map that maps ${\eta}_{k}$ to a subset of $X_{k}\subseteq X$. Since $(X,U,f,h,Y)$, and $X_{0}={\eta}_{0}$ are known, transitions for the quotient system can be described by induction as $X_{k+1}=\hat{X}_{k+1}(X_{k},u_{k})\cap H(y_{k+1})$, in which $\hat{X}_{k+1}=\bigcup_{x_{k}\in X_{k}}f(x_{k},u_{k})$ and $H(y_{k+1})\subseteq X$ is the set of all states that could yield $y_{k+1}$. By construction, ${\kappa}_{ndet}$ is sufficient. Suppose ${\kappa_{task}}:{\cal I}_{hist}/{\kappa}_{ndet}\rightarrow{\cal I}_{task}$ is a task labeling for localization that assigns each singleton a unique label and all the other subsets are labeled the same. Since the transition corresponding to $([X^{\prime}]_{{\kappa_{task}}},(u,y))$ in which $X^{\prime}\subseteq X$ is not a singleton can lead to multiple labels $[x^{\prime}]_{\kappa_{task}}$, in which $x^{\prime}$ is a singleton, ${\kappa_{task}}$ is not sufficient. Furthermore, ${\kappa}_{ndet}$ is a minimal sufficient refinement of ${\kappa_{task}}$ because it is sufficient and because there does not exist a sufficient ${\kappa}$ such that ${\kappa}_{ndet}\succ{\kappa}\succeq{\kappa_{task}}$. Suppose ${\kappa}$ exists, that would mean some equivalence classes can be merged. However, this is not possible because merging any of the non-singleton subsets violates sufficiency (as shown for ${\kappa_{task}}$) and merging singletons with others violates that it is a refinement. A policy can be described over ${\cal I}_{hist}/{\kappa}_{ndet}$; $u=(1,0)$ starting from $X_{0}$ until $y_{k}=1$ is obtained and applying $u=(0,1)$ starting from $X_{k}$ until $y_{n}=1$ is obtained, then it is found that $q=(k,n)$ and $E$ is the corridor with $l_{1}=k$, $l_{2}=n$. ## 6 Discussion We have introduced a mathematical framework for determining minimal feasible policies for robot systems, by comparing ITSs over I-spaces. The uniqueness and minimality results are quite general: $X$ and $U$ could be discrete, typical configuration spaces, or more exotic, such as the power set of all functions from an infinite-dimensional Hilbert space into an infinite- dimensional Banach space. Nevertheless, there are opportunities to expand the general theory. For example, we assumed that $u$ is both the output of a policy and the actuation stimulus in the physical world; more generally, we should introduce a mapping from an action symbol $\sigma\in\Sigma$ to a control function ${\tilde{u}}\in{\tilde{U}}$ so that plans are expressed as $\pi:{\cal I}\rightarrow\Sigma$ and each $\sigma=\pi(\iota)$ produces energy in the physical world via a mapping from $\Sigma$ to ${\tilde{U}}$. Another extension is to consider stochastic models, that amounts to an ITS with a probabilistic I-space, and in which ways it ties to the representations used in the literature such as _predictive state representations_ (PSRs) [10]. Other potential extensions include continuous-time “transitions” and novel logics to consider satisfiability questions over spaces of robot systems and tasks. An interesting direction is to consider the hardware and actuation models are variables, and fix other model components. A grand challenge remains: The results here are only a first step toward producing a more complete and unique theory of robotics that clearly characterizes the relationships between common tasks, robot systems, environments, and algorithms that perform filtering, planning, or learning. We should search for lattice structures that play a role similar to that of language class hierarchies in the theory of computation. This includes the structures of the current paper and the sensor lattices of [9, 13]. Many existing filtering, planning, and learning methods can be formally characterized within this framework, which would provide insights into relative complexity, completeness, minimality, and time/space/energy tradeoffs. ## References * [1] L. Bobadilla, O. Sanchez, J. Czarnowski, K. Gossman, and S. M. LaValle. Controlling wild bodies using linear temporal logic. In Proceedings Robotics: Science and Systems, 2011. * [2] Axel Brunnbauer, Luigi Berducci, Andreas Brandstätter, Mathias Lechner, Ramin M. Hasani, Daniela Rus, and Radu Grosu. Latent imagination facilitates zero-shot transfer in autonomous racing. https://arxiv.org/abs/2103.04909v2. * [3] G. E. Fainekos, A. Girard, H. Kress-Gazit, and G. J. Pappas. Temporal logic motion planning for dynamic mobile robots. Automatica, 45(2):343–352, 2009. * [4] K. Friston. The free-energy principle: A unified brain theory? Nature Reviews Neuroscience, 11(2):127–138, 2010. * [5] D. D. Hutto and E. Myin. Radicalizing enactivism: Basic minds without content. MIT Press, 2012. * [6] D. E. Koditschek. What is robotics? Why do we need it and how can we get it? Annual Review of Control, Robotics, and Autonomous Systems, 4:1–33, May 2021. * [7] Ü. Kotta, C. H. Moog, and M. Tõnso. Minimal realizations of nonlinear systems. Automatica, 95:207–212, 2018. * [8] S. M. LaValle. Planning Algorithms. Cambridge University Press, Cambridge, U.K., 2006. Also available at http://lavalle.pl/planning/. * [9] S. M. LaValle. Sensing and Filtering: A Fresh Perspective Based on Preimages and Information Spaces, volume 1:4 of Foundations and Trends in Robotics Series. Now Publishers, Delft, The Netherlands, 2012. * [10] Michael Littman and Richard S Sutton. Predictive representations of state. Advances in neural information processing systems, 14, 2001. * [11] J. M. O’Kane and D. A. Shell. Concise planning and filtering: Hardness and algorithms. IEEE Transactions on Automation Science and Engineering, 14(4):1666–1681, 2017. * [12] V. Weinstein, B. Sakcak, and S. M. LaValle. An enactivist-inspired mathematical model of cognition. under review, 2022. * [13] Y. Zhang and D. A. Shell. Lattices of sensors reconsidered when less information is preferred. In IEEE International Conference on Robotics and Automation, 2021\. *[I-space]: information space *[I-state]: information state *[ITS]: information transition system *[DITS]: deterministic information transition system *[NITS]: nondeterministic information transition system *[ I-space]: information space *[ I-state]: information state *[ ITS]: information transition system *[I-map]: information mapping *[PSRs]: predictive state representations
# Numerical simulations of inflationary dynamics: slow-roll and beyond Siddharth S. Bhatt Swagat S. Mishra Soumen Basak and Surya N. Sahoo ###### Abstract Cosmic inflation is a period of rapid accelerated expansion of space in the very early universe. During inflation, vacuum quantum fluctuations are amplified and stretched to cosmological scales which seed the fluctuations in the cosmic microwave background as well as the large-scale structure of our universe. Large quantum fluctuations may lead to the formation of primordial black holes (PBHs) in the post-inflationary universe. Numerical simulations of the inflationary dynamics are presented here for a single canonical scalar field minimally coupled to gravity. We spell out the basic equations governing the inflationary dynamics in terms of cosmic time $t$ and define a set of dimensionless variables convenient for numerical analysis. We then provide a link to our simple numerical Python code on GitHub that can be used to simulate the background dynamics as well as the evolution of linear perturbations during inflation. The code computes both scalar and tensor power spectra for a given inflaton potential $V(\phi)$. We discuss a concrete algorithm to use the code for various purposes, especially for computing the enhanced scalar power spectrum in the context of PBH formation. We intend to extend the framework to simulate the dynamics of a number of different quantities, including the computation of scalar-induced second-order tensor power spectrum in the revised version of this manuscript in the near future. ## 1 Introduction Cosmic inflation has emerged as the leading scenario for describing the very early universe prior to the commencement of the radiative hot Big Bang Phase [1, 2, 3, 4, 5, 6]. According to the inflationary paradigm, a transient epoch of at least 60-70 e-folds of rapid accelerated expansion suffices in setting natural initial conditions for the background space-time in the form of spatial flatness as well as statistical homogeneity and isotropy on large angular scales [2, 3, 4, 7]. Additionally, (and more significantly,) quantum fluctuations during inflation naturally generate a spectrum of almost scale- invariant initial scalar fluctuations which seed the temperature and polarisation fluctuations in the Cosmic Microwave Background (CMB) Radiation, and later, the formation of structure in the universe [8, 9, 10, 11, 7]. In addition to scalar perturbations, quantum fluctuations during inflation also create a spectrum of almost scale-invariant tensor perturbations which later become gravitational waves [12, 13]. The simplest models of inflation comprising of a single scalar field, called the ‘inflaton’, which is minimally coupled to gravity, makes several distinct predictions [14] (i.e an almost scale-invariant, nearly Gaussian, and adiabatic spectrum of scalar fluctuations) most of which have received spectacular observational confirmation, particularly from the latest CMB missions [15]. However, as mentioned earlier, inflation also generates tensor perturbations that later constitute the relic gravitational wave background (GW) which imprints a distinct signature on the CMB power spectrum in the form of the B-mode polarization [15]. The amplitude of these relic GWs provides us information about the inflationary energy scale while their spectrum enables us to access general properties of the epoch of reheating, being exceedingly sensitive to the post-inflationary equation of state [13, 16]. The amplitude of inflationary tensor fluctuations, relative to that of scalar fluctuations, is usually characterised by the tensor-to-scalar ratio $r$. Different models of inflation predict different values for $r$ which is sensitive to the gradient of the inflaton potential $V_{,\phi}(\phi)=\frac{dV(\phi)}{d\phi}$ relative to its height $V(\phi)$. Convex potentials predict large values for $r$, while concave potentials predict relatively small values of $r$. While the spectrum of inflationary tensor fluctuations has not yet been observed, current CMB observations are able to place an upper bound on the tensor-to- scalar ratio on large angular scales. In particular, the latest CMB observations of BICEP/Keck [17], combined with those of the PLANCK mission [15], place the strong upper bound $r\leq 0.036$ (at $95\%$ confidence). This most recent upper bound on $r$ has important consequences for single field canonical inflation. In particular, given $r\leq 0.036$, all monotonically increasing convex potentials, including the whole family of monomial potentials $V(\phi)\propto\phi^{p}$, are completely ruled out in the canonical framework. Among these strongly disfavoured models are the simplest classic inflaton potentials $\frac{1}{2}m^{2}\phi^{2}$ and $\lambda\phi^{4}$. Instead, the observational upper bound on $r$ appears to favour asymptotically-flat potentials possessing one or two plateau-like wings; see [19, 18]. Current observational data lead to a scenario in which the inflaton $\phi$ slowly rolls down a shallow potential $V(\phi)$ thereby giving rise to a quasi-de Sitter early stage of near-exponential expansion. A thorough analysis of the inflationary phase-space dynamics $\\{\phi,\dot{\phi}\\}$ for plateau potentials shows [20] that a large range of initial conditions leads to adequate inflation in these models. However, it is important to stress that the CMB window constitutes only a tiny part of the observationally available field space between the Hubble-exit of the largest scales in the sky and the smallest scale at the end of inflation. Consequently, a substantial period of the inflationary dynamics corresponding to potentially interesting small-scale primordial physics (which accounts roughly to the last 40–50 e-folds of accelerated expansion during inflation) remains observationally unexplored, being inaccessible to the CMB and LSS observations. Any departure from the slow-roll regime, that might be triggered by a change in the dynamics of the inflaton field, would lead to interesting observational consequences on small-scales. In particular, the presence of a feature at intermediate field values that might lead to large enough amplification of the small-scale scalar fluctuations, would facilitate the formation of Primordial Black Holes upon the Hubble re-entry of these modes during the post-inflationary epochs. Primordial Black Holes (PBHs) are extremely interesting compact objects which might have been formed from the collapse of large density fluctuations in the early universe [21, 22, 23, 24] and they constitute a potential candidate for dark matter [27, 26, 25, 28, 29, 30]. Seeds for such large fluctuations can be generated during inflation, as mentioned above. For instance, a feature in the inflaton potential in the form of a flat inflection point can further slow down the already slowly rolling inflaton field substantially, leading to an enhancement of the primordial scalar power $P_{\zeta}$. A number of different features for enhancing small-scale power during inflation have been proposed in the recent years [31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49]. Hence PBHs (and the associated induced relic GWs) are excellent probes of the small-scale primordial physics. In this preliminary version of our paper, we discuss a simple code developed for numerical simulations of the inflationary dynamics. We introduce the relevant dimensionless variables used in our numerical analysis and provide a link to the code in our GitHub account. We also discuss how to use the code in various scenarios, which include phase-space analysis of inflationary initial conditions, inflationary background dynamics and determining scalar and tensor power spectra both under slow-roll approximation and beyond. The latter case has important implications for PBH formation and we discuss how to use the code to simulate the Mukhanov-Sasaki equation mode by mode. We also discuss a number of important future directions that are to be included in the forthcoming version of our paper. The primary version of our numerical framework is quite simple and less compact. It is intended to provide a pedagogical guideline for researchers who are relatively new to numerical simulations of inflation. In the forthcoming version of our work, we will introduce a much more compact numerical framework that we are currently working on which will incorporate additional new features. This paper is organised as follows: we begin with a brief introduction of the inflationary scalar field dynamics in section 2 and quantum fluctuations in section 3. We then proceed to discuss numerical simulations of the background dynamics in section 4. This also includes studying the scalar and tensor fluctuations under the slow-roll approximations. Section 5 is dedicated to studying the inflationary quantum fluctuations by numerically solving the Mukhanov-Sasaki equation, and its application to inflaton potentials possessing a slow-roll violating feature. We also mention a number of future extensions of our numerical set-up in section 6, before concluding with a discussion section. We work in the units $c,\hbar=1$. The reduced Planck mass is defined to be $m_{p}\equiv 1/\sqrt{8\pi G}=2.43\times 10^{18}~{}{\rm GeV}$. We assume the background universe to be described by a spatially flat Friedmann-Lemaitre- Robertson-Walker (FLRW) metric with signature $(-,+,+,+)$. ## 2 Inflationary Dynamics The Action for a scalar field which couples minimally to gravity has the following general form $S[\phi]=\int d^{4}x\,\sqrt{-g}\;{\cal L}(F,\phi),$ (2.1) where the Lagrangian density ${\cal L}(\phi,F)$ is a function of the field $\phi$ and the kinetic term $F=\frac{1}{2}\partial_{\mu}\phi\;\partial^{\mu}\phi.$ (2.2) Varying (2.1) with respect to $\phi$ results in the equation of motion $\frac{\partial{\cal L}}{\partial\phi}-\left(\frac{1}{\sqrt{-g}}\right)\partial_{\mu}\left(\sqrt{-g}\frac{\partial{\cal L}}{\partial\left(\partial_{\mu}\phi\right)}\right)=0.$ (2.3) The energy-momentum tensor associated with the scalar field is $T^{\mu\nu}=\left(\frac{\partial{\cal L}}{\partial F}\right)\,\left(\partial^{\mu}\phi\;\partial^{\nu}\phi\right)-g^{\mu\nu}\,{\cal L}~{}.$ (2.4) Specializing to a spatially flat FRW universe and a homogeneous scalar field, one gets $ds^{2}=-dt^{2}+a^{2}(t)\left[dx^{2}+dy^{2}+dz^{2}\right]\,,$ (2.5) $T^{\mu}_{\;\>\;\nu}=\mathrm{diag}\left(-\rho_{{}_{\phi}},p_{{}_{\phi}},p_{{}_{\phi}},p_{{}_{\phi}}\right),$ (2.6) where the energy density, $\rho_{{}_{\phi}}$, and pressure, $p_{{}_{\phi}}$, are given by $\displaystyle\rho_{{}_{\phi}}$ $\displaystyle=\left(\frac{\partial{\cal L}}{\partial F}\right)\,(2\,F)-{\cal L}\,,$ (2.7) $\displaystyle p_{{}_{\phi}}$ $\displaystyle={\cal L}\,,$ (2.8) and $F=-({\dot{\phi}}^{2}/2)$. The evolution of the scale factor $a(t)$ is governed by the Friedmann equations: $\displaystyle\left(\frac{\dot{a}}{a}\right)^{2}\equiv H^{2}$ $\displaystyle=$ $\displaystyle\frac{1}{3m_{p}^{2}}\,\rho_{{}_{\phi}},$ (2.9) $\displaystyle\frac{\ddot{a}}{a}\equiv\dot{H}+H^{2}$ $\displaystyle=$ $\displaystyle-\frac{1}{6m_{p}^{2}}\,\left(\rho_{{}_{\phi}}+3\,p_{{}_{\phi}}\right),$ (2.10) where $H\equiv\dot{a}/a$ is the Hubble parameter and $\rho_{{}_{\phi}}$ satisfies the conservation equation ${\dot{\rho}_{{}_{\phi}}}=-3\,H\left(\rho_{{}_{\phi}}+p_{{}_{\phi}}\right)~{}.$ (2.11) In the standard single field inflationary paradigm, inflation is sourced by a minimally coupled canonical scalar field $\phi$ with a suitable potential $V(\phi)$ (see figure 1). For such a canonical scalar field ${\cal L}(F,\phi)=-F-V(\phi),$ (2.12) Figure 1: This figure schematically depicts a prototype inflation potential $V(\phi)$, plotted in solid green curve. The ‘CMB Window’ represents field values corresponding to the Hubble-exit epochs of scales $k\in\left[0.0005,0.5\right]~{}{\rm Mpc}^{-1}$ that are observable by the latest CMB missions. Substituting (2.12) into (2.7) and (2.8), we find $\displaystyle\rho_{{}_{\phi}}$ $\displaystyle=$ $\displaystyle\frac{1}{2}{\dot{\phi}}^{2}+\;V(\phi),$ $\displaystyle p_{{}_{\phi}}$ $\displaystyle=$ $\displaystyle\frac{1}{2}{\dot{\phi}}^{2}-\;V(\phi),~{}~{}$ (2.13) consequently the two Friedmann equations (2.9), (2.10) and the equation (2.11) become $\displaystyle H^{2}\equiv\frac{1}{3m_{p}^{2}}\,\rho_{\phi}$ $\displaystyle=$ $\displaystyle\frac{1}{3m_{p}^{2}}\left[\frac{1}{2}{\dot{\phi}}^{2}+V(\phi)\right]\,,$ (2.14) $\displaystyle\dot{H}\equiv\frac{\ddot{a}}{a}-H^{2}$ $\displaystyle=$ $\displaystyle-\frac{1}{2m_{p}^{2}}\,\dot{\phi}^{2}\,,$ (2.15) $\displaystyle{\ddot{\phi}}+3\,H{\dot{\phi}}+V_{,\phi}(\phi)$ $\displaystyle=$ $\displaystyle 0\,.$ (2.16) The epoch of inflation at any time $t<t_{\rm end}$ is conveniently marked by the number of e-folds before the end of inflation $N_{e}=\log_{e}{\frac{a_{\rm end}}{a(t)}}=\int_{t}^{t_{\rm end}}H(t^{\prime})dt^{\prime}\,,$ (2.17) where $H(t)$ is the Hubble parameter during inflation. $a(t)$ and $a_{\rm end}$ denote the scale factor at time $t$ and at the end of inflation respectively. Typically a period of quasi-de Sitter inflation lasting for at least 60-70 e-folds is required in order to address the problems of the standard hot Big Bang model. We denote $N_{*}$ as the number of e-folds (before the end of inflation) when the CMB pivot scale $k_{*}=(aH)_{*}=0.05~{}\rm Mpc^{-1}$ left the comoving Hubble radius during inflation. For convenience, we have chosen $N_{*}=60$ for the most part of this work, although the exact value of $N_{*}$ depends upon the particular detail of reheating history. The quasi-de Sitter like phase corresponds to the inflaton field rolling slowly down the potential $V(\phi)$. This slow-roll regime111It is well known that the slow-roll phase of the inflation is actually a local attractor for many different models of inflation, see [50, 20] and the references therein. of inflation, ensured by the presence of the Hubble friction term in the equation (2.16), is usually characterised by the first two kinematical Hubble slow-roll parameters $\epsilon_{H}$, $\eta_{H}$, defined by [7] $\displaystyle\epsilon_{H}$ $\displaystyle=$ $\displaystyle-\frac{\dot{H}}{H^{2}}=\frac{1}{2m_{p}^{2}}\,\frac{\dot{\phi}^{2}}{H^{2}},$ (2.18) $\displaystyle\eta_{H}$ $\displaystyle=$ $\displaystyle-\frac{\ddot{\phi}}{H\dot{\phi}}=\epsilon_{H}+\frac{1}{2\epsilon_{H}}\,\frac{d\epsilon_{H}}{dN_{e}}~{},$ (2.19) where the slow-roll regime of inflation corresponds to $\epsilon_{H},~{}\eta_{H}\ll 1~{}.$ (2.20) The slow-roll regime is also often characterised by the dynamical potential slow-roll parameters [7], defined by $\epsilon_{{}_{V}}=\frac{m_{p}^{2}}{2}\left(\frac{V_{,\phi}}{V}\right)^{2}~{},~{}~{}\eta_{{}_{V}}=m_{p}^{2}\,\left(\frac{V_{,\phi\phi}}{V}\right)~{}.$ (2.21) For small values of these parameters $\epsilon_{H},\,\eta_{H}\ll 1$, one finds $\epsilon_{H}\simeq\epsilon_{{}_{V}}$ and $\eta_{H}\simeq\eta_{{}_{V}}-\epsilon_{{}_{V}}$. Using the definition of Hubble parameter, $H=\dot{a}/a$, we have $\ddot{a}/a=\dot{H}+H^{2}=H^{2}(1+\dot{H}/H^{2})$. From the expression for $\epsilon_{H}$ in (2.18), it is easy to see that $\frac{\ddot{a}}{a}=\big{(}1-\epsilon_{H}\big{)}\,H^{2}~{}.$ (2.22) Which implies that the universe accelerates, ${\ddot{a}}>0$, when $\epsilon_{H}<1$. Using equation (2.14), the expression for $\epsilon_{H}$ in (2.18) reduces to $\epsilon_{H}\simeq\frac{3}{2}\frac{\dot{\phi}^{2}}{V}$ when ${\dot{\phi}}^{2}\ll V$. Figure 2: This figure schematically illustrates the evolution of the comoving Hubble radius $(aH)^{-1}$ with scale factor. During inflation $(aH)^{-1}$ decreases which causes physical scales to exit the Hubble radius. After inflation ends $(aH)^{-1}$ increases, and physical scales begin to re-enter the Hubble radius. The CMB pivot scale, as used by the Planck mission, is set at $k_{*}=0.05~{}{\rm Mpc}^{-1}$ and has been depicted by the dashed blue line. The ‘CMB Window’ corresponds to the scales $k\in\left[0.0005,0.5\right]~{}{\rm Mpc}^{-1}$ that are observable by the latest CMB missions. Note that $(aH)^{-1}\propto a$ during the radiative regime and $(aH)^{-1}\propto a^{-1}$ during inflation. ## 3 Quantum fluctuations during inflation In the standard scenario of a minimally coupled single canonical scalar field $\phi$ as the inflaton, two gauge-independent massless fields, one scalar, and one transverse traceless tensor, get excited during inflation and receive quantum fluctuations that are correlated over super-Hubble scales [51] at late times. ### Scalar fluctuations The evolution of the scalar degree of freedom, called the curvature perturbation222Note that the comoving curvature perturbation ${\cal R}$ is related to the curvature perturbation on uniform-density hypersurfaces, $\zeta$, and both are equal during slow-roll inflation as well as on super- Hubble scales, $k\ll aH$, in general (see [7]). $\zeta$ is described by the following quadratic Action [7, 52] $\boxed{S_{(2)}[\zeta]=\frac{1}{2}\int{\rm d}\tau{\rm d}x^{3}\,z^{2}\,\left[\,(\zeta^{\prime})^{2}-(\partial_{i}\zeta)^{2}\,\right]}~{},$ (3.1) which upon the change of variable $v\equiv z\,{\zeta}~{},\quad\mbox{with}\quad z=am_{p}\sqrt{2\epsilon_{H}}=a\,\frac{\dot{\phi}}{H}\,,$ (3.2) takes the form $S_{(2)}\left[v\right]=\frac{1}{2}\int{\rm d}\tau{\rm d}x^{3}\left[\left(v^{\prime}\right)^{2}-\left(\partial_{i}v\right)^{2}+\frac{z^{\prime\prime}}{z}v^{2}\right]~{},$ (3.3) where the $(^{\prime})$ denotes derivative with respect to conformal time $\tau=\int\frac{dt}{a(t)}$ ($\simeq\frac{-1}{aH}$ for quasi-de Sitter expansion). The variable $v$, which itself is a scalar quantum field like $\zeta$, is called the Mukhanov-Sasaki variable in literature. Its Fourier modes $v_{k}$ satisfy the famous Mukhanov-Sasaki equation given by [53, 54] $\boxed{v^{\prime\prime}_{k}\,+\,\left(k^{2}-\frac{z^{\prime\prime}}{z}\right)v_{k}=0}~{},$ (3.4) where the effective mass term is given by the following exact expression [55] $\displaystyle\boxed{\frac{z^{\prime\prime}}{z}=(aH)^{2}\left(2-\epsilon_{1}+\frac{3}{2}\epsilon_{2}+\frac{1}{4}\epsilon_{2}^{2}-\frac{1}{2}\epsilon_{1}\epsilon_{2}+\frac{1}{2}\epsilon_{2}\epsilon_{3}\right)}~{},$ (3.5) $\displaystyle\Rightarrow\boxed{\frac{z^{\prime\prime}}{z}=(aH)^{2}\left[2+2\epsilon_{H}-3\eta_{H}+2\epsilon_{H}^{2}+\eta_{H}^{2}-3\epsilon_{H}\eta_{H}-\frac{1}{aH}\,\eta^{\prime}_{H}\right]}~{},$ (3.6) with $\epsilon_{1}=\epsilon_{H}$ and where $\epsilon_{n+1}=-\frac{d\ln{\epsilon_{n}}}{dN_{e}}~{}$ (3.7) are the ‘Hubble flow’ parameters. Given a mode $k$, at sufficiently early times when it is sub-Hubble i.e $k\gg aH$, we can assume $v$ to be in the Bunch-Davies vacuum [56] satisfying $v_{k}\rightarrow\frac{1}{\sqrt{2k}}e^{-ik\tau}~{}.$ (3.8) During inflation as the comoving Hubble radius falls (see figure 2), modes start becoming super-Hubble i.e $k\ll aH$ and equation (3.4) dictates that $|v_{k}|\propto z$ and hence $\zeta_{k}$ approaches a constant value. By solving the Mukhanov-Sasaki equation we can estimate the dimensionless primordial power spectrum of $\zeta$ using the following relation [51] $\boxed{P_{\zeta}\equiv\frac{k^{3}}{2\pi^{2}}\,|{\zeta}_{k}|^{2}\Big{|}_{k\ll aH}=\frac{k^{3}}{2\pi^{2}}\,\frac{|v_{k}|^{2}}{z^{2}}\Big{|}_{k\ll aH}}~{}.$ (3.9) During slow-roll inflation, the factor $\frac{z^{\prime\prime}}{z}=\frac{\nu^{2}-0.25}{\tau^{2}}$ with $\nu\approx 1.5+\epsilon_{H}+\frac{\dot{\epsilon}_{H}}{2H\epsilon_{H}}$. Solving the Mukhanov-Sasaki equation with suitable Bunch-Davies vacuum conditions leads to the famous slow-roll approximation formula [7] $\boxed{P_{\zeta}=\frac{1}{8\pi^{2}}\left(\frac{H}{m_{p}}\right)^{2}\frac{1}{\epsilon_{H}}}~{}.$ (3.10) Note that one could also directly try to solve for the fourier modes of the comoving curvature perturbation $\zeta$ (instead of the Mukhanov-Sasaki variable $v$) which satisfies the equation $\boxed{{\zeta}^{\prime\prime}_{k}+2\left(\frac{z^{\prime}}{z}\right){\zeta}^{\prime}_{k}+k^{2}{\zeta}_{k}=0}$ (3.11) and implement the corresponding Bunch-Davies initial conditions for ${\zeta}_{k}$. The friction term in equation (3.11) is given by $\boxed{\frac{z^{\prime}}{z}=aH\,(1+\epsilon_{H}-\eta_{H})}~{}.$ (3.12) Before moving forward, we stress that the slow-roll regime of inflation necessarily requires both the slow-roll parameters to be small i.e $\epsilon_{H}\ll 1$ and $\eta_{H}\ll 1$. Violation of either of these conditions invalidates the above analytical treatment. When either of the slow-roll conditions is violated, which is the situation in the context of primordial black hole formation, a more accurate determination of $P_{\zeta}$ is provided by solving the Mukhanov-Sasaki equation (3.4) numerically. The computation of power spectrum when the slow-roll approximation is violated will be our primary focus. ### Tensor fluctuations The corresponding quadratic Action for tensor fluctuations is given by [7, 52] $\boxed{S_{(2)}[\gamma_{ij}]=\frac{1}{2}\int{\rm d}\tau{\rm d}^{3}x\,a^{2}\left(\frac{m_{p}}{2}\right)^{2}\left[(\gamma_{ij}^{\prime})^{2}-(\partial\gamma_{ij})^{2}\right]}\,.$ (3.13) The Mukhanov-Sasaki variable for tensor fluctuations are defined as $\frac{m_{p}}{2}\,a\,\gamma_{ij}\equiv\left(\begin{smallmatrix}h_{+}&h_{\times}&0\\\ h_{\times}&-h_{+}&0\\\ 0&0&0\end{smallmatrix}\right)$ (3.14) or, $\frac{m_{p}}{2}\,a\,\gamma_{ij}\equiv\sum_{s=+,\times}\Pi_{ij}^{s}\ h_{s}\,,$ (3.15) where $\Pi^{+}$ and $\Pi^{\times}$ are the 2 polarization modes, written as $\Pi^{+}_{ij}=\begin{pmatrix}1&0&0\\\ 0&-1&0\\\ 0&0&0\end{pmatrix}\quad\quad{\rm and}\quad\quad\Pi^{\times}_{ij}=\begin{pmatrix}0&1&0\\\ 1&0&0\\\ 0&0&0\end{pmatrix}\,.$ (3.16) Thus, $\gamma_{ij}=\frac{2}{m_{p}}\sum_{s=+,\times}\Pi_{ij}^{s}\,\frac{h_{s}}{a}\,.$ (3.17) The evolution equation for the mode functions (by dropping the ‘s’ subscript and remembering that it is summed over for 2 polarization states) is given by $\boxed{h_{k}^{\prime\prime}+\left(k^{2}-\frac{a^{\prime\prime}}{a}\right)h_{k}=0}\,.$ (3.18) The subsequent computation of tensor power spectrum $P_{T}(k)\equiv 2\times\frac{k^{3}}{2\pi^{2}}\,\frac{|h_{k}|^{2}}{a^{2}}\Big{|}_{k\ll aH}\,,$ (3.19) under quasi-de Sitter approximation leads to [7] $\boxed{P_{T}(k)=\frac{2}{\pi^{2}}\left(\frac{H}{m_{p}}\right)^{2}}\,.$ (3.20) Note that, unlike the Mukhanov-Sasaki equation (3.4) for scalar fluctuations, the tensor mode equation (3.18) does not depend upon $z$, rather it depends only upon the scale factor $a$. Hence, as long as the quasi-de Sitter approximation is valid, i.e $\epsilon_{H}\ll 1$, power spectrum of tensor fluctuations does not get affected by an appreciable amount even if slow-roll is violated. Although, this statement is true only at linear order in perturbation theory. Tensor fluctuations at second order in perturbation theory can be induced by large first-order scalar fluctuations [57, 58, 59, 60] (also see [61] and references therein). ### 3.1 Large scale primordial fluctuations On large cosmological scales which are accessible to CMB observations, the scalar power spectrum typically takes the form of a power law represented by $P_{\zeta}(k)=A_{{}_{S}}\left(\frac{k}{k_{*}}\right)^{n_{{}_{S}}-1},$ (3.21) where $A_{{}_{S}}=P_{\zeta}(k_{*})$ is the amplitude of the scalar power spectrum at the pivot scale $k=k_{*}$, given by333Note that in general, $k$ may correspond to any observable CMB scale in the range $k\in\left[0.0005,0.5\right]~{}{\rm Mpc}^{-1}$. However, in order to derive constraints on the inflationary observables $\\{n_{{}_{S}},r\\}$, we mainly focus on the CMB pivot scale, namely $k\equiv k_{*}=0.05~{}{\rm Mpc}^{-1}$. $\boxed{A_{{}_{S}}=\frac{1}{8\pi^{2}}\left(\frac{H}{m_{p}}\right)^{2}\frac{1}{\epsilon_{H}}\,\bigg{|}_{\phi=\phi_{*}}}~{},$ (3.22) where $\phi_{*}$ is the value of the inflaton field at the epoch of Hubble exit of the CMB pivot scale $k_{*}$. The scalar spectral tilt $n_{{}_{S}}$, in the slow-roll regime is given by [7] $\displaystyle\boxed{n_{{}_{S}}-1\equiv\frac{d\,\mathrm{ln}P_{\zeta}}{d\,\mathrm{ln}k}=2\eta_{H}-4\epsilon_{H}}~{}.$ (3.23) Similarly the tensor power spectrum, in the slow-roll limit, is represented by $P_{T}(k)=A_{{}_{T}}\left(\frac{k}{k_{*}}\right)^{n_{{}_{T}}},$ (3.24) with the amplitude of tensor power spectrum at the CMB pivot scale is given by [7, 51] $\boxed{A_{{}_{T}}\equiv P_{T}(k_{*})=\frac{2}{\pi^{2}}\left(\frac{H}{m_{p}}\right)^{2}\bigg{|}_{\phi=\phi_{*}}}~{},$ (3.25) and the tensor spectral index (with negligible running) is given by $\boxed{n_{{}_{T}}=-2\,\epsilon_{H}}~{}.$ (3.26) The tensor-to-scalar ratio $r$ is defined by $\boxed{r\equiv\frac{A_{{}_{T}}}{A_{{}_{S}}}=16\,\epsilon_{H}}~{},$ (3.27) yielding the single field consistency relation $\boxed{r=-8\,n_{{}_{T}}}~{}.$ (3.28) Hence the slow-roll parameters $\epsilon_{H}$ and $\eta_{H}$ play an important role in characterising the power spectra of scalar and tensor fluctuations during inflation. Before going forward, we briefly discuss the implications of the latest CMB observations for the slow-roll parameters as well as for other relevant inflationary observables. In order to relate the CMB observables to the inflaton potential $V(\phi)$, we work with the potential slow-roll parameters defined in equation (2.21). Consider a canonical scalar field minimally coupled to gravity and having the potential $V(\phi)=V_{0}\,f\left(\frac{\phi}{m_{p}}\right)~{}.$ (3.29) The potential slow-roll parameters (2.21) are given by $\displaystyle\epsilon_{{}_{V}}=\frac{m_{p}^{2}}{2}\left(\frac{f_{,\phi}}{f}\right)^{2}~{},$ (3.30) $\displaystyle\eta_{{}_{V}}=m_{p}^{2}\left(\frac{f_{,\phi\phi}}{f}\right)~{}.$ (3.31) In the slow-roll limit $\epsilon_{{}_{V}},\,\eta_{{}_{V}}\ll 1$, the scalar power spectrum is given by the expression (3.21) with the amplitude of scalar power at the CMB pivot scale $k\equiv k_{*}=0.05~{}{\rm Mpc}^{-1}$ expressed as [7] $A_{{}_{S}}\equiv P_{\cal\zeta}(k_{*})\simeq\frac{1}{24\pi^{2}}\frac{V_{0}}{m_{p}^{4}}\frac{f\left(\phi_{k}\right)}{\epsilon_{{}_{V}}(\phi_{k})}\bigg{|}_{k=k_{*}}~{},$ (3.32) and the scalar spectral index (with negligible running) is given by $n_{{}_{S}}\simeq 1+2\,\eta_{{}_{V}}(\phi_{*})-6\,\epsilon_{{}_{V}}(\phi_{*})~{},$ (3.33) Similarly the amplitude of tensor power spectrum at the CMB pivot scale is given by $A_{{}_{T}}\equiv P_{T}(k_{*})=\frac{2}{\pi^{2}}\left(\frac{H}{m_{p}}\right)^{2}\bigg{|}_{\phi=\phi_{*}}\simeq\frac{2}{3\pi^{2}}\frac{V_{0}}{m_{p}^{4}}f\left(\phi_{*}\right)~{},$ (3.34) and the tensor spectral index (3.26) becomes $n_{{}_{T}}\simeq-2\,\epsilon_{{}_{V}}(\phi_{*})~{},$ (3.35) and the tensor-to-scalar ratio (3.27) can be written as $r\simeq 16\,\epsilon_{{}_{V}}(\phi_{*})~{},$ (3.36) satisfying the single field consistency relation (3.28). From the CMB observations of Planck 2018 [15], we have $A_{{}_{S}}=2.1\times 10^{-9}~{},$ (3.37) while the $2\sigma$ constraint on the scalar spectral index is given by $n_{{}_{S}}\in[0.957,0.976]~{}.$ (3.38) Similarly the constraint on the tensor-to-scalar ratio $r$, from the latest combined observations of Planck 2018 [15] and BICEP/Keck [17], is given by $r\leq 0.036~{},$ (3.39) which translates into $A_{{}_{T}}\leq 3.6\times 10^{-2}\,A_{{}_{S}}$. Equation (3.34) helps place the following upper bound on the inflationary Hubble scale $H^{\rm inf}$ and the energy scale during inflation $E_{\inf}$ $\displaystyle H^{\rm inf}\leq 4.7\times 10^{13}~{}{\rm GeV}~{},$ (3.40) $\displaystyle E_{\inf}\equiv\left[\sqrt{3}\,m_{p}\,H^{\rm inf}\right]^{1/2}\leq 1.4\times 10^{16}~{}{\rm GeV}~{}~{}.$ (3.41) Similarly the CMB bound on $r$ when combined with (3.36) translates into an upper bound on the first slow-roll parameter $\epsilon_{H}\simeq\epsilon_{{}_{V}}\leq 0.00225~{},$ (3.42) rendering the tensor tilt from equation (3.35) to be negligibly small $|n_{{}_{T}}|\leq 0.0045~{}.$ (3.43) Given the upper limit on $\epsilon_{{}_{V}}$, using the CMB bound on $n_{{}_{S}}$ from (3.38) in (3.33), we infer that the second slow-roll parameter is negative and obtain interesting upper and lower limits on its magnitude, given by $|\eta_{H}|\in[0.0075,0.0215]~{}.$ (3.44) The EOS $w_{\phi}$ of the inflaton field is given by $w_{\phi}=\frac{\frac{1}{2}\dot{\phi}^{2}-V(\phi)}{\frac{1}{2}\dot{\phi}^{2}+V(\phi)}\simeq-1+\frac{2}{3}\epsilon_{{}_{V}}(\phi)\,,$ (3.45) Therefore one finds from (3.42) the following constraint on the inflationary EOS at the pivot scale $w_{\phi}\leq-0.9985\,,$ (3.46) implying that the expansion of the universe during inflation was near exponential (quasi-de Sitter like). Figure 3: This figure shows the Starobinsky potential (3.48) with CMB pivot scale $k_{*}=0.05~{}{\rm Mpc}^{-1}$ by a blue color star as well as the CMB window $k_{\rm CMB}\in\left[0.0005,0.5\right]~{}{\rm Mpc}^{-1}$ in grey color shade in the field space. From this figure, it is clear that CMB window constitutes only a tiny portion of the available field space in between $\phi_{\rm CMB}$ and end of inflation $\phi_{e}$. The CMB observations, in the context of single field slow-roll inflationary paradigm, favours asymptotically-flat potentials (featuring either one or two plateau wings) with $n_{{}_{S}}\simeq 0.965$ and $r\leq 0.036$. A typical plateau-potential is demonstrated in figure 1 and this is the standard/vanilla scenario. Given that power spectrum is almost scale-invariant with slightly red tilt, i.e $n_{{}_{S}}-1\lesssim 0$, large-scale fluctuations are more important while nothing drastic is expected to happen on smaller cosmological scales that are super-Hubble at the end of inflation. Before proceeding further to discuss small-scale inflationary fluctuations, let us make our nomenclature concrete (which is consistent with the standard nomenclature in the inflationary literature). • Quasi-de Sitter inflation corresponds to the condition $\epsilon_{H}\ll 1$. • Slow-roll inflation corresponds to $\epsilon_{H},\,\eta_{H}\ll 1$. This distinction will be important for the rest of the discussions in this paper. Under either of the aforementioned assumptions, the expression for conformal time is given by $\displaystyle\boxed{-\tau\simeq\frac{1}{aH}}\,.$ (3.47) ### 3.2 Small-scale primordial fluctuations As mentioned above, the recent CMB observations support the scenario of the inflaton field rolling slowly down an asymptotically-flat potential at the time when the observable CMB scales made their Hubble exit during inflation. However, the current CMB and LSS observations probe only about 7-8 e-folds of inflation around the Hubble exit time of the CMB pivot scale. We explicitly mention that CMB observations probe primordial fluctuations with comoving scales $k_{\rm CMB}\in\left[0.0005,0.5\right]~{}{\rm Mpc}^{-1}$ (which includes pivot scale $k_{*}=0.05~{}{\rm Mpc}^{-1}$) corresponding to multipole $l\in\left[2,2500\right]$ in the angular sky. Additionally, Lyman-$\alpha$ forest observations enforce constraints on the primordial power spectrum upto $k\simeq{\cal O}(1)~{}{\rm Mpc}^{-1}$ (see [29]). Hence a large portion of evolution during the inflationary phase that accounts roughly to about 50 e-folds of expansion, corresponding to scales smaller than those probed by CMB, remains observationally inaccessible at present. Consequently the associated dynamics of the inflaton field also remains unprobed. For example, figure 3 demonstrates that the CMB window constitutes only a tiny part of the observationally available field space between the largest scales in the sky and the smallest scale at the end of inflation for Starobinsky potential [1, 62] $\boxed{V(\phi)=V_{0}\,\left(1-e^{-\frac{2}{\sqrt{6}}\,\frac{\phi}{m_{p}}}\right)^{2}}\,.$ (3.48) Any deviation from the quasi-de Sitter expansion and/or departure from the slow-roll regime $\epsilon_{{}_{H}},\eta_{{}_{H}}\ll 1$ that might be triggered by a change in the dynamics of the inflaton field, would lead to interesting observational consequences on small scales. In particular if the inflaton potential possesses a near inflection point-like broad feature at some intermediate field values, then the scalar quantum fluctuations corresponding to scales becoming super-Hubble around the time when the inflaton rolls past such features, might receive enough amplification to facilitate the formation of Primordial Black Holes (PBHs) upon their Hubble re-entry during the radiative epoch. Figure 4: This figure schematically depicts a prototype plateau potential, plotted in solid green curve. The ‘CMB Window’ represents field values corresponding to the Hubble-exit epochs of scales $k\in\left[0.0005,0.5\right]~{}{\rm Mpc}^{-1}$ that are observable by the latest CMB missions. The potential exhibits a small-scale feature (shown in the salmon colour shading) in the form of a flat inflection point-like segment which results in ultra slow-roll (USR) inflation. After exiting the first slow-roll phase (SR-I) near the CMB window, the inflaton enters into an USR phase, during which the second slow-roll condition is violated, namely $\eta_{H}\simeq+3$. This leads to an enhancement of power spectrum at small- scales. Later, the inflaton emerges out of the USR to another slow-roll phase (SR-II) before the end of inflation. We note that, there are a plethora of possible features that would lead to deviation from standard scale-invariant power spectrum [63]. However, in this paper we will only focus on potentials with a tiny local bump/dip like feature [44] in order to illustrate the efficiency of our numerical analysis. Our code can be used to simulate a number of different types of features in the inflaton potential. PBH formation requires the enhancement of the inflationary power spectrum by roughly a factor of $10^{7}$ within less than 40 e-folds of expansion (on scales smaller than the pivot scale $k_{*}$) as illustrated in figure 5. Therefore the quantity $\Delta\ln{\epsilon_{H}}/\Delta N$, and hence also $|\eta_{H}|$, can grow to become of order unity, thereby violating the second slow-roll condition [64]. In fact the second Hubble slow-roll parameter $|\eta_{H}|$ becomes larger than unity even though $\epsilon_{H}$ itself remains much smaller than unity. As a result, equation (3.22) can no longer be trusted to compute the power spectrum and one must determine $P_{\cal R}$ by numerically integrating the Mukhanov-Sasaki equation (3.4). We proceed as follows. We first discuss the simulations of inflationary background dynamics in section 4, where we also discuss how to generate phase- space portrait $\\{\phi,\dot{\phi}\\}$ during inflation. In section 5.1, we introduce our numerical scheme for studying quantum fluctuations during slow- roll inflation. We work with convex as well as asymptotically-flat potentials. In section 5.2, we apply our numerical scheme to potentials featuring a local bump/dip like feature that facilitates the amplification of scalar power on small primordial scales. We demonstrate that slow-roll formula (3.22) underestimates both the location as well as the height of scalar power spectrum ${\cal P}_{\zeta}(k)$ for both type of aforementioned features and hence one must solve the Mukhanov-Sasaki equation (3.4) numerically to estimate the power spectrum accurately. We also demonstrate that the growth of the power spectrum obeys the steepest growth bounds discussed in [65, 66, 67, 68]. Figure 5: This figure schematically illustrates the typical amplification of inflationary primordial power spectrum at smaller length scales required for PBH formation. ## 4 Numerical analysis of inflationary background dynamics A complete analysis of the inflationary background dynamics can be obtained from the evolution of $\phi$, $\dot{\phi}$ and $H$. All of these quantities can be simulated by numerically solving equations (2.14), (2.15) and (2.16). The evolution of the scale factor follows directly from $H=\dot{a}/a$. Our system is defined by the following set of equations (as a function of cosmic time $t$) $\displaystyle H^{2}$ $\displaystyle=\frac{1}{3m_{p}^{2}}\left[\frac{1}{2}{\dot{\phi}}^{2}+V(\phi)\right]\,,$ (4.1) $\displaystyle\dot{H}$ $\displaystyle=-\frac{\dot{\phi}^{2}}{2m_{p}}\,,$ (4.2) $\displaystyle\ddot{\phi}$ $\displaystyle=-3H\dot{\phi}-V_{,\phi}(\phi)\,,$ (4.3) where the functional form of the potential $V(\phi)$ is given by the specific inflationary model. However, the rest of the algorithm is largely model- independent. We can re-write the potential as $V(\phi)=V_{0}\,f(\phi)\,.$ (4.4) In order to carry out numerical simulations, it is convenient to write down the dynamical equations in terms of dimensionless variables (which also ensures that we do not need to worry about keeping track of units). Furthermore, it is important to re-scale the time variable by a factor $S$ which can be suitably chosen according to the energy scale of the dynamics444Depending upon the potential, we usually choose the value of $S$ to be in the range $S\in[10^{-5},\,10^{-3}]$.. Our primary dimensionless variables are defined as $\displaystyle T$ $\displaystyle=\left(t\,m_{p}\right)\,S\,,$ (4.5) $\displaystyle x$ $\displaystyle=\frac{\phi}{m_{p}}\,,$ (4.6) $\displaystyle y$ $\displaystyle=\left(\frac{\dot{\phi}}{m_{p}^{2}}\right)\,\frac{1}{S}\,,$ (4.7) $\displaystyle z$ $\displaystyle=\left(\frac{H}{m_{p}}\right)\,\frac{1}{S}\,,$ (4.8) $\displaystyle A$ $\displaystyle=\left(a\,m_{p}\right)\,S\,.$ (4.9) In terms of these variables, the dynamical equations (to be simulated) take the form $\displaystyle\frac{{\rm d}x}{{\rm d}T}$ $\displaystyle=y\,,$ (4.10) $\displaystyle\frac{{\rm d}y}{{\rm d}T}$ $\displaystyle=-3\,z\,y-\frac{v_{0}}{S^{2}}\,f_{,x}(x)\,,$ (4.11) $\displaystyle\frac{{\rm d}z}{{\rm d}T}$ $\displaystyle=-\frac{1}{2}\,y^{2}\,,$ (4.12) $\displaystyle\frac{{\rm d}A}{{\rm d}T}$ $\displaystyle=A\,z\,.$ (4.13) We can also define the dimensionless potential to be $\frac{V(\phi)}{m_{p}^{4}}\equiv v_{0}\,f(x)=\frac{V_{0}}{m_{p}^{4}}\,f(x)\,.$ (4.14) We can solve the aforementioned set of equations with appropriate initial conditions. In our analysis, we use the odeint function provided in the scipy.integration package. By incorporating initial conditions $\\{x_{i},\,y_{i},\,z_{i},\,A_{i}\\}$ for the primary dynamical variables $\\{x,\,y,\,z,\,A\\}$, we simulate their time evolution during inflation. Accordingly, we determine the crucial derived/secondary (dimensionless) dynamical variables from the primary ones by555Note that the observables $A_{{}_{S}},\,A_{{}_{T}},\,n_{{}_{S}},\,n_{{}_{T}},\,{\rm and}\,r$ are related to inflationary scalar and tensor fluctuations and the expressions given here are under slow-roll approximations, i.e $\epsilon_{H},|\eta_{H}|\ll 1$, during which they can be determined purely from the dynamics of background quantities such as $H,\,\epsilon_{H},\,\eta_{H}$. Computation of inflationary power spectra when slow-roll is violated is described in section 5.2. $\displaystyle N$ $\displaystyle=\log{\frac{A}{A_{i}}}$ (4.15) $\displaystyle\epsilon_{H}$ $\displaystyle=\frac{1}{2}\,\frac{y^{2}}{z^{2}}~{},$ $\displaystyle\eta_{H}$ $\displaystyle=-\frac{1}{yz}\frac{{\rm d}y}{{\rm d}T}\,,$ (4.16) $\displaystyle A_{{}_{S}}$ $\displaystyle=\frac{1}{8\pi^{2}}\,\frac{\left(Sz\right)^{2}}{\epsilon_{H}}~{},$ $\displaystyle A_{{}_{T}}$ $\displaystyle=\frac{2}{\pi^{2}}\,\left(Sz\right)^{2}\,,$ (4.17) $\displaystyle n_{{}_{S}}$ $\displaystyle=1+2\,\eta_{H}-4\,\epsilon_{H}~{},$ $\displaystyle n_{{}_{T}}$ $\displaystyle=-2\,\epsilon_{H}\,,$ (4.18) $\displaystyle r$ $\displaystyle=16\,\epsilon_{H}\,.$ (4.19) We define $N_{T}$ to be the number of e-folds of accelerated expansion realised in between an arbitrary initial time and the end of inflation, which is marked by $\epsilon_{H}=1$. We then define the more important quantity $N_{e}=N_{T}-N$ as the number of e-folds before the end of inflation. Note that $N_{e}=0$ at the end of inflation while $N_{e}>0$ at early times. This will be our primary time variable against which we will be plotting the dynamics of different inflationary observables. In order to realise adequate amount of inflation, i.e. $N_{T}>60$, initial value of scalar field must be large enough (and is model dependent). For most large field potentials, this value is of the order $\phi_{i}\lesssim{\cal O}(10)\,m_{p}$. Composing the code involves typing down the dimensionless equations in the appropriate syntax and solving them by using an ODE solver. See our supplementary Python code 666https://github.com/bhattsiddharth/NumDynInflation for details. For a particular model of interest, we need to input the inflaton potential in the form $\frac{V(\phi)}{m_{p}^{4}}=v_{0}\,f(x)$. Since the slow- roll parameters and the duration $N_{T}$ do not strongly depend on $v_{0}$, we can initially set its value to roughly $v_{0}=10^{-10}$. We can later adjust the value of $v_{0}$ to yield the correct CMB normalised value of scalar power spectrum (3.37) at the pivot scale $N=N_{*}$. One can proceed in the following step-by-step algorithm. Figure 6: Time evolution of the number of e-folds (scale factor in the logarithm scale) of expansion of the universe is shown for Starobinsky potential (3.48). For the most part of inflation, the expansion is almost exponential (quasi-de Sitter) i.e $a\sim e^{Ht}$, leading to a rapid growth in the number of e-folds within a small amount of time. While after the end of inflation, the expansion is decelerated, leading to a much slower growth in the scale factor. 1. 1. After setting the parameters of the potential and defining the function $f(x)$, we need to incorporate initial conditions for the four primary variables $\\{x,\,y,\,z,\,A\\}$. We enter appropriate initial conditions $x_{i},\,y_{i}$ and $A_{i}$ in the following way. $A_{i}$ can be set arbitrarily in a spatially flat universe, however depending upon the energy scale of inflation, one can provide an appropriate value. We suggest a typical $A_{i}=1\times 10^{-3}$, although its precise value does not affect the dynamics. In regard to the initial value of $x$, we need to ensure that $x_{i}$ is large enough (or small enough if we are working with symmetry breaking hilltop type potentials) to yield adequate amount of inflation, i.e $N_{T}\geq 70$. As mentioned before, the typical value for large field models is $x_{i}\lesssim{\cal O}(10)$. Since we will be mostly working with potentials that exhibit slow-roll behaviour at initial times, and given that slow-roll trajectory is an attractor in relatively large field models [20], we can safely set $y_{i}=0$, as long as $x_{i}$ is large enough. One can also incorporate slow-roll initial conditions from the beginning, namely $y_{i}=-\frac{v_{0}}{3}\,\frac{f_{,x}}{Sz}$, as is usually done in practice777Note that for phase-space analysis, we need to incorporate arbitrary values of $x_{i},\,y_{i}$ (consistent with fixed initial $z_{i}$) which may be away from the slow-roll trajectory as discuss below.. Finally, the initial value of $z$ can be incorporated in terms of $x_{i},\,y_{i}$ using the dimensionless Friedmann equation $\displaystyle z_{i}=\sqrt{\frac{1}{6}\,y_{i}^{2}+\frac{1}{3}\,v_{0}\,f(x_{i})\,\frac{1}{S^{2}}}\,.$ (4.20) 2. 2. We then proceed to solve the system of equations by taking adequately small time steps $T$ in the appropriate range $T\in[T_{i}=0,\,T_{f}]$. We then plot $N$ vs $T$ as given in figure 6 for Starobinsky potential. Typically, $N$ grows linearly with $T$ during near exponential inflation and a substantial decrease in the rate of growth of $N$ indicates the end of inflation. 3. 3. In order to concretely determine the value of $N_{T}$, we plot $\epsilon_{H}$ vs $N$, and note the value of $N$ after which $\epsilon_{H}\geq 1$. By definition, initially $N=0$. If $N_{T}<70$, then we repeat this step by increasing the value of $x_{i}$, until we get $N_{T}\geq 70$. (Alternatively, if inflation has not ended888Note that if one simulates the cosmological equations in terms of number of e-folds $N$, rather than cosmic time $t$, this step can usually be avoided by simulating the system from $N=0$ to $N=70$. However, one has to adjust the value of $x_{i}$ in order to get enough inflation., i.e $\epsilon_{H}<1$, at the end of our simulation, then either one can increase the value of $T_{f}$ or decrease $x_{i}$. We suggest the latter.) We can then define the number of e-folds before the end of inflation to be $N_{e}=N_{T}-N$. Figure 7: This figure describes the evolution of inflaton field $\phi$, and its speed $\dot{\phi}$ in the left panel, while the Hubble parameter $H$ in the right panel as a function of the number of e-folds before the end of inflation $N_{e}$ for Starobinsky potential (3.48). Note that during slow-roll inflation, $\dot{\phi}$ and $H$ are nearly constant, while $\phi$ changes quite slowly. However, $\phi$ and $H$ begin to change rapidly towards the end of inflation. After inflation ends, $\phi$ and $\dot{\phi}$ start oscillating around the minimum of the potential (which is not shown in this figure). Figure 8: Evolution of the slow-roll parameters $\epsilon_{H}$ and $\eta_{H}$ is shown as a function of the number of e-folds before the end of inflation $N_{e}$ for Starobinsky potential (3.48). From this plot, it is easy to notice that at early times when $N_{e}\gg 1$, the slow-roll conditions are satisfied i.e $\epsilon_{H},\,|\eta_{H}|\ll 1$. However, the slow-roll conditions are violated towards the end of inflation (marked by $N_{e}=0$ and $\epsilon_{H}=1$). 4. 4. The pivot scale can then be fixed to an appropriate value, for example $N_{e}=60$, as used in this work. Figure 7 describes the evolution of $\phi,\,\dot{\phi},\,{\rm and}\,H$, while figure 8 illustrates the dynamics of slow-roll parameters $\epsilon_{H},\,|\eta_{H}|$ for Starobinsky potential, as determined from our code. As mentioned before, we usually plot the dynamics of inflation as a function of $N_{e}$. 5. 5. In order to accurately fix the value of $v_{0}$, we need to impose $A_{S}=2.1\times 10^{-9}$ at the pivot scale $N_{e}=60$. If $A_{S}$ is lower than expected for the given value of $v_{0}$, then we increase the value of $v_{0}$ or vice versa until we arrive at the correct value of $A_{S}$, and fix the corresponding value of $v_{0}$. Following the aforementioned algorithm, we can easily simulate the inflationary background dynamics and investigate the evolution of relevant quantities of our interest. Before going forward, we would like to stress that many of the aforementioned steps (in the present version of our code) are rather meant to be carried out manually by the user. While we are already developing an automated version of this code (which will be presented in the revised version of our paper), we believe that the present version will help the user to understand the inflationary dynamics much better. ### 4.1 Phase-space analysis Phase-space analysis of inflationary dynamics is usually carried out to determine the set of initial conditions that results in adequate amount of inflation, and hence it is important to access the generality of initial conditions for inflation [69, 50, 20]. For a spatially flat background, the phase-space portrait consists of trajectories of $\\{\phi,\,\dot{\phi}\\}$ for different initial conditions, with fixed $H_{i}$. The standard algorithm to generate such a plot is the following. 1. 1. The initial energy scale of inflation is kept constant by fixing the value of initial Hubble parameter in the phase-space portrait simulations. A typical value often used is $H_{i}\leq m_{p}$ (see [20] for detail). Hence, the user is expected to incorporate an appropriate value of $z_{i}$. 2. 2. One can then input a suitable value of $x_{i}$ and determine the value of $y_{i}$ for a given potential function $f(x)$ from the dimensionless Friedmann equation (4.20) as $y_{i}=\pm\sqrt{6}\,\sqrt{z_{i}^{2}-\frac{1}{3}\,v_{0}\,f(x_{i})\,\frac{1}{S^{2}}}~{}.$ (4.21) 3. 3. With these initial conditions, one can then simulate the system of dimensionless differential equations for $\\{x,\,y,\,z,\,A\\}$ from $T_{i}=0$ till an appropriate $T_{f}$. One can then repeat the same step by incorporating a number of different values of of $x_{i}$ in order to generate the phase-space portrait for the given potential. We provide the GitHub link to our phase-space portrait framework here 999https://github.com/bhattsiddharth/NumDynInflation/blob/main/inf_dyn_phase.py. The phase-space portraits for Starobinsky potential (3.48), and quadratic potential $V(\phi)\propto\phi^{2}$ are illustrated in the left and right panels of figure 9 respectively. Figure 9: The phase-space portrait $\\{\phi,\,\dot{\phi}\\}$ of the inflaton field has been illustrated for Starobinsky potential (3.48) in the left panel, and for quadratic potential $V(\phi)=\frac{1}{2}\,m^{2}\,\phi^{2}$ in the right panel corresponding to different initial conditions $\\{\phi_{i},\,\dot{\phi_{i}}\\}$ (plotted in solid black colour) with a fixed initial scale $H_{i}$. The figure demonstrates that trajectories commencing from a large class of initial field values (including those with large initial velocities $\dot{\phi_{i}}$) quickly converge towards the slow-roll attractor separatrix $\dot{\phi}=-V_{,\phi}/3H\simeq\mathrm{const.}$ (plotted in green colour) as can be seen from the rapid decline in the inflaton speed until they meet the green colour curve. After the end of inflation, the inflaton begins to oscillate around the minimum of the potential. In order to determine the degree of generality of inflation, we need to define a measure for the distribution of $\\{\phi_{i},\,\dot{\phi_{i}}\\}$. The correct choice for the measure might depend on the quantum theory of gravity. However, a uniform measure is usually considered in the literature. Interested readers are referred to [69, 20] for further detail. ### 4.2 Quantum fluctuations under the slow-roll approximation In section 3, we described the expressions for a number of inflationary observables such as $A_{{}_{S}},\,A_{{}_{T}},\,n_{{}_{S}},\,n_{{}_{T}},\,{\rm and}\,r$ associated with the scalar and tensor power spectra which can be determined purely from the dynamics of background quantities such as $H,\,\epsilon_{H},\,\eta_{H}$ under the slow-roll approximation. Hence they can be conveniently determined from our background dynamics code as discussed earlier in section 4. For example, the scalar and tensor power spectra for Starobinsky inflation have been plotted in figure 10 as a function $N_{e}$. Similarly, one can plot the spectral indices101010In the standard literature, one usually plots $r$ vs $n_{{}_{S}}$ for a given inflaton potential for a range of possible values of $N_{*}\in[50,60]$ which can also be done easily using our code. $n_{{}_{S}}-1$ and $n_{{}_{T}}$ and determine their values at the pivot scale $N_{*}$. The spectral indices for Starobinsky potential have been plotted in figure 11. Figure 10: The power spectra of scalar and tensor quantum fluctuations (computed using the slow-roll formulae (3.10) and (3.25) respectively) are shown for comoving modes exiting the Hubble radius at different number of e-folds $N_{e}$ before the end of inflation for Starobinsky potential (3.48). The CMB window (shown in grey shaded region) corresponds to comoving modes in the range $k_{\rm CMB}\in[0.0005,0.5]~{}{\rm Mpc}^{-1}$ that are being probed by the current CMB missions. Fluctuations over larger scales (shown in red shaded region) are outside the observable universe at present and those over smaller scales remain to be (potentially) probed by a plethora of upcoming missions, from GW observatories to PBHs. Figure 11: The scalar and tensor spectral indices $n_{{}_{S}}$ and $n_{{}_{T}}$ are shown as a function of $N_{e}$ for Starobinsky potential (3.48) as determined by their slow-roll approximated formulae (3.23) and (3.26) respectively. Around the pivot scale, they take the approximate values $n_{{}_{S}}\simeq 0.967\text{ and }n_{{}_{T}}\simeq-0.0004$. Note that we have plotted $n_{{}_{S}}-1$ (instead of $n_{{}_{S}}$) since it is the correct scalar spectral index. The tensor-to-scalar ratio is given by $r\simeq-8\,n_{{}_{T}}$. ## 5 Numerical analysis for quantum fluctuations during inflation In the previous section we used slow-roll approximated formulae to study the spectra of inflationary fluctuations in terms of background quantities such as $H,\,\epsilon_{H},\,\eta_{H}$. Hence, we only had to simulate the background dynamics for a given potential in order to plot the relevant inflationary observables. However, if we want to analyze the behaviour of quantum fluctuations more accurately, especially in situations where one or both the slow-roll conditions (2.20) are violated, we need to numerically solve the Mukhanov-Sasaki equation (3.4) corresponding to each comoving scale $k$. For this purpose, we first rewrite the Mukhanov-Sasaki equation (3.4) in cosmic time as $\frac{{\rm d}^{2}v_{k}}{{\rm d}t^{2}}+H\frac{{\rm d}v_{k}}{{\rm d}t}+\left[\frac{k^{2}}{a^{2}}-\frac{1}{a^{2}}\frac{z^{\prime\prime}}{z}\right]v_{k}=0\,.$ (5.1) Note that here $z$ is not the dimensionless Hubble parameter used in our numerical code, rather it is the variable $z=am_{p}\sqrt{2\epsilon_{H}}$ in the Mukhanov-Sasaki equation (3.4). The effective mass term $z^{\prime\prime}/z$ in (3.6) can be re-written as $\frac{z^{\prime\prime}}{z}=a^{2}\left[\frac{5}{2}\frac{\dot{\phi}^{2}}{m_{p}^{2}}+2\frac{\dot{\phi}\ddot{\phi}}{Hm_{p}^{2}}+2H^{2}+\frac{1}{2}\frac{\dot{\phi}^{4}}{H^{2}m_{p}^{4}}-V_{,\phi\phi}(\phi)\right]\,.$ (5.2) Since $v_{k}$ is a complex valued function, it is convenient to split it into its real and imaginary parts to study their evolution separately for the numerical analysis. While both will follow the same evolution equation, they will be supplied with different initial conditions in the form of the real and imaginary parts of the Bunch-Davies vacuum (3.8). Writing the Mukhanov-Sasaki equation for scalar fluctuations in terms of dimensionless variables, we obtain $\boxed{\frac{{\rm d}^{2}v_{k}}{{\rm d}T^{2}}+z\,\frac{{\rm d}v_{k}}{{\rm d}T}+\left[\frac{k^{2}}{A^{2}}-\frac{5}{2}\,y^{2}+2\,\frac{y}{z}\left(3\,z\,y+\frac{v_{0}}{S^{2}}\,f_{,x}\right)-2\,z^{2}-\frac{1}{2}\,\frac{y^{4}}{z^{2}}+\frac{v_{0}}{S^{2}}\,f_{,xx}\right]v_{k}=0}\,.$ (5.3) Figure 12: Evolution of scalar power $\frac{k^{3}}{2\pi^{2}}|\zeta_{k}|^{2}$ is plotted by numerically solving the Mukhanov-Sasaki equation (5.3) for a mode exiting the Hubble radius at about 60 e-folds before the end of inflation (for Starobinsky potential). At early times when the mode is sub-Hubble, i.e. $k\gg aH$, the power decreases as ${\cal P}_{\zeta}\sim(aH)^{-2}$ as expected. After the Hubble-exit, the power freezes to a constant in the super-Hubble regime when $k\ll aH$. We note down its value after the mode-freezing as the super-Hubble scale power corresponding to that mode. Repeating the procedure for a range of scales $k$ yields us the power spectrum of scalar fluctuations. The same numerical analysis can be carried out for tensor fluctuations. Our primary goal in this section is to numerically solve equation (5.3) for the Fourier modes $v_{k}$ corresponding to each comoving scale $k$ and plot the frozen value of the scalar power spectrum of $\zeta_{k}$ given by (3.9) after the mode becomes super-Hubble. We can conveniently relate a comoving scale $k$ to its Hubble-exit epoch by $k=aH$. Since we are only interested in the super-Hubble power spectra, we only need to simulate the system to evolve $v_{k}$ for a small duration of time around the Hubble-exit of scale $k$ (which should be sufficiently early enough to impose Bunch-Davies initial conditions and sufficiently late enough for the mode to be frozen outside the Hubble radius). In the following, we discuss the algorithm to solve the Mukhanov-Sasaki equation (5.3) and determine the scalar power spectrum (3.9) numerically. We also discuss how to solve the corresponding equation (3.18) for the tensor power spectrum (at linear order in perturbation theory)111111Second-order tensor fluctuations which are induced by first-order scalar fluctuations will be discussed in the revised version of our manuscript.. The dimensionless Mukhanov-Sasaki equation for tensor fluctuations is given by $\boxed{\frac{{\rm d}^{2}h_{k}}{{\rm d}T^{2}}+z\,\frac{{\rm d}h_{k}}{{\rm d}T}+\left[\frac{k^{2}}{A^{2}}+\frac{1}{2}\,y^{2}-2\,z^{2}\right]h_{k}=0}\,.$ (5.4) We explicitly write down the Mukhanov-Sasaki equations for scalar and tensor fluctuations in terms of dimensionless variables (as used in our code) in the following way $\displaystyle v_{k,T}$ $\displaystyle=\frac{{\rm d}v_{k}}{{\rm d}T}\,,$ (5.5) $\displaystyle\frac{{\rm d}v_{k,T}}{{\rm d}T}$ $\displaystyle=-z\,v_{k,T}-\left[\frac{k^{2}}{A^{2}}-\frac{5}{2}\,y^{2}+2\,\frac{y}{z}\left(3\,z\,y+\frac{v_{0}}{S^{2}}\,f_{,x}\right)-2\,z^{2}-\frac{1}{2}\,\frac{y^{4}}{z^{2}}+\frac{v_{0}}{S^{2}}\,f_{,xx}\right]v_{k}\,;$ (5.6) $\displaystyle h_{k,T}$ $\displaystyle=\frac{{\rm d}h_{k}}{{\rm d}T}\,,$ (5.7) $\displaystyle\frac{{\rm d}h_{k,T}}{{\rm d}T}$ $\displaystyle=-z\,h_{k,T}-\left(\frac{k^{2}}{A^{2}}+\frac{1}{2}\,y^{2}-2\,z^{2}\right)\,h_{k}\,.$ (5.8) In our numerical set up, we split $v_{k}$ and $h_{k}$ into their real and imaginary parts and simulate them separately with appropriate Bunch-Davies initial conditions. We begin with a discussion of numerical simulations of the Mukhanov-Sasaki equations (5.3) and (5.4) for a purely slow-roll potential (which we choose to be the Starobinsky potential (3.48) as usual), before moving forward to discuss the same for a potential with a slow-roll violating feature. This latter case is of the primary focus of our paper. In particular, we will illustrate our numerical scheme for the case of a base slow-roll potential possessing a tiny local bump feature, which was proposed in [44] in the context of PBH formation. Figure 13: The super-Hubble power spectra of scalar fluctuations (in green colour) and tensor fluctuations (in red colour) are plotted for modes exiting the Hubble radius at different number of e-folds $N_{e}$ before the end of inflation for Starobinsky potential (3.48). The solid curves represent power spectra computed under the slow-roll approximation (3.10), while the dotted curves represent the power computed by numerically solving the Mukhanov-Sasaki equation (5.3). We conclude that for Starobinsky model, since slow-roll conditions $\epsilon_{H},\,|\eta_{H}|\ll 1$ are easily satisfied for most part of inflation, the power spectra computed under the slow-roll approximation match quite well with their numerically determined counterparts. ### 5.1 Numerical analysis for slow-roll potentials 1. 1. As the first step, we numerically solve the background dynamics for a given potential, determine the values of all relevant parameters of the potential and the evolution of relevant primary dynamical variables $\\{x,\,y,\,z,\,A\\}$ as well as the derived quantities such as $\\{N_{e},\,\epsilon_{H},\,\eta_{H}\\}$ (as discussed in section 4). 2. 2. We then proceed to identify different comoving scales $k$. This can be done by determining their Hubble-exit epochs in the following way. For example, we plot $aH$ (in log scale) against $N_{e}$ and identify the value of $aH$ at $N_{e}=N_{*}$ to be the CMB pivot scale $k_{p}$. As mentioned before, we take $N_{*}=60$ in all our analysis. Similarly, we associate a corresponding value of $N_{e}$ to each comoving scale $k$ by the value of $aH$ at its Hubble-exit epoch. This step ensures that we have a one-to-one correspondence between $k$ and $N_{e}$ in our analysis and we can use them interchangeably. 3. 3. We intend to impose Bunch-Davies initial conditions for a given mode $v_{k}$ at an epoch when it is sub-Hubble. As it tuns out, for most potentials, the Bunch-Davies initial conditions can be safely imposed as long as $k\geq 100\,aH$. Hence, rather than simulating the Mukhanov-Sasaki equation for each mode (making Hubble-exit at the corresponding value of $N_{e}$) all through the inflationary history (starting from $\phi_{i}>\phi_{*}$), we actually impose the initial conditions from the background solutions for $\\{x,\,y,\,z,\,A\\}$ at around $5$ e-folds before the Hubble-exit of that mode. This step greatly reduces the running-time of the code. We then incorporate the initial value of scale factor $A_{i}$ at the same epoch, namely $A_{i}\exp{(N_{T}-N_{e}-5)}$ and do the same for the initial values of the field $x_{i}$, and its derivative $y_{i}$. The initial conditions for the mode functions $v_{k}$ and their derivatives $\dot{v_{k}}$ can then be safely taken to be of Bunch-Davies type. 4. 4. We solve the set of cosmological equations with these initial conditions for a period of time $T=T_{i}\to T=T_{f}$ such that the mode becomes super-Hubble and its power ($k^{3}|\zeta_{k}|^{2}/2\pi^{2}$) is frozen to a constant value (see figure 12), which is typically within $5$ e-folds after Hubble-exit in the kind of models we are interested in. We note down this frozen value as the value of the power spectrum of that mode. While we have been discussing about scalar fluctuations mostly, the same can be done for tensor fluctuations which we have incorporated in our code. 5. 5. We then select another mode that leaves the Hubble radius at some epoch $N_{e}$ and repeat the procedure until we have collected the frozen super- Hubble power spectra of a range of scales that we are interested in (see figure 13). Figure 14: This is a schematic plot of an asymptotically-flat inflationary potential with a tiny bump/dip feature (5.9) near intermediate field values $\phi\simeq\phi_{\rm PBH}$ that leads to an enhancement of the scalar power spectrum. The full potential asymptotes to the base (slow-roll) potential near the CMB window $\phi\simeq\phi_{*}$, thus satisfying observational constraints on large cosmological scales. Slow-roll is violated around the feature, whose position $\phi_{\rm PBH}$ dictates the range of moving scales $k$ that receive amplification of power (which accordingly determines the mass and abundance of formed PBHs). Note that the feature has been greatly exaggerated for illustration purpose. In most realistic models, both the height and the width of the feature are too small to be seen (without zooming-in considerably). From the above numerical analysis of featureless vanilla potentials which exhibit slow-roll dynamics until close to the end of inflation, we observe that the power spectra of scalar and tensor fluctuations are nearly scale- invariant (with small red-tilt) and their behaviour (as obtained from numerically solving the Mukhanov-Sasaki equation) matches quite well with the analytical predictions under the slow-roll approximations (see figure 13). However, for potentials exhibiting a small-scale feature at intermediate field values $\phi<\phi_{*}$, there might exist a short period of slow-roll violating phase before the end of inflation during which slow-roll approximations break down. In particular, as we will see, while the first slow-roll parameter remains small $\epsilon_{H}\ll 1$, the second slow-roll parameter might become $\eta_{H}\sim{\cal O}(1)$. Hence, a numerical analysis of the Mukhanov-Sasaki equation is desired in order to determine the scalar power spectrum more accurately. This will be the main focus of discussion in the next subsection. Figure 15: Evolution of the field value $\phi$ for the KKLT potential with a tiny bump (5.10) is shown in solid green curve as a function of number of e-folds $N_{e}$ before the end of inflation. We note that at CMB scales, $\phi$ is much smaller than the corresponding value in the base model (KKLT) (shown in dashed black curve). At intermediate scales when the inflaton evolves across the bump feature in the potential, we gain a lot of extra e-folds of expansion $\Delta N_{e}\simeq 15$ with little change in the field value. After crossing the feature, evolution of $\phi$ mimics its corresponding value in the base model. Figure 16: Evolution of the slow-roll parameters $\epsilon_{H}$ and $\eta_{H}$ is shown in solid green and solid red curves respectively for the KKLT potential with a tiny bump (5.10). Both $\epsilon_{H}$ and $\eta_{H}$ are close to their corresponding values for the base KKLT potential (shown in dashed curves) at early times near the CMB window. At $N_{e}\simeq 30$, the value of $\epsilon_{H}$ starts decreasing rapidly leading to an increase in $\eta_{H}$ from $|\eta_{H}|\ll 1$ to a higher and positive value $\eta_{H}\simeq+3.3$ (almost USR phase). Thereafter, the inflaton enters a phase of constant-roll inflation (where $\eta_{H}\simeq-0.37$) before returning to the final slow-roll phase. ### 5.2 Numerical analysis for potentials with a local bump/dip feature In order to facilitate PBH formation, we need a large amplification of scalar power spectrum at smaller scales during inflation. This can be achieved by introducing a small-scale feature in the potential which leads to a transient period of slow-roll violating phase (including a short almost-USR phase). Adequate amplification in the super-Hubble scalar power spectrum results in a large density contrast in the post-inflationary universe (upon the Hubble- entry of the corresponding modes) which in turn can collapse to form PBHs. Usually, such a PBH feature in the potential leads to an increase in the value of the second slow-roll parameter $\eta_{H}$ from near-zero to a positive value of $\eta_{H}\sim{\cal O}(1)$. A number of models with different types of features have been proposed in the recent literature (as mentioned before) that facilitate the amplification of scalar power spectrum at small scales. The most common amongst them is an inflection point-like feature. However, we choose the model proposed in [44] in which the base inflaton potential $V_{b}(\phi)$ possesses a tiny local bump or dip $\pm\varepsilon(\phi,\phi_{0})$ at an intermediate field value $\phi_{0}$ of the form $V(\phi)=V_{b}(\phi)\,\left[1\pm\varepsilon(\phi,\phi_{0})\right]\,,$ (5.9) where we assume $V_{b}(\phi)$ to be a symmetric or an anti-symmetric asymptotically-flat potential in order to satisfy CMB constraints at large cosmological scales. Such a potential has been schematically illustrated in figure 14. To be specific, in this paper we choose the base potential to be the D-brane KKLT potential [70, 71, 72, 73] with a tiny Gaussian bump of the form [44] $\boxed{V(\phi)=V_{0}\,\frac{\phi^{2}}{m^{2}+\phi^{2}}\,\left[1+A\,\exp{\left({-\frac{1}{2}\,\frac{(\phi-\phi_{0})^{2}}{\sigma^{2}}}\right)}\right]}~{},$ (5.10) where $m$ is a mass scale in the KKLT model, while $A$ and $\sigma$ represent the height and the width of the tiny bump respectively. We use this particular model to demonstrate our numerical framework because of its simplicity and efficiency. However, one can choose any model of their interest. Values of all the parameters appearing in (5.10), which we use in our numerical analysis, have been explicitly shown in figure 17. The height and the width of the (bump/dip) feature required to facilitate adequate amount of power amplification are quite small, and hence the feature is tiny and local (in contrast to inflection point-like features). This ensures that the feature does not significantly affect the CMB observables. However, since we gain a lot of extra e-folds of expansion $\Delta N_{e}\simeq 15$ with little change in the field value (as shown in figure 15) when the inflaton crosses the feature, the CMB pivot scale gets shifted towards smaller values as compared to the same for the base potential. Figure 17: The super-Hubble power spectra of scalar fluctuations are plotted for modes exiting the Hubble radius at different number of e-folds $N_{e}$ before the end of inflation for KKLT potential with a tiny bump (5.10). The solid green curve represents the power spectrum computed under the slow-roll approximation (3.10), while the dotted red curve represents the power computed by numerically solving the Mukhanov-Sasaki equation (5.3). Since the second slow-roll condition is violated due to the presence of the bump (leading to $\eta_{H}>1$), the slow-roll approximation underestimates the value of power spectrum near the peak (as well as the position of the peak) for modes exiting the Hubble radius near the epoch when the inflaton crosses the local maximum around the bump feature. This leads to an incorrect estimation of the mass fraction as well as the central mass of the PBHs formed in this scenario. Since slow-roll is violated in these models (as shown in figure 16), we need to solve the Mukhanov-Sasaki equation numerically in order to accurately compute the scalar power spectrum. This has been explicitly demonstrated in figure 17. Note that the inflationary dynamics in such models contains a number of phases that include an early slow-roll phase SR-I near the CMB window, a transition T-I from the early SR-I to the subsequent almost ultra slow-roll phase USR and a transition T-II back to the next slow-roll phase SR- II after passing through an intermediate constant-roll phase CR (shown in figure 16). Figure 18: Evolution of the effective mass term in the Mukhanov-Sasaki equation is shown here in solid green curve for the KKLT potential with a tiny bump (5.10). Important transient phases of the scalar field dynamics are also highlighted following the behaviour of the second slow-roll parameter $\eta_{H}$ (plotted in dashed blue curve). There is a sharp dip in the effective mass term when the field transitions from the first slow-roll phase (SR-I) to an almost ultra slow-roll phase (USR). The inflaton later makes a transition to a phase of constant-roll inflation with $\eta_{H}\simeq-0.37$, before reaching a final slow-roll phase until the end of inflation. The effective mass term $z^{\prime\prime}/z$ in the Mukhanov-Sasaki equation (3.4), which primarily governs the dynamics of scalar fluctuations, has been shown in figure 18, and the resultant Hubble-exit behaviour of different modes is described in figure 19. As the inflaton approaches the PBH feature, $\eta_{H}$ starts to increase rapidly. This is accompanied by an initial sharp dip in the effective mass term $z^{\prime\prime}/z$, which then increases to a higher plateau as $\eta_{H}$ approaches its maximum in the USR type phase. The modes that leave the Hubble radius slightly before the transition already start receiving power amplification on super-Hubble scales as shown by the orange color curve in figure 19. It is worth noting that the power spectrum exhibits a dip which corresponds to very narrow range of scales $k\simeq k_{\rm dip}$ that leave the Hubble radius a few e-folds before the USR phase (shown by the blue color curve in figure 19). The maximum rate of growth observed in this model is consistent with the steepest growth bound discussed in [65]. Near the USR phase when $\eta_{H}\gtrsim 3$, $z^{\prime\prime}/z$ saturates to a constant value. Modes leaving the Hubble radius around this USR epoch receive maximal amplification in their super-Hubble power spectrum. Figure 19: This figure demonstrates the horizon exit behaviour of different modes i.e the evolution of $\sqrt{\cal{P}_{\zeta}}$ for different modes as they cross the Hubble radius in KKLT model with a Gaussian bump (5.10). The dot on each curve corresponds to its Hubble-exit epoch. The sharp dip in the power spectrum (figure 17) corresponds to the mode $k_{\mathrm{dip}}$ (plotted in blue color) that exits the Hubble radius a few e-folds before the commencement of the USR phase. The mode $k_{\mathrm{PBH}}$ which exits the Hubble radius during the USR phase receives a maximal amplification of power (plotted in green color). As the field crosses the maximum of the bump feature, $\eta_{H}$ decreases to a constant negative value. It stays in this constant-roll phase until the inflaton meets the base potential eventually and approaches the final slow- roll phase in its dynamics before the end of inflation. The dynamics of scalar fluctuations in the aforementioned phases are quite rich, and interesting. However, since the main aim of this paper is to illustrate how to use our numerical code with an example, we do not discuss these phases and their impact on the power spectrum (some of which have been explicitly shown in figures 18, 19, 20), and refer the interested readers to [63] for more detail. Figure 20: The super-Hubble power spectrum of scalar fluctuations obtained by numerically solving the Mukhanov-Sasaki equation (5.3) is plotted here for the KKLT potential with a tiny bump (5.10). At large scales (around the CMB window), the power spectrum matches that of the base KKLT potential. The power spectrum (after exhibiting a sharp dip) receives a large amplification at intermediate scales $k\sim k_{\rm PBH}$ around the USR phase. The maximum rate of growth observed in this model is consistent with the steepest growth bound discussed in [65]. After reaching the peak, the power then decreases at a steady rate during the constant-roll phase before finally asymptoting towards its base slow-roll value towards the end of inflation. ## 6 Future extension of our numerical framework In preceding sections, we described the relevant cosmological equations governing the inflationary dynamics in terms of dimensionless variables. We also introduced our numerical code (written in terms of cosmic time) that can easily simulate the inflationary dynamics both at the background level in section 4 and at linear order in perturbation theory in section 5. However, with minimal to moderate extension, our numerical code can be used to simulate a number of different scenarios associated with scalar field dynamics both during inflation as well as in the post-inflationary universe. We have already started working on some of these aspects which will appear in the revised version of our manuscript. In the following we discuss some of the important future extensions of our code that we intend to include in our revised version. 1. 1. As stressed in section 4, the present version of our numerical code, although quite fast and neat, contains segments that require the user to carry out a number of tasks manually. While we believe that the present version will definitely help a user (who is relatively new to the field) to understand the inflationary dynamics much better, we are already developing an automated version of this code that is much more compact and requires substantially less manual involvement of the user. We also plan to make the code even faster. We will present the updated version of our code in the revised version of our paper and refer to it in the same GitHub link 121212https://github.com/bhattsiddharth/NumDynInflation. 2. 2. In section 5, we briefly discussed how to solve the evolution equation for the tensor fluctuations at linear order in perturbation theory. However, first- order scalar fluctuations induce tensor fluctuations at second order in perturbation theory which might be significant when slow-roll is violated, especially in the scenario where scalar power spectrum is largely amplified in order to source PBH formation. We will incorporate the computation of such scalar-induced Gravitational Waves in the updated version of our code. 3. 3. In our analysis, we used the Mukhanov-Sasaki variable $v_{k}$ with its corresponding evolution equation (5.1) in order to simulate scalar fluctuations at linear order. Our code runs quite quickly and generates the power spectrum without requiring a large computational time. However, one can study the scalar power spectrum by using other variables proposed in the literature. We are particularly interested in two such variables. The first one is the curvature perturbation $\zeta_{k}$ itself using the corresponding evolution equation (3.11). For example, it was claimed in [74] that numerical simulations are more stable in terms of $\zeta_{k}$ since it is explicitly frozen well outside the Hubble radius. Similarly, authors of [75] have suggested a change in the variable $v_{k}$ in the form $g_{k}\equiv\frac{v_{k}}{z}\,e^{ik\tau}$ that is supposed to make the simulations much more stable since it removes the early time oscillations131313We are thankful to Christian Byrnes for bringing this to our attention.. In the revised version of our work, we plan on carrying out a comprehensive numerical analysis to compare both stability as well speed of the numerical simulations using all three of the aforementioned variables $v_{k}$, $\zeta_{k}$, and $g_{k}$. 4. 4. It is easy to extend our numerical analysis to incorporate the dynamics of more than one scalar fields during inflation, at least in the background level. In the future, we are going to provide an updated numerical code to study both the background dynamics as well as quantum fluctuations in two- field inflationary dynamics (where the second field might also source inflation, or act as a spectator field). 5. 5. Additionally, our code can be extended to simulate the post-inflationary dynamics of the inflaton field as well as to study parametric resonance by simulating the evolution of different Fourier modes of the inflaton fluctuations. During the post-inflationary oscillations, it is usually advisable to make a change in the Mukhanov-Sasaki variable of the form ${\tilde{v}_{k}}=a^{1/2}\,v_{k}$ as suggested in [76, 77]. The code can also be extended to study the dynamics of scalar field dark matter and quintessence by suitably redefining the dimensionless variables as per the energy scale of the dynamics. ## 7 Discussion In the present version of our manuscript, we introduced our numerical approach to simulate the cosmological equations in order to study the inflationary dynamics at the level of both background as well as linear-order in perturbation theory. We provided the link to our open-source GitHub repository where we have supplied a Python-based simple numerical code to simulate inflationary dynamics in terms of cosmic time $t$. We explicitly demonstrated how to use the code to study the inflationary background dynamics in section 4 that includes plotting the phase-space portrait of inflation as well as to characterise quantum fluctuations during inflation using the simulations of the background dynamics. Section 5 was dedicated to study quantum fluctuations during inflation (without using slow-roll approximated expressions) by numerically solving the mode function equations of scalar and tensor fluctuations. For a featureless slow-roll inflaton potential, the difference between the results obtained numerically and those obtained under slow-roll approximations were negligible until close to the end of inflation, as expected. We used the Starobinsky potential (3.48) as an example to illustrate our analysis. Our primary focus was the numerical evaluation of the scalar power spectrum ${\cal P}_{\zeta}(k)$ for potentials that exhibit a slow-roll violating feature. In particular, we used the example of an asymptotically-flat base inflationary potential that possesses a tiny local bump feature (5.9) to illustrate our numerical scheme in section 5.2. By suitably choosing the parameters of the potential (5.10), one can achieve a large enough amplification of the scalar power spectrum in order to facilitate the formation of PBHs in the post- inflationary epoch. In our numerical analysis for the case of potentials with a slow-roll violating feature, we explicitly chose the parameters in order to amplify the small-scale scalar power by a factor of $\sim 10^{7}$ with respect to the corresponding power at large cosmological scales, as is usually assumed in the literature to facilitate the formation of of PBHs in the subsequent radiation dominated epoch. However, it is important to stress that a significant growth in the scalar power spectrum at small-scales might engender the dynamics to enter into non-perturbative regime. For example, a careful computation of loop corrections to the two-point scalar fluctuations demonstrates [78, 79] that contribution from $1$-loop effects becomes of the same order as the tree-level computation (which we carried out in this work) if ${\cal P}_{\zeta}(k)\sim 10^{-2}$ indicating a breakdown of the perturbative analysis. Moreover, the mechanism of PBH formation in the context of single field models of inflation involves additional intricacies that demand for a non- perturbative analysis of primordial fluctuations. Firstly, a sharp drop in the classical drift speed of the inflaton due to the presence of the PBH-producing feature often incites the system to enter into a phase where stochastic quantum diffusion effects become non-negligible, at times even significant. More importantly, since PBHs form from rare extreme peaks, and hence are determined by the tail of the probability distribution function (PDF) $P[\zeta]$ of the primordial fluctuations, perturbative computations based on only power-spectrum lead to an inaccurate estimation of the PBH mass fraction. Consequently, determination of the full primordial PDF becomes crucial, which can be computed non-perturbatively using the stochastic inflation framework [80, 81, 82, 83, 84, 85, 86, 88, 87, 89, 90, 91, 92, 94, 95, 93, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105], which usually predicts a non-Gaussian exponential tail [86, 90]. Tail of the primordial PDF can also be computed by using semi-classical techniques discussed in [106]. Determining the tail of the primordial PDF is an important and active topic of research [90, 106, 107, 108, 109, 110, 111] at present, which is beyond the scope of our perturbative analysis presented here. Before concluding, let us mention that we also stressed upon various important future extensions of our numerical scheme in section 6 that will result in making our code more efficient, and enable us to simulate the scalar field dynamics in a number of interesting scenarios. This includes updating our code to make it more compact and automated, as well as extending it to study the spectrum of scalar-induced gravitational waves, inflaton dynamics in the post- inflationary epoch, multi-field inflationary dynamics, and even scalar field models of dark matter and dark energy. We will incorporate most of these additional features in the revised version of our manuscript. In the meantime, we welcome constructive comments and suggestions as well as queries from interested readers which will help us in improving the quality of our numerical work and presentation in the updated version of the paper. ## 8 Acknowledgements S.S.M thanks Satadru Bag, Shabbir Shaikh, and Varun Sahni for crucial inputs during the early stages of development of this code. The authors are grateful to Parth Bhargava and Sanket Dave for stimulating discussions on various topics related to numerical dynamics discussed in this paper. S.S.M. is supported as a postdoctoral Research Associate at the School of Physics and Astronomy, University of Nottingham by the STFC funded consolidated grant, UK. S.S.B was supported by the INSPIRE scholarship of the Department of Science and Technology (DST), Govt. of India during his Master’s thesis work during which a significant portion of this work was carried out. ## References * [1] A. A. Starobinsky, “A New Type of Isotropic Cosmological Models Without Singularity,” Phys. Lett. B 91, 99-102 (1980). * [2] A. H. Guth, “The Inflationary Universe: A Possible Solution to the Horizon and Flatness Problems,” Phys. Rev. D 23, 347 (1981). * [3] A. D. Linde, “A New Inflationary Universe Scenario: A Possible Solution of the Horizon, Flatness, Homogeneity, Isotropy and Primordial Monopole Problems”, Phys. Lett. B 108, 389 (1982). * [4] A. Albrecht and P. J. Steinhardt, “Cosmology for Grand Unified Theories with Radiatively Induced Symmetry Breaking,” Phys. Rev. Lett. 48, 1220 (1982). * [5] A. D. Linde, “Chaotic Inflation,” Phys. Lett. B 129, 177-181 (1983). * [6] A.D. Linde, Particle Physics and Inflationary Cosmology, Harwood, Chur, Switzerland (1990). * [7] D. Baumann, “TASI Lectures on Inflation”, [arXiv:0907.5424]. * [8] V. F. Mukhanov and G. V. Chibisov, JETP Lett. 33, 532 (1981). * [9] S. W. Hawking, Phys. Lett. B 115, 295 (1982). * [10] A. A. Starobinsky, Phys. Lett. B 117, 175 (1982). * [11] A. H. Guth and S. -Y. Pi, Phys. Rev. Lett. 49, 1110 (1982). * [12] A.A. Starobinsky, JETP Lett., 30 , 682 (1979). * [13] V. Sahni, Phys. Rev. D42, 453 (1990). * [14] M. Tegmark, “What does inflation really predict?,” JCAP 04 (2005), 001 [arXiv:astro-ph/0410281 [astro-ph]]. * [15] Y. Akrami et al. [Planck], “Planck 2018 results. X. Constraints on inflation” Astron. Astrophys. 641, A10 (2020) [arXiv:1807.06211 [astro-ph.CO]]. * [16] S. S. Mishra, V. Sahni and A. A. Starobinsky, “Curing inflationary degeneracies using reheating predictions and relic gravitational waves,” JCAP 05, 075 (2021) [arXiv:2101.00271 [gr-qc]]. * [17] P. A. R. Ade et al. [BICEP and Keck], “Improved Constraints on Primordial Gravitational Waves using Planck, WMAP, and BICEP/Keck Observations through the 2018 Observing Season,” Phys. Rev. Lett. 127, no.15, 151301 (2021) [arXiv:2110.00483 [astro-ph.CO]]. * [18] S. S. Mishra and V. Sahni, “Canonical and Non-canonical Inflation in the light of the recent BICEP/Keck results,” [arXiv:2202.03467 [astro-ph.CO]]. * [19] R. Kallosh and A. Linde, “BICEP/Keck and cosmological attractors,” JCAP 12, no.12, 008 (2021) [arXiv:2110.10902 [astro-ph.CO]]. * [20] S. S. Mishra, V. Sahni and A. V. Toporensky, “Initial conditions for Inflation in an FRW Universe,” Phys. Rev. D 98, no.8, 083538 (2018) [arXiv:1801.04948 [gr-qc]]. * [21] S. Hawking, “Gravitationally collapsed objects of very low mass,” Mon. Not. Roy. Astron. Soc. 152 (1971), 75 * [22] B. J. Carr and S. W. Hawking, “Black holes in the early Universe,” Mon. Not. Roy. Astron. Soc. 168 (1974), 399-415 * [23] B. J. Carr, “The Primordial black hole mass spectrum,” Astrophys. J. 201 (1975), 1-19 * [24] M. Sasaki, T. Suyama, T. Tanaka and S. Yokoyama, “Primordial black holes—perspectives in gravitational wave astronomy,” Class. Quant. Grav. 35 (2018) no.6, 063001 [arXiv:1801.05235 [astro-ph.CO]]. * [25] G. F. Chapline, “Cosmological effects of primordial black holes,” Nature 253 (1975) no.5489, 251-252 * [26] P. Meszaros, “Primeval black holes and galaxy formation,” Astron. Astrophys. 38 (1975), 5-13 * [27] P. Ivanov, P. Naselsky and I. Novikov, “Inflation and primordial black holes as dark matter,” Phys. Rev. D 50 (1994), 7173-7178 * [28] B. Carr, F. Kuhnel and M. Sandstad, “Primordial Black Holes as Dark Matter,” Phys. Rev. D 94 (2016) no.8, 083504 [arXiv:1607.06077 [astro-ph.CO]]. * [29] A. M. Green and B. J. Kavanagh, “Primordial Black Holes as a dark matter candidate,” J. Phys. G 48, no.4, 043001 (2021) [arXiv:2007.10722 [astro-ph.CO]]. * [30] B. Carr and F. Kuhnel, “Primordial Black Holes as Dark Matter: Recent Developments,” Ann. Rev. Nucl. Part. Sci. 70 (2020), 355-394 [arXiv:2006.02838 [astro-ph.CO]]. * [31] E. Bugaev and P. Klimai, “Large curvature perturbations near horizon crossing in single-field inflation models,” Phys. Rev. D 78 (2008), 063515 [arXiv:0806.4541 [astro-ph]]. * [32] C. Germani and T. Prokopec, “On primordial black holes from an inflection point,” Phys. Dark Univ. 18 (2017), 6-10 [arXiv:1706.04226 [astro-ph.CO]]. * [33] J. M. Ezquiaga, J. Garcia-Bellido and E. Ruiz Morales, “Primordial Black Hole production in Critical Higgs Inflation,” Phys. Lett. B 776 (2018), 345-349 doi:10.1016/j.physletb.2017.11.039 [arXiv:1705.04861 [astro-ph.CO]]. * [34] J. Garcia-Bellido and E. Ruiz Morales, “Primordial black holes from single field models of inflation,” Phys. Dark Univ. 18 (2017), 47-54 [arXiv:1702.03901 [astro-ph.CO]]. * [35] I. Dalianis, A. Kehagias and G. Tringas, “Primordial black holes from $\alpha$-attractors,” JCAP 01 (2019), 037 [arXiv:1805.09483 [astro-ph.CO]]. * [36] O. Özsoy, S. Parameswaran, G. Tasinato and I. Zavala, “Mechanisms for Primordial Black Hole Production in String Theory,” JCAP 07 (2018), 005 [arXiv:1803.07626 [hep-th]]. * [37] M. Cicoli, V. A. Diaz and F. G. Pedro, “Primordial Black Holes from String Inflation,” JCAP 06 (2018), 034 [arXiv:1803.02837 [hep-th]]. * [38] M. P. Hertzberg and M. Yamada, “Primordial Black Holes from Polynomial Potentials in Single Field Inflation,” Phys. Rev. D 97 (2018) no.8, 083509 [arXiv:1712.09750 [astro-ph.CO]]. * [39] G. Ballesteros and M. Taoso, “Primordial black hole dark matter from single field inflation,” Phys. Rev. D 97 (2018) no.2, 023501 [arXiv:1709.05565 [hep-ph]]. * [40] H. Di and Y. Gong, “Primordial black holes and second order gravitational waves from ultra-slow-roll inflation,” JCAP 07 (2018), 007 [arXiv:1707.09578 [astro-ph.CO]]. * [41] H. Motohashi, S. Mukohyama and M. Oliosi, “Constant Roll and Primordial Black Holes,” JCAP 03 (2020), 002 [arXiv:1910.13235 [gr-qc]]. * [42] R. Mahbub, “Primordial black hole formation in inflationary $\alpha$-attractor models,” Phys. Rev. D 101 (2020) no.2, 023533 [arXiv:1910.10602 [astro-ph.CO]]. * [43] N. Bhaumik and R. K. Jain, “Primordial black holes dark matter from inflection point models of inflation and the effects of reheating,” JCAP 01 (2020), 037 [arXiv:1907.04125 [astro-ph.CO]]. * [44] S. S. Mishra and V. Sahni, “Primordial Black Holes from a tiny bump/dip in the Inflaton potential,” JCAP 04, 007 (2020) [arXiv:1911.00057 [gr-qc]]. * [45] K. Kefala, G. P. Kodaxis, I. D. Stamou and N. Tetradis, “Features of the inflaton potential and the power spectrum of cosmological perturbations,” Phys. Rev. D 104 (2021) no.2, 023506 [arXiv:2010.12483 [astro-ph.CO]]. * [46] H. V. Ragavendra, P. Saha, L. Sriramkumar and J. Silk, “Primordial black holes and secondary gravitational waves from ultraslow roll and punctuated inflation,” Phys. Rev. D 103 (2021) no.8, 083510 [arXiv:2008.12202 [astro-ph.CO]]. * [47] O. Özsoy and Z. Lalak, “Primordial black holes as dark matter and gravitational waves from bumpy axion inflation,” JCAP 01 (2021), 040 [arXiv:2008.07549 [astro-ph.CO]]. * [48] R. Zheng, J. Shi and T. Qiu, “On Primordial Black Holes and secondary gravitational waves generated from inflation with solo/multi-bumpy potential,” [arXiv:2106.04303 [astro-ph.CO]]. * [49] K. Inomata, E. McDonough and W. Hu, “Primordial black holes arise when the inflaton falls,” Phys. Rev. D 104 (2021) no.12, 123553 [arXiv:2104.03972 [astro-ph.CO]]. * [50] R. Brandenberger, “Initial conditions for inflation — A short review,” Int. J. Mod. Phys. D 26 (2016) no.01, 1740002 [arXiv:1601.01918 [hep-th]]. * [51] D. Baumann, PoS TASI 2017, 009 (2018) [arXiv:1807.03098 [hep-th]]. * [52] J. M. Maldacena, “Non-Gaussian features of primordial fluctuations in single field inflationary models,” JHEP 05, 013 (2003) [arXiv:astro-ph/0210603 [astro-ph]]. * [53] M. Sasaki, “Large Scale Quantum Fluctuations in the Inflationary Universe,” Prog. Theor. Phys. 76, 1036 (1986). * [54] V. F. Mukhanov, “Quantum Theory of Gauge Invariant Cosmological Perturbations,” Sov. Phys. JETP 67, 1297 (1988) [Zh. Eksp. Teor. Fiz. 94N7, 1 (1988)]. * [55] H. Motohashi, A. A. Starobinsky and J. Yokoyama, “Inflation with a constant rate of roll,” JCAP 1509, 018 (2015) [arXiv:1411.5021 [astro-ph.CO]]. * [56] T. S. Bunch and P. C. W. Davies, “Quantum Field Theory in de Sitter Space: Renormalization by Point Splitting,” Proc. Roy. Soc. Lond. A 360, 117 (1978). * [57] S. Matarrese, S. Mollerach and M. Bruni, “Second order perturbations of the Einstein-de Sitter universe,” Phys. Rev. D 58 (1998), 043504 * [58] D. Baumann, P. J. Steinhardt, K. Takahashi and K. Ichiki, “Gravitational Wave Spectrum Induced by Primordial Scalar Perturbations,” Phys. Rev. D 76 (2007), 084019 [arXiv:hep-th/0703290 [hep-th]]. * [59] N. Bartolo, V. De Luca, G. Franciolini, A. Lewis, M. Peloso and A. Riotto, “Primordial Black Hole Dark Matter: LISA Serendipity,” Phys. Rev. Lett. 122 (2019) no.21, 211301 [arXiv:1810.12218 [astro-ph.CO]]. * [60] R. g. Cai, S. Pi and M. Sasaki, “Gravitational Waves Induced by non-Gaussian Scalar Perturbations,” Phys. Rev. Lett. 122 (2019) no.20, 201101 [arXiv:1810.11000 [astro-ph.CO]]. * [61] G. Domènech, “Scalar Induced Gravitational Waves Review,” Universe 7 (2021) no.11, 398 [arXiv:2109.01398 [gr-qc]]. * [62] B. Whitt, “Fourth Order Gravity as General Relativity Plus Matter,” Phys. Lett. B 145 (1984), 176-178 * [63] A. Karam, N. Koivunen, E. Tomberg, V. Vaskonen and H. Veermäe, “Anatomy of single-field inflationary models for primordial black holes,” [arXiv:2205.13540 [astro-ph.CO]]. * [64] H. Motohashi and W. Hu, “Primordial Black Holes and Slow-Roll Violation,” Phys. Rev. D 96 (2017) no.6, 063503 doi:10.1103/PhysRevD.96.063503 [arXiv:1706.06784 [astro-ph.CO]]. * [65] C. T. Byrnes, P. S. Cole and S. P. Patil, “Steepest growth of the power spectrum and primordial black holes,” JCAP 06, 028 (2019) [arXiv:1811.11158 [astro-ph.CO]]. * [66] P. Carrilho, K. A. Malik and D. J. Mulryne, “Dissecting the growth of the power spectrum for primordial black holes,” Phys. Rev. D 100, no.10, 103529 (2019) [arXiv:1907.05237 [astro-ph.CO]]. * [67] O. Özsoy and G. Tasinato, “On the slope of the curvature power spectrum in non-attractor inflation,” JCAP 04, 048 (2020) [arXiv:1912.01061 [astro-ph.CO]]. * [68] P. S. Cole, A. D. Gow, C. T. Byrnes and S. P. Patil, “Steepest growth re-examined: repercussions for primordial black hole formation,” [arXiv:2204.07573 [astro-ph.CO]]. * [69] V. A. Belinsky, I. M. Khalatnikov, L. P. Grishchuk and Y. B. Zeldovich, “INFLATIONARY STAGES IN COSMOLOGICAL MODELS WITH A SCALAR FIELD,” Phys. Lett. B 155 (1985), 232-236. * [70] S. Kachru, R. Kallosh, A. D. Linde and S. P. Trivedi, “De Sitter vacua in string theory,” Phys. Rev. D 68, 046005 (2003) [arXiv:hep-th/0301240 [hep-th]]. * [71] S. Kachru, R. Kallosh, A. D. Linde, J. M. Maldacena, L. P. McAllister and S. P. Trivedi, “Towards inflation in string theory,” JCAP 0310, 013 (2003) [hep-th/0308055]. * [72] R. Kallosh and A. Linde, “CMB targets after the latest Planck data release,” Phys. Rev. D 100, no.12, 123523 (2019) [arXiv:1909.04687 [hep-th]]. * [73] J. Martin, C. Ringeval and V. Vennin, “Encyclopædia Inflationaris,” Phys. Dark Univ. 5-6 (2014), 75-235 [arXiv:1303.3787 [astro-ph.CO]]. * [74] W. J. Handley, A. N. Lasenby, H. V. Peiris and M. P. Hobson, “Bayesian inflationary reconstructions from Planck 2018 data,” Phys. Rev. D 100 (2019) no.10, 103511 [arXiv:1908.00906 [astro-ph.CO]]. * [75] S. Rasanen and E. Tomberg, “Planck scale black hole dark matter from Higgs inflation,” JCAP 01 (2019), 038 [arXiv:1810.12608 [astro-ph.CO]]. * [76] F. Finelli and R. H. Brandenberger, “Parametric amplification of gravitational fluctuations during reheating,” Phys. Rev. Lett. 82 (1999), 1362-1365 [arXiv:hep-ph/9809490 [hep-ph]]. * [77] K. Jedamzik, M. Lemoine and J. Martin, “Collapse of Small-Scale Density Perturbations during Preheating in Single Field Inflation,” JCAP 09 (2010), 034 [arXiv:1002.3039 [astro-ph.CO]]. * [78] K. Inomata, M. Braglia and X. Chen, “Questions on calculation of primordial power spectrum with large spikes: the resonance model case,” [arXiv:2211.02586 [astro-ph.CO]]. * [79] J. Kristiano and J. Yokoyama, “Ruling Out Primordial Black Hole Formation From Single-Field Inflation,” [arXiv:2211.03395 [hep-th]]. * [80] A. A. Starobinsky, “STOCHASTIC DE SITTER (INFLATIONARY) STAGE IN THE EARLY UNIVERSE,” Lect. Notes Phys. 246, 107-126 (1986) doi:10.1007/3-540-16452-9_6 * [81] D. S. Salopek and J. R. Bond, “Stochastic inflation and nonlinear gravity,” Phys. Rev. D 43, 1005-1031 (1991) doi:10.1103/PhysRevD.43.1005 * [82] A. A. Starobinsky and J. Yokoyama, “Equilibrium state of a selfinteracting scalar field in the De Sitter background,” Phys. Rev. D 50, 6357-6368 (1994) [arXiv:astro-ph/9407016 [astro-ph]]. * [83] T. Fujita, M. Kawasaki, Y. Tada and T. Takesako, “A new algorithm for calculating the curvature perturbations in stochastic inflation,” JCAP 12, 036 (2013) [arXiv:1308.4754 [astro-ph.CO]]. * [84] T. Fujita, M. Kawasaki and Y. Tada, “Non-perturbative approach for curvature perturbations in stochastic $\delta N$ formalism,” JCAP 10, 030 (2014) [arXiv:1405.2187 [astro-ph.CO]]. * [85] V. Vennin and A. A. Starobinsky, Eur. Phys. J. C 75, 413 (2015) [arXiv:1506.04732]. * [86] C. Pattison, V. Vennin, H. Assadullahi and D. Wands, JCAP 10, 046 (2017) [arXiv:1707.00537]. * [87] J. M. Ezquiaga and J. García-Bellido, “Quantum diffusion beyond slow-roll: implications for primordial black-hole production,” JCAP 08 (2018), 018 [arXiv:1805.06731 [astro-ph.CO]]. * [88] M. Biagetti, G. Franciolini, A. Kehagias and A. Riotto, “Primordial Black Holes from Inflation and Quantum Diffusion,” JCAP 07 (2018), 032 [arXiv:1804.07124 [astro-ph.CO]]. * [89] C. Pattison, V. Vennin, H. Assadullahi and D. Wands, “Stochastic inflation beyond slow roll,” JCAP 07, 031 (2019) [arXiv:1905.06300 [astro-ph.CO]]. * [90] J. M. Ezquiaga, J. García-Bellido and V. Vennin, “The exponential tail of inflationary fluctuations: consequences for primordial black holes,” JCAP 03 (2020), 029 [arXiv:1912.05399 [astro-ph.CO]]. * [91] H. Firouzjahi, A. Nassiri-Rad and M. Noorbala, “Stochastic Ultra Slow Roll Inflation,” JCAP 01 (2019), 040 [arXiv:1811.02175 [hep-th]]. * [92] V. Vennin, “Cosmological Inflation: Theoretical Aspects and Observational Constraints,” (PhD Thesis) tel-01094199. * [93] G. Ballesteros, J. Rey, M. Taoso and A. Urbano, “Stochastic inflationary dynamics beyond slow-roll and consequences for primordial black hole formation,” JCAP 08 (2020), 043 [arXiv:2006.14597 [astro-ph.CO]]. * [94] K. Ando and V. Vennin, “Power spectrum in stochastic inflation,” JCAP 04, 057 (2021) [arXiv:2012.02031 [astro-ph.CO]]. * [95] A. De and R. Mahbub, “Numerically modeling stochastic inflation in slow-roll and beyond,” Phys. Rev. D 102, no.12, 123509 (2020) [arXiv:2010.12685 [astro-ph.CO]]. * [96] D. G. Figueroa, S. Raatikainen, S. Rasanen and E. Tomberg, “Non-Gaussian Tail of the Curvature Perturbation in Stochastic Ultraslow-Roll Inflation: Implications for Primordial Black Hole Production,” Phys. Rev. Lett. 127, no.10, 101302 (2021) [arXiv:2012.06551 [astro-ph.CO]]. * [97] D. Cruces and C. Germani, Phys. Rev. D 105, no.2, 023533 (2022) [arXiv:2107.12735]. * [98] G. Rigopoulos and A. Wilkins, JCAP 12, no.12, 027 (2021) [arXiv:2107.05317]. * [99] C. Pattison, V. Vennin, D. Wands and H. Assadullahi, “Ultra-slow-roll inflation with quantum diffusion,” JCAP 04, 080 (2021) [arXiv:2101.05741 [astro-ph.CO]]. * [100] E. Tomberg, “A numerical approach to stochastic inflation and primordial black holes,” J. Phys. Conf. Ser. 2156, no.1, 012010 (2021) [arXiv:2110.10684 [astro-ph.CO]]. * [101] D. G. Figueroa, S. Raatikainen, S. Rasanen and E. Tomberg, “Implications of stochastic effects for primordial black hole production in ultra-slow-roll inflation,” [arXiv:2111.07437 [astro-ph.CO]]. * [102] Y. Tada and V. Vennin, “Statistics of coarse-grained cosmological fields in stochastic inflation,” JCAP 02, no.02, 021 (2022) [arXiv:2111.15280 [astro-ph.CO]]. * [103] R. Mahbub and A. De, “Smooth coarse-graining and colored noise dynamics in stochastic inflation,” [arXiv:2204.03859 [astro-ph.CO]]. * [104] N. Ahmadi, M. Noorbala, N. Feyzabadi, F. Eghbalpoor and Z. Ahmadi, “Quantum Diffusion in Sharp Transition to Non-Slow-Roll Phase,” JCAP 08 (2022), 078 [arXiv:2207.10578 [gr-qc]]. * [105] J. H. P. Jackson, H. Assadullahi, K. Koyama, V. Vennin and D. Wands, “Numerical simulations of stochastic inflation using importance sampling,” JCAP 10 (2022), 067 [arXiv:2206.11234 [astro-ph.CO]]. * [106] M. Celoria, P. Creminelli, G. Tambalo and V. Yingcharoenrat, “Beyond perturbation theory in inflation,” JCAP 06, 051 (2021) doi:10.1088/1475-7516/2021/06/051 [arXiv:2103.09244 [hep-th]]. * [107] S. Hooshangi, M. H. Namjoo and M. Noorbala, “Rare events are nonperturbative: Primordial black holes from heavy-tailed distributions,” Phys. Lett. B 834 (2022), 137400 [arXiv:2112.04520 [astro-ph.CO]]. * [108] Y. F. Cai, X. H. Ma, M. Sasaki, D. G. Wang and Z. Zhou, “One small step for an inflaton, one giant leap for inflation: A novel non-Gaussian tail and primordial black holes,” Phys. Lett. B 834 (2022), 137461 [arXiv:2112.13836 [astro-ph.CO]]. * [109] J. M. Ezquiaga, J. García-Bellido and V. Vennin, “Could ”El Gordo” be hinting at primordial quantum diffusion?,” [arXiv:2207.06317 [astro-ph.CO]]. * [110] Y. F. Cai, X. H. Ma, M. Sasaki, D. G. Wang and Z. Zhou, “Highly non-Gaussian tails and primordial black holes from single-field inflation,” [arXiv:2207.11910 [astro-ph.CO]]. * [111] A. D. Gow, H. Assadullahi, J. H. P. Jackson, K. Koyama, V. Vennin and D. Wands, “Non-perturbative non-Gaussianity and primordial black holes,” [arXiv:2211.08348 [astro-ph.CO]].
# EBHI-Seg: A Novel Enteroscope Biopsy Histopathological Haematoxylin and Eosin Image Dataset for Image Segmentation Tasks Liyu Shi Xiaoyan Li<EMAIL_ADDRESS>Weiming Hu Haoyuan Chen Jing Chen Zizhen Fan Minghe Gao Yujie Jing Guotao Lu Deguo Ma Zhiyu Ma Qingtao Meng Dechao Tang Hongzan Sun Marcin Grzegorzek Shouliang Qi Yueyang Teng Chen Li<EMAIL_ADDRESS>Microscopic Image and Medical Image Analysis Group, College of Medicine and Biological Information Engineering, Northeastern University, Shenyang 110169, China Department of Pathology, Cancer Hospital of China Medical University, Liaoning Cancer Hospital and Institute, Shengyang 110042, China Shengjing Hospital, China Medical University, Shenyang 110122, China Institute of Medical Informatics, University of Luebeck, Luebeck 23538, Germany Department of Knowledge Engineering, University of Economics in Katowice, Bogucicka 3, 40287 Katowice, Poland ###### Abstract Background and Purpose: Colorectal cancer is a common fatal malignancy, the fourth most common cancer in men, and the third most common cancer in women worldwide. Timely detection of cancer in its early stages is essential for treating the disease. Currently, there is a lack of datasets for histopathological image segmentation of rectal cancer, which often hampers the assessment accuracy when computer technology is used to aid in diagnosis. Methods: This present study provided a new publicly available _Enteroscope Biopsy Histopathological Hematoxylin and Eosin Image Dataset for Image Segmentation Tasks_ (EBHI-Seg). To demonstrate the validity and extensiveness of EBHI-Seg, the experimental results for EBHI-Seg are evaluated using classical machine learning methods and deep learning methods. Results: The experimental results showed that deep learning methods had a better image segmentation performance when utilizing EBHI-Seg. The maximum accuracy of the Dice evaluation metric for the classical machine learning method is 0.948, while the Dice evaluation metric for the deep learning method is 0.965. Conclusion: This publicly available dataset contained 5,170 images of six types of tumor differentiation stages and the corresponding ground truth images. The dataset can provide researchers with new segmentation algorithms for medical diagnosis of colorectal cancer, which can be used in the clinical setting to help doctors and patients. EBHI-Seg is publicly available at: https://doi.org/10.6084/m9.figshare.21540159.v1 ###### keywords: Colorectal Histopathology , Enteroscope Biopsy , Image Dataset , Image segmentation ††journal: Journal of LaTeX Templates ## 1 Introduction Colon cancer is a common deadly malignant tumor, the fourth most common cancer in men, and the third most common cancer in women worldwide. Colon cancer is responsible for $10\%$ of all cancer cases [1]. According to prior research, colon and rectal tumors share many of the same or similar characteristics. Hence, they are often classified collectively. The present study categorized rectal and colon cancers into one colorectal cancer category [2]. Histopathological examination of the intestinal tract is both the gold standard for the diagnosis of colorectal cancer and a prerequisite for disease treatment [3]. The advantage of using the intestinal biopsy method to remove a part of the intestinal tissue for histopathological analysis, which is used to determine the true status of the patient, is that it considerably reduces damage to the body and rapid wound healing [4]. The histopathology sample is then sectioned and processed with Hematoxylin and Eosin (H&E). Treatment with H&E is a common approach when staining tissue sections to show the inclusions between the nucleus and cytoplasm and highlight the fine structures between tissues [5, 6]. When a pathologist performs an examination of the colon, they first examine the histopathological sections for eligibility and find the location of the lesion. The pathology sections are then examined and diagnosed using a low magnification microscope. If finer structures need to be observed, the microscope is adjusted to use high magnification for further analysis. However, the following problems usually exist in the diagnostic process: the diagnostic results become more subjective and varied due to different doctors reasons; doctors can easily overlook some information in the presence of a large amount of test data; it is difficult to analyze large amounts of previously collected data [7]. Therefore, it is a necessary to address these issues effectively. With the development and popularization of computer-aided diagnosis (CAD), the pathological sections of each case can be accurately and efficiently examined with the help of computers [8]. Now, CAD is widely used in many biomedical image analysis tasks, such as microorganism image analysis [9, 10, 11, 12, 13, 14, 15, 16, 17], COVID-19 image analysis [18], histopatholgical image analysis [19, 20, 21, 22, 23, 24, 25, 26], cytopathological image analysis [27, 28, 29, 30] and sperm video analysis [31, 32]. Therefore, the application of computer vision technology for colorectal cancer CAD provides a new direction in this research field [33]. One of the fundamental features of CAD is the aspect of image segmentation, the results of which can be used as key evidence in the pathologists’ diagnostic processes. Along with the rapid development of medical image segmentation methodology, there is a wide demand for its application to identify benign and malignant tumors, tumor differentiation stages, and other related fields [34]. Therefore, a multi-class image segmentation method is needed to obtain high segmentation accuracy and good robustness [35]. The present study presents a novel _Enteroscope Biopsy Histopathological H_ &_E Image Dataset for Image Segmentation Tasks_ (EBHI-Seg), which contains 5,710 electron microscopic images of histopathological colorectal cancer sections that encompass six tumor differentiation stages: normal, polyp, low- grade intraepithelial neoplasia, high-grade intraepithelial neoplasia, serrated adenoma, and adenocarcinoma. The segmentation coefficients and evaluation metrics are obtained by segmenting the images of this dataset using different classical machine learning methods and novel deep learning methods. ## 2 Related Work The present study analyzed and compared the existing colorectal cancer biopsy dataset and provided an in-depth exploration of the currently known research findings. The limitations of the presently available colorectal cancer dataset were also pointed out. The following conclusions were obtained in the course of the study. For existing datasets, the data types can be grouped into two major categories: Multi and Dual Categorization datasets. Multi Categorization datasets contain tissue types at all stages from Normal to Neoplastic. In [36], a dataset called “Collection of textures in colorectal cancer histology” is described. It includes 5,000 patches of size 74 $\mu$m $\times$ 74 $\mu$m and contains seven categories. However, because there were only 10 images, it is too small for a data sample and lacked generalization capability. In [22], a dataset called “NCT-CRC-HE-100K” is proposed. This is a set of 100,000 non-overlapping image patches of histological human colorectal cancer (CRC) and normal tissue samples stained with (H&E) that was presented by the National Center for Tumor Diseases (NCT). These image patches are from nine different tissues with an image size of 224 $\times$ 224 pixels. The nine tissue categories are adipose, background, debris, lymphocytes, mucus, smooth muscle, normal colon mucosa, cancer-associated stroma, and colorectal adenocarcinoma epithelium. This dataset is publicly available and commonly used. However, because the image sizes are all 224 $\times$ 224 pixels, the dataset underperformed in some global details that need to be observed in individual categories. Two datasets are utilized in [37]: one containing colonic H&E-stained biopsy sections (CRC dataset) and the other consisting of prostate cancer H&E-stained biopsy sections (PCa dataset). The CRC dataset contains 1,133 colorectal biopsy and polypectomy slides grouped into three categories and labelled as non- neoplastic, low-grade and high-grade lesions. In [38], a dataset named “MICCAI 2016 gland segmentation challenge dataset (GlaS)” is used. This dataset contained 165 microscopic images of H&E-stained colon glandular tissue samples, including 85 training and 80 test datasets. Each dataset is grouped into two parts: benign and malignant tumors. The image size is 775 $\times$ 522 pixels. Since this dataset has only two types of data and the number of data is too little, so that it performs poorly on some multi-type training. Dual Categorization datasets usually contain only two types of tissue types: Normal and Neoplastic. In [39], a dataset named “FFPE” is proposed. This dataset obtained its images by extracting 328 Formalin-fixed Paraffin-embedded (FFPE) whole-slide images of colorectal polyps classified into two categories of : hyperplastic polyps (HPs) and sessile serrated adenomas (SSAs). This dataset contained 3,125 images with an image size of 224 $\times$ 224 pixels and is small in type and number. In [40], two datasets named “UHCW” and “TCGA” are proposed. The first dataset is a colorectal cancer biopsy sequence developed at the University Hospital of Coventry and Warwickshire (UHCW) for internal validation of the rectal biopsy trial. The second dataset is the Cancer Genome Atlas (TCGA) for external validation of the trial. This dataset is commonly used as a publicly available cancer dataset and stores genomic data for more than 20 types of cancers. The two dataset types are grouped into two categories: Normal and Neoplastic. The first dataset contains 4,292 slices, and the second dataset contained 731 slices with an image size of 224 $\times$ 224 pixels. All of the information for the existing datasets is summarized in Table 1. The issues associated with the dataset mentioned above included fewer data types, small amount of data, inaccurate dataset ground truth, etc. The current study required an open-source multi-type colonoscopy biopsy image dataset. Table 1: A dataset for the pathological classification of colorectal cancer. Dataset Name | Multi Categorization | Amount | Size | Year ---|---|---|---|--- Collection of textures in colorectal cancer histology | lymphoid follicles, mucosal glands, debris, adipose, tumor epithelium simple stroma, complex stroma, background patches with no tissue | 5000 | 74$\mu$m $\times$ 74$\mu$m (0.495 micrometre per pixel) | 2016 HE-NCT-CRC-100K | MUS, NORM, STR, TUM ADI, BACK, DEB, LYM, MUC | 100000 | 224$\times$224 pixels | 2016 MICCAI’16 gland seg- mentation challenge dataset | benign tumors, malignant tumors | 85 | 775$\times$522 pixels | 2017 CRC dataset | non-neoplastic, low-grade, high-grade lesions | 1133 | 512$\times$512 pixels | 2021 Dataset Name | Dual Categorization | Amount | Size | Year FFPE | HPs, SSAs | 3152 | 224$\times$224 pixels | 2021 The Cancer Genome Atlas dataset | Normal, Neoplastic | 731 | 224$\times$224 pixels | 2021 University Hospitals Coventry and Warwick- shire dataset | Normal, Neoplastic | 4292 | 224$\times$224 pixels | 2021 ## 3 Basic Information for EBHI-Seg ### 3.1 Dataset Overview The dataset in the present study contained 5,710 histopathology images, including 2,855 histopathology section images and 2,855 ground truth images. The basic information for the dataset is described in detail below. EBHI-Seg is publicly available at: https://doi.org/10.6084/m9.figshare.21540159.v1 In the present paper, H&E-treated histopathological sections of colon tissues are used as data for evaluating image segmentation. The dataset is obtained from two histopathologists at the Cancer Hospital of China Medical University (proved by “Research Project Ethics Certification” (No. 202229) ). It is prepared by 12 biomedical researchers according to the following rules: Firstly, if there is only one differentiation stage in the image and the rest of the image is intact, then the differentiation stage became the image label; Secondly, if there is more than one differentiation stage in the image, then the most obvious differentiation is selected as the image label; In general, the most severe and prominent differentiation in the image was used as the image label. Intestinal biopsy was used as the sampling method in this dataset. The magnification of the data slices is $400\times$, with an eyepiece magnification of $10\times$ and an objective magnification of $40\times$. A Nissan Olympus microscope and NewUsbCamera acquisition software are used. The image input size is $224\times 224$ pixels, and the format is *.png. The data are grouped into five types described in detail in section 2.2. ### 3.2 Data Type Description Normal: Colorectal tissue sections of the standard category are made-up of consistently ordered tubular structures and that does not appear infected when viewed under a light microscope [41]. Section images with the corresponding ground truth images are shown in Figure 1(a). Polyp: Colorectal polyps are similar in shape to the structures in the normal category, but have a completely different histological structure. A polyp is a redundant mass that grows on the surface of the body’s cells. Modern medicine usually refers to polyps as unwanted growths on the mucosal surface of the body [42]. The pathological section of the polyp category also has an intact luminal structure with essentially no nuclear division of the cells. Only the atomic mass is slightly higher than that in the normal category. The polyp category and corresponding ground truth images are shown in Figure 1(b). Intraepithelial neoplasia: Intraepithelial neoplasia (IN) is the most critical precancerous lesion. Compared to the normal category, its histological images show increased branching of adenoid structures, dense arrangement, and different luminal sizes and shapes. In terms of cellular morphology, the nuclei are enlarged and vary in size, while nuclear division increases [43]. The standard Padova classification currently classifies intraepithelial neoplasia into low-grade and high-grade INs. High-grade IN demonstrate more pronounced structural changes in the lumen and nuclear enlargement compared to low-grade IN. The images and ground truth diagrams of high-grade and low-grade INs are shown in Figure 1(c)(d). Adenocarcinoma: Adenocarcinoma is a malignant digestive tract tumor with a very irregular distribution of luminal structures. It is difficult to identify its border structures during observation, and the nuclei are significantly enlarged at this stage [44]. An adenocarcinoma with its corresponding ground truth diagram is shown in Figure 1(e). Serrated adenoma: Serrated adenomas are uncommon lesions, accounting for 1% of all colonic polyps [45]. The endoscopic surface appearance of serrated adenomas is not well characterized but is thought to be similar to that of colonic adenomas with tubular or cerebral crypt openings [46]. The image of a serrated adenoma with a corresponding ground truth diagram is shown in Figure 1(f). Figure 1: An example of histopathological images database: (a) Normal and ground truth, (b) Polyp and ground truth, (c) High-grade Intraepithelial Neoplasia and ground truth, (d) Low-grade Intraepithelial Neoplasia and ground truth, (e) Adenocarcinoma and ground truth, (f) Serrated adenoma and ground truth. ## 4 Evaluation of EBHI-Seg ### 4.1 Image Segmentation Evaluation Metric Six evaluation metrics are commonly used for image segmentation tasks. The Dice ratio metric is a standard metric used in medical images that is often utilized to evaluate the performance of image segmentation algorithms. It is a validation method based on spatial overlap statistics that measures the similarities between the algorithm segmentation output and ground truth [47]. The Dice ratio is defined in Eq. (1). $\centering\rm DiceRatio=\frac{2\left|X\cap Y\right|}{\left|X\right|+\left|Y\right|}.\@add@centering$ (1) In Eq. (1), for a segmentation task, $X$ and $Y$ denote the ground truth and segmentation mask prediction, respectively. The range of the calculated results is [0,1], and the larger the result the better. The Jaccard index is a classical set similarity measure with many practical applications in image segmentation. The Jaccard index measures the similarity of a finite set of samples: the ratio between the intersection and concatenation of the segmentation results and ground truth [48]. The Jaccard index is defined in Eq. (2). $\rm JaccardIndex=\frac{\left|X\cap Y\right|}{\left|X\cup Y\right|}.$ (2) The range of the calculated results is [0,1], and the larger the result the better. Recall and precision are the recall and precision rates, respectively. The range of the calculated results is [0,1]. A higher output indicates a better segmentation result. Recall and precision are defined in Eq. (3) and Eq. (4), $\rm Precison=\frac{TP}{TP+FP},$ (3) $\rm Recall=\frac{TP}{TP+FN},$ (4) where TP, FP, TN, and FN are defined in table 2. Table 2: Confusion Matrix Ground truth | Predict mask ---|--- Positive | Negative Positive | TP | TN Negative | FP | FN The conformity coefficient (Confm Index) is a consistency coefficient, which is calculated by putting the binary classification result of each pixel from [$-\infty$,1] into continuous interval [$-\infty$,1] to calculate the ratio of the number of incorrectly segmented pixels to the number of correctly segmented pixels to measure the consistency between the segmentation result and ground truth. The conformity coefficient is defined in Eq. (5)(6), $\rm ConfmIndex=(1-\frac{\theta_{AE}}{\theta_{TP}}),\theta_{TP}>0,$ (5) $\rm ConfmIndex=Failure,\theta_{TP}=0,$ (6) where $\theta_{AE}$= $\theta_{FP}$+$\theta_{FN}$ represents all errors of the fuzzy segmentation results. $\theta_{TP}$ is the number of correctly classified pixels. Mathematically, ConfmIndex can be negative infinity if $\theta_{TP}$=0. Such a segmentation result is definitely inadequate and treated as failure without the need of any further analysis. ### 4.2 Classical Machine Learning Methods Image segmentation is one of the most commonly used methods for classifying image pixels in decision-oriented applications [49]. It groups an image into regions high in pixel similarity within each area and has a significant contrast between different regions [50]. Machine learning methods for segmentation distinguish the image classes using image features. (1) $k$-means algorithm is a classical division-based clustering algorithm, where image segmentation means segmenting the image into many disjointed regions. The essence is the clustering process of pixels, and the $k$-means method is one of the simplest clustering methods [51]. Image segmentation of the present study dataset is performed using the classical machine learning method described above. (2) Markov random field (MRF) is a powerful stochastic tool that models the joint probability distribution of an image based on its local spatial action [52]. It can extract the texture features of the image and model the image segmentation problem. (3) OTSU algorithm is a global adaptive binarized threshold segmentation algorithm that uses the maximum inter-class variance between the image background and the target image as the selection criterion [53]. The image is grouped into foreground and background parts based on its grayscale characteristics independent of the brightness and contrast. (4) Watershed algorithm is a region-based segmentation method, that takes the similarity between neighboring pixels as a reference and connects those pixels with similar spatial locations and grayscale values into a closed contour to achieve the segmentation effect [54]. (5) Sobel algorithm has two operators, where one detects horizontal edges and the other detects vertical flat edges. An image is the final result of its operation. Sobel edge detection operator is a set of directional operators that can be used to perform edge detection from different directions [55]. The segmentation results are shown in Figure 2. Figure 2: Five types of data segmentation results obtained by different classical machine learning methods. The performance of EBHI-Seg for different machine learning methods is observed by comparing the images segmented using classical machine learning methods with the corresponding ground truth. The segmentation evaluation metrics results are shown in Table 3. The Dice ratio algorithm is a similarity measure, usually used to compare the similarity of two samples. The value of one for this metric is c onsidered to indicate the best effect, while the value of the worst impact is zero. The Table 3 shows that $k$-means has a good Dice ratio algorithm value of up to 0.650 in each category. The MRF and Sobel segmentation results also achieved a good Dice ratio algorithm value of around 0.6. In terms of image precision and recall segmentation coefficients, $k$-means is maintained at approximately 0.650 in each category. In the classical machine learning methods, $k$-means has the best segmentation results, followed by MRF and Sobel. OTSU has a general effect, while the watershed algorithm has various coefficients that are much lower than those in the above methods. Moreover,there are apparent differences in the segmentation results when using the above methods. Table 3: Evaluation metrics for five different segmentation methods based on classical machine learning. | | DiceRatio | JaccardIndex | ConformityCoefficient | Precision | Recall ---|---|---|---|---|---|--- | $k$-means | 0.648 | 0.488 | -0.184 | 0.646 | 0.663 | MRF | 0.636 | 0.473 | -0.230 | 0.637 | 0.658 Normal | OTSU | 0.410 | 0.265 | -2.871 | 0.515 | 0.351 | Watershed | 0.461 | 0.300 | -1.375 | 0.668 | 0.356 | Sobel | 0.652 | 0.487 | -0.102 | 0.763 | 0.579 | $k$-means | 0.592 | 0.430 | -0.528 | 0.546 | 0.663 | MRF | 0.511 | 0.362 | -2.133 | 0.540 | 0.502 Polyp | OTSU | 0.400 | 0.259 | -3.108 | 0.413 | 0.399 | Watershed | 0.433 | 0.277 | -1.675 | 0.551 | 0.362 | Sobel | 0.583 | 0.416 | -0.499 | 0.626 | 0.562 | $k$-means | 0.626 | 0.478 | -0.467 | 0.650 | 0.620 | MRF | 0.550 | 0.441 | -30.85 | 0.614 | 0.526 High-grade IN | OTSU | 0.249 | 0.150 | -12.06 | 0.373 | 0.191 | Watershed | 0.472 | 0.309 | -1.258 | 0.738 | 0.350 | Sobel | 0.634 | 0.469 | -0.200 | 0.728 | 0.577 | $k$-means | 0.650 | 0.492 | -0.172 | 0.651 | 0.663 | MRF | 0.554 | 0.404 | -1.808 | 0.643 | 0.504 Low-grade IN | OTSU | 0.886 | 0.811 | 0.6998 | 0.832 | 0.979 | Watershed | 0.464 | 0.303 | -1.345 | 0.676 | 0.357 | Sobel | 0.656 | 0.492 | -0.079 | 0.771 | 0.582 | $k$-means | 0.633 | 0.481 | -0.414 | 0.655 | 0.645 | MRF | 0.554 | 0.404 | -1.808 | 0.643 | 0.504 Adenocarcinoma | OTSU | 0.336 | 0.215 | -5.211 | 0.454 | 0.282 | Watershed | 0.458 | 0.298 | -1.437 | 0.700 | 0.349 | Sobel | 0.553 | 0.388 | -0.733 | 0.692 | 0.484 | $k$-means | 0.636 | 0.473 | -0.230 | 0.637 | 0.658 | MRF | 0.571 | 0.419 | -0.898 | 0.656 | 0.547 Serrated adenoma | OTSU | 0.393 | 0.248 | -2.444 | 0.565 | 0.315 | Watershed | 0.449 | 0.290 | -1.494 | 0.656 | 0.345 | Sobel | 0.698 | 0.541 | 0.7484 | 0.662 | 0.572 In summary, EBHI-Seg has significantly different results when using different classical machine learning segmentation methods. Different classical machine learning methods have an obvious differentiation according to the image segmentation evaluation metrics. Therefore, EBHI-Seg can effectively evaluate the segmentation performance of different segmentation methods. ### 4.3 Deep Learning Methods Besides the classical macine learning metheds tested above, some popular deep learning methods are also tested. (1) Seg-Net is an open source project for image segmentation [56]. The network is identical to the convolutional layer of VGG-16, with the removal of the fully-connected hierarchy and the addition of max-pooling indices resulting in improved boundary delineation. Seg-Net performs better in large datasets. (2) U-Net network structure was first proposed in 2015 [57] for medical imaging. U-Net is lightweight, and its simultaneous detection of local and global information is helpful for both information extraction and diagnostic results from clinical medical images. (3) MedT is a network published in 2021, which is a transformer structure that applies an attention mechanism based on medical image segmentation [58]. The segmentation results are shown in Figure 3. Figure 3: Three types of data segmentation results obtained by different deep learning methods. The segmentation effect is test on the present dataset using three deep learning models. In the experiments, each model is trained using the ratio of the training set, validation set, and test set of $4:4:2$. The model learning rate is set to $3e-6$, epochs are set to 100, and batch-size is set to 1. The dataset segmentation results of using three different models are shown in Figure 3. The experimental segmentation evaluation metrics are shown in Table 4. Overall, deep learning performs much better than classical machine learning methods. Among them, the evaluation indexes of the training results using the U-Net and Seg-Net models can reach 0.90 on average. The evaluation results of the MedT model are slightly worse at a level, between 0.70 and 0.80. The training time is longer for MedT and similar for U-Net and Seg-Net. Table 4: Evaluation metrics for three different segmentation methods based on deep learning. | | DiceRatio | JaccardIndex | ConformityCoefficient | Precision | Recall ---|---|---|---|---|---|--- | U-Net | 0.411 | 0.263 | -2.199 | 0.586 | 0.328 Normal | Seg-Net | 0.777 | 0.684 | -0.607 | 0.895 | 0.758 | MedT | 0.676 | 0.562 | -0.615 | 0.874 | 0.610 | U-Net | 0.965 | 0.308 | -1.514 | 0.496 | 0.470 Polyp | Seg-Net | 0.937 | 0.886 | 0.858 | 0.916 | 0.965 | MedT | 0.771 | 0.643 | 0.336 | 0.687 | 0.920 | U-Net | 0.895 | 0.816 | 0.747 | 0.847 | 0.961 High-grade IN | Seg-Net | 0.894 | 0.812 | 0.757 | 0.881 | 0.913 | MedT | 0.824 | 0.707 | 0.556 | 0.740 | 0.958 | U-Net | 0.911 | 0.849 | 0.773 | 0.879 | 0.953 Low-grade IN | Seg-Net | 0.924 | 0.864 | 0.826 | 0.883 | 0.977 | MedT | 0.889 | 0.808 | 0.730 | 0.876 | 0.916 | U-Net | 0.887 | 0.808 | 0.718 | 0.850 | 0.950 Adenocarcinoma | Seg-Net | 0.865 | 0.775 | 0.646 | 0.792 | 0.977 | MedT | 0.735 | 0.595 | 0.197 | 0.662 | 0.864 | U-Net | 0.938 | 0.886 | 0.865 | 0.899 | 0.983 Serrated adenoma | Seg-Net | 0.907 | 0.832 | 0.794 | 0.859 | 0.963 | MedT | 0.670 | 0.509 | -0.043 | 0.896 | 0.544 Based on the above results, EBHI-Seg achieved a clear differentiation using deep learning image segmentation methods. Image segmentation metrics for different deep learning methods are significantly different so that EBHI-Seg can evaluate their segmentation performance. ## 5 Discussion ### 5.1 Discussion of Image Segmentation Results Using Classical Machine Learning Methods Six types of tumor differentiation stage data in EBHI-Seg were analyzed using classical machine learning methods to obtain the results in Table 3. Base on the Dice ratio metrics, $k$-means, MRF and Sobel show no significant differences among the three methods around 0.55. In contrast, Watershed metrics are $\sim$0.45 on average, which is lower than the above three metrics. OTSU index is around $\sim$0.40 because the foreground-background is blurred in some experimental samples and OTSU had a difficulty extracting a suitable segmentation threshold, which resulted in undifferentiated test results. Precision and Recall evaluation indexes for k-means, MRF, and Sobel are also around 0.60, which is higher than those for OTSU and Watershed methods by about 0.20. In these three methods, $k$-means and MRF are higher than Sobel in the visual performance of the images. Although Sobel is the same as these two methods in terms of metrics, it is difficult to distinguish foreground and background images in real images.The segmentation results for MRF are obvious but the running time for MRF is too long in comparison with other classical learning methods. Since classical machine learning methods have a rigorous theoretical foundation and simple ideas, they have been shown to perform well when used for specific problems. However, the performance of different methods varied in the present study. ### 5.2 Discussion of Image Segmentation Results Using Deep Learning Methods In general, deep learning models are considerably superior to classical machine learning methods, and even the lowest MedT performance is still higher than the highest accuracy of classical machine learning methods. In EBHI-Seg, the Dice ratio evaluation index of MedT reaches $\sim$0.75. However, the MedT model size was larger and as a result the training time was too long. U-Net and Seg-Net have higher evaluation indexes than MedT, both of about 0.88. Among them, Seg-Net has the least training time and the lowest training model size. Because the normal category has fewer sample images than other categories, the evaluation metrics of the three deep learning methods in this category are significantly lower than those in other categories. The evaluation metrics of the three segmentation methods are significantly higher in the other categories, with Seg-Net averaging above 0.90 and MedT exceeding 0.80. ## 6 Conclusion and Future Work The present stduy introduced a publicly available colorectal pathology image dataset containing 5,710 magnified $400\times$ pathology images of six types of tumor differentiation stages. EBHI-Seg has high segmentation accuracy as well as good robustness. In the classical machine learning approach, segmentation experiments were performed using different methods and evaluation metrics analysis was carried out utilizing segmentation results. The highest and lowest Dice ratios are 0.65 and 0.30, respectively. The highest Precision and Recall values are 0.70 and 0.90, respectively, while the lowest values are 0.50 and 0.35, respectively. All three models performed well when using the deep learning method, with the highest Dice ratio reaching above 0.95 and both Precision and Recall values reaching above 0.90. The segmentation experiments using EBHI-Seg show that this dataset effectively perform the segmentation task in each of the segmentation methods. Furthermore, there are significant differences among the segmentation evaluation metrics. Therefore, EBHI-Seg is practical and effective in performing image segmentation tasks. ## Data Availability Statement EBHI-Seg is proved by “Research Project Ethics Certification” (No. 202229) from Cancer Hospital of China Medical University, Shenyang, China. EBHI-Seg is publicly available at: https://doi.org/10.6084/m9.figshare.21540159.v1 ## Author Contributions L. Shi: Data preparation, experiment, result analysis, paper writing; X. Li: Corresponding author, data collection, medical knowledge; W. Hu: Data collection, data preparation, paper writing; H. Chen: Data preparation, paper writing; J. Chen, Z. Fan, M. Gao, Y. Jing, G. Lu, D. Ma, Z. Ma, Q. Meng, D. Tang: Data preparation; H. Sun: Medical knowledge; M. Grzegorzek: Result analysis; S. Qi: Method; Y. Teng; Result analysis; C. Li; Corresponding author, data collection, method, experiment, result analysis, paper writing, proofreading. ## Funding This work is supported by the “National Natural Science Foundation of China” (No. 82220108007) and “Beijing Xisike Clinical Oncology Research Foundation” (No. Y-tongshu2021/1n-0379). ## Acknowledgements We thank Miss. Zixian Li and Mr. Guoxian Li for their important discussion in this work. ## Conflict of Interest The authors declare that they have no conflict of interest in this paper. ## References * [1] H. Sung, J. Ferlay, R. L. Siegel, M. Laversanne, I. Soerjomataram, A. Jemal, F. Bray, Global cancer statistics 2020: Globocan estimates of incidence and mortality worldwide for 36 cancers in 185 countries, CA: a cancer journal for clinicians 71 (3) (2021) 209–249. * [2] V. Pamudurthy, N. Lodhia, V. J. Konda, Advances in endoscopy for colorectal polyp detection and classification, in: Baylor University Medical Center Proceedings, Vol. 33, Taylor & Francis, 2020, pp. 28–35. * [3] J. Thijs, A. Van Zwet, W. Thijs, H. Oey, A. Karrenbeld, F. Stellaard, D. Luijt, B. Meyer, J. Kleibeuker, Diagnostic tests for helicobacter pylori: a prospective evaluation of their accuracy, without selecting a single test as the gold standard., American Journal of Gastroenterology (Springer Nature) 91 (10) (1996). * [4] R. Labianca, B. Nordlinger, G. Beretta, S. Mosconi, M. Mandalà, A. Cervantes, D. Arnold, Early colon cancer: Esmo clinical practice guidelines for diagnosis, treatment and follow-up, Annals of oncology 24 (2013) vi64–vi72. * [5] A. H. Fischer, K. A. Jacobson, J. Rose, R. Zeller, Hematoxylin and eosin staining of tissue and cell sections, Cold spring harbor protocols 2008 (5) (2008) pdb–prot4986. * [6] J. K. Chan, The wonderful colors of the hematoxylin–eosin stain in diagnostic surgical pathology, International journal of surgical pathology 22 (1) (2014) 12–32. * [7] V. Gupta, M. Vasudev, A. Doegar, N. Sambyal, Breast cancer detection from histopathology images using modified residual neural networks, Biocybernetics and Biomedical Engineering 41 (4) (2021) 1272–1287. * [8] T. Mathew, J. R. Kini, J. Rajan, Computational methods for automated mitosis detection in histopathology images: A review, Biocybernetics and Biomedical Engineering 41 (1) (2021) 64–82. * [9] C. Li, K. Wang, N. Xu, A survey for the applications of content-based microscopic image analysis in microorganism classification domains, Artificial Intelligence Review 51 (4) (2019) 577–646. * [10] J. Zhang, C. Li, Y. Yin, J. Zhang, M. Grzegorzek, Applications of artificial neural networks in microorganism image analysis: a comprehensive review from conventional multilayer perceptron to popular convolutional neural network and potential visual transformer, Artificial Intelligence Review (2022) 1–58. * [11] J. Zhang, C. Li, S. Kosov, M. Grzegorzek, K. Shirahama, T. Jiang, C. Sun, Z. Li, H. Li, Lcu-net: A novel low-cost u-net for environmental microorganism image segmentation, Pattern Recognition 115 (2021) 107885. * [12] P. Zhao, C. Li, M. M. Rahaman, H. Xu, H. Yang, H. Sun, T. Jiang, M. Grzegorzek, A comparative study of deep learning classification methods on a small environmental microorganism image dataset (emds-6): from convolutional neural networks to visual transformers, Frontiers in Microbiology 13 (2022). * [13] F. Kulwa, C. Li, J. Zhang, K. Shirahama, S. Kosov, X. Zhao, T. Jiang, M. Grzegorzek, A new pairwise deep learning feature for environmental microorganism image analysis, Environmental Science and Pollution Research (2022) 1–18. * [14] P. Ma, C. Li, M. M. Rahaman, Y. Yao, J. Zhang, S. Zou, X. Zhao, M. Grzegorzek, A state-of-the-art survey of object detection techniques in microorganism image analysis: from classical methods to deep learning approaches, Artificial Intelligence Review (2022) 1–72. * [15] F. Kulwa, C. Li, M. Grzegorzek, M. M. Rahaman, K. Shirahama, S. Kosov, Segmentation of weakly visible environmental microorganism images using pair-wise deep learning features, Biomedical Signal Processing and Control 79 (2023) 104168. * [16] J. Zhang, C. Li, M. M. Rahaman, Y. Yao, P. Ma, J. Zhang, X. Zhao, T. Jiang, M. Grzegorzek, A comprehensive survey with quantitative comparison of image analysis methods for microorganism biovolume measurements, Archives of Computational Methods in Engineering (2022). * [17] J. Zhang, C. Li, M. M. Rahaman, Y. Yao, P. Ma, J. Zhang, X. Zhao, T. Jiang, M. Grzegorzek, A comprehensive review of image analysis methods for microorganism counting: from classical image processing to deep learning approaches, Artificial Intelligence Review (2021) 1–70. * [18] M. M. Rahaman, C. Li, Y. Yao, F. Kulwa, M. A. Rahman, Q. Wang, S. Qi, F. Kong, X. Zhu, X. Zhao, Identification of covid-19 samples from chest x-ray images using deep learning: A comparison of transfer learning approaches, Journal of X-ray Science and Technology 28 (5) (2020) 821–839. * [19] H. Chen, C. Li, G. Wang, X. Li, M. M. Rahaman, H. Sun, W. Hu, Y. Li, W. Liu, C. Sun, et al., Gashis-transformer: A multi-scale visual transformer approach for gastric histopathological image detection, Pattern Recognition 130 (2022) 108827\. * [20] Y. Li, X. Wu, C. Li, X. Li, H. Chen, C. Sun, M. M. Rahaman, Y. Yao, Y. Zhang, T. Jiang, A hierarchical conditional random field-based attention mechanism approach for gastric histopathology image classification, Applied Intelligence (2022) 1–22. * [21] W. Hu, C. Li, X. Li, M. M. Rahaman, J. Ma, Y. Zhang, H. Chen, W. Liu, C. Sun, Y. Yao, et al., Gashissdb: A new gastric histopathology image dataset for computer aided diagnosis of gastric cancer, Computers in biology and medicine (2022) 105207. * [22] H. Chen, C. Li, X. Li, M. M. Rahaman, W. Hu, Y. Li, W. Liu, C. Sun, H. Sun, X. Huang, et al., Il-mcam: An interactive learning and multi-channel attention mechanism-based weakly supervised colorectal histopathology image classification approach, Computers in Biology and Medicine 143 (2022) 105265. * [23] W. Hu, H. Chen, W. Liu, X. Li, H. Sun, X. Huang, M. Grzegorzek, C. Li, A comparative study of gastric histopathology sub-size image classification: from linear regression to visual transformer, Frontiers in Medicine (2022). * [24] Y. Li, C. Li, X. Li, K. Wang, M. M. Rahaman, C. Sun, H. Chen, X. Wu, H. Zhang, Q. Wang, A comprehensive review of markov random field and conditional random field approaches in pathology image analysis, Archives of Computational Methods in Engineering 29 (1) (2022) 609–639. * [25] C. Sun, C. Li, J. Zhang, M. M. Rahaman, S. Ai, H. Chen, F. Kulwa, Y. Li, X. Li, T. Jiang, Gastric histopathology image segmentation using a hierarchical conditional random field, Biocybernetics and Biomedical Engineering 40 (4) (2020) 1535–1555. * [26] X. Li, C. Li, M. M. Rahaman, H. Sun, X. Li, J. Wu, Y. Yao, M. Grzegorzek, A comprehensive review of computer-aided whole-slide image analysis: from datasets to feature extraction, segmentation, classification and detection approaches, Artificial Intelligence Review (2022) 1–70. * [27] M. M. Rahaman, C. Li, X. Wu, Y. Yao, Z. Hu, T. Jiang, X. Li, S. Qi, A survey for cervical cytopathology image analysis using deep learning, IEEE Access 8 (2020) 61687–61710. * [28] M. Mamunur Rahaman, C. Li, Y. Yao, F. Kulwa, X. Wu, X. Li, Q. Wang, Deepcervix: A deep learning-based framework for the classification of cervical cells using hybrid deep feature fusion techniques, Computers in Biology and Medicine 136 (2021) 104649. * [29] W. Liu, C. Li, M. M. Rahaman, T. Jiang, H. Sun, X. Wu, W. Hu, H. Chen, C. Sun, Y. Yao, et al., Is the aspect ratio of cells important in deep learning? a robust comparison of deep learning methods for multi-scale cytopathology cell image classification: From convolutional neural networks to visual transformers, Computers in biology and medicine (2021) 105026. * [30] W. Liu, C. Li, N. Xu, T. Jiang, M. M. Rahaman, H. Sun, X. Wu, W. Hu, H. Chen, C. Sun, et al., Cvm-cervix: A hybrid cervical pap-smear image classification framework using cnn, visual transformer and multilayer perceptron, Pattern Recognition (2022) 108829. * [31] A. Chen, C. Li, S. Zou, M. M. Rahaman, Y. Yao, H. Chen, H. Yang, P. Zhao, W. Hu, W. Liu, et al., Svia dataset: A new dataset of microscopic videos and images for computer-aided sperm analysis, Biocybernetics and Biomedical Engineering (2022). * [32] S. Zou, C. Li, H. Sun, P. Xu, J. Zhang, P. Ma, Y. Yao, X. Huang, M. Grzegorzek, Tod-cnn: An effective convolutional neural network for tiny object detection in sperm videos, Computers in Biology and Medicine 146 (2022) 105543. * [33] I. Pacal, D. Karaboga, A. Basturk, B. Akay, U. Nalbantoglu, A comprehensive review of deep learning in colon cancer, Computers in Biology and Medicine 126 (2020) 104003. * [34] E. Miranda, M. Aryuni, E. Irwansyah, A survey of medical image classification techniques, in: 2016 International Conference on Information Management and Technology (ICIMTech), IEEE, 2016, pp. 56–61. * [35] H. Kotadiya, D. Patel, Review of medical image classification techniques, in: Third International Congress on Information and Communication Technology, Springer, 2019, pp. 361–369. * [36] E. Trivizakis, G. S. Ioannidis, I. Souglakos, A. H. Karantanas, M. Tzardi, K. Marias, A neural pathomics framework for classifying colorectal cancer histopathology images based on wavelet multi-scale texture analysis, Scientific Reports 11 (1) (2021) 1–10. * [37] S. P. Oliveira, P. C. Neto, J. Fraga, D. Montezuma, A. Monteiro, J. Monteiro, L. Ribeiro, S. Gonçalves, I. M. Pinto, J. S. Cardoso, Cad systems for colorectal cancer from wsi are still not ready for clinical acceptance, Scientific Reports 11 (1) (2021) 1–15. * [38] T. Kausar, A. Kausar, M. A. Ashraf, M. F. Siddique, M. Wang, M. Sajid, M. Z. Siddique, A. U. Haq, I. Riaz, Sa-gan: Stain acclimation generative adversarial network for histopathology image analysis, Applied Sciences 12 (1) (2021) 288. * [39] J. Wei, A. Suriawinata, B. Ren, X. Liu, M. Lisovsky, L. Vaickus, C. Brown, M. Baker, M. Nasir-Moin, N. Tomita, et al., Learn like a pathologist: curriculum learning by annotator agreement for histopathology image classification, in: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2021, pp. 2473–2483. * [40] M. Bilal, Y. W. Tsang, M. Ali, S. Graham, E. Hero, N. Wahab, K. Dodd, H. Sahota, W. Lu, M. Jahanifar, et al., Ai based pre-screening of large bowel cancer via weakly supervised learning of colorectal biopsy histology images, medRxiv (2022). * [41] M. P. De Leon, C. Di Gregorio, Pathology of colorectal cancer, Digestive and Liver Disease 33 (4) (2001) 372–388. * [42] H. S. Cooper, L. M. Deppisch, E. I. Kahn, R. Lev, P. N. Manley, R. R. Pascal, A. H. Qizilbash, R. R. Rickert, J. F. Silverman, J. A. Wirman, Pathology of the malignant colorectal polyp, Human pathology 29 (1) (1998) 15–26. * [43] W. Ren, J. Yu, Z.-M. Zhang, Y.-K. Song, Y.-H. Li, L. Wang, Missed diagnosis of early gastric cancer or high-grade intraepithelial neoplasia, World Journal of Gastroenterology: WJG 19 (13) (2013) 2092. * [44] J. R. Jass, L. H. Sobin, Histological typing of intestinal tumours, Springer Science & Business Media, 2012. * [45] K. J. Spring, Z. Z. Zhao, R. Karamatic, M. D. Walsh, V. L. Whitehall, T. Pike, L. A. Simms, J. Young, M. James, G. W. Montgomery, et al., High prevalence of sessile serrated adenomas with braf mutations: a prospective study of patients undergoing colonoscopy, Gastroenterology 131 (5) (2006) 1400–1407. * [46] S. C. Li, L. Burgart, Histopathology of serrated adenoma, its variants, and differentiation from conventional adenomatous and hyperplastic polyps, Archives of pathology & laboratory medicine 131 (3) (2007) 440–445. * [47] K. H. Zou, S. K. Warfield, A. Bharatha, C. M. Tempany, M. R. Kaus, S. J. Haker, W. M. Wells III, F. A. Jolesz, R. Kikinis, Statistical validation of image segmentation quality based on a spatial overlap index1: scientific reports, Academic radiology 11 (2) (2004) 178–189. * [48] P. Jaccard, Étude comparative de la distribution florale dans une portion des alpes et des jura, Bull Soc Vaudoise Sci Nat 37 (1901) 547–579. * [49] S. Naz, H. Majeed, H. Irshad, Image segmentation using fuzzy clustering: A survey, in: 2010 6th international conference on emerging technologies (ICET), IEEE, 2010, pp. 181–186. * [50] N. M. Zaitoun, M. J. Aqel, Survey on image segmentation techniques, Procedia Computer Science 65 (2015) 797–806. * [51] N. Dhanachandra, K. Manglem, Y. J. Chanu, Image segmentation using k-means clustering algorithm and subtractive clustering algorithm, Procedia Computer Science 54 (2015) 764–771. * [52] H. Deng, D. A. Clausi, Unsupervised image segmentation using a simple mrf model with a new implementation scheme, Pattern recognition 37 (12) (2004) 2323–2335. * [53] C. Huang, X. Li, Y. Wen, An otsu image segmentation based on fruitfly optimization algorithm, Alexandria Engineering Journal 60 (1) (2021) 183–188. * [54] M. S. H. Khiyal, A. Khan, A. Bibi, Modified watershed algorithm for segmentation of 2d images., Issues in Informing Science & Information Technology 6 (2009). * [55] H. Zhang, Q. Zhu, X.-f. Guan, Probe into image segmentation based on sobel operator and maximum entropy algorithm, in: 2012 International Conference on Computer Science and Service System, IEEE, 2012, pp. 238–241. * [56] V. Badrinarayanan, A. Kendall, R. Cipolla, Segnet: A deep convolutional encoder-decoder architecture for image segmentation, IEEE transactions on pattern analysis and machine intelligence 39 (12) (2017) 2481–2495. * [57] O. Ronneberger, P. Fischer, T. Brox, U-net: Convolutional networks for biomedical image segmentation, in: International Conference on Medical image computing and computer-assisted intervention, Springer, 2015, pp. 234–241. * [58] J. M. J. Valanarasu, P. Oza, I. Hacihaliloglu, V. M. Patel, Medical transformer: Gated axial-attention for medical image segmentation, in: International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2021, pp. 36–46.
Thermalization of long range Ising model in different dynamical regimes: a full counting statistics approach. Nishan Ranabhat1,2$\star$, Mario Collura1 1 SISSA - International School for Advanced Studies, via Bonomea 265, 34136 Trieste, Italy 2 The Abdus Salam International Centre for Theoretical Physics, Strada Costiera 11, 34151 Trieste, Italy ⋆<EMAIL_ADDRESS> ## Abstract We study the thermalization of the transverse field Ising chain with a power law decaying interaction $\sim 1/r^{\alpha}$ following a global quantum quench of the transverse field in two different dynamical regimes. The thermalization behavior is quantified by comparing the full probability distribution function (PDF) of the evolving states with the corresponding thermal state given by the canonical Gibbs ensemble (CGE). To this end, we used the matrix product state (MPS)-based Time Dependent Variational Principle (TDVP) algorithm to simulate both real time evolution following a global quantum quench and the finite temperature density operator. We observe that thermalization is strongly suppressed in the region with strong confinement for all interaction strengths $\alpha$, whereas thermalization occurs in the region with weak confinement. ###### Contents 1. 1 Introduction 2. 2 Model and Methods 3. 3 Numerical details 1. 3.1 Real and imaginary time evolution 2. 3.2 Extraction of effective temperature of a global quench 4. 4 Results 1. 4.1 Quench to dynamical ferromagnetic regime 2. 4.2 Quench to dynamical paramagnetic regime 5. 5 Conclusion 6. A Exact results for smaller systems 7. B Simulations details 1. B.1 Simulation of finite temperature density operator 2. B.2 Calculating full counting statistics with MPS 3. B.3 Errors and data convergence 8. C Thermal phase transition in long range Ising model 9. D Confinement dynamics in different regimes ## 1 Introduction The investigation of non-equilibrium dynamics in isolated many-body systems has garnered significant attention in recent decades, owing to advancements in the manipulation of synthetic quantum systems in laboratory settings [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11] and the development of analytical and numerical techniques [12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25]. An enduring question in quantum many-body dynamics pertains to the potential thermalization of a closed system that has been perturbed from equilibrium. Thermalization implies that the long-term behavior of a dynamic system can be anticipated using the principles of statistical mechanics. Generally, in the case of a non-integrable closed system, one would expect thermalization in accordance with the Eigenstate Thermalization Hypothesis (ETH) [26, 27, 28, 29, 30]. Nonetheless, certain studies have presented contradictory evidence, at least within their specific regime and time scales of investigation [31, 32, 33, 34]. In a scenario where a closed system is initially prepared in a generic state, denoted as $\ket{\psi_{i}}$ (which is not an eigenstate of the Hamiltonian), and subsequently evolved using unitary dynamics with a non- integrable Hamiltonian, $\hat{H}$, thermalization is said to occur if the local observables eventually relax to an equilibrium state that corresponds to the predictions of the thermal ensemble, $\langle\hat{O}\rangle_{t|t\rightarrow\infty}\rightarrow\langle\hat{O}\rangle_{\text{eq}}=\langle\hat{O}\rangle_{\text{MCE}}=\frac{1}{N_{E_{i},\delta\mathcal{E}}}\sum_{\absolutevalue{E_{n}-E_{i}}<\delta\mathcal{E}}\matrixelement{\psi_{n}}{\hat{O}}{\psi_{n}},$ (1) $\langle\hat{O}\rangle_{t|t\rightarrow\infty}\rightarrow\langle\hat{O}\rangle_{\text{eq}}=\langle\hat{O}\rangle_{\text{CGE}}=\frac{\text{Tr}(\rho_{\beta}\hat{O})}{\text{Tr}(\rho_{\beta})}.$ (2) Equation (1) indicates thermalization according to the microcanonical ensemble (MCE). The summation encompasses all the eigenstates of the Hamiltonian within a narrow energy range $\delta\mathcal{E}$ centered around the initial energy $E_{i}=\matrixelement{\psi_{i}}{\hat{H}}{\psi_{i}}$. The normalization factor $N_{E_{i},\delta\mathcal{E}}$ tallies the energy eigenstates within this range, spanning $2\delta\mathcal{E}$. This approach requires either full diagonalization [28, 35, 36] or partial diagonalization centered on the initial energy density [37] of the Hamiltonian. Consequently, computational limitations arise as a function of the system size. Equation (2) implies thermalization in accordance with the canonical Gibbs ensemble (CGE). The trace is performed over the density operator $\hat{\rho}_{\beta}$, defined as the inverse temperature $\beta=1/T$, which is determined by the system’s initial energy $E_{i}=\frac{\text{Tr}(\hat{\rho}_{\beta}\hat{H})}{\text{Tr}(\hat{\rho}_{\beta})}$. Further elaboration on how to extract $\beta$ and $\hat{\rho}_{\beta}$ is provided in section 3.2. Equations (1) and (2) are recognized as the conditions for strong thermalization. An alternative weak thermalization condition occurs when the time-averaged local order parameter converges to the thermal prediction [33, 34]. Disordered systems demonstrating many-body localization impede thermalization [38, 39, 40, 41]. In clean systems, dynamical confinement hinders information propagation and the thermalization process [42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53]. The Long Range Ising model (LRIM) exhibits confinement [44, 45] as outlined in Appendix D. A recent study [4] observed the suppression of thermalization in the confined regime of LRIM simulated with trapped ions. Furthermore, employing high-scale exact diagonalization, the validity of ETH has been examined for various interaction strength parameters, denoted as $\alpha$ (see section 2), in LRIM. The results indicate that a strong ETH typically holds at least within the range $\alpha\geq 0.6$ [35]. In our previous work [54], we explored the relaxation of order parameter statistics following a global quench, revealing two distinct dynamical regimes based on the gaussification of the full counting statistics (FCS) of subsystem magnetization. Building on this foundation, the present study investigates the thermalization of LRIM under the CGE framework following a global quench into different dynamical regimes. We evaluate thermalization using the most rigorous criteria by comparing the FCS of the time-evolving state post-global quench with that of the corresponding thermal state. The remainder of this paper is organized as follows: In Section 2, we introduce the model, the order parameter, its distribution, and a metric for quantifying the proximity of the time-evolved state to the thermal state. Section 3 provides comprehensive details of the numerical methods, quench protocol, and extraction of effective temperatures associated with the global quench. Specific details on simulating finite temperature density operators and calculating the full counting statistics of the order parameter is provided in Appendix B.1 and Appendix B.2 respectively. Section 4 presents the outcomes of our study. Appendix B.3 details the error analysis of the numerical results and Appendices C and D provide supplementary information on thermal phase transitions and correlation propagation in the Long Range Ising model. Finally, we conclude by summarizing our findings and suggesting potential avenues for future research in Section 5. ## 2 Model and Methods We investigate the ferromagnetic long-range Ising model (LRIM) described by the Hamiltonian in Equation (3), $\hat{H}(J,\alpha,h)=-\frac{1}{\mathcal{K}(\alpha)}\sum_{i<j}^{N}\frac{\absolutevalue{J}}{|i-j|^{\alpha}}\hat{s}^{x}_{i}\hat{s}^{x}_{j}-h\sum_{i=1}^{N}\hat{s}^{z}_{i}$ (3) where $\hat{s}^{\mu}_{i},\mu={x,y,z}$ is the spin one-half matrices at site $i$. We consider open boundary condition, that is relevant to existing experimental setups. The ferromagnetic interaction between two spins falls as the inverse power of the distance between them and is parameterized by the interaction strength parameter $\alpha$. For $\alpha\leq 1$, the inverse power-law interaction series diverges with the lattice size and is normalized using the Kac normalization constant, as defined in Equation (4). $\mathcal{K}(\alpha)=\frac{1}{N-1}\sum_{i<j}^{N}\frac{1}{|i-j|^{\alpha}}=\frac{1}{N-1}\sum_{n=1}^{N}\frac{N-n}{n^{\alpha}}.$ (4) This normalization ensures the intensivity of the energy density in the regime $\alpha\leq 1$. The static and dynamic behaviors of this model are strongly influenced by the interaction strength parameter $\alpha$. At $\alpha=\infty$, the model simplifies to the transverse field Ising model (TFIM), which can be solved exactly by mapping it to a system of spinless fermions through Jordan- Wigner transformations [55]. TFIM exhibits a quantum phase transition from the ferromagnetic phase to the paramagnetic phase at $h=J/2$. This quantum phase transition persists as $\alpha$ decreases, with the transition point shifting towards higher values of the magnetic field $h$ [56, 57, 5]. At the opposite extreme of $\alpha=0$, we have a fully connected regime that is amenable to analytical treatment for both the static and dynamic properties [58, 59, 54]. For $\alpha<2$, this model displays a long-range ferromagnetic order at low finite temperatures [60, 61]. Given the absence of spontaneous symmetry breaking in finite systems and the $\mathbb{Z}_{2}$ symmetry of the model in Equation (3), the finite-temperature states with ferromagnetic order also exhibit $\mathbb{Z}_{2}$ symmetry (see appendix C). The regime with $\alpha<2$ is particularly intriguing and features a wealth of exotic phenomena such as prethermalization [11], nonlinear propagation of light cones[62, 63], dynamical phase transitions [64, 65, 66, 2, 3, 54, 67, 68, 69], and dynamical confinement [44, 45, 4]. Furthermore, this model has garnered significant attention owing to its experimental relevance, particularly in systems involving trapped ions with adjustable transverse field strengths and interaction ranges [1, 2, 3, 4, 11]. The complete information of a generic time evolving quantum state, expanded in the computational basis $\ket{\psi_{t}}=\sum_{\\{\sigma_{i}\\}}C_{\\{\sigma_{i}\\}}(t)\ket{\sigma_{1},\sigma_{2},\ldots,\sigma_{i},\ldots,\sigma_{N}}$, is encapsulated within the set of time dependent coefficients $\\{C_{\\{\sigma_{i}\\}}(t)\\}$. In many-body systems, these coefficients scales exponentially with the number of spins, rendering their study exceedingly challenging. A common approach for investigating dynamics in such systems is to monitor the evolution of the expectation value of a local observable $\matrixelement{\psi_{t}}{\hat{O}}{\psi_{t}}$, such as the order parameter in systems exhibiting order-disorder transitions. A more robust strategy involves tracking the full probability distribution function (PDF) of this observable, which provides comprehensive information on quantum fluctuations in the system, including all moments and cumulants. Specifically, when the operator $\hat{O}$ is diagonal in computational basis, the corresponding PDF is defined as $P(O,t)=\sum_{\\{\sigma_{i}\\}}\absolutevalue{C_{\\{\sigma_{i}\\}}(t)}^{2}\hskip 14.22636pt\text{for all }\\{\sigma_{i}\\}:\matrixelement{\sigma_{1},\sigma_{2},\ldots,\sigma_{i},\ldots,\sigma_{N}}{\hat{O}}{\sigma_{1},\sigma_{2},\ldots,\sigma_{i},\ldots,\sigma_{N}}=O,$ (5) which represents the histogram of the squared coefficients of the many-body wave function within the range of possible outcomes of the measurements of $\hat{O}$. The PDF allows for straightforward calculation of moments (or cumulants) of any order. In this study, we employ the eigenvectors of the total spin operator in the longitudinal direction, that is, $\hat{S}^{x}=\sum_{i=1}^{N}\hat{s}_{i}^{x}$, as our computational basis. In this context, the order parameter of interest is the longitudinal magnetization defined for a subsystem of size $l$ within a system of $N$ spins: $\hat{M}_{l}=\sum_{i=1}^{l}\hat{s}^{x}_{i}.$ (6) The observable $\hat{M}_{l}$ is suitable because it typically relaxes to a stationary state [70], ultimately approaching a stationary statistical distribution in a subsystem of dimension $l$. In addition, because $\hat{M}_{l}$ is diagonal in computational basis, the definition in Equation (5) applies. The probability distribution function of the subsystem magnetization $\hat{M}_{l}$ within a generic state $\hat{\rho}_{t}$ (whether pure or mixed) is given by $P_{l}(m,t)=\text{Tr}(\hat{\rho}_{t}\delta(\hat{M}_{l}-m)),$ (7) which can be Fourier-transformed into an integral form: $P_{l}(m,t)=\int_{-\pi}^{\pi}\frac{d\theta}{2\pi}e^{-i\theta m}\text{Tr}\big{(}\hat{\rho}_{t}e^{i\theta\hat{M}_{l}}\big{)},$ (8) where $G_{l}(\theta,t)=\text{Tr}\big{(}\hat{\rho}_{t}e^{i\theta\hat{M}_{l}}\big{)}$ denotes the moment-generating function. Given that the Hamiltonian (3) involves a system of spin one-half particles, the values of $m$ span either integers or half-integers within the range $m\in\Big{\\{}-\frac{l}{2},-\frac{l}{2}+1,\ldots,\frac{l}{2}-1,\frac{l}{2}\Big{\\}}$ depending on whether $l$ is even or odd. In Appendix B.2 we illustrate the detailed calculation of $G_{l}(\theta,t)$ with matrix product state (MPS) representation. Historically, the PDF has been studied as the full counting statistic (FCS) of electron fluctuations in mesoscopic systems [71, 72, 73]. More recently, FCS has been explored in quantum many-body systems in both equilibrium and non-equilibrium scenarios [18, 74, 75, 19, 76, 43, 54, 77, 78, 79]. To assess thermalization, we introduce a metric called "Distance to Thermalization", $\text{DT}(t)$, initially introduced in [80]. This metric quantifies the Euclidean distance between the probability distribution function (PDF) of the order parameter at time $t$ following a quantum quench, denoted as $P_{l}(m,t)$, and the corresponding thermal PDF, represented as $P_{l}^{\text{TH}}(m)$. Mathematically, it is defined as $\text{DT}(t)=\sqrt{\sum_{m}\big{[}P_{l}(m,t)-P_{l}^{\text{TH}}(m)\big{]}^{2}}.$ (9) It is noteworthy that the convergence of the PDF provides a more rigorous criterion for thermalization than the convergence of the expectation value. This is because the former implies the latter, whereas the reverse is not necessarily true. A similar approach has been employed in previous studies to investigate thermalization dynamics [80, 37, 81, 79]. Comprehensive details of how to extract the thermal state corresponding to a global quantum quench are discussed in Section 3.2. In cases where the system undergoes thermalization, $\text{DT}(t)$ is expected to converge to zero in the long time limit. Figure 1: (a) Global Quench protocol: System is initialized as the ground state of a trivial hamiltonian $\hat{H}_{i}$ (in our case the initial state is a Greenberger-Horne-Zeilinger (GHZ) state), at time $t=0$ the system is suddenly quenched to a final Hamiltonian $\hat{H}_{f}$ and the initial state is unitarily evolved with the final Hamiltonian. (b) The nonequilibrium state following a global quantum quench can exhibit different relaxation behavior; Path 1: a direct relaxation to thermal equilibrium with a single time scale, Path 2: a quick relaxation to a long lived prethermal state eventually followed a relaxation to thermal equilibrium, Path 3: a strong retention of initial memory and suppression of relaxation to thermal equilibrium. ## 3 Numerical details ### 3.1 Real and imaginary time evolution The numerical simulations in this study are classified into two distinct categories: * • real time evolution of pure state following a global quench. * • simulation of finite temperature density operators. For both of these simulation tasks, we employ the MPS-based Time Dependent Variational Principle (TDVP) algorithm [23, 24] with second order integration scheme. This choice affords us a significant advantage over exact diagonalization methods, allowing us to simulate systems of much larger sizes than can be accommodated by the current exact diagonalization techniques. The quench protocol implemented in this study, with energy rescaled to $\absolutevalue{J}=1$, is as follows: At time $t=0$, the system is prepared in the ground state of the Hamiltonian $\hat{H}_{i}(\alpha,0)$, which takes the form of a $\mathbb{Z}_{2}$ symmetric Greenberger-Horne-Zeilinger (GHZ) state oriented along the longitudinal direction: $\ket{\psi_{i}}=\frac{1}{\sqrt{2}}(\ket{\rightarrow,\ldots\rightarrow,\rightarrow,\rightarrow\ldots,\rightarrow}_{x}+\ket{\leftarrow,\ldots\leftarrow,\leftarrow,\leftarrow\ldots,\leftarrow}_{x}),$ (10) GHZ state is characterized by the PDF $P^{\text{GHZ}}_{l}(m)=\frac{\delta_{m,\absolutevalue{l}/2}}{2}$ which is sharply bimodal with two peaks at $m=l/2$ and $m=-l/2$ respectively. Equation (10) can be explicitly represented as an exact Matrix Product State (MPS) with bond dimension $\chi=2$. Subsequently, a global quench is initiated along the transverse field $h$ to a final Hamiltonian $\hat{H}_{f}(\alpha,h_{f})$ and the system is evolved unitarily using the expression $\ket{\psi_{t+dt}}=e^{-idt\hat{H}_{f}}\ket{\psi_{t}}$. The evolution is monitored by calculating the Full Counting Statistics (FCS) of the order parameter at each time step. The details of the quench protocol is pictorially represented in Figure 1 (a). The finite temperature density operator is simulated by an imaginary time evolution starting from a maximally mixed state at infinite temperature. Additional details pertaining to the calculation of the thermal density operator are provided in Appendix B.1. For both sets of simulations, we maintain a fixed maximum bond dimension of the MPS at $\chi_{max}=128$. Furthermore, Trotter time steps of $dt=0.05$ and $d\beta=0.001$ are used for real and imaginary time evolution, respectively. There is a finite time-step error of $O(dt^{3})$ per time step and $O(dt^{2})$ per unit time [82]. In Appendix B.3 we access the accuracy of the TDVP data by comparing the TDVP results with the exact results obtained by the full diagonalization of a system of size $N=14$. Furthermore, we test the convergence of the data by calculating the relative error for three increasing bond dimensions. ### 3.2 Extraction of effective temperature of a global quench A global quantum quench $\hat{H}_{i}(\alpha,0)\xrightarrow{}\hat{H}_{f}(\alpha,h)$ in an isolated system adds an extensive amount of energy to the system. Consequently, the system relaxes to a state at a higher energy level than the ground state of the post-quench Hamiltonian [70], $\lim_{N\to\infty}\frac{1}{N}\frac{\matrixelement{\psi_{t}}{\hat{H}_{f}}{\psi_{t}}}{\innerproduct{\psi_{t}}{\psi_{t}}}>\lim_{N\to\infty}\frac{1}{N}\frac{\matrixelement{\psi_{0}}{\hat{H}_{f}}{\psi_{0}}}{\innerproduct{\psi_{0}}{\psi_{0}}}$ (11) where $\ket{\psi_{0}}$ is the ground state of the post-quench Hamiltonian $\hat{H}_{f}(\alpha,h)$. The left hand side of Equation (11) is a conserved quantity because the real time evolution of $\ket{\psi_{t}}$ is unitary. For every global quantum quench we can attribute an effective temperature $\beta_{\text{eff}}$ which is the temperature at which the thermal energy density above the ground state of the post-quench Hamiltonian matches the conserved energy density of the system, $\frac{1}{N}\frac{\matrixelement{\psi_{t}}{\hat{H}_{f}}{\psi_{t}}}{\innerproduct{\psi_{t}}{\psi_{t}}}=\frac{1}{N}\frac{\text{Tr}(\hat{\rho}_{\beta}\hat{H}_{f})}{\text{Tr}(\hat{\rho}_{\beta})}.$ (12) Figure 2: Numerical extraction of $\beta_{\text{eff}}$ corresponding to a global quantum quench. The horizontal black dashed line represent the energy density attributed to the quench. The colored lines represents the energy density as the function of inverse temperature $\beta$ for the corresponding post-quench parameter (in legend). The point at which the colored lines intersects the black dashed lines represents $\beta_{\text{eff}}$ for the corresponding post-quench parameters (represented by vertical colored lines). The effective temperature is extracted by solving Equation (12). The left hand side of the equation is trivially calculated as $\matrixelement{\psi_{t}}{\hat{H}_{f}}{\psi_{t}}=\matrixelement{\psi_{i}}{e^{it\hat{H}_{f}}\hat{H}_{f}e^{-it\hat{H}_{f}}}{\psi_{i}}=\matrixelement{\psi_{i}}{\hat{H}_{f}}{\psi_{i}}$, and the right hand side can be calculated for a series of $\beta$ by numerically solving Equation (23) and calculating the energy density at each instance. The precision of $\beta_{\text{eff}}$ depends on the trotter steps $d\beta$ in the solution to equation (23). In Fig. 2 we plot the numerical solution of equation (12). The energy density attributed to quench (represented by the black dashed line) in our setup is independent of the post-quench parameters because the spin-spin interaction term in the Hamiltonian (3) is normalized with the Kac normalization (4), whereas the expectation value $h\matrixelement{\psi_{i}}{\sum_{j}\hat{s}^{z}_{j}}{\psi_{i}}$ taken over the transverse field term is trivially zero. If we extend the simulation to a larger $\beta$ (i.e., lower temperature), all curves will converge to the ground state energy density of $\hat{H}_{f}$ at the corresponding post-quench parameters. Once $\beta_{\text{eff}}$ is extracted, we can calculate the corresponding thermal PDF, $P^{\text{TH}}(m)=P^{\beta_{\text{eff}}}(m)$, using equation (8). ## 4 Results The global quantum quench, as discussed in Section 3.1, induces a dynamical quantum phase transition (DQPT) [83, 84] in LRIM, which has garnered extensive attention in recent years. This transition falls into two distinct categories: the first, known as DQPT-I, is characterized by distinctive behaviors in the time-averaged local order parameter following a global quench across the dynamical critical point [54, 64, 2], and the second, DQPT-II, is marked by non-analytic cusps in the Loschmidt echo rate [67, 68, 69, 85, 3]. In LRIM, the dynamical critical points for DQPT-I and DQPT-II coincide at approximately $h^{\text{dyn}}_{c}\approx 0.5$ for $\alpha\leq 2$ [64, 85]. When $h_{f}<h^{\text{dyn}}_{c}$, the system is in the dynamical ferromagnetic phase, which strongly retains the ferromagnetic order of the initial GHZ state following a global quantum quench. This is evident from the persistent oscillation of $P_{l}(m,t)$ around $P^{\text{GHZ}}_{l}(m)$. Conversely, when $h_{f}>h^{\text{dyn}}_{c}$, the system transits to the dynamical paramagnetic phase, characterized by the rapid dissolution of the initial ferromagnetic order of the initial GHZ state following a global quantum quench. This is signified by the Gaussification of $P_{l}(m,t)$ [54, 64]. The comprehensive dynamical phase diagram and universality behavior related to the dynamical phase transition in LRIM remain active areas of investigation [86, 87, 88, 66]. Figure 3: First row: Time evolution of the metric $\text{DT}(t)$ following a global quantum quench to three interaction strength values $\alpha\in\\{0.0,1.5,1.9\\}$ and transverse field $h_{f}=0.3$ at three different subsystem sizes $l=\\{20,60,100\\}$. All three points are in dynamical ferromagnetic phases[64, 54]. Second row: $P_{l}(m)$ versus $m$ for $m\in[0,l/2]$ with $l=100$ at four time different slices $t=\\{2,6,20,50\\}$. $P_{l}(m)$ versus $m$ for $m\in[-l/2,0)$ is its mirror image. The black dashed curve represents the thermal PDF, $P^{\text{TH}}(m)$ attributed to the corresponding global quantum quenches. Our primary objective is to examine the convergence behavior of the metric $\text{DT}(t)$ in two distinct dynamical phases of the LRIM. The convergence of $\text{DT}(t)$ towards zero is an indicator of thermalization within a particular phase under consideration. We initialize the system as $\mathbb{Z}_{2}$ symmetric GHZ state presented in Equation (10), which represents the ground state of the Hamiltonian given in Equation (3). This choice is made because the model (3) undergoes a thermal transition from a paramagnetic phase, characterized by a Gaussian probability density function (PDF) at high temperatures, to a ferromagnetic phase with a $\mathbb{Z}_{2}$ symmetric bimodal PDF, as comprehensively detailed in Appendix C. We maintain that $\alpha<2$ is crucial as this region exhibits an interesting landscape encompassing both finite temperature phase transitions [89, 90] and dynamic phase transitions [64, 65, 66, 2, 3, 54, 67, 68, 69]. Specifically, we consider three distinct values for interaction strength, namely $\alpha\in{0.0,1.5,1.9}$. At $\alpha=0.0$, the system exhibits integrability because of its full connectivity and complete permutation symmetry, thereby leading to a lack of thermalization[17]. On the other hand, the choices of $\alpha=1.5$ and $\alpha=1.9$ are motivated by the relatively faster equilibration and Gaussification of the PDF following a quench in the dynamical paramagnetic phase, as previously observed [54]. Figure 4: First row: Time evolution of the metric $\text{DT}(t)$ following a global quantum quench to three interaction strength values $\alpha\in\\{0.0,1.5,1.9\\}$ and transverse field $h_{f}=0.6$ at three different subsystem sizes $l=\\{20,60,100\\}$. All three points are in dynamical paramagnetic phases[64, 54]. Second row: $P_{l}(m)$ versus $m$ for $m\in[0,l/2]$ with $l=100$ at four time different slices $t=\\{2,6,20,50\\}$. $P_{l}(m)$ versus $m$ for $m\in[-l/2,0)$ is its mirror image. The black dashed curve represents the thermal PDF, $P^{\text{TH}}(m)$ attributed to the corresponding global quantum quenches. ### 4.1 Quench to dynamical ferromagnetic regime Figure 3 shows the temporal evolution of $\text{DT}(t)$ following a global quantum quench of the transverse field to $h_{f}=0.3$ with $\alpha=0.0,1.5,1.9$ for subsystem sizes $l=20,60,100$. Notably, all these points belong to the dynamical ferromagnetic phase [64, 54]. For all three quenches, a persistent oscillation in $\text{DT}(t)$ is evident, indicating that the initial ferromagnetic order is strongly retained and thermalization is suppressed. This behavior aligns with the relaxation mode represented by Path 3 in figure 1(b). Specifically, $\alpha=0.0$ is in the integrable regime; therefore, thermalization is expected to be absent [26] whereas we anticipate thermalization for quenches with $\alpha=1.5$ and $\alpha=1.9$. The apparent suppression of thermalization can be attributed to the confinement behavior. The long-range interaction of the model effectively confines low-energy domain wall kinks into heavier quasiparticles that typically travel slower than free quasiparticles, thereby suppressing the spread of correlations in the system [44, 45]. Consequently, thermalization is still expected but only at significantly longer time scales [47]. Appendix D details the confinement behavior in LRIM where we observe the spreading of connected correlation function $\expectationvalue{\hat{s}^{x}_{k}\hat{s}^{x}_{k+\Delta}}_{c}$ for $\alpha=1.9$ and $h_{f}=0.3$ shows a strong temporal suppression. In figure 3(d), (e), and (f), the colored scattered plots depict $P_{l}(m)$ as a function of $m$ at four distinct time intervals post-quench. The black dashed curve represents the Probability Density Function (PDF) of the expected thermal state, $P_{l}^{\text{TH}}(m)$. We observe that the time-evolving $P_{l}(m)$ oscillates persistently around $P_{l}^{\text{TH}}(m)$. Of particular importance is the observation that, in all three cases, the thermal PDFs are bimodal, indicating the presence of long-range ferromagnetic order. This observation suggests that if the system eventually thermalizes for these post-quench parameters at extended time scales, it would exhibit a long-range ferromagnetic order. This finding further strengthens the argument that this is indeed a dynamical ferromagnetic phase. ### 4.2 Quench to dynamical paramagnetic regime Figure 4 illustrates the temporal evolution of $\text{DT}(t)$ following a global quantum quench of the transverse field to $h_{f}=0.6$, with $\alpha=0.0,1.5,1.9$ for subsystem sizes $l=20,60,100$. These points are located within the dynamical paramagnetic phase [64, 54]. Notably, these quenches exhibit a distinct relaxation behavior of $\text{DT}(t)$ compared with the previous cases. In Figure 4(a), we observe rapid equilibration for all values of $l$. However, it is essential to highlight that $\text{DT}(t)$ remains at or above the order of $O(10^{-1})$ following equilibration, which suggests a lack of thermalization. This behavior aligns with expectations, because $\alpha=0$ represents an integrable point. For $\alpha=1.5$, $\text{DT}(t)$ does not exhibit stable equilibration (see figure 4 (b)); Finally, when $\alpha=1.9$, we observe equilibration for $l=60,100$ (see figure 4 (c)). $\text{DT}(t)$ exhibits a stable oscillation around a constant value of approximately $O(10^{-3})$. A more comprehensive picture is shown in Fig. 4(f), where the late-time PDF perfectly overlaps with the corresponding thermal PDF represented by a black dashed curve. This is indicative of thermalization of the corresponding quench. Although we observed signatures of thermalization, the system is not in a de-confined phase [4, 91]. In Appendix D, the connected correlation function $\expectationvalue{\hat{s}^{x}_{k}\hat{s}^{x}_{k+\Delta}}_{c}$ for $\alpha=1.9$ and $h_{f}=0.6$ still exhibits weaker temporal suppression. A recent study observed a de-confinement transition for a system of up to 31 spins for a much higher value of the transverse field [4]. This suggests that, although strong confinement suppresses thermalization, signatures of thermalization can be observed in the presence of weak confinement. This suggests that while strong confinement suppresses thermalization, signatures of thermalization can still be detected in the presence of weak confinement. Figure 5: Time evolution of domain wall kinks, $\langle\hat{k}\rangle$, following a global quantum quench to three interaction strength values $\alpha\in\\{0.0,1.5,1.9\\}$ (Panels (a), (b), and (c) respectively) and $h_{f}=0.3$ and $h_{f}=0.6$. The dashed horizontal lines represent the expected thermal value of domain wall kinks, $\langle\hat{k}\rangle^{\text{TH}}$ corresponding to the quenches. To further support this observation we study the post quench temporal evolution of domain wall kinks defined as, $\hat{k}=\sum_{j=1}^{l-1}\frac{1-\hat{s}^{x}_{i}\hat{s}^{x}_{i+1}}{2}.$ (13) $\hat{k}$ counts the number of nearest neighbor kinks in the $\hat{x}$ direction within subsystem $l$. Because confinement bounds the domain walls kinks into heavier quasiparticles, it is a relevant parameter to study. In Figure 5, we illustrate the temporal evolution of the average domain-wall kinks, $\langle\hat{k}\rangle$, following a global quantum quench. As anticipated, quenches to the dynamical ferromagnetic phase with $h_{f}=0.3$ display persistent oscillations around the thermal value, indicating a lack of thermalization. Conversely, for quenches to the dynamical paramagnetic phase with $h_{f}=0.6$, we observe distinctly different post-quench behavior. In the case of $\alpha=0$, the domain wall kinks equilibrate to a stable value that differs from the expected thermal value, as expected because it is an integrable point. This observation complements the post-quench behavior of DT, as depicted in Figure 4 (a). Although $\alpha=1.5$ is a non-integrable point, thermalization is not observed within the simulation time. At later times, a stable prethermal plateau, close but distinct from the expected thermal value, becomes apparent. Conversely, for a quench corresponding to $\alpha=1.9$, the average domain wall kinks converge to the expected thermal value. Notably, before reaching the thermal value, the kink density exhibits a relatively stable prethermal plateau until time $t\simeq 35$. This relaxation mode, which is characterized by two time scales, is represented by Path 2 in Figure 1(b). This discovery provides another robust indicator of thermalization in weakly confined regimes. ## 5 Conclusion We investigate the relaxation dynamics of the long-range Ising model subsequent to a global quantum quench of the transverse field, assessing the thermalization on a computationally viable time scale according to the canonical Gibbs ensemble (CGE). The model is non-integrable for all values of $\alpha$ except at the extremes ($\alpha={0.0,\infty}$), where we anticipate thermalization following a global quantum quench. However, the long-range Ising model exhibits confinement, which suppresses correlation spreading and eventually impedes thermalization. Starting from the Greenberger-Horne- Zeilinger (GHZ) state, we subject the system to two distinct dynamical regimes. As anticipated, robust confinement hinders thermalization for smaller quenches, specifically in the dynamical ferromagnetic region, where the metric DT exhibits persistent oscillations characteristic of the masses of bound mesons. Conversely, for quenches to the dynamical paramagnetic region, a notably different behavior emerges. The persistent oscillation diminishes and the DT relaxes more rapidly. Although conclusive evidence of thermalization for $\alpha=1.5$ is not observed within the simulation time, compelling indications of the thermalization surface for $\alpha=1.9$ are based on the relaxation of DT. This observation gains additional support from the convergence of the domain wall kinks to the expected thermal value. ## Acknowledgements Nishan Ranabhat thanks Alvise Bastianello for fruitful discussion and suggesting the future extension of the work. The numerical simulation of this project was performed at the Ulysses v2 cluster at SISSA. ## Appendix A Exact results for smaller systems For small systems we can calculate the time evolution of FCS and other relevant order parameters by exact diagonalization of the post-quench Hamiltonian. We begin from our initial state, $\mathbb{Z}_{2}$ symmetric GHZ state, $\ket{\psi_{0}}=\frac{1}{\sqrt{2}}(\ket{\rightarrow,\ldots\rightarrow,\rightarrow,\rightarrow\ldots,\rightarrow}+\ket{\leftarrow,\ldots\leftarrow,\leftarrow,\leftarrow\ldots,\leftarrow}).$ (14) The time evolved state is given by $\ket{\psi_{t}}=e^{-i\hat{H}t}\ket{\psi_{0}}$, where $\hat{H}$ is the post- quench Hamiltonian 3. We proceed by expanding $\ket{\psi_{0}}$ in the eigenbasis, of the post-quench Hamiltonian, $\ket{\psi_{0}}=\sum_{j=0}^{2^{N}-1}q_{j}\ket{E_{j}},$ (15) where $q_{j}=\innerproduct{E_{j}}{\psi_{0}}$. Further expanding $\ket{E_{j}}$ in the computational basis, $\ket{E_{j}}=\sum_{n}c^{j}_{n}\ket{n}$, we can derive the expression for $q_{j}$ as, $q_{j}=\innerproduct{E_{j}}{\psi_{0}}=\frac{\Big{(}c_{\ket{\rightarrow,\ldots,\rightarrow}}^{j}\Big{)}^{*}+\Big{(}c_{\ket{\leftarrow,\ldots,\leftarrow}}^{j}\Big{)}^{*}}{\sqrt{2}}.$ (16) The post-quench state is $\ket{\psi_{t}}=\sum_{n}X_{n}(t)\ket{n}$ (17) where $X_{n}(t)=\sum_{j=0}^{N}q_{j}c_{n}^{j}e^{-iE_{j}t}$. We can now calculate the time evolution of the expectation value of a generic parameter $\hat{O}$ as, $\displaystyle O(t)=\bra{\psi_{t}}\hat{O}\ket{\psi_{t}}=\sum_{n,\tilde{n}}X^{\dagger}_{\tilde{n}}(t)X_{n}(t)\bra{\tilde{n}}\hat{O}\ket{n}.$ (18) If $\ket{n}$ is the simultaneous eigenket of the order parameter $\hat{O}$ then 18 becomes, $O(t)=\sum_{n}|X_{n}(t)|^{2}O_{n}.$ (19) Finally, with the full eigenvalues of hamiltonian at hand we can also calculate the energy density corresponding to a thermal density matrix $\hat{\rho}_{\beta}$, $\epsilon_{\beta}=\frac{\sum_{j}E_{j}e^{\beta E_{j}}}{\sum_{j}e^{\beta E_{j}}}.$ (20) ## Appendix B Simulations details In this section we present the details of numerical simulation complementary to the results in the main text. ### B.1 Simulation of finite temperature density operator The finite temperature states can be simulated by casting the density operator as locally purified tensors [92, 93]. The thermal density operator is defined by Gibbs distribution $\hat{\rho}_{\beta}=\frac{e^{-\beta\hat{H}}}{Tr[e^{-\beta\hat{H}}]}$ where $\beta=\frac{1}{T}$ is the inverse temperature. At $\beta=0$ (infinite temperature) the state is maximally mixed and is given as the tensor product of local identities $\hat{\rho}_{0}=\bigotimes_{i=1}^{N}\mathbf{1}^{\sigma^{\prime}_{i},\sigma_{i}}=\mathbb{1}$, where each $\mathbf{1}^{\sigma^{\prime}_{i},\sigma_{i}}$ is a unit matrix of size $(d,d)$, i.e. $\mathbf{1}^{\sigma^{\prime}_{i},\sigma_{i}}=[\delta_{\sigma^{\prime}_{i},\sigma_{i}}]_{d\times d}$ and $d$ is the dimension of the physical space (for spin $\frac{1}{2}$, $d=2$). The density operator for any finite temperature (non-zero $\beta$) is $\displaystyle\hat{\rho}_{\beta}\propto e^{-\beta\hat{H}}$ $\displaystyle=e^{-\frac{\beta}{2}\hat{H}}\mathbb{1}e^{-\frac{\beta}{2}\hat{H}}$ (21a) $\displaystyle\propto e^{-\frac{\beta}{2}\hat{H}}\hat{\rho}_{0}e^{-\frac{\beta}{2}\hat{H}}$ (21b) We keep the density operator operator in locally purified form $\hat{\rho}=\mathbb{X}\mathbb{X}^{\dagger}$ at each stage where $\mathbb{X}$ is represented as tensor $\mathbb{X}^{\sigma_{1},\sigma_{2},\ldots\sigma_{i},\ldots,\sigma_{N}}_{k_{1},k_{2},\ldots,k_{i},\ldots,k_{N}}=\mathbf{X}^{\sigma_{1},k_{1}}_{c_{0},c_{1}}\mathbf{X}^{\sigma_{2},k_{2}}_{c_{1},c_{2}}\ldots\mathbf{X}^{\sigma_{i},k_{i}}_{c_{i-1},c_{i}}\ldots\mathbf{X}^{\sigma_{N},k_{N}}_{c_{N-1},c_{N}}$ (22) where $\sigma_{i}=d$, $k_{i}=d$ are the physical index and the Kraus index are are fixed through out and $1\leq c_{i}\leq\chi_{max}$ is the bond index and $\chi_{max}$ is the maximum value of bond dimension. The density operator initialized at infinite temperature can now be purified to a finite temperature in trotterized steps $\displaystyle\hat{\rho}_{\beta+d\beta}$ $\displaystyle=e^{-\frac{d\beta}{2}\hat{H}}\hat{\rho}_{\beta}e^{-\frac{d\beta}{2}\hat{H}}$ (23a) $\displaystyle=e^{-\frac{d\beta}{2}\hat{H}}\mathbb{X}\mathbb{X}^{\dagger}e^{-\frac{d\beta}{2}\hat{H}}$ (23b) $\displaystyle=e^{-\frac{d\beta}{2}\hat{H}}\mathbb{X}[e^{-\frac{d\beta}{2}\hat{H}}\mathbb{X}]^{\dagger}$ (23c) Equation (23) can be simulated using imaginary time TDVP ($-idt\rightarrow-d\beta$) in only the half section of the density operator operator and never contracting the $X$ and $X^{\dagger}$ layer during the evolution, thus strictly preserving the locally purified form. Figure 6: Maximally mixed density operator at $\beta=0$ as the tensor product of local identities. Figure 6 shows the tensor notation of the infinite temperature density operator $\hat{\rho}_{\beta=0}$ which is a tensor product of identity matrices of size $(d,d)$, where $d$ is the physical dimension. Rather than working with the density operator as an MPO we represent the density operator in the locally purified form [93, 94] which is positive semi-definite by construction and keep it in locally purified form at every stage of the thermal purification process. In Fig. 7 we represent $\hat{\rho}_{\beta=0}$ in the locally purified form $\mathbb{X}_{\beta=0}\mathbb{X}_{\beta=0}^{\dagger}$, where the index in purple is an auxiliary index called the Krauss index. Figure 7: Representing $\hat{\rho}_{\beta=0}$ in the locally purified form. we can now evolve one of the halves ($\mathbb{X}$ or $\mathbb{X}^{\dagger}$) as shown in equation (23) and the evolution on the other half is its trivial conjugate. This approach is computationally efficient as we can work with cheaper MPS instead of more expensive MPDO. In Fig. 8 one half of the $\hat{\rho}_{\beta=0}$ in locally purified form is shown, form here on we will only work with this half. Figure 8: One half of the $\hat{\rho}_{\beta=0}$ in the locally purified form. Algebraically, $\mathbb{X}_{\beta=0}$ can be written as $\mathbb{X}^{\sigma_{1},k_{1},\ldots\sigma_{i},k_{i},\ldots,\sigma_{N},k_{N}}=\mathbf{X}^{\sigma_{1},k_{1}}\otimes\ldots\mathbf{X}^{\sigma_{i},k_{i}}\ldots\otimes\mathbf{X}^{\sigma_{N},k_{N}}$ (24) For the system of spin one-half particles we choose $A$ as $\mathbf{X}^{\sigma_{i},k_{i}}=\frac{1}{\sqrt{2}}\begin{pmatrix}1&0\\\ 0&1\end{pmatrix}\hskip 28.45274pt\forall\hskip 2.84544pti\in\\{1,2,\ldots,N\\}$ (25) Figure 9: Choice of $\mathbf{X}^{\sigma_{i},k_{i}}$ to preserve the trace of $\hat{\rho}$. as shown in Fig. 9. This particular choice is taken to preserve the trace of the density operator, $\sum_{k}\mathbf{X}^{\sigma,k}[\mathbf{X}^{\sigma^{\prime},k}]^{*}=\frac{1}{2}\begin{pmatrix}1&0\\\ 0&1\end{pmatrix}$ (26) Finally, we reshape $\mathbb{X}_{\beta=0}$ from a string of $2\times 2$ matrices to a string of four legged tensors of shape $(1,2,2,1)$ as shown in Fig. 10, which is a MPS of bond dimension $1$. Figure 10: $\mathbb{X}_{\beta=0}$ in MPS form. Now that we have our initial state as an MPS, we can simulate a finite temperature density operator by solving the equation (23), Figure 11: Expectation of the local operator $\hat{O}_{i}$ in thermal density operator $\hat{\rho}_{\beta}$ Numerically, equation (23) can be solved for long-range spin systems through imaginary time evolution, (where $idt$ is transformed into $d\beta$) using the Time-Dependent Variational Principle (TDVP). The TDVP algorithm employed for simulating the thermal state remains fundamentally identical to that used for the real-time evolution of the pure state, with the distinction of an additional auxiliary Krauss index. However, in the thermal purification of a closed system, the Krauss index becomes obsolete because all physical operators act solely on the physical index, and the Krauss indices are contracted among themselves [93]. Figure 11 illustrates the tensor network diagram for computing the expectation value of a two point operator $\hat{O}_{i}\hat{O_{j}}$ acting on site $i$ and $j$ within the thermal state $\hat{\rho}_{\beta}$. ### B.2 Calculating full counting statistics with MPS The Central object in the calculating the full probability distribution function of an order parameter is the moment generating function $G_{l}(\theta,t)=\text{Tr}(\hat{\rho}_{t}e^{i\theta\hat{M}_{l}})$. For pure state, the density matrix can be written as $\hat{\rho}_{t}=|\phi_{t}\rangle\langle\phi_{t}|$ such that $G_{l}(\theta,t)=\langle\phi_{t}|e^{i\theta\hat{M}_{l}}|\phi_{t}\rangle=\langle\phi_{t}|\prod_{j=i}^{i+l-1}e^{i\theta\hat{s}^{x}_{j}}|\phi_{t}\rangle$. The state $|\phi_{t}\rangle$ can be represented as a matrix product state (MPS) [95], and the single-site operator $e^{i\theta\hat{s}^{x}_{j}}$ can be expressed as a two-by-two matrix. Utilizing this representation, the moment generating function can be computed by sandwiching the operators between the matrix product states, as depicted in Figure 12. By obtaining $G_{l}(\theta,t)$, the complete probability distribution $P_{l}(m,t)$ is computed numerically by discretizing the Fourier integral in equation 8. Figure 12: Computing the generating function $G_{l}(\theta,t)$ in matrix product state representation. The site $i$ is chosen such that the subsystem of size $l$ is in the center of the full system. ### B.3 Errors and data convergence Figure 13: Absolute error in the energy density,$|\epsilon_{\beta}^{\text{ED}}-\epsilon_{\beta}^{\text{TDVP}}|$, of thermal states - (a). Absolute errors in the evolution domain wall kinks $|\langle\hat{k}\rangle^{\text{ED}}-\langle\hat{k}\rangle^{\text{TDVP}}|$ following a quantum quench - (b),(c),(d). The numerically exact results are calculated using equations 19 and 20 as detailed in A. TDVP results are obtained with bond dimension $\chi=128$. The system size considered is N=14. Figure 14: Convergence of the TDVP data for $\text{DT}(t)$ with increasing bond dimensions, $\chi=60,90,128$, for six different post-quench parameters considered in the main text. The black dashed line is for visual guidance. We conducted two types of error analysis to assess the accuracy of the numerical results. In figure 13, we assess the absolute error of the TDVP algorithm in comparison with the numerically exact full diagonalization results for a system with size $N=14$ and various post-quench parameters. Figure 13, panel (a), shows the absolute error in the energy density of the thermal states, defined as $|\epsilon_{\beta}^{\text{ED}}-\epsilon_{\beta}^{\text{TDVP}}|$. The absolute error remains of the order $O(10^{-5})$ or smaller across the entire temperature range under consideration. Figures 13, (b), (c), and (d) show the absolute error in domain wall kinks, defined as $|\langle\hat{k}\rangle^{\text{ED}}-\langle\hat{k}\rangle^{\text{TDVP}}|$, following a quantum quench to various post-quench parameters. $\langle\hat{k}\rangle$ is defined in equation 13 and the computational basis $\\{\ket{n}\\}$ is its simultaneous eigenbasis. Notably, the error rapidly converge and is of order $O(10^{-6})$ or smaller for all the cases studied. In Figure 14, we investigate the convergence of the TDVP data for $\text{DT}(t)$ by computing the relative errors $|\text{DT}^{\chi_{1}}(t)-\text{DT}^{\chi_{2}}(t)|$ for three increasing bond dimensions. Our observations reveal that the relative error eventually converges and consistently remains in the order $O(10^{-3})$ or smaller for all cases. It is noteworthy that the error for $\alpha=0.0$ is several orders of magnitude smaller than that for the other values of $\alpha$. This is attributed to $\alpha=0.0$ being an integrable point with an extensive number of conserved quantities, and therefore has a smaller Hilbert space to be explored compared to non-integrable points. Furthermore, for $\alpha=\\{1.5,1.9\\}$, the error for $h_{f}=0.3$ is approximately two orders of magnitude smaller than that for $h_{f}=0.6$. This discrepancy arises because the former case exhibits dynamical confinement, which effectively suppresses the spread of correlations and constrains the total Hilbert space that can be explored during time evolution. ## Appendix C Thermal phase transition in long range Ising model Figure 15: Thermal phase transition of long range Ising model at four different points in parameter space. The initial state in all cases is the maximally mixed state at infinite temperature represented by $\rho_{\beta=0}$, refer to 6. The color coding from red to blue signifies decreasing temperature . For values of $\alpha>2$, the long-range Ising model falls within the regime of short-range interactions and does not exhibit any finite-temperature phase transitions [89]. Extensive investigations into the critical properties of the thermal phase transition in the quantum long-range Ising model have been conducted using numerically exact path integral Monte Carlo methods [90]. The thermal phase transition is qualitatively depicted in Figures 15 for specific parameter values: $\alpha=1.5,1.9$ and $h=0.3,0.6$. As described in Section B.1, the simulation begins with a maximally mixed state at $\beta=0$. This initial state is characterized by a sharply peaked Gaussian distribution of $P(m)$ centered around $m=0$, which signifies a strongly paramagnetic phase. As the system is gradually cooled by increasing $\beta$, the distribution gradually widens, eventually becoming nearly flat around the critical temperature. A further reduction in temperature leads to the emergence of a bimodal distribution of $P(m)$, which is indicative of the ferromagnetic phase. Notably, this transition from a unimodal Gaussian distribution to a bimodal distribution highlights the $\mathbb{Z}_{2}$ symmetry that is inherent in the long-range Ising Hamiltonian. ## Appendix D Confinement dynamics in different regimes Figure 16: Real time dynamics of half chain connected correlation function $\expectationvalue{\hat{s}^{x}_{k}\hat{s}^{x}_{k+\Delta}}_{c}$after a global quantum quench of the transverse field starting from a fully polarized initial state. The dashed black lines is $v_{max}=2h$ line for nearest neighbor transverse field Ising model[42]. Confinement phenomena in the long-range Ising model result from ferromagnetic interactions extending over long distances between the interacting spins. However, the strength of confinement varies within different regions of the phase space [44, 45]. In this section, we present comprehensive numerical results pertaining to the temporal spreading of correlations in the long-range Ising chain following a sudden quench to various post-quench Hamiltonians starting from a fully polarized initial state denoted as $\ket{\psi_{i}}=\ket{\leftarrow,\leftarrow,\ldots,\leftarrow,\ldots,\leftarrow,\leftarrow}_{x}$. Figure 15 illustrates the time evolution of the half chain connected correlation function $\expectationvalue{\hat{s}^{x}_{k}\hat{s}^{x}_{k+\Delta}}_{c}=\expectationvalue{\hat{s}^{x}_{k}\hat{s}^{x}_{k+\Delta}}-\expectationvalue{\hat{s}^{x}_{k}}\expectationvalue{\hat{s}^{x}_{k+\Delta}}$ in a chain of 200 spins, where $k$ is kept fixed at the center of the chain. In panels (a), (b), and (c), we examine a fixed value of $\alpha=1.9$ while varying the transverse field $h={0.3,0.6,0.8}$. Notably, panel (a) shows a pronounced signature of confinement, which gradually diminishes as the value of $h$ increases, as shown in panels (b) and (c). This behavior is expected because the transverse field competes with long-range interactions and weakens the confinement effect. In panel (d), we observe a linear light cone spreading of the correlation with the maximum possible velocity, $v_{\text{max}}=2h$ [42]. ## References * [1] R. Islam, E. Edwards, K. Kim, S. Korenblit, C. Noh, H. Carmichael, G.-D. Lin, L.-M. Duan, C.-C. Joseph Wang, J. Freericks and C. Monroe, _Onset of a quantum phase transition with a trapped ion quantum simulator_ , Nature Communications 2 (2011), https://doi.org/10.1038/ncomms1374. * [2] J. Zhang, G. Pagano, P. W. Hess, A. Kyprianidis, P. Becker, H. Kaplan, A. V. Gorshkov, Z.-X. Gong and C. Monroe, _Observation of a many-body dynamical phase transition with a 53-qubit quantum simulator_ , Nature 551, 601 (2017), https://doi.org/10.1038/nature24654. * [3] P. Jurcevic, H. Shen, P. Hauke, C. Maier, T. Brydges, C. Hempel, B. P. Lanyon, M. Heyl, R. Blatt and C. F. Roos, _Direct observation of dynamical quantum phase transitions in an interacting many-body system_ , Phys. Rev. Lett. 119, 080501 (2017), 10.1103/PhysRevLett.119.080501. * [4] W. L. Tan, P. Becker, F. Liu, G. Pagano, K. S. Collins, A. De, L. Feng, H. B. Kaplan, A. Kyprianidis, R. Lundgren, W. Morong, S. Whitsitt _et al._ , _Domain-wall confinement and dynamics in a quantum simulator_ , Nature Physics 17, 742747 (2021), 10.1038/s41567-021-01194-3. * [5] S. Fey and K. P. Schmidt, _Critical behavior of quantum magnets with long-range interactions in the thermodynamic limit_ , Phys. Rev. B 94, 075156 (2016), 10.1103/PhysRevB.94.075156. * [6] J. Bohnet, B. Sawyer, J. Britton, M. Wall, A. Rey, M. Foss-Feig and J. Bollinger, _Quantum spin dynamics and entanglement generation with hundreds of trapped ions_ , Science 352, 1297 (2016), 10.1126/science.aad9958. * [7] T. Kinoshita, T. Wenger and D. Weiss, _A quantum newton’s cradle_ , Nature 440, 900 (2006), https://doi.org/10.1038/nature04693. * [8] T. Langen, R. Geiger and J. Schmiedmayer, _Ultracold atoms out of equilibrium_ , Annual Review of Condensed Matter Physics 6(1), 201 (2015), 10.1146/annurev-conmatphys-031214-014548. * [9] F. Meinert, M. J. Mark, E. Kirilov, K. Lauber, P. Weinmann, A. J. Daley and H.-C. Nägerl, _Quantum quench in an atomic one-dimensional ising chain_ , Phys. Rev. Lett. 111, 053003 (2013), 10.1103/PhysRevLett.111.053003. * [10] C. Gross and I. Bloch, _Quantum simulations with ultracold atoms in optical lattices_ , Science 357, 995 (2017), 10.1126/science.aal3837. * [11] B. Neyenhuis, J. Zhang, P. W. Hess, J. Smith, A. C. Lee, P. Richerme, Z.-X. Gong, A. V. Gorshkov and C. Monroe, _Observation of prethermalization in long-range interacting spin chains_ , Science Advances 3 (2017), 10.1126/sciadv.1700672. * [12] P. Calabrese, F. H. L. Essler and M. Fagotti, _Quantum quench in the transverse field ising chain: I. time evolution of order parameter correlators_ , Journal of Statistical Mechanics: Theory and Experiment 2012(07), P07016 (2012), 10.1088/1742-5468/2012/07/P07016. * [13] P. Calabrese, F. H. L. Essler and M. Fagotti, _Quantum quench in the transverse-field ising chain_ , Phys. Rev. Lett. 106, 227203 (2011), 10.1103/PhysRevLett.106.227203. * [14] P. Calabrese and J. Cardy, _Quantum quenches in extended systems_ , Journal of Statistical Mechanics: Theory and Experiment 2007(06), P06008 (2007), 10.1088/1742-5468/2007/06/P06008. * [15] P. Calabrese and J. Cardy, _Evolution of entanglement entropy in one-dimensional systems_ , Journal of Statistical Mechanics: Theory and Experiment 2005(04), P04010 (2005), 10.1088/1742-5468/2005/04/P04010. * [16] P. Calabrese and J. Cardy, _Time dependence of correlation functions following a quantum quench_ , Phys. Rev. Lett. 96, 136801 (2006), 10.1103/PhysRevLett.96.136801. * [17] J.-S. Caux and F. H. L. Essler, _Time evolution of local observables after quenching to an integrable model_ , Phys. Rev. Lett. 110, 257203 (2013), 10.1103/PhysRevLett.110.257203. * [18] S. Groha, F. H. L. Essler and P. Calabrese, _Full counting statistics in the transverse field Ising chain_ , SciPost Phys. 4, 043 (2018), 10.21468/SciPostPhys.4.6.043. * [19] M. Collura, _Relaxation of the order-parameter statistics in the Ising quantum chain_ , SciPost Phys. 7, 072 (2019), 10.21468/SciPostPhys.7.6.072. * [20] A. J. Daley, C. Kollath, U. Schollwöck and G. Vidal, _Time-dependent density-matrix renormalization-group using adaptive effective hilbert spaces_ , Journal of Statistical Mechanics: Theory and Experiment 2004(04), P04005 (2004), 10.1088/1742-5468/2004/04/P04005. * [21] S. R. White and A. E. Feiguin, _Real-time evolution using the density matrix renormalization group_ , Phys. Rev. Lett. 93, 076401 (2004), 10.1103/PhysRevLett.93.076401. * [22] G. Vidal, _Efficient simulation of one-dimensional quantum many-body systems_ , Phys. Rev. Lett. 93, 040502 (2004), 10.1103/PhysRevLett.93.040502. * [23] J. Haegeman, J. I. Cirac, T. J. Osborne, I. Pižorn, H. Verschelde and F. Verstraete, _Time-dependent variational principle for quantum lattices_ , Phys. Rev. Lett. 107, 070601 (2011), 10.1103/PhysRevLett.107.070601. * [24] J. Haegeman, C. Lubich, I. Oseledets, B. Vandereycken and F. Verstraete, _Unifying time evolution and optimization with matrix product states_ , Phys. Rev. B 94, 165116 (2016), 10.1103/PhysRevB.94.165116. * [25] G. Lami, G. Carleo and M. Collura, _Matrix product states with backflow correlations_ , Phys. Rev. B 106, L081111 (2022), 10.1103/PhysRevB.106.L081111. * [26] M. Srednicki, _Chaos and quantum thermalization_ , Phys. Rev. E 50, 888 (1994), 10.1103/PhysRevE.50.888. * [27] J. M. Deutsch, _Quantum statistical mechanics in a closed system_ , Phys. Rev. A 43, 2046 (1991), 10.1103/PhysRevA.43.2046. * [28] H. Kim, T. N. Ikeda and D. A. Huse, _Testing whether all eigenstates obey the eigenstate thermalization hypothesis_ , Phys. Rev. E 90, 052105 (2014), 10.1103/PhysRevE.90.052105. * [29] R. Steinigeweg, J. Herbrych and P. Prelovšek, _Eigenstate thermalization within isolated spin-chain systems_ , Phys. Rev. E 87, 012118 (2013), 10.1103/PhysRevE.87.012118. * [30] M. Rigol, V. Dunjko and M. Olshanii, _Thermalization and its mechanism for generic isolated quantum systems_ , Nature 452, 854858 (2008), 10.1038/nature06838. * [31] C. Kollath, A. M. Läuchli and E. Altman, _Quench dynamics and nonequilibrium phase diagram of the bose-hubbard model_ , Phys. Rev. Lett. 98, 180601 (2007), 10.1103/PhysRevLett.98.180601. * [32] S. R. Manmana, S. Wessel, R. M. Noack and A. Muramatsu, _Strongly correlated fermions after a quantum quench_ , Phys. Rev. Lett. 98, 210405 (2007), 10.1103/PhysRevLett.98.210405. * [33] F. Chen, Z.-H. Sun, M. Gong, Q. Zhu, Y.-R. Zhang, Y. Wu, Y. Ye, C. Zha, S. Li, S. Guo, H. Qian, H.-L. Huang _et al._ , _Observation of strong and weak thermalization in a superconducting quantum processor_ , Phys. Rev. Lett. 127, 020602 (2021), 10.1103/PhysRevLett.127.020602. * [34] M. C. Bañuls, J. I. Cirac and M. B. Hastings, _Strong and weak thermalization of infinite nonintegrable quantum systems_ , Phys. Rev. Lett. 106, 050405 (2011), 10.1103/PhysRevLett.106.050405. * [35] S. Sugimoto, R. Hamazaki and M. Ueda, _Eigenstate thermalization in long-range interacting systems_ , Phys. Rev. Lett. 129, 030602 (2022), 10.1103/PhysRevLett.129.030602. * [36] A. Russomanno, M. Fava and M. Heyl, _Quantum chaos and ensemble inequivalence of quantum long-range ising chains_ , Phys. Rev. B 104, 094309 (2021), 10.1103/PhysRevB.104.094309. * [37] K. R. Fratus and M. Srednicki, _Eigenstate thermalization and spontaneous symmetry breaking in the one-dimensional transverse-field ising model with power-law interactions_ , 10.48550/ARXIV.1611.03992 (2016). * [38] D. A. Abanin, E. Altman, I. Bloch and M. Serbyn, _Colloquium: Many-body localization, thermalization, and entanglement_ , Rev. Mod. Phys. 91, 021001 (2019), 10.1103/RevModPhys.91.021001. * [39] C. Gogolin, M. P. Müller and J. Eisert, _Absence of thermalization in nonintegrable systems_ , Phys. Rev. Lett. 106, 040401 (2011), 10.1103/PhysRevLett.106.040401. * [40] R. Nandkishore and D. A. Huse, _Many-body localization and thermalization in quantum statistical mechanics_ , Annual Review of Condensed Matter Physics 6(1), 15 (2015), 10.1146/annurev-conmatphys-031214-014726, https://doi.org/10.1146/annurev-conmatphys-031214-014726. * [41] M. Fava, R. Fazio and A. Russomanno, _Many-body dynamical localization in the kicked bose-hubbard chain_ , Phys. Rev. B 101, 064302 (2020), 10.1103/PhysRevB.101.064302. * [42] M. Kormos, M. Collura, G. Takács and P. Calabrese, _Real-time confinement following a quantum quench to a non-integrable model_ , Nature Physics 13, 249246 (2017), 10.1038/nphys3934. * [43] R. J. V. Tortora, P. Calabrese and M. Collura, _Relaxation of the order-parameter statistics and dynamical inement_ , EPL (Europhysics Letters) 132(5), 50001 (2020), http://dx.doi.org/10.1209/0295-5075/132/50001. * [44] A. Lerose, B. Žunkovič, A. Silva and A. Gambassi, _Quasilocalized excitations induced by long-range interactions in translationally invariant quantum spin chains_ , Phys. Rev. B 99, 121112 (2019), 10.1103/PhysRevB.99.121112. * [45] F. Liu, R. Lundgren, P. Titum, G. Pagano, J. Zhang, C. Monroe and A. V. Gorshkov, _Confined quasiparticle dynamics in long-range interacting quantum spin chains_ , Phys. Rev. Lett. 122, 150601 (2019), 10.1103/PhysRevLett.122.150601. * [46] S. Scopa, P. Calabrese and A. Bastianello, _Entanglement dynamics in confining spin chains_ , Phys. Rev. B 105, 125413 (2022), 10.1103/PhysRevB.105.125413. * [47] S. Birnkammer, A. Bastianello and M. Knap, _Prethermalization in one-dimensional quantum many-body systems with confinement_ , Nature Communications 13 (2022), 10.1038/s41467-022-35301-6. * [48] M. Rigobello, S. Notarnicola, G. Magnifico and S. Montangero, _Entanglement generation in $(1+1)\mathrm{D}$ qed scattering processes_, Phys. Rev. D 104, 114501 (2021), 10.1103/PhysRevD.104.114501. * [49] R. Verdel, F. Liu, S. Whitsitt, A. V. Gorshkov and M. Heyl, _Real-time dynamics of string breaking in quantum spin chains_ , Phys. Rev. B 102, 014308 (2020), 10.1103/PhysRevB.102.014308. * [50] R. C. Myers, M. Rozali and B. Way, _Holographic quenches in a confined phase_ , Journal of Physics A: Mathematical and Theoretical 50(49), 494002 (2017), 10.1088/1751-8121/aa927c. * [51] A. C. Cubero and N. J. Robinson, _Lack of thermalization in (1+1)-d quantum chromodynamics at large nc_ , Journal of Statistical Mechanics: Theory and Experiment 2019(12), 123101 (2019), 10.1088/1742-5468/ab4e8d. * [52] J. Vovrosh and J. Knolle, _Confinement and entanglement dynamics on a digital quantum computer_ , Scientific Reports 11 (2021), 10.1038/s41467-022-35301-6. * [53] G. Lagnese, F. M. Surace, M. Kormos and P. Calabrese, _False vacuum decay in quantum spin chains_ , Phys. Rev. B 104, L201106 (2021), 10.1103/PhysRevB.104.L201106. * [54] N. Ranabhat and M. Collura, _Dynamics of the order parameter statistics in the long range Ising model_ , SciPost Phys. 12, 126 (2022), 10.21468/SciPostPhys.12.4.126. * [55] E. Lieb, T. Schultz and D. Mattis, _Two soluble models of an antiferromagnetic chain_ , Annals of Physics 16(3), 407 (1961), https://doi.org/10.1016/0003-4916(61)90115-4. * [56] T. Koffel, M. Lewenstein and L. Tagliacozzo, _Entanglement entropy for the long-range ising chain in a transverse field_ , Phys. Rev. Lett. 109, 267203 (2012), 10.1103/PhysRevLett.109.267203. * [57] M. Gabbrielli, L. Lepori and L. Pezzè, _Multipartite-entanglement tomography of a quantum simulator_ , New Journal of Physics 21(3), 033039 (2019), 10.1088/1367-2630/aafb8c. * [58] B. Žunkovič, A. Silva and M. Fabrizio, _Dynamical phase transitions and Loschmidt echo in the infinite-range XY model_ , Phil. Trans. R. Soc. A. 374 (2016), 10.1098/rsta.2015.0160. * [59] A. Das, K. Sengupta, D. Sen and B. K. Chakrabarti, _Infinite-range ising ferromagnet in a time-dependent transverse magnetic field: Quench and ac dynamics near the quantum critical point_ , Phys. Rev. B 74, 144423 (2006), 10.1103/PhysRevB.74.144423. * [60] A. Dutta and J. K. Bhattacharjee, _Phase transitions in the quantum ising and rotor models with a long-range interaction_ , Phys. Rev. B 64, 184106 (2001), 10.1103/PhysRevB.64.184106. * [61] E. Gonzalez-Lazo, M. Heyl, M. Dalmonte and A. Angelone, _Finite-temperature critical behavior of long-range quantum Ising models_ , SciPost Phys. 11, 76 (2021), 10.21468/SciPostPhys.11.4.076. * [62] M. Foss-Feig, Z.-X. Gong, C. W. Clark and A. V. Gorshkov, _Nearly linear light cones in long-range interacting quantum systems_ , Phys. Rev. Lett. 114, 157201 (2015), 10.1103/PhysRevLett.114.157201. * [63] A. S. Buyskikh, M. Fagotti, J. Schachenmayer, F. Essler and A. J. Daley, _Entanglement growth and correlation spreading with variable-range interactions in spin and fermionic tunneling models_ , Phys. Rev. A 93, 053620 (2016), 10.1103/PhysRevA.93.053620. * [64] B. Žunkovič, M. Heyl, M. Knap and A. Silva, _Dynamical quantum phase transitions in spin chains with long-range interactions: Merging different concepts of nonequilibrium criticality_ , Phys. Rev. Lett. 120, 130601 (2018), 10.1103/PhysRevLett.120.130601. * [65] J. C. Halimeh and V. Zauner-Stauber, _Dynamical phase diagram of quantum spin chains with long-range interactions_ , Phys. Rev. B 96, 134427 (2017), 10.1103/PhysRevB.96.134427. * [66] R. Khasseh, A. Russomanno, M. Schmitt, M. Heyl and R. Fazio, _Discrete truncated wigner approach to dynamical phase transitions in ising models after a quantum quench_ , Phys. Rev. B 102, 014303 (2020), 10.1103/PhysRevB.102.014303. * [67] J. C. Halimeh, V. Zauner-Stauber, I. P. McCulloch, I. de Vega, U. Schollwöck and M. Kastner, _Prethermalization and persistent order in the absence of a thermal phase transition_ , Phys. Rev. B 95, 024302 (2017), 10.1103/PhysRevB.95.024302. * [68] I. Homrighausen, N. O. Abeling, V. Zauner-Stauber and J. C. Halimeh, _Anomalous dynamical phase in quantum spin chains with long-range interactions_ , Phys. Rev. B 96, 104436 (2017), 10.1103/PhysRevB.96.104436. * [69] J. Lang, B. Frank and J. C. Halimeh, _Concurrence of dynamical phase transitions at finite temperature in the fully connected transverse-field ising model_ , Phys. Rev. B 97, 174401 (2018), 10.1103/PhysRevB.97.174401. * [70] F. H. L. Essler and M. Fagotti, _Quench dynamics and relaxation in isolated integrable quantum spin chains_ , Journal of Statistical Mechanics: Theory and Experiment 2016(6), 064002 (2016). * [71] L. S. Levitov, H. Lee and G. B. Lesovik, _Electron counting statistics and coherent states of electric current_ , Journal of Mathematical Physics 37(4845) (1996), https://doi.org/10.1063/1.531672. * [72] M. Esposito, U. Harbola and S. Mukamel, _Nonequilibrium fluctuations, fluctuation theorems, and counting statistics in quantum systems_ , Rev. Mod. Phys. 81, 1665 (2009), 10.1103/RevModPhys.81.1665. * [73] M. Campisi, P. Hänggi and P. Talkner, _Colloquium: Quantum fluctuation relations: Foundations and applications_ , Rev. Mod. Phys. 83, 771 (2011), 10.1103/RevModPhys.83.771. * [74] P. Calabrese, M. Collura, G. D. Giulio and S. Murciano, _Full counting statistics in the gapped xxz spin chain_ , Europhysics Letters 129(6), 60007 (2020), 10.1209/0295-5075/129/60007. * [75] M. Collura, F. H. L. Essler and S. Groha, _Full counting statistics in the spin-1/2 heisenberg xxz chain_ , Journal of Physics A: Mathematical and Theoretical 50(41), 414002 (2017), 10.1088/1751-8121/aa87dd. * [76] M. Collura and F. H. L. Essler, _How order melts after quantum quenches_ , Phys. Rev. B 101, 041110 (2020), 10.1103/PhysRevB.101.041110. * [77] D. J. Luitz, N. Laflorencie and F. Alet, _Many-body localization edge in the random-field heisenberg chain_ , Phys. Rev. B 91, 081103 (2015), 10.1103/PhysRevB.91.081103. * [78] R. Singh, J. H. Bardarson and F. Pollmann, _Signatures of the many-body localization transition in the dynamics of entanglement and bipartite fluctuations_ , New Journal of Physics 18(2), 023046 (2016), 10.1088/1367-2630/18/2/023046. * [79] M. Gring, M. Kuhnert, T. Langen, T. Kitagawa, B. Rauer, M. Schreitl, I. Mazets, D. A. Smith, E. Demler and J. Schmiedmayer, _Relaxation and prethermalization in an isolated quantum system_ , Science 337(6100), 1318 (2012), 10.1126/science.1224953. * [80] Y. Tang, W. Kao, K.-Y. Li, S. Seo, K. Mallayya, M. Rigol, S. Gopalakrishnan and B. L. Lev, _Thermalization near integrability in a dipolar quantum newton’s cradle_ , Phys. Rev. X 8, 021030 (2018), 10.1103/PhysRevX.8.021030. * [81] B. Blaß and H. Rieger, _Test of quantum thermalization in the two-dimensional transverse-field ising model_ , Scientific Reports 6 (2016), 10.1038/srep38185. * [82] S. Paeckel, T. Köhler, A. Swoboda, S. R. Manmana, U. Schollwöck and C. Hubig, _Time-evolution methods for matrix-product states_ , Annals of Physics 411, 167998 (2019), 10.1016/j.aop.2019.167998. * [83] A. A. Zvyagin, _Dynamical quantum phase transitions_ (2017), 1701.08851. * [84] T. Mori, T. N. Ikeda, E. Kaminishi and M. Ueda, _Thermalization and prethermalization in isolated quantum systems: a theoretical overview_ , Journal of Physics B: Atomic, Molecular and Optical Physics 51(11), 112001 (2018), 10.1088/1361-6455/aabcdf. * [85] N. Defenu, T. Enss and J. C. Halimeh, _Dynamical criticality and domain-wall coupling in long-range hamiltonians_ , Phys. Rev. B 100, 014434 (2019), 10.1103/PhysRevB.100.014434. * [86] P. Titum, J. T. Iosue, J. R. Garrison, A. V. Gorshkov and Z.-X. Gong, _Probing ground-state phase transitions through quench dynamics_ , Phys. Rev. Lett. 123, 115701 (2019), 10.1103/PhysRevLett.123.115701. * [87] P. Titum and M. F. Maghrebi, _Nonequilibrium criticality in quench dynamics of long-range spin models_ , Phys. Rev. Lett. 125, 040602 (2020), 10.1103/PhysRevLett.125.040602. * [88] J. C. Halimeh, D. Trapin, M. Van Damme and M. Heyl, _Local measures of dynamical quantum phase transitions_ , Phys. Rev. B 104, 075130 (2021), 10.1103/PhysRevB.104.075130. * [89] A. Dutta and J. K. Bhattacharjee, _Phase transitions in the quantum ising and rotor models with a long-range interaction_ , Phys. Rev. B 64, 184106 (2001), 10.1103/PhysRevB.64.184106. * [90] E. G. Lazo, M. Heyl, M. Dalmonte and A. Angelone, _Finite-temperature critical behavior of long-range quantum ising models_ , SciPost Phys. 11, 076 (2021), 10.21468/SciPostPhys.11.4.076. * [91] N. Ranabhat, A. Santini, E. Tirrito and M. Collura, _Dynamical deconfinement transition driven by density of excitations_ (2023), 2310.02320. * [92] F. Verstraete, J. J. García-Ripoll and J. I. Cirac, _Matrix product density operators: Simulation of finite-temperature and dissipative systems_ , Phys. Rev. Lett. 93, 207204 (2004), 10.1103/PhysRevLett.93.207204. * [93] A. H. Werner, D. Jaschke, P. Silvi, M. Kliesch, T. Calarco, J. Eisert and S. Montangero, _Positive tensor network approach for simulating open quantum many-body systems_ , Phys. Rev. Lett. 116, 237201 (2016), 10.1103/PhysRevLett.116.237201. * [94] D. Jaschke, S. Montangero and L. D. Carr, _One-dimensional many-body entangled open quantum systems with tensor network methods_ , Quantum Science and Technology 4(1), 013001 (2018), 10.1088/2058-9565/aae724. * [95] U. Schollwöck, _The density-matrix renormalization group in the age of matrix product states_ , Annals of Physics 326(1), 96 (2011), https://doi.org/10.1016/j.aop.2010.09.012, January 2011 Special Issue.
# On the geometry of uniform meandric systems Jacopo Borga Stanford University Ewain Gwynne University of Chicago Minjae Park University of Chicago ###### Abstract A meandric system of size $n$ is the set of loops formed from two arc diagrams (non-crossing perfect matchings) on $\\{1,\dots,2n\\}$, one drawn above the real line and the other below the real line. A uniform random meandric system can be viewed as a random planar map decorated by a Hamiltonian path (corresponding to the real line) and a collection of loops (formed by the arcs). Based on physics heuristics and numerical evidence, we conjecture that the scaling limit of this decorated random planar map is given by an independent triple consisting of a Liouville quantum gravity (LQG) sphere with parameter $\gamma=\sqrt{2}$, a Schramm-Loewner evolution (SLE) curve with parameter $\kappa=8$, and a conformal loop ensemble (CLE) with parameter $\kappa=6$. We prove several rigorous results which are consistent with this conjecture. In particular, a uniform meandric system admits loops of nearly macroscopic graph-distance diameter with high probability. Furthermore, a.s., the uniform infinite meandric system with boundary has no infinite path. But, a.s., its boundary-modified version has a unique infinite path whose scaling limit is conjectured to be chordal SLE6. ###### Contents 1. 1 Introduction 1. 1.1 Meandric systems 2. 1.2 Conjectures for scaling limit and largest loop exponent 3. 1.3 Macroscopic loops in finite meandric systems 4. 1.4 The uniform infinite meandric system (UIMS) 5. 1.5 The uniform infinite half-plane meandric system (UIHPMS) 2. 2 Percolation interpretation 1. 2.1 Non-crossing perfect matchings and non-crossing integer partitions 2. 2.2 Meandric systems as boundaries of clusters of open edges 3. 2.3 Meandric systems and critical percolation on random planar maps 4. 2.4 Box crossings 5. 2.5 The lack of positive association and two open problems 3. 3 Existence of macroscopic loops in the UIMS 4. 4 Bounding distances via the mated-CRT map 1. 4.1 The mated-CRT map 2. 4.2 Proof of lower bounds for graph distances 3. 4.3 Proof of Proposition 1.4, Theorem 1.5, and Theorem 1.9 5. 5 Estimate for the mated-CRT map via SLE and LQG 1. 5.1 SLE/LQG description of the mated-CRT map 2. 5.2 Proof of lower bounds for mated-CRT graph distances 6. 6 Proofs for the UIHPMS 1. 6.1 Equivalence of the definitions of the UIHPMS 2. 6.2 Aperiodicity via good boundary points 3. 6.3 No semi-infinite paths started from the boundary 4. 6.4 Infinitely many paths separating the origin from $\infty$ 5. 6.5 Proofs of Theorem 1.12 and Proposition 1.14 7. 7 Justification for Conjectures 1.2 and 1.3 1. 7.1 SLE8 on $\sqrt{2}$-LQG via mating of trees 2. 7.2 Physics argument 3. 7.3 Predictions for exponents via KPZ 4. 7.4 Simulations Acknowledgments. We thank two anonymous referees for helpful comments on an earlier version of this article. We thank Ahmed Bou-Rabee, Valentin Féray, Gady Kozma, Ron Peled, and Xin Sun for helpful discussions. E.G. was partially supported by a Clay research fellowship. M.P. was partially supported by an NSF grant DMS 2153742. ## 1 Introduction ### 1.1 Meandric systems Throughout this paper, we identify $\mathbbm{R}$ (resp. $\mathbbm{Z}$) with the set $\mathbbm{R}\times\\{0\\}$ (resp. $\mathbbm{Z}\times\\{0\\}$). ###### Definition 1.1. A meandric system of size $n\in\mathbbm{N}$ is a configuration ${\mathfrak{S}}_{n}$ consisting of a finite collection of simple loops in $\mathbbm{R}^{2}$ with the following properties: * • No two loops of ${\mathfrak{S}}_{n}$ intersect each other. * • Each loop of ${\mathfrak{S}}_{n}$ intersects the real line $\mathbbm{R}$ at least twice, and does not intersect $\mathbbm{R}$ without crossing it. * • The total number of intersection points between the loops of ${\mathfrak{S}}_{n}$ and $\mathbbm{R}$ is equal to $2n$. We view such configurations as being defined modulo orientation-preserving homeomorphisms from $\mathbbm{R}^{2}$ to $\mathbbm{R}^{2}$ which take $\mathbbm{R}$ to $\mathbbm{R}$. Figure 1: A meandric system of size $n=6$. The two corresponding arc diagrams are shown in red and blue. The graphs of the simple walk excursions corresponding to these arc diagrams are shown in black. The walk excursions are translated (in time) by a $1/2$-factor to the right so that each step of the walks is centered with the corresponding integer point in the real line. We made this choice (here and in all our graphical representations of meandric systems) to highlight better the correspondence between walks and arcs. See (1.1) and the surrounding text for further explanation. See Figure 1 for an illustration of a meandric system. If ${\mathfrak{S}}_{n}$ is a meandric system of size $n$, then by applying a homeomorphism, we can always arrange the set of intersection points of arcs in ${\mathfrak{S}}_{n}$ with $\mathbbm{R}$ so that it is equal to $[1,2n]\cap\mathbbm{Z}$. We will make this assumption throughout the paper. There have been several recent works in probability and combinatorics which studied meandric systems (see, e.g., [CKST19, FN19, GNP20, Kar20, FT22]). These works were in part motivated by the study of decorated planar maps and by the connection between meandric systems, non-crossing partitions, and meanders, as we discuss just below. Additional motivations come from the fact that random meandric systems are equivalent to a certain percolation-type model on a random planar map (Section 2) and to a version of the fully packed $O(0\times 1)$ loop model on a random planar map (Section 7.2). A meander of size $n$ is a meandric system of size $n$ with a single loop. Each of the loops in a meandric system can be viewed as a meander by forgetting the other loops. However, a typical loop in a uniformly sampled meandric system of size $n$ is not the same as a uniformly sampled meander [FT22, Section 4]. The study of meanders dates back to at least the work of Poincaré in 1912 [Poi12] and is connected to a huge number of different areas of math and physics. See [La 03, Zvo21] for surveys of results on meanders. Most features of meanders are notoriously difficult to analyze mathematically. For example, determining the $n\rightarrow\infty$ asymptotics of the total number of meanders of size $n$ is a long-standing open problem (but see [DFGG00] for a conjecture). Meandric systems are significantly more tractable than meanders. The main reason for this is that meandric systems are in bijection with pairs of arc diagrams (non-crossing perfect matchings). An arc diagram of size $n\in\mathbbm{N}$ is a collection of arcs in the upper half-plane $\mathbbm{R}\times[0,\infty)$, each of which joint two points in $[1,2n]\cap\mathbbm{Z}$, subject to the condition that no two of the arcs cross. If ${\mathfrak{S}}_{n}$ is a meandric system of size $n$, then the segments of loops in ${\mathfrak{S}}_{n}$ above (resp. below) the real line form an arc diagram. Conversely, any two arc diagrams of size $n$ give rise to a meandric system by drawing one above and one below the real line, and considering the set of loops that they form. It is well known that arc diagrams of size $n$ are counted by the Catalan number $\operatorname{Cat}_{n}=\frac{1}{n+1}\binom{2n}{n}$. Consequently, the number of meandric systems of size $n$ is $\operatorname{Cat}_{n}^{2}$. Furthermore, arc diagrams of size $n$ are in bijection with $2n$-step simple walks on $\mathbbm{Z}_{\geq 0}$ from $0$ to $0$, often called (non-negative) simple walk excursions or Dyck paths. If $\mathcal{X}:[0,2n]\cap\mathbbm{Z}\rightarrow\mathbbm{Z}$ is a $2n$-step simple walk excursion, then the corresponding arc diagram is defined as follows. Two points $x_{1},x_{2}\in[1,2n]\cap\mathbbm{Z}$ with $x_{1}<x_{2}$ are joined by an arc if and only if $\mathcal{X}_{x_{1}-1}=\mathcal{X}_{x_{2}}<\min_{y\in[x_{1},x_{2}-1]\cap\mathbbm{Z}}\mathcal{X}_{y};$ (1.1) see Figure 1. So, one can sample a uniform random meandric system of size $n$ by sampling two independent simple random walk excursions with $2n$ steps, drawing one of the two corresponding arc diagrams above the real line and the other below the real line, then looking at the loops formed by the union of the two arc diagrams. Let $\mathfrak{S}_{n}$ be a uniform meandric system of size $n$. There are a number of natural questions about the large-scale geometry of $\mathfrak{S}_{n}$, e.g., the following: 1. 1. How many loops does $\mathfrak{S}_{n}$ typically have? 2. 2. What is the size of the largest loop of $\mathfrak{S}_{n}$, in terms of the number of intersection points with $\mathbbm{R}$? What about for other notions of size, e.g., graph-distance diameter in the 4-regular graph whose vertices are the intersection points of loops with $\mathbbm{R}$, and whose edges are the segments of loops and the segments of $\mathbbm{R}$ between these vertices? 3. 3. Is there typically a single loop of $\mathfrak{S}_{n}$ which is much larger (in some sense) than the other loops, or are there multiple large loops of comparable size? 4. 4. Is there some sort of scaling limit of $\mathfrak{S}_{n}$ as $n\rightarrow\infty$? Due to the bijection between meandric systems and pairs of $2n$-step simple walk excursions, questions of the above type, in principle, can be reduced to questions about simple random walks on $\mathbbm{Z}$. However, the encoding of the meandric system loops in terms of the pair of walk excursions is complicated, so the answers to the above questions are far from trivial. Question 1 was largely solved by Féray and Thévenin [FT22], who showed that there is a constant $c>0$ (expressed in terms of a sum over meanders) such that the number of loops in $\mathfrak{S}_{n}$ is asymptotic to $cn$ as $n\rightarrow\infty$. Regarding Question 2, Kargin [Kar20] showed that the number of intersection points with $\mathbbm{R}$ of the largest loop is at least constant times $\log n$, and presented some numerical simulations which suggested that this quantity in fact behaves like $n^{\alpha}$ for $\alpha\approx 4/5$. Question 3 is closely related to the question of whether there exists a so-called “infinite noodle”, i.e., an infinite path in the infinite-volume limit of a uniform meandric system. It was shown in [CKST19] that there is at most one such path. The existence is still open, but it is conjectured in [CKST19] that such a path does not exist. See Section 1.4 for further discussion. In this paper, we present conjectures for the answers to each of Questions 2, 3, and 4 (see Conjectures 1.2 and 1.3). In particular, if $k\in\mathbbm{N}$ is fixed, then as $n\rightarrow\infty$ the number of intersection points with $\mathbbm{R}$ of the $k$th largest loop should grow like $n^{\alpha}$ where $\alpha=\frac{1}{2}(3-\sqrt{2})\approx 0.7929$. Moreover, the scaling limit of $\mathfrak{S}_{n}$ should be described by a $\sqrt{2}$-Liouville quantum gravity sphere, a Schramm-Loewner evolution curve with parameter $\kappa=8$, and a conformal loop ensemble with parameter $\kappa=6$. We also prove several rigorous results in the direction of the above questions. We show that a uniform meandric system admits loops of nearly macroscopic graph-distance diameter (Theorem 1.5). This leads to an explicit power-law lower bound for the size of the largest loop in a uniform meandric system (Corollary 1.6). We also construct the _uniform infinite half-plane meandric system (UIHPMS)_ and show that it does not admit any infinite paths of arcs (Theorem 1.12). But, a minor modification of the UIHPMS admits a unique infinite path of arcs which should converge to SLE6 (Proposition 1.14). Most of our proofs use only elementary discrete arguments, but we need to use the theory of Liouville quantum gravity at one step in the proof, namely in Section 5. ### 1.2 Conjectures for scaling limit and largest loop exponent We want to state a conjecture for the scaling limit of the uniformly sampled meandric system $\mathfrak{S}_{n}$ as $n\rightarrow\infty$. To formulate this conjecture, we let $\mathcal{M}_{n}$ be the planar map whose vertices are the $2n$ intersection points of the loops $\ell$ in $\mathfrak{S}_{n}$ with the real line, whose edges are the segments of the loops or the line $\mathbbm{R}$ between these intersection points (we consider the two infinite rays of $\mathbbm{R}$ as being an edge from the leftmost to the rightmost intersection point), and whose faces are the connected components of $\mathbbm{R}^{2}\setminus\mathopen{}\mathclose{{}\left(\bigcup_{\ell\in\mathfrak{S}_{n}}\ell\cup\mathbbm{R}}\right)$. The planar map $\mathcal{M}_{n}$ is equipped with a Hamiltonian path $P_{n}:[1,2n]\cap\mathbbm{Z}\rightarrow\\{\text{vertices of $\mathcal{M}_{n}$}\\}$ which traverses the vertices of $\mathcal{M}_{n}$ and the segments of $\mathbbm{R}$ between these vertices in left-right numerical order. The planar map is also equipped with a collection of loops $\Gamma_{n}$ (simple cycles in $\mathcal{M}_{n}$) corresponding to the loops in $\mathfrak{S}_{n}$. See Figure 2 for a simulation of $(\mathcal{M}_{n},P_{n},\Gamma_{n})$. To talk about convergence, we can, e.g., view $(\mathcal{M}_{n},P_{n},\Gamma_{n})$ as a metric measure space (equipped with the graph metric and the counting measure on vertices) decorated by a path and a collection of loops. We can then ask whether this decorated metric measure space has a scaling limit with respect to the generalization of the Gromov- Hausdorff topology for metric measure spaces decorated by curves and/or loops [GM17, GHS21]. Alternatively, we could embed $(\mathcal{M}_{n},P_{n},\Gamma_{n})$ into $\mathbbm{C}$ in some manner (e.g., Tutte embedding [GMS21] as in Figure 2 or circle packing [Ste03]) and ask whether the resulting metric, measure, curve, and collection of loops in $\mathbbm{C}$ have a joint scaling limit in law. We now state a conjecture for the scaling limit of $(\mathcal{M}_{n},P_{n},\Gamma_{n})$. The limiting object is described in terms of several random objects whose definitions we will not write down explicitly. * • The Liouville quantum gravity (LQG) sphere with parameter $\gamma\in(0,2]$ is a random fractal surface with the topology of the sphere first introduced in [DMS21, DKRV16]. A $\gamma$-LQG sphere can be described by a random metric and a random measure on the Riemann sphere $\mathbbm{C}\cup\\{\infty\\}$. LQG spheres (and other types of LQG surfaces) describe the scaling limits of various types of random planar maps. See [Gwy20, She22, BP] for expository articles on LQG. * • Schramm-Loewner evolution (SLEκ) with parameter $\kappa>0$ is a random fractal curve introduced in [Sch00]. The curve is simple for $\kappa\in(0,4]$, has self-intersections but not self-crossings for $\kappa\in(4,8)$, and is space- filling for $\kappa\geq 8$. * • The conformal loop ensemble (CLEκ) with parameter $\kappa\in(8/3,8)$ is a random countable collection of loops which do not cross themselves or each other and which locally look like SLEκ curves [She09]. We allow our CLE loops to be nested (i.e., we do not restrict attention to the outermost loops). CLE on the whole plane was first defined in [MWW16] for $\kappa\in(4,8)$ and in [KW16] for $\kappa\in(8/3,4]$. Figure 2: Simulation of a uniform meandric system with boundary of size $n=10^{6}$ (see Section 7.4 for a precise definition and for the details of simulations). The left picture shows the corresponding arc diagrams. The right picture shows the associated planar map $\mathcal{M}_{n}$, embedded in the disk via the Tutte embedding [GMS21], together with some of the loops in $\Gamma_{n}$. The largest 300 loops in $\Gamma_{n}$ (in terms of number of vertices) are each shown in color, as indicated by the color bar. Smaller loops and edges between consecutive vertices of $\mathbbm{R}$ are shown in gray. Note that the distribution of colors in the arc diagram picture is rather chaotic – this is consistent with the fact that the meandric system loops are a complicated functional of the arc diagrams. According to Conjecture 1.2, the embedded planar map $\mathcal{M}_{n}$ together with the path $P_{n}$ and the loops in $\Gamma_{n}$ should converge to $\sqrt{2}$-LQG decorated by SLE8 and CLE6. ###### Conjecture 1.2. Let $(\mathcal{M}_{n},P_{n},\Gamma_{n})$ be the random planar map decorated by a Hamiltonian path and a collection of loops associated to a uniform meandric system of size $n$, as described just above. Then $(\mathcal{M}_{n},P_{n},\Gamma_{n})$ converges under an appropriate scaling limit to an independent triple consisting of a $\sqrt{2}$-LQG sphere, a whole- plane SLE8 from $\infty$ to $\infty$, and a whole-plane CLE6. In the setting of Conjecture 1.2, the metric and measure on $\mathbbm{C}$ corresponding to the $\sqrt{2}$-LQG sphere, the SLE8 curve (viewed modulo time parametrization), and the CLE6 are independent. At first glance, this may be surprising since $(\mathcal{M}_{n},P_{n})$ and $(\mathcal{M}_{n},\Gamma_{n})$ each determine each other. However, we expect that the function which goes from $(\mathcal{M}_{n},P_{n})$ to $(\mathcal{M}_{n},\Gamma_{n})$ depends only on microscopic features of $(\mathcal{M}_{n},P_{n})$ which are not seen in the scaling limit, and the same is true for the function which goes in the opposite direction. This independence is numerically justified by Figure 18 (Right), using the discussion in Section 7.3. An equivalent formulation of the conjecture is that $(\mathcal{M}_{n},P_{n},\Gamma_{n})$ should be in the same universality class as a uniform triple consisting of a planar map decorated by a spanning tree (represented by its associated discrete Peano curve) and a critical Bernoulli percolation configuration (represented by the loops which describe the interfaces between open and closed clusters). Conjecture 1.2 is based on a combination of rigorous results, physics heuristics, and numerical simulations. We will explain the reasoning leading to the conjecture in Section 7. A similar scaling limit conjecture for meanders, rather than meandric systems, is stated as [BGS22, Conjecture 1.3] (building on [DFGG00]). In the meander case, one has $\gamma=\sqrt{\frac{1}{3}\mathopen{}\mathclose{{}\left(17-\sqrt{145}}\right)}$ instead of $\gamma=\sqrt{2}$ and there are two SLE8 curves instead of an SLE8 and a CLE6. The heuristic justification for Conjecture 1.2 is similar to the argument leading to this meander conjecture. In fact, as we will explain in Section 7.2, Conjecture 1.2 may be viewed as a special case of a conjecture for the $O(n\times m)$ loop model on a random planar map from [DFGG00]. See also [DDGG22] for additional related predictions. As we will explain in Section 7.3, Conjecture 1.2 together with the KPZ formula [KPZ88] leads to the following conjectural answer to Questions 2 and 3 above. ###### Conjecture 1.3. Let $\mathfrak{S}_{n}$ be a uniform meandric system of size $n$. For each fixed $k\in\mathbbm{N}$, it holds with probability tending to 1 as $n\rightarrow\infty$ that $\\#\\{\text{vertices of $k$th largest loop of $\mathfrak{S}_{n}$}\\}=n^{\alpha+o(1)},\quad\text{where}\quad\alpha=\frac{1}{2}\mathopen{}\mathclose{{}\left(3-\sqrt{2}}\right)\approx 0.7929.$ (1.2) Conjecture 1.3 for $k=1$ is consistent with the numerical study of [Kar20, Section 3], which suggested that the size of the largest loop in $\mathfrak{S}_{n}$ is of order $n^{\alpha}$ for $\alpha$ close to $4/5$. We have also run some numerical simulations of our own which are consistent with Conjecture 1.3 for $k=1,2,3,4,5$. See Section 7.4 for more details. ### 1.3 Macroscopic loops in finite meandric systems Let $d=d_{\sqrt{2}}$ be the Hausdorff dimension of the $\sqrt{2}$-Liouville quantum gravity metric space (this quantity is well-defined thanks to [GP22, Corollary 1.7]). A reader not familiar with Liouville quantum gravity can simply think of $d$ as a certain constant. The number $d$ is not known explicitly, but fairly good rigorous upper and lower bounds are available. In particular, it was shown in [GP19, Corollary 2.5], building on [DG18, Theorem 1.2], that $3.5504\approx 2(9+3\sqrt{5}-\sqrt{3})(4-\sqrt{15})\leq d\leq\frac{2}{3}(3+\sqrt{6})\approx 3.6330.$ (1.3) The following proposition can be proven using previously known techniques for bounding distances in random planar maps [GHS20, GP21]. ###### Proposition 1.4. Let $\mathcal{M}_{n}$ be the planar map associated with a uniform meandric system of size $n$. For each $\zeta\in(0,1)$, it holds except on an event of probability decaying faster than any negative power of $n$ that the graph- distance diameter of $\mathcal{M}_{n}$ is between $n^{1/d-\zeta}$ and $n^{1/d+\zeta}$. Proposition 1.4 is proven via a coupling with a so-called mated-CRT map, a certain type of random planar map which is directly connected to Liouville quantum gravity. See Section 4.3 for details. It is possible to prove Proposition 1.4 via exactly the same argument as in [GP21, Theorem 1.9], which gives an analogous statement for spanning-tree decorated random planar maps. But, we will give a more self-contained proof in Section 4.3. Our first main result tells us that a uniform meandric system admits loops whose graph-distance diameter is nearly of the same order as the graph- distance diameter of $\mathcal{M}_{n}$, in the following sense, c.f. Proposition 1.4. ###### Theorem 1.5. Let $\mathcal{M}_{n}$ be the planar map associated with a uniform meandric system of size $n$ and let $\Gamma_{n}$ be the associated collection of loops on $\mathcal{M}_{n}$. For each $\zeta\in(0,1)$, it holds except on an event of probability decaying faster than any negative power of $n$ that the following is true. There is a loop in $\Gamma_{n}$ which has $\mathcal{M}_{n}$-graph- distance diameter at least $n^{1/d-\zeta}$. The proof of Theorem 1.5 is based on a combination of two results. The first input is a purely discrete argument, based on a parity trick, which shows that the infinite-volume analog of $(\mathcal{M}_{n},\Gamma_{n})$ admits with positive probability loops which are “macroscopic” in a certain sense (Theorem 3.2). The second input is a lower-bound for certain graph distances in $\mathcal{M}_{n}$ which is proven via a combination of discrete arguments and SLE/LQG techniques (Proposition 4.1). The continuum part of the argument, given in Section 5, is short and simple, but very far from elementary since it relies on both the mating of trees theorem [DMS21] and the existence of the LQG metric [DDDF20, GM21b]. The reason why we have an error of order $n^{\zeta}$ in Proposition 1.4 and Theorem 1.5 is that we are only able to estimate graph distances in $\mathcal{M}_{n}$ up to $o(1)$ errors in the exponent. If we had up-to- constants bounds for graph distances in $\mathcal{M}_{n}$, then our arguments would show that with high probability, there exist loops in $\Gamma_{n}$ whose $\mathcal{M}_{n}$-graph-distance diameter is comparable, up to constants, to the graph-distance diameter of $\mathcal{M}_{n}$. Hence, Theorem 1.5 suggests that the scaling limit of $(\mathcal{M}_{n},\Gamma_{n})$ should be non- degenerate, in the sense that the loops of $\Gamma_{n}$ do not collapse to points. This is consistent with Conjecture 1.2. Theorem 1.5 is similar in spirit to the recent work [DCGPS21], which proves the existence of macroscopic loops for the critical $O(n)$ loop model on the hexagonal lattice when $1\leq n\leq 2$ (see also [CGHP20] for a similar result for a different range of parameter values). However, our proof of Theorem 1.5 is very different from the arguments in [DCGPS21, CGHP20]. Part of the reason for this is that we are not aware of any positive association (FKG) inequality in our setting (see Question 2.2), in addition to the fundamental difference that we are working on a random lattice. Loops in $\Gamma_{n}$ are connected subsets of $\mathcal{M}_{n}$, so a loop of graph-distance diameter at least $n^{1/d-\zeta}$ must hit at least $n^{1/d-\zeta}$ vertices of $\mathcal{M}_{n}$. We therefore have the following corollary of Theorem 1.5. ###### Corollary 1.6. Let $\mathfrak{S}_{n}$ be a uniform meandric system of size $n$. For each $\zeta\in(0,1)$, it holds except on an event of probability decaying faster than any negative power of $n$ that there is a loop in $\mathfrak{S}_{n}$ which crosses the real line at least $n^{1/d-\zeta}$ times. We note that the bounds for $d$ from (1.3) show that $0.2753\approx\frac{1}{2}(3-\sqrt{6})\leq\frac{1}{d}\leq\frac{1}{72-38\sqrt{3}+30\sqrt{5}-18\sqrt{15}}\approx 0.2817.$ (1.4) Corollary 1.6 gives a power-law lower bound for the number of vertices of the largest loop in a typical meandric system. To our knowledge, the best lower bound for this quantity prior to our work is [Kar20, Theorem 3.4], which shows that the number of vertices in the largest loop is typically at least a constant times $\log n$. If Conjecture 1.3 is correct, then the lower bound of Corollary 1.6 is far from optimal. Nevertheless, it would require substantial new ideas to get any exponent larger than $1/d$ for the number of vertices in the largest loop of $\mathfrak{S}_{n}$. Indeed, to do this one would need to show that the largest loop in $\mathfrak{S}_{n}$ is much longer than an $\mathcal{M}_{n}$-graph distance geodesic. ### 1.4 The uniform infinite meandric system (UIMS) The uniform infinite meandric system (UIMS) is the local limit (in the sense of Benjamini-Schramm [BS01]) of a uniform meandric system of size $n$ based at a uniform vertex. It is shown in [FT22, Proposition 5] that this local limit exists (in a quenched sense; see [FT22, Section 1.2] for further explanations) and is the same as the object studied in [CKST19]. Figure 3: A subset of the uniform infinite meandric system (UIMS) together with the graphs of the corresponding increments of the encoding walks $\mathcal{L}$ and $\mathcal{R}$. Each horizontal line below the graph of $\mathcal{L}$ (resp. $\mathcal{R}$) corresponds to an arc above (resp. below) the real line. See (1.5) and the surrounding text for further explanation. We now define the UIMS, following [CKST19]. Let111 The reason why we write $\mathcal{L}$ and $\mathcal{R}$ for the walks is that $\mathcal{L}$ (resp. $\mathcal{R}$) describes the arcs which lie to the left (resp. right) of the real line when we traverse the real line from left to right. This notation is chosen to be consistent with the mating of trees literature [DMS21, GHS23, GHS20]. $\mathcal{L}$ and $\mathcal{R}$ be independent two-sided simple random walks on $\mathbbm{Z}$ with $\mathcal{L}_{0}=\mathcal{R}_{0}=0$. See Figure 3 for an illustration. We define two infinite arc diagrams (non-crossing perfect matchings of $\mathbbm{Z}$), one above $\mathbbm{R}$ and one below $\mathbbm{R}$, as follows. For $x_{1},x_{2}\in\mathbbm{Z}$ with $x_{1}<x_{2}$, we draw an arc above (resp. below) the real line joining $x_{1}$ and $x_{2}$ if and only if $\mathcal{L}_{x_{1}-1}=\mathcal{L}_{x_{2}}<\min_{y\in[x_{1},x_{2}-1]\cap\mathbbm{Z}}\mathcal{L}_{y}\qquad\text{(resp.}\>\mathcal{R}_{x_{1}-1}=\mathcal{R}_{x_{2}}<\min_{y\in[x_{1},x_{2}-1]\cap\mathbbm{Z}}\mathcal{R}_{y}\>\text{)};$ (1.5) c.f. (1.1). It is easy to see that there is exactly one arc above (resp. below) the real line incident to each $x\in\mathbbm{Z}$, and that the arcs above (resp. below) the real line can be taken to be non-intersecting. We then define the UIMS to be the set of the loops and bi-infinite paths formed by the union of the arcs in the above two arc diagrams. It is also possible to recover the walks $\mathcal{L}$ and $\mathcal{R}$ from the arc diagrams. Indeed, for each $x\in\mathbbm{Z}$, the increment $\mathcal{L}_{x}-\mathcal{L}_{x-1}$ is $+1$ (resp. $-1$) if and only if the arc above $\mathbbm{R}$ which is incident to $x$ has its other endpoint greater than (resp. less than) $x$. A similar statement holds for $\mathcal{R}$. The UIMS is easier to work with than a finite uniform meandric system since there is no conditioning on the walks. Most of our proofs will be in the setting of the UIMS. It is shown in [CKST19, Theorem 1] that either a.s. the collection of loops (and possibly bi-infinite paths) in the UIMS has a unique infinite path, or a.s. it has no infinite paths. This infinite path, if it exists, is called the infinite noodle in [CKST19]. It is not known rigorously which of these two possibilities holds. But, the following is conjectured in [CKST19]. ###### Conjecture 1.7 (​[CKST19]). Almost surely, the UIMS has no infinite path. Exactly as in the case of a finite meandric system, we can associate to the UIMS an infinite planar map $\mathcal{M}$ decorated by a bi-infinite Hamiltonian path $P$ and a collection of loops (and possibly bi-infinite paths) $\Gamma$. Namely, the vertex set of $\mathcal{M}$ is $\mathbbm{Z}$; the edges of $\mathcal{M}$ are the arcs above and below the real line together with the segments $[x-1,x]$ for $x\in\mathbbm{Z}$; the path $P$ traverses the vertex set $\mathbbm{Z}$ in left-right numerical order; and $\Gamma$ is the set of loops (and possibly bi-infinite paths) formed by the two arc diagrams as above. We will now state the infinite-volume analog of Conjecture 1.2. The $\sqrt{2}$-quantum cone is the most natural $\sqrt{2}$-LQG surface with the topology of the plane. It arises as the local limit of the $\sqrt{2}$-quantum sphere based at a point sampled from its associated area measure [DMS21, Proposition 4.13(ii)]. ###### Conjecture 1.8. Let $(\mathcal{M},P,\Gamma)$ be the infinite random planar map decorated by a bi-infinite Hamiltonian path and a collection of loops (and possibly bi- infinite paths) associated to the UIMS, as described just above. Then $(\mathcal{M},P,\Gamma)$ converges under an appropriate scaling limit to an independent triple consisting of a $\sqrt{2}$-LQG cone, a whole-plane SLE8 from $\infty$ to $\infty$, and a whole-plane CLE6. Conjecture 1.8 is consistent with Conjecture 1.7, since all of the loops in a whole-plane CLE6 are compact sets [MWW16]. Our main result concerning the UIMS gives a polynomial lower tail bound for the graph-distance diameter (and hence also the number of vertices) of the loop containing the origin. It turns out to be an easy consequence of one of the intermediate results in the proof of Theorem 1.5 (see, in particular, Proposition 4.4). ###### Theorem 1.9. Let $(\mathcal{M},\Gamma)$ be the planar map and collection of loops associated with the UIMS and let $\ell_{0}$ be the loop (or infinite path) in $\Gamma$ which passes through $0\in\mathbbm{Z}$. Then with $d$ as in (1.3), $\mathbbm{P}\mathopen{}\mathclose{{}\left[\mathopen{}\mathclose{{}\left(\text{$\mathcal{M}$-graph- distance diameter of $\ell_{0}$}}\right)\geq k}\right]\geq k^{-(d-1)-o(1)},\quad\text{as $k\rightarrow\infty$}.$ (1.6) Unlike in Theorem 1.5, we expect that the exponent in Theorem 1.9 is not optimal. Rather, we have the following conjecture. ###### Conjecture 1.10. Let $\ell_{0}$ be as in Theorem 1.9. Then with $\alpha=\frac{1}{2}(3-\sqrt{2})$ as in Conjecture 1.3, $\mathbbm{P}\mathopen{}\mathclose{{}\left[\mathopen{}\mathclose{{}\left(\text{$\mathcal{M}$-graph- distance diameter of $\ell_{0}$}}\right)\geq k}\right]\geq k^{-d(1-\alpha)-o(1)},\quad\text{as $k\rightarrow\infty$},$ (1.7) and $\mathbbm{P}\mathopen{}\mathclose{{}\left[\mathopen{}\mathclose{{}\left(\text{number of vertices in $\ell_{0}$}}\right)\geq k}\right]\geq k^{-(1/\alpha-1)-o(1)},\quad\text{as $k\rightarrow\infty$}.$ (1.8) If Conjecture 1.3 were true, then, at least heuristically, the same arguments as in the proof of Theorem 1.9 would lead to a proof of Conjecture 1.10. ### 1.5 The uniform infinite half-plane meandric system (UIHPMS) The uniform infinite half-plane meandric system (UIHPMS) is a natural variant of the UIMS which is half-plane like in the sense that it has a bi-infinite sequence of distinguished “boundary vertices” (points in $\mathbbm{Z}$ which are not disconnected from $\infty$ below the real line by any path in the UIHPMS), but “most” vertices are not boundary vertices. We will now give two equivalent definitions of this object, see Figure 4. _By cutting:_ Start with a sample of the UIMS on $\mathbb{Z}$, represented by a pair of arc diagrams as in Section 1.4. We assume that the arcs of the two arc diagrams are drawn in $\mathbb{R}^{2}$ in such a way that the arcs do not cross each other and each arc with endpoints $x<y$ is contained in the vertical strip $[x,y]\times\mathbbm{R}$ (the arc diagrams in all of the figures in this paper have this property). We cut all of the arcs below the real line which intersect the vertical ray $\\{1/2\\}\times(-\infty,0]$, leaving two unmatched ends corresponding to each arc. Then, we rewire successive pairs of unmatched ends to form new arcs. That is, we enumerate the unmatched ends from left to right as $\\{e_{j}\\}_{j\in\mathbbm{Z}}$, with positive indices corresponding to ends to the right of $\\{1/2\\}\times(-\infty,0]$ and non-positive indices corresponding to ends to the left. Then, we link up $e_{2j-1}$ and $e_{2j}$ for each $j\in\mathbbm{Z}$. See Figure 4 (Left). The resulting infinite collection of the loops (and possibly bi-infinite paths) is the UIHPMS. A point $x\in\mathbbm{Z}$ is called a boundary point if it can be connected to the ray $\\{1/2\\}\times(-\infty,0]$ by some continuous path without crossing any arc or the real line. Each endpoint of arcs which were cut is a boundary point, but there are also other boundary points. See Figure 4 (Left) for an illustration. It is clear from the random walk description of the meandric system (1.5) that each time $x\in\mathbbm{Z}$ at which the walk $\mathcal{R}$ attains a running infimum when run forward or backward from time 0 gives rise to a boundary point of the UIHPMS (but not every boundary point arises in this way). Hence, there are infinitely many boundary points. Denote by $J_{k}\in\mathbbm{Z}$ (resp. $J_{-k+1}\in\mathbbm{Z}$) the $k$th positive (resp. non-positive) boundary point. Note that in the UIHPMS we always have that $J_{0}=0$ and $J_{1}=1$ and that $J_{0}$ and $J_{1}$ are not joined by an arc. _By random walks:_ We also have a random walk description for the UIHPMS similar to that of the UIMS. The only difference is that we use a _reflected_ simple random walk instead of an ordinary simple random walk for the arc diagram below the real line. Let $\mathcal{L}$ and $\mathcal{R}$ be independent two-sided simple random walks on $\mathbb{Z}$ with $\mathcal{L}_{0}=\mathcal{R}_{0}=0$. For $x_{1},x_{2}\in\mathbbm{Z}$ with $x_{1}<x_{2}$, we draw an arc above (resp. below) the real line joining $x_{1}$ and $x_{2}$ if and only if $\mathcal{L}_{x_{1}-1}=\mathcal{L}_{x_{2}}<\min_{y\in[x_{1},x_{2}-1]\cap\mathbbm{Z}}\mathcal{L}_{y}\quad\text{(resp.}\>|\mathcal{R}|_{x_{1}-1}=|\mathcal{R}|_{x_{2}}<\min_{y\in[x_{1},x_{2}-1]\cap\mathbbm{Z}}|\mathcal{R}|_{y}\>\text{)};$ (1.9) c.f. (1.5). A point $x\in\mathbbm{Z}$ is called a boundary point if $|\mathcal{R}|$ crosses $1/2$ in between time $x-1$ and $x$, i.e., $\mathcal{R}_{x-1},\mathcal{R}_{x}\in\\{-1,0,1\\}$. As before, denote by $J_{k}\in\mathbbm{Z}$ (resp. $J_{-k+1}\in\mathbbm{Z}$) the $k$th positive (resp. non-positive) boundary point. See Figure 4 (Right) for an illustration. Since $\mathcal{R}_{0}=0$, we always have $J_{0}=0$, $J_{1}=1$, and $J_{0}$ not matched with $J_{1}$ as in the previous definition. Figure 4: Left: The UIHPMS constructed by cutting. The dash-dotted arcs straddling $0$ below the real line in the UIMS (defined in Section 1.4) are cut by the gray ray $\\{1/2\\}\times(-\infty,0]$, and these arcs are rewired successively to result in the UIHPMS. Boundary points are labeled as $\\{J_{k}\\}_{k\in\mathbbm{Z}}$ and are colored green together with their connecting arcs. Note that the points $J_{-2}$ and $J_{-3}$ are boundary points but are not endpoints of any cut arcs. Right: The UIHPMS constructed by random walks. We show a subset of the UIHPMS together with the graphs of the corresponding increments of the encoding walks $\mathcal{L}$ and $|\mathcal{R}|$. Each horizontal line below the graph of $\mathcal{L}$ (resp. |$\mathcal{R}$|) corresponds to an arc above (resp. below) the real line. The boundary points and arcs incident to them are colored green. The following lemma will be proven in Section 6.1 using a discrete version of Lévy’s theorem [Sim83] which relates (roughly speaking) the zero set and the running minimum times of a random walk. ###### Lemma 1.11. The law of the UIHPMS is the same under the above two definitions. Just as in the case of other types of meandric systems, we can view the UIHPMS as a planar map $\mathcal{M}^{\prime}$ decorated by a collection of loops (plus possibly bi-infinite paths) $\Gamma^{\prime}$ and a Hamiltonian path $P^{\prime}$. The law of the UIHPMS is invariant under even translations along the boundary. That is, if $\\{J_{k}\\}_{k\in\mathbbm{Z}}$ is the set of boundary points as above, then $(\mathcal{M}^{\prime}-J_{2k},P^{\prime}-J_{2k},\Gamma^{\prime}-J_{2k})\overset{d}{=}(\mathcal{M}^{\prime},P^{\prime},\Gamma^{\prime}),\quad\forall k\in\mathbbm{Z}.$ (1.10) This is clear from the random walk description. As in Conjecture 1.7, we can also ask if there exists an infinite path in the UIHPMS. The answer should be no if the scaling limit conjecture is true. Unlike in the whole-plane setting, we are able to prove this in the half-plane setting. ###### Theorem 1.12. Almost surely, the set $\Gamma^{\prime}$ for the UIHPMS has no infinite path. Theorem 1.12 will be proven in Section 6 via a purely discrete argument. See the beginning of Section 6 for an outline. The proof is completely independent of the proofs of our theorems for finite meandric systems and for the UIMS (stated in Sections 1.3 and 1.4). So, the proofs can be read in any order. In Conjecture 1.8, we discussed how the UIMS is conjecturally related to the $\sqrt{2}$-quantum cone. Another natural $\sqrt{2}$-LQG surface – with the half-plane topology instead of whole-plane topology – is the $\sqrt{2}$-quantum wedge. See [DMS21, Section 1.4] or [GHS23, Section 3.4] for a rigorous definition. In the continuum setting, cutting a $\sqrt{2}$-quantum cone by an independent whole-plane SLE2 from 0 to $\infty$ yields the $\sqrt{2}$-quantum wedge [DMS21, Theorem 1.5]. This is a continuum analog of the above cutting description of the UIHPMS. That is, in the scaling limit, the gray ray in Figure 4 should converge to some simple fractal curve from 0 to $\infty$. Cutting the previous loop-decorated whole-plane along this curve yields the half-plane (or whole-plane with a slit) topology. The points located immediately on the left or right side of the curve become boundary points. Thus, Conjecture 1.8 has a natural half-plane version as follows. ###### Conjecture 1.13. Let $(\mathcal{M}^{\prime},P^{\prime},\Gamma^{\prime})$ be the infinite random planar map decorated by a bi-infinite Hamiltonian path and a collection of loops associated to the UIHPMS, as described just above. Then $(\mathcal{M}^{\prime},P^{\prime},\Gamma^{\prime})$ converges under an appropriate scaling limit to an independent triple consisting of a $\sqrt{2}$-LQG wedge, a space-filling SLE8 from $\infty$ to $\infty$ in the half-plane, and a CLE6 in the half-plane. In the setting of Conjecture 1.13, the boundary vertices of $\mathcal{M}^{\prime}$ should correspond to boundary points for the $\sqrt{2}$-LQG wedge. For example, if $(\mathcal{M}^{\prime},P^{\prime},\Gamma^{\prime})$ is embedded into the half- plane appropriately (e.g., via some version of Tutte embedding or circle packing), in such a way that the boundary vertices of $\mathcal{M}^{\prime}$ are mapped to points on the real line, then one should have the convergence of the embedded objects toward the $\sqrt{2}$-LQG wedge together with SLE8 and CLE6. Starting from a CLE6 in the half-plane, one can construct a chordal SLE6 curve from 0 to $\infty$ by concatenating certain arcs of boundary-touching CLE6 loops (see the proof of [She09, Theorem 5.4]). The analogous path can be also constructed in the UIHPMS by leaving exactly one boundary point unmatched as follows. We define the pointed infinite half-plane meandric system (PIHPMS) by rewiring lower arcs between non-positive boundary points in the UIHPMS as follows. Start with the UIHPMS and recall that $J_{0}=0$ is not matched with $J_{1}=1$. Remove the arcs below the real line which join $J_{2k-1}$ and $J_{2k}$ for each $k\leq 0$. Then, add new arcs joining $J_{2k-2}$ and $J_{2k-1}$ for each $k\leq 0$, in such a way that the new arcs do not cross any other arcs. Note that now $J_{0}=0$ is unmatched. This new configuration is defined to be the PIHPMS and we think of it as pointed at $J_{0}=0$. One can also describe the PIHPMS in terms of random walks by replacing $|\mathcal{R}|_{\cdot}$ in (1.9) with the modified walk $\mathcal{R}^{\circ}_{\cdot}:=|\mathcal{R}|_{\cdot}-\textbf{1}_{\\{\cdot\geq 0\\}}.$ (1.11) As there is no $x<0$ such that $\mathcal{R}^{\circ}_{x-1}=\mathcal{R}^{\circ}_{0}=-1$, we have a special boundary point $J_{0}=0$ which is not incident to any arc below the real line as desired. See Figure 5 (Top Left) for an illustration. As per usual, we view the PIHPMS as a planar map $\mathcal{M}^{\circ}$ decorated by a Hamiltonian path $P^{\circ}$ and a collection of loops (plus possibly infinite paths) $\Gamma^{\circ}$. Note that the path started from $J_{0}=0$ must be an infinite path because no arc below the real line is incident to the origin, so the path cannot come back to its starting point to form a loop. We prove that this is the unique infinite path. ###### Proposition 1.14. Consider the PIHPMS as defined above. Almost surely, there is a unique infinite path $\gamma^{\circ}\in\Gamma^{\circ}$. This path starts at $J_{0}=0$ and hits infinitely many points in each of $\\{J_{k}\\}_{k<0}$ and $\\{J_{k}\\}_{k>0}$. The proof of Proposition 1.14 is given in Section 6, and uses many of the same ideas as in the proof of Theorem 1.12. Another conjecture follows accordingly. Figure 5: Top left: The random walks encoding a sample of the PIHPMS obtained from the previous example in Figure 4 (Right). The yellow strip in the PIHPMS highlights the special boundary point $J_{0}=0$ which is not incident to any arc below the real line as a consequence of the fact that there is no $x<0$ such that $\mathcal{R}^{\circ}_{x-1}=\mathcal{R}^{\circ}_{0}=-1$; see the yellow strip in the walk. Bottom left/right: A simulation of a finite-volume version (i.e., with the topology of the disk) of the PIHPMS with two marked points, of size $n=9\cdot 10^{6}$. The bottom left picture shows the arc diagrams and the right picture shows the associated planar map, drawn in the disk via the Tutte embedding [GMS21]. The red curve is conjectured to converge to a chordal SLE6 between the two marked boundary points. See Section 7.4 for the details of simulations. ###### Conjecture 1.15. Let $(\mathcal{M}^{\circ},P^{\circ},\Gamma^{\circ})$ be the infinite random planar map decorated by a bi-infinite Hamiltonian path and a collection of loops (plus possibly infinite paths) associated to the PIHPMS, as described just above. Also, let $\gamma^{\circ}$ be the path started from 0 in $\Gamma^{\circ}$ (which is a.s. the unique infinite path by Proposition 1.14). Then $(\mathcal{M}^{\circ},P^{\circ},\gamma^{\circ},\Gamma^{\circ}\setminus\gamma^{\circ})$ converges under an appropriate scaling limit to a $\sqrt{2}$-LQG wedge, a space-filling SLE8 from $\infty$ to $\infty$ in the half-plane, a chordal SLE6 from $0$ to $\infty$ in the half-plane, and a CLE6 in the complement of the SLE6 curve (i.e., a union of conditionally independent CLE6s in the complementary connected components). ###### Remark 1.16. One can also consider finite or infinite meandric systems constructed from a pair of random walks which are correlated rather than independent, i.e., the step distribution of the pair of walks assigns different weights to steps in $\\{(-1,-1),(1,1)\\}$ and steps in $\\{(-1,1),(1,-1)\\}$. We do not have a simple combinatorial description for the random meandric systems obtained in this way, so we do not emphasize them in this paper. Based on mating of trees theory [DMS21, GHS23], it is natural to conjecture that the associated decorated planar map converges to $\gamma$-LQG decorated by a space-filling SLE${}_{16/\gamma^{2}}$ and an independent CLE6, where $\gamma\in(0,2)$ is chosen so that the correlation of the encoding walks is $-\cos(4\pi/\gamma^{2})$. We have run some numerical simulations (similar to Section 7.4) which are consistent with this. Our proofs of Theorems 1.5 and 1.9 work verbatim in the case of correlated walks, with $d$ replaced by the dimension of $\gamma$-LQG. However, the upper bound for the diameter of $\mathcal{M}_{n}$ in Proposition 1.4 uses estimates from [GP21] which are only proven for $\gamma=\sqrt{2}$. Most of our proofs for the UIHPMS do not work in the case of correlated walks since we cannot apply the discrete Lévy theorem (6.2) to one coordinate of the walk independently from the other. ## 2 Percolation interpretation The main goal of this section is to explain how meandric systems can be viewed as a model of (critical) percolation on planar maps. Throughout the section we assume that the reader is familiar with the basic terminology and results in percolation theory. This section is included only for intuition and motivation: it is not needed for the proofs of our main results. ### 2.1 Non-crossing perfect matchings and non-crossing integer partitions We start by recalling a classical bijection between non-crossing perfect matchings and non-crossing integer partitions, see for instance [GNP20, Section 3]. We write a partition $\pi$ of a finite set $A\subset\mathbbm{R}$ as $\pi=\\{V_{1},\dots,V_{k}\\}$, where $V_{1},\dots,V_{k}$ are called the blocks of $\pi$ and are non-empty, pairwise disjoint sets, with $\cup_{i}V_{i}=A$. We say that a partition $\pi$ is non-crossing when it is not possible to find two distinct blocks $V,W\in\pi$ and numbers $a<b<c<d$ in $A$ such that $a,c\in V$ and $b,d\in W$. Equivalently, $\pi$ is non-crossing if it can be plotted as a planar graph with the vertices in $A$ arranged on the real line, so that the blocks of $\pi$ are the connected components of the planar graph drawn in the upper-half plane; see the orange planar graph in Figure 6 for an example. Figure 6: We plot in orange the planar graph associated with the non-crossing partition $\pi=\\{\\{2-1/2,10-1/2,14-1/2,16-1/2\\},\\{4-1/2,6-1/2,8-1/2\\},\\{12-1/2\\}\\}$ of the points $2\mathbbm{Z}\cap[1,16]-1/2$ and in black the corresponding planar graph of the non-crossing perfect matching $\psi(\pi)$ of the points $[1,16]\cap\mathbbm{Z}$. Note that the black arcs are chosen so that they “block” new potential orange edges. We can now describe the aforementioned bijection. Given a non-crossing partition $\pi=\\{V_{1},\dots,V_{k}\\}$ of the points $2\mathbbm{Z}\cap[1,2n]-1/2$, we consider the unique non-crossing perfect matching $\psi(\pi)$ of $[1,2n]\cap\mathbbm{Z}$ such that for every pair of points $z<w\in 2\mathbbm{Z}\cap[1,2n]-1/2$ not contained in the same block $V_{i}$, there exists a pair of matched points $x,y\in[1,2n]\cap\mathbbm{Z}$ in $\psi(\pi)$ such that either $x<z<y<w$ or $z<x<w<y$. See Figure 6 for a graphical interpretation. Conversely, given a non-crossing perfect matching $m$ of $[1,2n]\cap\mathbbm{Z}$, we construct a non-crossing partition $\psi^{-1}(m)$ of the points $2\mathbbm{Z}\cap[1,2n]-1/2$ as follows: for every pair of points $z,w\in 2\mathbbm{Z}\cap[1,2n]-1/2$ we say that $z\sim_{m}w$ if there is no pair of matched points $x,y$ in $m$ with either $x<z<y<w$ or $z<x<w<y$. We then set the equivalent classes of $(2\mathbbm{Z}\cap[1,2n]-1/2,\sim_{m})$ to be the blocks of $\psi^{-1}(m)$. Note that the fact that $\psi$ and $\psi^{-1}$ are inverse maps is immediate. ### 2.2 Meandric systems as boundaries of clusters of open edges Given a non-crossing perfect matching $m$ of $[1,2n]\cap\mathbbm{Z}$, one can also consider (in the previous description of the inverse map $\psi^{-1}$) the points $2\mathbbm{Z}\cap[0,2n]+1/2$ (see the green points in Figure 7). Then, as before, one determines a second non-crossing integer partition $\widetilde{\psi}^{-1}(m)$ of the points $2\mathbbm{Z}\cap[0,2n]+1/2$ (see for instance the green planar graph on the upper-half plane in Figure 7). We further notice that any element of the triple $(m,\psi^{-1}(m),\widetilde{\psi}^{-1}(m))$ determines the two other elements of the triple. Figure 7: A meandric system $m=(m_{1},m_{2})$ plotted in black together with its two pairs of non-crossing integer partitions $(\psi^{-1}(m_{1}),\psi^{-1}(m_{2}))$ and $(\widetilde{\psi}^{-1}(m_{1}),\widetilde{\psi}^{-1}(m_{2}))$ in orange and green, respectively. Consider now a meandric system of size $n$, which is a pair $(m_{1},m_{2})$ of non-crossing perfect matchings of $[1,2n]\cap\mathbbm{Z}$; the first above and the second below the real line. Using the bijections $\psi$ and $\widetilde{\psi}$ described above, one can associate to the meandric system $(m_{1},m_{2})$ two pairs of non-crossing integer partitions of $2\mathbbm{Z}\cap[1,2n]-1/2$ and $2\mathbbm{Z}\cap[0,2n]+1/2$: the first pair (resp. the second pair) is obtained applying the map $\psi^{-1}$ (resp. the map $\widetilde{\psi}^{-1}$) to $m_{1}$ and $m_{2}$, obtaining the orange (resp. the green) planar graph in Figure 7. From now on we will refer to these two pairs of non-crossing integer partitions as the orange and the green planar graphs (associated to a meandric system). We now present a new way of thinking about a meandric system of size $n$ closer to percolation models (in what follows we explain what are the classical quantities in percolation models in our setting; c.f. Figure 7). _Main lattice:_ This is the planar map whose vertices are the union of the green and orange vertices, and whose edges are the union of the green and orange edges and the edges corresponding to the real line (between consecutive vertices of different colors). _Dual lattice:_ This is the planar map that we previously associated with a meandric system, i.e. the map whose vertices are the black vertices, and whose edges are the black edges and the edges corresponding to the real line (between consecutive black vertices). _Open edges in the main lattice:_ These are the orange edges. _Closed edges in the main lattice:_ These are the green edges. _Boundary of clusters of open edges:_ These are the black loops in the meandric system. Note that if one considers a uniform meandric system (or equivalently a uniform pair of non-crossing integer partitions), then one obtains (through the interpretation above) a model of percolation on random planar maps. Analogously, if one considers the UIMS or the UIHPMS (Sections 1.4 and 1.5), then one obtains (through the obvious adaptations of the constructions above222We do not give the details of these constructions since they are not needed to continue our discussion.) some natural model of percolation on infinite random planar maps. One of the main difficulties in the study of these models is that the randomness for the environment (i.e. the planar map) and the randomness for the percolation are strongly coupled. See also the beginning of Section 2.5 for a second major difficulty. ### 2.3 Meandric systems and critical percolation on random planar maps Our interpretation of meandric systems as a percolation model described in the previous section, gives also some natural new interpretations of our results. In particular, Theorem 1.5, which states that a uniform meandric system admits loops whose graph distance diameter is nearly of the same order as the graph distance diameter of the associated planar map, implies that our percolation model admits macroscopic clusters. Hence, it does not behave like subcritical Bernoulli percolation on a fixed lattice. Due to Corollary 3.3 below, a.s. the UIMS has a bi-infinite path if and only if our above percolation model has an infinite open cluster. Theorem 1.12 implies that there is no infinite cluster for the percolation model on the UIHPMS. Due to the construction of the UIHPMS by cutting the UIMS along an infinite ray (Section 1.5), this is roughly analogous to the statement that there is no infinite cluster for, say, Bernoulli percolation on $\mathbbm{Z}^{2}\setminus(\\{0\\}\times\mathbbm{Z}_{+})$. In particular, our percolation model does not behave like supercritical Bernoulli percolation on a fixed lattice. By combining the preceding two paragraphs, we see that our percolation model is in some sense “critical” (see also Proposition 2.1 for further evidence). Determining whether there is a bi-infinite path in the UIMS (Conjecture 1.7) is therefore analogous to determining whether there is percolation at criticality. We note that the interpretation of meandric systems as a critical percolation model is also consistent with Conjectures 1.2, 1.8 and 1.13, which state that the scaling limit of the loops of a meandric system should be the conformal loop ensemble with parameter $\kappa=6$. Indeed, the latter is also conjectured to be the scaling limit of the cluster interfaces for critical percolation models on various deterministic discrete lattices. This conjecture has been proved in the case of critical site percolation on the triangular lattice [Smi01, CN06, CN08]. ### 2.4 Box crossings A classical result for critical Bernoulli ($p=1/2$) bond percolation on $\mathbb{Z}^{2}$ is that given a box $B_{n}=\mathbbm{Z}^{2}\cap\\{[0,n]\times[0,n+1]\\}$, then the probability that there exists a path of open edges connecting the top side of $B_{n}$ to the bottom side of $B_{n}$ is always equal to $1/2$, independently of the size $n$ of the box. We prove the analogous result in the context of the UIMS. We start by defining a natural notion of box for the UIMS. Given the UIMS together with its orange and green planar graphs as in the left-hand side of Figure 8, we define the box of size $n$ rooted at $x\in\mathbb{Z}+\frac{1}{2}$ to be the collection of (black/orange/green) vertices in $[x,x+2n-1]$ and (black/orange/green) edges with at least one endpoint in $(x,x+2n-1)$. We denote such box by $B_{n}(x)$. See Figure 8 for an example. Note that with this definition any box of size $n$ contains the same number of green and orange vertices. Figure 8: Left: A portion of the UIMS together with its orange and green planar graphs. A box of size $2$ rooted at $1/2$ is highlighted in yellow. The bottom-green edge between $-3.5$ and $2.5$ and the top-green edge between $2.5$ and $4.5$ form a bottom-to-top green crossing of the box $B_{2}(1/2)$. Note that $B_{2}(1/2)$ contains also a top-to-bottom green crossing of the box. Right: An example of another box. Note that both the top and bottom green edges from $0.5$ to $2.5$ crosses both the top-left boundary and the bottom- left boundary of the box $B_{2}(1/2)$, according to our definition below. Given a box $B_{n}(x)$, we say that an edge crosses the top-left (resp. bottom-left, top-right, bottom-right) boundary of the box, if that edge touches the vertical line $\\{x\\}\times[0,\infty)$ (resp. $\\{x\\}\times(-\infty,0]$, $\\{x+2n-1\\}\times[0,\infty)$, $\\{x+2n-1\\}\times(-\infty,0]$), where we highlight that the point $(x,0)$ (resp. $(x+2n-1,0)$) is included in the line. See Figure 8 for an example. We then say that there exists a top-to-bottom orange crossing (resp. bottom-to- top orange crossing) of $B_{n}(x)$ if there exists a path of all orange edges such that the first edge in the path crosses the top-left (resp. bottom-left) boundary of $B_{n}(x)$, the last edge in the path crosses the bottom-right (resp. top-right) boundary of $B_{n}(x)$, and all the other edges have both extremities in $B_{n}(x)$. We similarly define top-to-bottom and bottom-to-top green crossings. We can now state our analogous result for “the probability that there exists a top-to-bottom crossing of open edges in a box is 1/2”. ###### Proposition 2.1. Consider the UIMS together with its green and orange planar graphs. For all $x\in\mathbb{Z}+\frac{1}{2}$ and all $n>0$, the probability that there exists a top-to-bottom orange crossing of the box $B_{n}(x)$ is 1/2. ###### Proof. Fix $x\in\mathbb{Z}+\frac{1}{2}$ and $n>0$. Note that by the construction of the orange and green planar graphs (recall from Section 2.2 how the orange arcs determine the green arcs), exactly one of the following two events holds: either $O=\\{\text{There exists a top-to-bottom orange crossing of the box }B_{n}(x)\\}$ or $G=\\{\text{There exists a bottom-to-top green crossing of the box }B_{n}(x)\\}.$ Moreover, since $B_{n}(x)$ contains the same number of green and orange vertices and the events $O$ and $G$ are symmetric, then they must must have the same probability. Hence $\mathbb{P}(O)=\mathbb{P}(G)=1/2$. ∎ ### 2.5 The lack of positive association and two open problems Proposition 2.1 is a further evidence that meandric systems behave like a critical percolation model on a planar map and also suggests that it might be possible to prove analogs of other standard results in the theory of percolation models. Unfortunately, as already mentioned in Section 1.3, we are not aware of any positive association (FKG) inequality in our setting and for the moment all the notions of monotonicity that we tried do not satisfy FKG. This discussion naturally leads to the following open problem. ###### Question 2.2. Is there a natural notion of monotonicity in infinite meandric configurations such that some type of positive association (FKG) inequality holds (for instance) for our notion of top-to-bottom orange crossings of proper boxes? Another interesting question, which might be relevant to Conjecture 1.2 (or one of its variants), is to compute the following loop crossing probability. Recall the definition of the box $B_{n}(x)$ given at the beginning of Section 2.4. We define a top-to-bottom loop crossing of $B_{n}(x)$ in the same way as we defined a top-to-bottom orange crossing of $B_{n}(x)$, where in the definition orange edges are replaced by black edges. ###### Question 2.3. Consider the UIMS. What is the asymptotic probability as $n\to\infty$ that there exists a top-to-bottom loop crossing of $B_{n}(0)$? Conjecture 1.8 gives a natural candidate for the answer to the question above, which we now explain (assuming a certain familiarity with SLEs). Building on Conjecture 1.8, the points in $B_{n}(0)$ are expected to converge under an appropriate scaling limit to the points visited by the whole-plane SLE8 $\eta$ between time $0$ and $1$.333Here we assume that $\eta$ is parametrized by $\sqrt{2}$-LQG area with respect to an independent unit area quantum cone, so that $\eta:\mathbbm{R}\rightarrow\mathbbm{R}$. The set $\eta((-\infty,0])\cap\eta([0,\infty))$ is the union of two disjoint SLE2-type curves [DMS21, Footnote 4], one called the left-boundary of $\eta((-\infty,0])$ and the other one called the right-boundary of $\eta((-\infty,0])$. The same holds for $\eta((-\infty,1])\cap\eta([1,\infty))$. Moreover, the left-boundaries (resp. right-boundaries) of $\eta((-\infty,0])$ and $\eta((-\infty,1])$ a.s. merge into each other. As a consequence, the set $\eta([0,1])$ forms a topological rectangle. We refer to the piece of the left boundary of $\eta((-\infty,1])$ not in common with the left boundary of $\eta((-\infty,0])$ (resp. the piece of the right boundary of $\eta((-\infty,0])$ not in common with the left boundary of $\eta((-\infty,1])$) as the top-left boundary (resp. bottom-right boundary) of $\eta([0,1])$. We expect that the probability in Question 2.3 converges to the probability that there exists a continuous portion of a loop in the whole-plane CLE6 of Conjecture 1.8 crossing $\eta([0,1])$ from its top-left boundary to its bottom-right boundary, without leaving $\eta([0,1])$. We highlight that the latter crossing probability is not known explicitly, and that computing it is an interesting problem in its own right. ## 3 Existence of macroscopic loops in the UIMS Throughout this section, we let $(\mathcal{M},\Gamma)$ be the infinite planar map (with vertex set $\mathbbm{Z}$) and collection of loops (and possibly bi- infinite paths) associated with the UIMS, as defined in Section 1.4. ###### Definition 3.1. Let $A,B\subset\mathbbm{R}^{2}$ and let $\ell$ be a loop or a path $\mathbbm{R}^{2}\cup\\{\infty\\}$. We say that $\ell$ disconnects $A$ from $B$ if every path from $A$ to $B$ hits $\ell$. The loops (and possibly bi-infinite paths) of the UIMS can be viewed as loops in $\mathbbm{R}^{2}$ which hit $\mathbbm{R}$ only at integer points and which are defined modulo orientation-preserving homeomorphisms from $\mathbbm{R}^{2}$ to $\mathbbm{R}^{2}$ which fix $\mathbbm{R}$. If $A,B\subset\mathbbm{R}\cup\\{\infty\\}$, such a homeomorphism does not alter whether a loop disconnects $A$ from $B$. Hence it makes sense to talk about loops in $\Gamma$ disconnecting subsets of $\mathbbm{R}\cup\\{\infty\\}$. The goal of this section is to prove the following theorem, which can roughly speaking be thought of as saying that $\Gamma$ has a positive chance to admit macroscopic loops or paths at all scales. This theorem will eventually be combined with estimates for distances in the infinite planar map $\mathcal{M}$ (which we prove in Sections 4 and 5.1) to prove Theorems 1.5 and 1.9, see Section 4.3. ###### Theorem 3.2. For each sufficiently large $n\in\mathbbm{N}$, it holds for each $C>1$ that with probability at least $5-2\sqrt{6}\approx 0.1010$, at least one of the following two conditions is satisfied: 1. A. There is a loop in $\Gamma$ which disconnects $[-n,n]$ from $\infty$ and which hits a point of $[n,(2C+1)n+3]\cap\mathbbm{Z}$. 2. B. There is a loop or an infinite path in $\Gamma$ which hits a point in each of $[-n,n]\cap\mathbbm{Z}$ and $\mathbbm{Z}\setminus[-Cn,Cn]$. When we apply Theorem 3.2, we will take $C$ large but fixed independently of $n$. For such a choice of $C$, a loop satisfying either of the two conditions of Theorem 3.2 should be thought of as being “macroscopic”, in the sense that it should give rise to a non-trivial loop when we send $n\rightarrow\infty$ and pass to an appropriate scaling limit (c.f. Conjecture 1.8). Theorem 3.2 implies the following independently interesting corollary. We note that a similar statement for the $O(n)$ loop model on the hexagonal lattice, for a certain range of parameter values, is proven in [CGHP20, Theorem 1]. ###### Corollary 3.3. Exactly one of the following two conditions occurs with probability one, and the other occurs with probability zero: 1. A′. There is an infinite path in $\Gamma$. 2. B′. For each $x\in\mathbbm{Z}$, there are infinitely many loops in $\Gamma$ which disconnect $x$ from $\infty$. It is possible to prove Corollary 3.3 via a more direct argument which does not use Theorem 3.2, but we will deduce it from Theorem 3.2 for convenience. ###### Proof of Corollary 3.3. By the zero-one law for translation invariant events, the event that $\Gamma$ has an infinite path has probability zero or one (see also [CKST19, Theorem 1]). Loops and infinite paths in $\Gamma$ cannot cross each other, so for any $x\in\mathbbm{Z}$ which is hit by an infinite path in $\Gamma$, there is no loop in $\Gamma$ which disconnects $x$ from $\infty$. Therefore, to prove the corollary it suffices to assume that $\Gamma$ a.s. has no infinite paths and show that this implies that for each $x\in\mathbbm{Z}$, a.s. there are infinitely many loops in $\Gamma$ which disconnect $x$ from $\infty$. The definition (1.5) of $(\mathcal{M},\Gamma)$ implies that translating by a fixed $x\in\mathbbm{Z}$ preserves the law of $(\mathcal{M},\Gamma)$. So, we can restrict attention to the case $x=0$. Let $n\in\mathbbm{N}$ be large enough so that the conclusion of Theorem 3.2 is satisfied. Since we are assuming that $\Gamma$ has no infinite paths, as $C\rightarrow\infty$ ($n$ fixed) the probability that condition B of Theorem 3.2 tends to zero. Hence, the theorem implies that with probability at least $5-2\sqrt{6}$, there is a loop in $\Gamma$ which disconnects $[-n,n]$ from $\infty$. Let $G_{n}$ be the event that this is the case, so that $\mathbbm{P}[G_{n}]\geq 5-2\sqrt{6}$ for each large enough $n\in\mathbbm{N}$. Then $\mathbbm{P}\mathopen{}\mathclose{{}\left[\bigcap_{m=1}^{\infty}\bigcup_{n=m}^{\infty}G_{n}}\right]\geq 5-2\sqrt{6},$ i.e., with probability at least $5-2\sqrt{6}$, there are infinitely many loops in $\Gamma$ which disconnect 0 from $\infty$. By the zero-one law for translation invariant events, this in fact holds with probability one. ∎ We now give an overview of the proof of Theorem 3.2. If $\Gamma$ admits an infinite path, it is straightforward to check that condition B in the proposition statement holds with high probability when $n$ is large. So, we can assume without loss of generality that there is no infinite path. The key idea of the proof is that a meandric system satisfies rather rigid parity properties. In particular, any distinct $x,y\in\mathbbm{Z}$ such that $x-1/2$ and $y-1/2$ are not separated by a loop or an infinite path have to have the same parity (Lemma 3.4). Under the assumption that there is no infinite path in $\Gamma$, this allows us to force the existence of a macroscopic loop as follows. Fix $C>1$ and let $E_{n}$ be the event that there is a loop in $\Gamma$ which hits only vertices of $[-Cn,Cn]\cap\mathbbm{Z}$ and which disconnects $[-n,n]$ from $\infty$ (see Figure 10). If $\mathbbm{P}[E_{n}]$ is bounded below by an $n$-independent constant, then condition A in the theorem statement holds with uniformly positive probability. So, we can assume that $\mathbbm{P}[E_{n}]$ is small, i.e., $\mathbbm{P}[E_{n}^{c}]$ is large. The event $E_{n}$ depends only on the restriction of the encoding walk $(\mathcal{L},\mathcal{R})$ to $[-Cn,Cn]\cap\mathbbm{Z}$ (Lemma 3.5). Therefore, if $x\in\mathbbm{Z}$ with $x>2Cn$, then the probability that $E_{n}^{c}$ occurs, and also $E_{n}^{c}$ occurs with the translated map $\mathcal{M}-x$ in place of $\mathcal{M}$, is $\mathbbm{P}[E_{n}^{c}]^{2}$. If we choose $x$ to be odd, then, using the definition of $E_{n}$, we can show that with probability at least $\mathbbm{P}[E_{n}^{c}]^{2}/8$, there are pairs of points $(y,y_{x})$ with $y\in[-n,n]$ and $y_{x}\in[x-n,x+n]$ which have opposite parity and which are not disconnected from $\infty$ by loops whose vertex sets are contained in $[-Cn,Cn]$ and $[x-Cn,x+Cn]$, respectively. Since $y$ and $y_{x}$ have opposite parity, there has to be a loop which disconnects $y$ from $y_{x}$ (Lemma 3.7 and Figure 11). It is then straightforward to check that this loop has to satisfy one of the two conditions of Theorem 3.2. The probability $5-2\sqrt{6}$ in the proposition statement comes from considering the “worst case” possibility for $\mathbbm{P}[E_{n}]$. We now commence with the proof, starting with the requisite parity lemma. Figure 9: Illustration of the proof of Lemma 3.4. We prove the following statement: suppose $x-1/2$ and $y-1/2$ are not disconnected by any loop or infinite path in $\Gamma$, and each point $z-1/2$ with $z\in[x+1,y-1]\cap\mathbbm{Z}$ is disconnected from $x$ and $y$ by a loop or an infinite path in $\Gamma$. Then any loop or infinite path in $\Gamma$ must hit $[x,y-1]\cap\mathbbm{Z}$ an even number of times (examples of a loop and an infinite path are shown in purple and brown, respectively). This implies that $y-x$ is even, which gives the contrapositive of the lemma statement. ###### Lemma 3.4. Let $x\in\mathbbm{Z}$ be even and let $y\in\mathbbm{Z}$ be odd. There is a loop or an infinite path in $\Gamma$ which disconnects $x-1/2$ from $y-1/2$. ###### Proof. See Figure 9 for an illustration. Define an equivalence relation on $\mathbbm{Z}$ by $x\sim y$ if there is no loop or infinite path in $\Gamma$ which disconnects $x-1/2$ from $y-1/2$. Let $X$ be any equivalence class. If suffices to show that every element of $X$ has the same parity. By considering the elements of $X$ in left-right numerical order, it suffices to show the following. If $x$ and $y$ are two consecutive elements of $X$ (i.e., $x,y\in X$, $x<y$, and there is no element of $X$ in $[x,y]$) then $x$ and $y$ have the same parity. Since $x\sim y$, there exists a path $P$ in $\mathbbm{R}^{2}$ from $x-1/2$ to $y-1/2$ which does not hit any loop or infinite path in $\Gamma$. By erasing loops made by $P$, we can take $P$ to be a simple path. Since loops in $\Gamma$ do not intersect $\mathbbm{R}$ except at integer points, we can also arrange that $P$ hits $(x-1/2,x)$ and $(y-1/2,y)$ only at their endpoints. Since $x$ and $y$ are consecutive elements of $X$, the path $P$ does not hit $[x,y-1]$ (otherwise, there would be an element of $X$ between $x$ and $y$). Therefore, $[x-1/2,y-1/2]\cup P$ is a simple closed loop in $\mathbbm{R}^{2}$. By the Jordan curve theorem, there are exactly two connected components of $\mathbbm{R}^{2}\setminus([x-1/2,y-1/2]\cup P)$ whose common boundary is $[x-1/2,y-1/2]\cup P$. Let $U$ and $V$ be these two connected components. Consider a loop $\ell\in\Gamma$ which hits a point of $[x,y-1]\cap\mathbbm{Z}$. We traverse $\ell$ counterclockwise, say, started from a point of $\ell\setminus\mathbbm{R}$. Since $\ell$ cannot intersect $\mathbbm{R}$ without crossing it (by the definition of a meandric system) and $\ell$ cannot hit $P$ (by our choice of $P$), the number of times that $\ell$ intersects $[x,y-1]\cap\mathbbm{Z}$ is equal to the number of times that $\ell$ crosses from $U$ to $V$ or from $V$ to $U$. Since $\ell$ starts and ends at the same point, the number of times that $\ell$ crosses from $U$ to $V$ is equal to the number of times that $\ell$ crosses from $V$ to $U$. Hence the number of times that $\ell$ intersects $[x,y-1]\cap\mathbbm{Z}$ is even. Similarly, if $\Gamma$ has an infinite path, then the number of times that this infinite path intersects $[x,y-1]\cap\mathbbm{Z}$ is even. Since every point of $[x,y-1]\cap\mathbbm{Z}$ is hit by either a loop or an infinite path of $\Gamma$, we get that $[x,y-1]\cap\mathbbm{Z}$ is even. Hence, $y-x$ is even. ∎ Figure 10: Illustration of the event $E_{n}$, which depends on the constant $C>1$. Fix $C>1$ and for $n\in\mathbbm{N}$, let $E_{n}$ be the event that the following is true: 1. $(\dagger)$ There exists a loop in $\Gamma$ which hits only vertices in $[-Cn,Cn]\cap\mathbbm{Z}$ and which disconnects $[-n,n]$ from $\infty$. Note that $E_{n}$ depends on $C$. See Figure 10 for an illustration. ###### Lemma 3.5. The event $E_{n}$ is determined by the encoding walk increment $(\mathcal{L},\mathcal{R})|_{[-Cn,Cn]\cap\mathbbm{Z}}$. ###### Proof. The event $E_{n}$ depends only on the arcs of the upper and lower arc diagrams for $\mathcal{M}$ which join points of $[-Cn,Cn]\cap\mathbbm{Z}$. These arcs are determined by $(\mathcal{L},\mathcal{R})|_{[-Cn,Cn]\cap\mathbbm{Z}}$ by the relationship between the arc diagrams and the walks (1.5). ∎ It is clear that the conclusion of Theorem 3.2 is satisfied if $\mathbbm{P}[E_{n}]\geq 5-2\sqrt{6}$. So, we need to show that the conclusion of the theorem is also true if $\mathbbm{P}[E_{n}^{c}]\geq 1-(5-2\sqrt{6})$. The following elementary topological lemma, in conjunction with Lemma 3.4, will help us do so. ###### Lemma 3.6. If $E_{n}^{c}$ occurs, then there exists $y\in[-n+1,n]\cap\mathbbm{Z}$ such that $y-1/2$ is not disconnected from $\infty$ by any loop in $\Gamma$ which hits only vertices in $[-Cn,Cn]\cap\mathbbm{Z}$. ###### Proof. Let $\Gamma_{Cn}$ be the set of loops in $\Gamma$ which hit only vertices in $[-Cn,Cn]\cap\mathbbm{Z}$. Then $\Gamma_{Cn}$ is a finite collection of simple loops in $\mathbbm{R}^{2}$ which do not intersect each other. Let $\Gamma_{Cn}^{\prime}$ be the set of outermost loops in $\Gamma_{Cn}$ (i.e., those which are not disconnected from $\infty$ by any other loop in $\Gamma_{Cn}$). For $\ell\in\Gamma_{Cn}^{\prime}$, let $U_{\ell}$ be the open region disconnected from $\infty$ by $\ell$. Then the closures $\overline{U}_{\ell}$ of the sets $U_{\ell}$ for $\ell\in\Gamma_{Cn}^{\prime}$ are disjoint (since the loops $\ell\in\Gamma_{Cn}^{\prime}$ are disjoint and non-nested) and their union is the same as the set of points which are disconnected from $\infty$ by the union of the loops in $\Gamma_{Cn}$. By the definition ($\dagger$) of $E_{n}$, if $E_{n}^{c}$ occurs then $[-n,n]$ is not contained in $\overline{U}_{\ell}$ for any $\ell\in\Gamma_{Cn}^{\prime}$. Since $[-n,n]$ is connected and the sets $\overline{U}_{\ell}$ for $\ell\in\Gamma_{Cn}^{\prime}$ are closed and disjoint, it follows that $[-n,n]$ is not contained in the union of the sets $\overline{U}_{\ell}$ for $\ell\in\Gamma_{Cn}^{\prime}$. Hence, there must be a point $z\in[-n,n]$ which is not contained in $\overline{U}_{\ell}$ for any $\ell\in\Gamma_{Cn}^{\prime}$. The set of such $z$ is an open subset of $[-n,n]$, so we can take $z\in[-n,n]\setminus\mathbbm{Z}$. The point $z$ is not disconnected from $\infty$ by any loop in $\Gamma_{Cn}$. Since loops in $\Gamma$ hit $\mathbbm{R}$ only at integer points, if $y\in[-n+1,n]\cap\mathbbm{Z}$ is chosen so that $z\in(y-1,y)$, then also $y-1/2$ is not disconnected from $\infty$ by any loop in $\Gamma_{Cn}$. ∎ The following lemma is the main step in the proof of Theorem 3.2. Figure 11: Illustration of the proof of Lemma 3.7. Let $x\in[2Cn+1,2Cn+3]$, so that $[-Cn,Cn]\cap[x-Cn,x+Cn]=\emptyset$. Note that we are considering two symmetric intervals of size $2Cn$ centered at $0$ and $x$ (recall Figure 10). If (i) $E_{n}^{c}$ occurs, (ii) the analogous event for $[x-Cn,x+Cn]$ occurs, and (iii) the “exposed” points $y$ and $y_{x}$ given by Lemma 3.6 have opposite parity, then Lemma 3.4 (and our assumption that there is no infinite path) tells us that there is a loop in $\Gamma$ which disconnects $y-1/2$ from $y_{x}-1/2$ (purple). By symmetry, we can take this loop to disconnect $y-1/2$ from $\infty$. By our choice of $y$ (from Lemma 3.6), this loop must hit a vertex outside of $[-Cn,Cn]\cap\mathbbm{Z}$. Since the loop disconnects $y$ from $y_{x}$, it must also hit a vertex in $[y,y_{x}]\cap\mathbbm{Z}\subset[y,(2C+1)n+3]\cap\mathbbm{Z}$. We note that the loop given by Lemma 3.7 can have many possible behaviors besides what is shown in the figure. For example, it could disconnect $[-n,n]$ from $\infty$ and/or it could hit $(-\infty,-Cn]$. ###### Lemma 3.7. Assume that there is no infinite path in $\Gamma$. With probability at least $\mathbbm{P}[E_{n}^{c}]^{2}/8$, there is a loop $\ell\in\Gamma$ with the following properties: 1. $(i)$ $\ell$ disconnects $y-1/2$ from $\infty$ for some $y\in[-n+1,n]\cap\mathbbm{Z}$. 2. $(ii)$ $\ell$ hits a point of $\mathbbm{Z}\setminus[-Cn,Cn]$. 3. $(iii)$ $\ell$ hits a point of $[y,(2C+1)n+3]\cap\mathbbm{Z}$. ###### Proof. See Figure 11 for an illustration. Write $p:=\mathbbm{P}[E_{n}^{c}]$. Recall from Lemma 3.5 that $E_{n}$ is determined by $(\mathcal{L},\mathcal{R})|_{[-Cn,Cn]\cap\mathbbm{Z}}$. On $E_{n}^{c}$, let $y\in[-n+1,n]\cap\mathbbm{Z}$ be a point as in Lemma 3.6, chosen in some manner which depends only on $(\mathcal{L},\mathcal{R})|_{[-Cn,Cn]\cap\mathbbm{Z}}$ (on $E_{n}$, we arbitrarily set $y=0$). Then one of the events $E_{n}^{c}\cap\mathopen{}\mathclose{{}\left\\{\text{$y$ is even}}\right\\}\quad\text{or}\quad E_{n}^{c}\cap\mathopen{}\mathclose{{}\left\\{\text{$y$ is odd}}\right\\}$ (3.1) has probability at least $p/2$. We will assume that $\mathbbm{P}\mathopen{}\mathclose{{}\left[E_{n}^{c}\cap\mathopen{}\mathclose{{}\left\\{\text{$y$ is even}}\right\\}}\right]\geq\frac{p}{2}.$ (3.2) The other case is treated in an identical manner. Let $x\in[2Cn+1,2Cn+3]\cap\mathbbm{Z}$ be an odd integer (so that $[-Cn,Cn]\cap[x-Cn,x+Cn]=\emptyset$). Define the event $E_{n}(x)$ in the same manner as the event $E_{n}$ from just above Lemma 3.5, but with the translated meandric system $\mathcal{M}-x$ in place of $\mathcal{M}$. Also let $y_{x}\in[x-n+1,x-n]\cap\mathbbm{Z}$ be defined in the same manner as the point $y$ above but with $\mathcal{M}-x$ in place of $\mathcal{M}$. By (3.2) and since $x$ is odd, $\mathbbm{P}\mathopen{}\mathclose{{}\left[E_{n}^{c}(x)\cap\mathopen{}\mathclose{{}\left\\{\text{$y_{x}$ is odd}}\right\\}}\right]\geq\frac{p}{2}.$ (3.3) The translated meandric system $\mathcal{M}-x$ is encoded by the translated pair of walks $j\mapsto(\mathcal{L}_{j+x}-\mathcal{L}_{x},\mathcal{R}_{j+x}-\mathcal{R}_{x})$ in the same manner that $\mathcal{M}$ is encoded by $(\mathcal{L},\mathcal{R})$. By Lemma 3.5 and the definition of $y_{x}$, the event $E_{n}(x)$ and the point $y_{x}$ are determined by the walk increment $(\mathcal{L}-\mathcal{L}_{x},\mathcal{R}-\mathcal{R}_{x})|_{[x-Cn,x+Cn]\cap\mathbbm{Z}}$. Since $[-Cn,Cn]\cap[x-Cn,x+Cn]=\emptyset$, the pairs $(E_{n},y)$ and $(E_{n}(x),y_{x})$ are independent. From this together with (3.2) and (3.3), we obtain $\mathbbm{P}\mathopen{}\mathclose{{}\left[E_{n}^{c}\cap\mathopen{}\mathclose{{}\left\\{\text{$y$ is even}}\right\\}\cap E_{n}^{c}(x)\cap\mathopen{}\mathclose{{}\left\\{\text{$y_{x}$ is odd}}\right\\}}\right]\geq\frac{p^{2}}{4}.$ (3.4) Recall that we are assuming that there is no infinite path in $\Gamma$. By Lemma 3.4, if the event in (3.4) occurs, then there is a loop $\ell\in\Gamma$ which disconnects $y-1/2$ from $y_{x}-1/2$. The loop $\ell$ has precisely two complementary connected components (by the Jordan curve theorem), so $\ell$ must disconnect exactly one of $y-1/2$ or $y_{x}-1/2$ from $\infty$. Since the law of $\mathcal{M}$ is invariant under integer translation and reflection about the origin, symmetry considerations show that the probability that $\ell$ disconnects $y-1/2$ from $\infty$ is at least $p^{2}/8$. By our choice of $y$ (recall Lemma 3.6), the loop $\ell$ must hit a point of $\mathbbm{Z}\setminus[-Cn,Cn]$. Since $\ell$ disconnects $y$ from $y_{x}$, $\ell$ must hit an integer point in the interval $[y,y_{x}]\subset[y,x+n]\subset[y,(2C+1)n+3].$ Therefore, $\ell$ satisfies the conditions in the lemma statement. ∎ ###### Proof of Theorem 3.2. First assume that there is an infinite path $\mathfrak{P}$ in $\Gamma$. It holds with probability tending to 1 as $n\rightarrow\infty$ that $\mathfrak{P}$ intersects $[-n,n]\cap\mathbbm{Z}$, so since $\mathfrak{P}$ is infinite we get that condition B is satisfied with probability444Here and throughout this paper, if $f,g:\mathbbm{N}\to(0,\infty)$ we write $g(n)=o_{n}(f(n))$ (resp. $g(n)=O_{n}(f(n))$) if $g(n)/f(n)\to 0$ (resp. $g(n)/f(n)$ remains bounded above) as $n\to\infty$. $1-o_{n}(1)$. We choose $n$ large enough so that this probability is at least $5-2\sqrt{6}$. Now assume that there is no infinite path in $\Gamma$. Since $\mathbbm{P}[E_{n}^{c}]=1-\mathbbm{P}[E_{n}]\in[0,1]$, $\max\mathopen{}\mathclose{{}\left\\{\mathbbm{P}[E_{n}],\mathbbm{P}[E_{n}^{c}]^{2}/8}\right\\}\geq 5-2\sqrt{6}.$ (3.5) By the definition ($\dagger$) of $E_{n}$, if $E_{n}$ occurs then condition A is satisfied, so condition A is satisfied with probability at least $\mathbbm{P}[E_{n}]$. In light of (3.5), it remains to show that the probability that either A or B is satisfied is at least $\mathbbm{P}[E_{n}^{c}]^{2}/8$. By Lemma 3.7, it holds with probability at least $\mathbbm{P}[E_{n}^{c}]^{2}/8$ that there is a loop $\ell\in\Gamma$ satisfying the three properties in that lemma. If $\ell$ disconnects $[-n,n]$ from $\infty$, then property (iii) of Lemma 3.7 shows that condition A in the theorem statement is satisfied. If $\ell$ does not disconnect $[-n,n]$ from $\infty$, then since $\ell$ disconnects some point of $[-n,n]\cap\mathbbm{Z}$ from $\infty$ by property (i), we get that $\ell$ must hit a point of $[-n,n]\cap\mathbbm{Z}$. By property (ii), this implies that condition B in the theorem statement is satisfied. ∎ ## 4 Bounding distances via the mated-CRT map Recall that $\mathcal{M}$ denotes the planar map associated with the UIMS. In this section, we will prove a lower bound for certain graph distances in $\mathcal{M}$ (Proposition 4.1 just below), conditional on an estimate (Proposition 4.3) which will be proven in Section 5 using Liouville quantum gravity techniques. Then, in Section 4.3, we will combine Proposition 4.1 and Theorem 3.2 to deduce our main results on the sizes of loops in meandric systems (Proposition 1.4, Theorem 1.5, and Theorem 1.9). To state our lower bound for distances in $\mathcal{M}$, we introduce some notation. We define the submaps $\mathcal{M}[a,b]:=\mathopen{}\mathclose{{}\left(\text{submap of $\mathcal{M}$ induced by $[a,b]\cap\mathbbm{Z}$}}\right),\quad\forall-\infty<a<b<\infty.$ (4.1) For a graph $G$, we also define $\operatorname{dist}_{G}(x,y):=\mathopen{}\mathclose{{}\left(\text{$G$-graph distance from $x$ to $y$}}\right),\quad\text{$\forall$ vertices $x,y\in G$}.$ (4.2) If $H$ is a subgraph of $G$ and $x,y\in H$, then $\operatorname{dist}_{G}(x,y)\leq\operatorname{dist}_{H}(x,y)$. The inequality can be strict since the minimal-length path in $G$ between $x$ and $y$ may not be contained in $H$. We will frequently use this fact without comment, often with $H=M[a,b]$ as in (4.1). ###### Proposition 4.1. For each $\zeta\in(0,1)$ and each $p\in(0,1)$, there exists $C=C(p,\zeta)>3$ such that for each large enough $n\in\mathbbm{N}$, it holds with probability at least $p$ that the following is true: 1. $(i)$ $\operatorname{dist}_{\mathcal{M}}\mathopen{}\mathclose{{}\left([-n,n]\cap\mathbbm{Z},\mathbbm{Z}\setminus[-Cn,Cn]}\right)\geq n^{1/d-\zeta}$. 2. $(ii)$ $\operatorname{dist}_{\mathcal{M}}\mathopen{}\mathclose{{}\left([-3Cn,3Cn]\cap\mathbbm{Z},\mathbbm{Z}\setminus[-C^{2}n,C^{2}n]}\right)\geq n^{1/d-\zeta}$. 3. $(iii)$ $\operatorname{dist}_{\mathcal{M}}\mathopen{}\mathclose{{}\left([-C^{2}n,C^{2}n]\cap\mathbbm{Z},\mathbbm{Z}\setminus[-C^{3}n,C^{3}n]}\right)\geq n^{1/d-\zeta}$. 4. $(iv)$ There are two paths $\Pi_{1},\Pi_{2}$ in $\mathcal{M}$ each going from a vertex of $[-n,n]\cap\mathbbm{Z}$ to a vertex of $\mathbbm{Z}\setminus[-C^{2}n,C^{2}n]$ which lie at $\mathcal{M}[-C^{3}n,C^{3}n]$-graph distance at least $n^{1/d-\zeta}$ from each other. For non-uniform random planar maps, it appears to be quite difficult to estimate graph distances directly. So, to prove Proposition 4.1, we will use an indirect approach which was introduced in [GHS20]. The idea is as follows. In Section 4.1, we will define the mated-CRT map $\mathcal{G}$, a random planar map constructed from a pair of independent two-sided Brownian motions via a semicontinuous analog of the construction of $\mathcal{M}$ from a pair of independent two-sided random walks. We will also state a comparison result for distances in $\mathcal{M}$ and distances in $\mathcal{G}$ (Theorem 4.2) which follows from a more general result in [GHS20]. This comparison result allows us to reduce Proposition 4.1 to a similar estimate for the mated-CRT map (Proposition 4.3). The proof of this latter estimate is given in Section 5. Due to the results of [DMS21], the mated-CRT map admits an alternative description in terms of $\sqrt{2}$-LQG decorated by SLE8 (see Section 5.1). Using this description, the needed estimate for the mated-CRT map turns out to be an easy consequence of known results for SLEs and LQG. ### 4.1 The mated-CRT map Let $(L,R)$ be a pair of standard linear two-sided Brownian motions, with $L_{0}=R_{0}=0$. The mated-CRT map associated with $(L,R)$ is the graph $\mathcal{G}$ with vertex set $\mathbbm{Z}$ and edge set defined as follows. Two integers $x_{1},x_{2}\in\mathbbm{Z}$ with $x_{1}<x_{2}$ are joined by an edge if and only if either $\displaystyle\max\mathopen{}\mathclose{{}\left\\{\inf_{t\in[x_{1}-1,x_{1}]}L_{t},\,\inf_{t\in[x_{2}-1,x_{2}]}L_{t}}\right\\}\leq\inf_{t\in[x_{1},x_{2}-1]}L_{t}\quad\operatorname{or}\quad$ $\displaystyle\max\mathopen{}\mathclose{{}\left\\{\inf_{t\in[x_{1}-1,x_{1}]}R_{t},\,\inf_{t\in[x_{2}-1,x_{2}]}R_{t}}\right\\}\leq\inf_{t\in[x_{1},x_{2}-1]}R_{t}.$ (4.3) We note that a.s. both conditions in (4.1) hold whenever $x_{2}=x_{1}+1$. If both conditions in (4.1) hold and $x_{2}\geq x_{1}+2$, we declare that $x_{1}$ and $x_{2}$ are joined by two edges. The edge set of $\mathcal{G}$ naturally splits into three subsets: * • Trivial edges, which join $x$ and $x+1$ for $x\in\mathbbm{Z}$. * • Upper edges, which join $x_{1},x_{2}\in\mathbbm{Z}$ with $x_{2}\geq x_{1}+2$ and arise from the first condition (the one involving $L$) in (4.1). * • Lower edges, which join $x_{1},x_{2}\in\mathbbm{Z}$ with $x_{2}\geq x_{1}+2$ and arise from the second condition (the one involving $R$) in (4.1). We can assign a planar map structure to $\mathcal{G}$ by associating each trivial edge with the line segment from $x$ to $x+1$ in $\mathbbm{R}^{2}$, each upper edge $\\{x_{1},x_{2}\\}$ with an arc from $x_{1}$ to $x_{2}$ in the upper half-plane, and each lower edge $\\{x_{1},x_{2}\\}$ with an arc from $x_{1}$ to $x_{2}$ in the lower half-plane. In fact, $\mathcal{G}$ is a triangulation when equipped with this planar map structure. See Figure 12. Figure 12: Left: To construct the mated-CRT map $\mathcal{G}$ geometrically, one can draw the graph of $C-L$ (red) and the graph of $R$ (blue) for some large constant $C>0$ chosen so that the parts of the graphs over some time interval of interest do not intersect. Here, this time interval is $[0,12]$. One then divides the region between the graphs into vertical strips (boundaries shown in orange). Each vertical strip corresponds to the vertex $x\in\mathbbm{Z}$ which is the horizontal coordinate of its rightmost points. Vertices $x_{1},x_{2}\in\mathbbm{Z}$ are connected by an edge if and only if the corresponding vertical strips are connected by a horizontal line segment which lies above the graph of $C-L$ or below the graph of $R$. For each pair of vertices for which the condition holds for $C-L$ (resp. $R$), we have drawn the highest (resp. lowest) segment which joins the corresponding vertical strips in red (resp. blue). Equivalently, for each $x\in\mathbbm{Z}$, we let $t_{x}$ be the time in $[x-1,x]$ at which $L$ attains its minimum value and we draw in red the longest horizontal segment above the graph of $C-L$ which contains $(t_{x},C-L_{t_{x}})$; and we perform a similar procedure for $R$. Note that consecutive vertices are always joined by an edge. Right: One can draw the graph $\mathcal{G}$ in the plane by connecting two non-consecutive vertices $x_{1},x_{2}\in\mathbbm{Z}$ by an arc above (resp. below) the real line if the corresponding vertical strips are connected by a horizontal segment above (resp. below) the graph of $C-L$ (resp. $R$); and connecting each pair of consecutive vertices of $\mathbbm{Z}$ by an edge. This gives $\mathcal{G}$ a planar map structure under which it is a triangulation. A similar figure and caption appeared in [GMS19]. Analogously to (4.1), we define $\mathcal{G}[a,b]:=\mathopen{}\mathclose{{}\left(\text{submap of $\mathcal{G}$ induced by $[a,b]\cap\mathbbm{Z}$}}\right),\quad\forall-\infty<a<b<\infty.$ (4.4) The definition of $\mathcal{G}$ is similar to the construction of the UIMS $\mathcal{M}$ from the pair of bi-infinite simple random walks $(\mathcal{L},\mathcal{R})$. In particular, by (1.5), two vertices $x_{1},x_{2}\in\mathbbm{Z}$ are joined by an edge of $\mathcal{M}$ if and only if $\displaystyle x_{2}=x_{1}+1\quad\text{or}\quad\mathcal{L}_{x_{1}-1}=\mathcal{L}_{x_{2}}<\min_{y\in[x_{1},x_{2}-1]}\mathcal{L}_{y}\quad\operatorname{or}\quad\mathcal{R}_{x_{1}-1}=\mathcal{R}_{x_{2}}<\min_{y\in[x_{1},x_{2}-1]}\mathcal{R}_{y}$ (4.5) which is a continuous time analog of (4.1). We note, however, that for $x\in\mathbbm{Z}$ the number of arcs of $\mathcal{G}$ in the upper (resp. lower) half-plane incident to $x$ can be any non-negative integer, whereas for $\mathcal{M}$ the number of such arcs is always one. Using the KMT strong coupling theorem for random walk and Brownian motion [KMT76, Zai98] and an elementary geometric argument, it was shown in [GHS20, Theorem 2.1] that one can couple $(L,R)$ and $(\mathcal{L},\mathcal{R})$ so that the graph distances in their corresponding planar maps $\mathcal{M}$ and $\mathcal{G}$ are close (actually, [GHS20] considers a more general class of pairs of random walks). ###### Theorem 4.2 ([GHS20]). For each $\alpha>0$, there exists $A=A(\alpha)>0$ and a coupling of $(L,R)$ with $(\mathcal{L},\mathcal{R})$ such that the following is true with probability $1-O_{n}(n^{-\alpha})$. Let $\mathcal{G}$ be the mated-CRT map constructed from $(L,R)$ as in (4.1) and let $\mathcal{M}$ be the infinite planar map associated with the UIMS, constructed from $(\mathcal{L},\mathcal{R})$ as in Section 1.4. For each $x,y\in[-n,n]\cap\mathbbm{Z}$, $A^{-1}(\log n)^{-3}\operatorname{dist}_{\mathcal{G}[-n,n]}(x,y)\leq\operatorname{dist}_{\mathcal{M}[-n,n]}(x,y)\leq A(\log n)^{3}\operatorname{dist}_{\mathcal{G}[-n,n]}(x,y),$ (4.6) where here we use the notations (4.1) and (4.4). ### 4.2 Proof of lower bounds for graph distances In this section we prove Proposition 4.1. We will deduce it from the combination of Theorem 4.2 and the following analogous estimate for the mated- CRT map. ###### Proposition 4.3. For each $\zeta\in(0,1)$ and each $p\in(0,1)$, there exists $C=C(p,\zeta)>3$ such that for each large enough $n\in\mathbbm{N}$, it holds with probability at least $p$ that the following is true: 1. $(a)$ $\operatorname{dist}_{\mathcal{G}}\mathopen{}\mathclose{{}\left([-n,n]\cap\mathbbm{Z},\mathbbm{Z}\setminus[-Cn,Cn]}\right)\geq n^{1/d-\zeta}$. 2. $(b)$ $\operatorname{dist}_{\mathcal{G}}\mathopen{}\mathclose{{}\left([-3Cn,3Cn]\cap\mathbbm{Z},\mathbbm{Z}\setminus[-C^{2}n,C^{2}n]}\right)\geq n^{1/d-\zeta}$. 3. $(c)$ $\operatorname{dist}_{\mathcal{G}}\mathopen{}\mathclose{{}\left([-C^{2}n,C^{2}n]\cap\mathbbm{Z},\mathbbm{Z}\setminus[-C^{3}n,C^{3}n]}\right)\geq n^{1/d-\zeta}$. 4. $(d)$ There are two paths $\Pi_{1},\Pi_{2}$ in $\mathcal{G}$ each going from a vertex of $[-n,n]\cap\mathbbm{Z}$ to a vertex of $\mathbbm{Z}\setminus[-C^{2}n,C^{2}n]$ which lie at $\mathcal{G}$-graph distance at least $n^{1/d-\zeta}$ from each other. The proof of Proposition 4.3 is given in Section 5, using the relationship between the mated-CRT map and SLE-decorated LQG. We now deduce Proposition 4.1 from Proposition 4.3. ###### Proof of Proposition 4.1. We divide the proof in three main steps. Step 1: Regularity event. We first define a high-probability event which we will work on throughout the rest of the proof. Let $A_{0}>0$ be as in Theorem 4.2 with $\alpha=1$, say, and let $A:=8A_{0}$. By Theorem 4.2 with $n^{2}$ in place of $n$, we can couple $\mathcal{M}$ and $\mathcal{G}$ so that with probability tending to 1 as $n\rightarrow\infty$, it holds for each $x,y\in[-n^{2},n^{2}]\cap\mathbbm{Z}$ that $A^{-1}(\log n)^{-3}\operatorname{dist}_{\mathcal{G}[-n^{2},n^{2}]}(x,y)\leq\operatorname{dist}_{M[-n^{2},n^{2}]}(x,y)\leq A(\log n)^{3}\operatorname{dist}_{\mathcal{G}[-n^{2},n^{2}]}(x,y).$ (4.7) In order to make sure that we can compare distances in $\mathcal{M}$ and $\mathcal{G}$ from points in $[-C^{2}n,C^{2}n]\cap\mathbbm{Z}$ to the set $\mathbbm{Z}\setminus[-C^{2}n,C^{2}n]$, we also need to impose a further regularity condition. By the adjacency conditions (4.5) and (4.1) for $\mathcal{M}$ and $\mathcal{G}$ in terms of $(\mathcal{L},\mathcal{R})$ and $(L,R)$, respectively, together with basic estimates for random walk and Brownian motion, for any fixed $C>1$ it holds with probability tending to 1 as $n\rightarrow\infty$ that $\text{$\forall$ edges $\\{x,y\\}\in\mathcal{M}$ such that $x\in[-C^{3}n,C^{3}n]\cap\mathbbm{Z}$, we have $y\in[-n^{2},n^{2}]$};$ (4.8) and the same is true with $\mathcal{G}$ in place of $\mathcal{M}$. Note that the quantity $n^{2}$ is unimportant in (4.8): the same would be true with $n^{2}$ replaced by any function of $n$ which goes to $\infty$ as $n\rightarrow\infty$. We now take $C=C(1-(1-p)/2,\zeta/2)>0$ as in Proposition 4.3 with $1-(1-p)/2$ instead of $p$ and $\zeta/2$ instead of $\zeta$. By Proposition 4.3 and our above estimates, for each large $n\in\mathbbm{N}$ it holds with probability at least $p$ that (4.7) and (4.8) both hold and the numbered conditions in Proposition 4.3 hold with $\zeta/2$ in place of $\zeta$. Henceforth assume that this is the case. The rest of the argument is deterministic. Step 2: Proofs of (i), (ii), and (iii). Consider a path $P$ in $\mathcal{M}$ from a point $x\in[-n,n]\cap\mathbbm{Z}$ to a point of $\mathbbm{Z}\setminus[-Cn,Cn]$. By (4.8), $P$ hits a vertex in $\mathbbm{Z}\cap([-n^{2},n^{2}]\setminus[-Cn,Cn])$. Let $y$ be the first such vertex. Then the segment $P^{\prime}$ of $P$ between its starting point and the first time it hits $y$ is a path in $\mathcal{M}[-n^{2},n^{2}]$. Using our above estimates, we now get that the length of $P^{\prime}$ satisfies $\displaystyle|P^{\prime}|$ $\displaystyle\geq A^{-1}(\log n)^{-3}\operatorname{dist}_{\mathcal{G}[-n^{2},n^{2}]}(x,y)\quad\text{(by~{}\eqref{eqn- use-ghs-coupling})}$ $\displaystyle\geq A^{-1}(\log n)^{-3}\operatorname{dist}_{\mathcal{G}}\mathopen{}\mathclose{{}\left([-n,n]\cap\mathbbm{Z},\mathbbm{Z}\setminus[-Cn,Cn]}\right)\quad\text{(choice of $x$ and $y$)}$ $\displaystyle\geq A^{-1}(\log n)^{-3}n^{1/d-\zeta/2}\quad\text{(\eqref{item-mcrt-in} of Proposition~{}\ref{prop-mcrt-estimate} with $\zeta/2$ instead of $\zeta$)}.$ (4.9) Since $P^{\prime}$ is a sub-path of $P$, we have $|P|\geq|P^{\prime}|$. Taking the infimum over all $P$ now shows that $\operatorname{dist}_{\mathcal{M}}\mathopen{}\mathclose{{}\left([-n,n]\cap\mathbbm{Z},\mathbbm{Z}\setminus[-Cn,Cn]}\right)$ is at least the right side of (4.2), which is at least $n^{1/d-\zeta}$ if $n$ is large enough. Thus (i) in the proposition statement holds. The proofs of (ii) and (iii) are identical to the proof of (i), except that we use (b) and (c) from Proposition 4.3 instead of (a). Step 3: Proof of (iv). Let $\widetilde{\Pi}_{1}$ and $\widetilde{\Pi}_{2}$ be the paths in $\mathcal{G}$ as in (d) of Proposition 4.3 with $\zeta/2$ instead of $\zeta$. By possibly replacing $\widetilde{\Pi}_{1}$ and $\widetilde{\Pi}_{2}$ by sub-paths (which can only increase the graph distance between them), we can assume that each of these paths intersects $[-n,n]\cap\mathbbm{Z}$ and $\mathbbm{Z}\setminus[-C^{2}n,C^{2}n]$ only at its endpoints. We view $\widetilde{\Pi}_{1}$ as a function $\widetilde{\Pi}_{1}:[0,N]\cap\mathbbm{Z}\rightarrow\mathbbm{Z}$. For $i=1,\dots,N$, we apply (4.7) with $(\widetilde{\Pi}_{1}(i-1),\widetilde{\Pi}_{1}(i))$ in place of $(x,y)$ to get $N$ new paths in $\mathcal{M}[-n^{2},n^{2}]$ (the ones realizing $\operatorname{dist}_{\mathcal{M}[-n^{2},n^{2}]}(\widetilde{\Pi}_{1}(i-1),\widetilde{\Pi}_{1}(i))$), each of length at most $A(\log n)^{3}$. Then we concatenate the $N$ new paths in $\mathcal{M}[-n^{2},n^{2}]$. This results in a path $\Pi_{1}$ in $\mathcal{M}[-n^{2},n^{2}]$ with the same endpoints as $\widetilde{\Pi}_{1}$ with the property that each point of $\Pi_{1}$ lies at $\mathcal{M}[-n^{2},n^{2}]$-graph distance at most $A(\log n)^{3}$ from a point of $\widetilde{\Pi}_{1}$. We similarly construct a path $\Pi_{2}$ in $\mathcal{M}[-n^{2},n^{2}]$ but started from $\widetilde{\Pi}_{2}$ instead of $\widetilde{\Pi}_{1}$. Since $\Pi_{1},\Pi_{2}$ have the same endpoints as $\widetilde{\Pi}_{1},\widetilde{\Pi}_{2}$, these paths each go from a vertex of $[-n,n]\cap\mathbbm{Z}$ to a vertex of $\mathbbm{Z}\setminus[-C^{2}n,C^{2}n]$. Furthermore, by the distance estimate in the preceding paragraph and the triangle inequality, the $\mathcal{M}[-n^{2},n^{2}]$-graph distance from any point of $\Pi_{1}$ to any point of $\Pi_{2}$ is at least $\operatorname{dist}_{\mathcal{M}[-n^{2},n^{2}]}(\widetilde{\Pi}_{1},\widetilde{\Pi}_{2})-2A(\log n)^{3}$. By this and (4.7), $\displaystyle\operatorname{dist}_{\mathcal{M}[-C^{3}n,C^{3}n]}(\Pi_{1},\Pi_{2})$ $\displaystyle\geq\operatorname{dist}_{\mathcal{M}[-n^{2},n^{2}]}(\Pi_{1},\Pi_{2})$ $\displaystyle\geq\operatorname{dist}_{\mathcal{M}[-n^{2},n^{2}]}(\widetilde{\Pi}_{1},\widetilde{\Pi}_{2})-2A(\log n)^{3}$ $\displaystyle\geq A^{-1}(\log n)^{-3}\operatorname{dist}_{\mathcal{G}}(\widetilde{\Pi}_{1},\widetilde{\Pi}_{2})-2A(\log n)^{3}.$ (4.10) By our initial choice of $\widetilde{\Pi}_{1}$ and $\widetilde{\Pi}_{2}$ the right side of (4.2) is at least $A^{-1}(\log n)^{-3}n^{1/d-\zeta/2}-2A(\log n)^{3}$, which is at least $n^{1/d-\zeta}$ for each large enough $n\in\mathbbm{N}$. ∎ ### 4.3 Proof of Proposition 1.4, Theorem 1.5, and Theorem 1.9 Combining Theorem 3.2 and Proposition 4.1 leads to the following result. ###### Proposition 4.4. For each $\zeta\in(0,1)$ and each sufficiently large $n\in\mathbbm{N}$ (depending on $\zeta$), it holds with probability at least $1/10$ that the following is true. There is a segment $\overline{\ell}$ of a loop or infinite path in $\Gamma$ which is contained in $[-n,n]\cap\mathbbm{Z}$ such that the $\mathcal{M}[-n,n]$-graph-distance diameter of $\overline{\ell}$ and the $\mathcal{M}$-graph distance from $\overline{\ell}$ to $\mathbbm{Z}\setminus[-n,n]$ are each at least $n^{1/d-\zeta}$. ###### Proof. Fix $\zeta\in(0,1)$. By Theorem 3.2, for any choice of $C>1$, for each large enough $n\in\mathbbm{N}$ it holds with probability at least $5-2\sqrt{6}>1/10$ that at least one of A or B in the statement of Theorem 3.2 is satisfied. By Proposition 4.1, there is a universal constant $C>1$ so that for each large enough $n$, it holds with probability at least $1-(5-2\sqrt{6}-1/10)$ that all four of the numbered conditions in the statement of Proposition 4.1 are satisfied. Henceforth assume that the events of Theorem 3.2 and Proposition 4.1 occur (both with this same choice of $C$), which happens with probability at least $1/10$ if $n$ is large enough. We will show that 1. $(*)$ there is a segment $\overline{\ell}$ of a loop or infinite path in $\Gamma$ which is contained in $[-C^{3}n,C^{3}n]\cap\mathbbm{Z}$ such that the $\mathcal{M}[-C^{3}n,C^{3}n]$-graph-distance diameter of $\overline{\ell}$ and the $\mathcal{M}$-graph distance from $\overline{\ell}$ to $\mathbbm{Z}\setminus[-C^{3}n,C^{3}n]\cap\mathbbm{Z}$ are each at least $n^{1/d-\zeta}$. To extract the proposition statement from ($*$), we can apply ($*$) with $\lfloor n/C^{3}\rfloor$ in place of $n$, then slightly shrink $\zeta$ in order to absorb a factor of $1/C^{3\zeta}$ into a power of $n$. To prove ($*$), we will treat the two possible scenarios in Theorem 3.2 separately. First suppose that 1. A. there is a loop $\ell$ in $\Gamma$ which disconnects $[-n,n]$ from $\infty$ and which hits a point of $[n,(2C+1)n+3]\cap\mathbbm{Z}$. If $\ell$ is contained in $[-C^{2}n,C^{2}n]\cap\mathbbm{Z}$, then by planarity and since $\ell$ disconnects $[-n,n]$ from $\infty$, the loop $\ell$ must intersect every path in $\mathcal{M}$ from $[-n,n]\cap\mathbbm{Z}$ to $\mathbbm{Z}\setminus[-C^{2}n,C^{2}n]$. In particular, $\ell$ must intersect the paths $\Pi_{1}$ and $\Pi_{2}$ from (iv) of Proposition 4.1. Hence, the $\mathcal{M}[-C^{3}n,C^{3}n]$-graph-distance diameter of $\ell$ is at least $n^{1/d-\zeta}$. Furthermore, by (iii) of Proposition 4.1, the $\mathcal{M}$-graph distance from $\ell$ to $\mathbbm{Z}\setminus[-C^{3}n,C^{3}n]\cap\mathbbm{Z}$ is at least $n^{1/d-\zeta}$. So, we can take $\overline{\ell}=\ell$. If $\ell$ is not contained in $[-C^{2}n,C^{2}n]\cap\mathbbm{Z}$, then since $\ell$ hits $[n,(2C+1)n+3]\cap\mathbbm{Z}$, there is a segment $\overline{\ell}$ of $\ell$ which is a path in $\mathcal{M}$ from a point of $[n,(2C+1)n+3]\cap\mathbbm{Z}\subset[-3Cn,3Cn]\cap\mathbbm{Z}$ to a point of $\mathbbm{Z}\setminus[-C^{2}n,C^{2}n]$. By possibly replacing $\overline{\ell}$ with a sub-path, we can assume that $\overline{\ell}$ is contained in $[-C^{2}n,C^{2}n]\cap\mathbbm{Z}$ except for its terminal endpoint. By (ii) of Proposition 4.1, the $\mathcal{M}[-C^{3}n,C^{3}n]$-graph-distance diameter of $\overline{\ell}$ is at least $n^{1/d-\zeta}$. By (iii) of Proposition 4.1, the $\mathcal{M}$-graph distance from $\overline{\ell}$ to $[-C^{3}n,C^{3}n]\cap\mathbbm{Z}$ is at least $n^{1/d-\zeta}$. Next suppose that 1. B. there is a loop or an infinite path in $\Gamma$ which hits a point of each of $[-n,n]\cap\mathbbm{Z}$ and $\mathbbm{Z}\setminus[-Cn,Cn]$. Let $\overline{\ell}$ be a segment of this loop or infinite path which is a path from a point of $[-n,n]\cap\mathbbm{Z}$ to a point of $\mathbbm{Z}\setminus[-Cn,Cn]$. By possibly replacing $\overline{\ell}$ with a sub-path, we can assume that $\overline{\ell}$ is contained in $[-Cn,Cn]\cap\mathbbm{Z}$ except for its terminal endpoint. Then (i) of Proposition 4.1 shows that the $\mathcal{M}[-C^{3}n,C^{3}n]$-graph-distance diameter of $\overline{\ell}$ is at least $n^{1/d-\zeta}$ and (iii) of Proposition 4.1 shows that the $\mathcal{M}$-graph distance from $\overline{\ell}$ to $[-C^{3}n,C^{3}n]\cap\mathbbm{Z}$ is at least $n^{1/d-\zeta}$. ∎ Using Proposition 4.4, we obtain our lower tail bound for the diameter of the origin-containing loop in the UIMS. ###### Proof of Theorem 1.9. Let $\zeta\in(0,1)$, which we will eventually send to zero. For $k\in\mathbbm{N}$, let $X=X_{k}$ be sampled uniformly from $[-k^{d+\zeta},k^{d+\zeta}]\cap\mathbbm{Z}$, independently from everything else, and let $\ell_{X}$ be the loop in $\Gamma$ which hits $X$. By the translation invariance of the law of $(\mathcal{M},\Gamma)$, the translated planar map / loop pair $(\mathcal{M}-X,\Gamma-X)$ has the same law as $(\mathcal{M},\Gamma)$. Hence, $\mathbbm{P}\mathopen{}\mathclose{{}\left[\mathopen{}\mathclose{{}\left(\text{$\mathcal{M}$-graph- distance diameter of $\ell_{0}$}}\right)\geq k}\right]=\mathbbm{P}\mathopen{}\mathclose{{}\left[\mathopen{}\mathclose{{}\left(\text{$\mathcal{M}$-graph- distance diameter of $\ell_{X}$}}\right)\geq k}\right].$ (4.11) We will now lower-bound the second quantity in (4.11). By Proposition 4.4 (with $\lfloor k^{d+\zeta}\rfloor$ in place of $n$ and with $\zeta$ possibly replaced by a smaller positive number), if $k$ is large enough then it holds with probability at least $1/10$ that there is a segment $\overline{\ell}$ of a loop or infinite path in $\Gamma$ which is contained in $[-k^{d+\zeta},k^{d+\zeta}]\cap\mathbbm{Z}$ such that the $\mathcal{M}[-k^{d+\zeta},k^{d+\zeta}]$-graph-distance diameter of $\overline{\ell}$ and the $\mathcal{M}$-graph distance from $\overline{\ell}$ to $\mathbbm{Z}\setminus[-k^{d+\zeta},k^{d+\zeta}]$ are each at least $k$. Let $E_{k}$ be the event that such an $\overline{\ell}$ exists. On $E_{k}$ let $\overline{\ell}$ be a path as in the definition of $E_{k}$, chosen in some manner which is measurable with respect to $\sigma(\mathcal{M},\Gamma)$. We claim that if $E_{k}$ occurs, then the $\mathcal{M}$-graph-distance diameter of $\overline{\ell}$, and hence also the number of vertices of $\mathcal{M}$ hit by $\overline{\ell}$, are each at least $k$. Indeed, by definition, there are two vertices $x,y\in[-k^{d+\zeta},k^{d+\zeta}]\cap\mathbbm{Z}$ hit by $\overline{\ell}$ which lie at $\mathcal{M}[-k^{d+\zeta},k^{d+\zeta}]$-graph distance at least $k$ from each other. Any path in $\mathcal{M}$ from $x$ to $y$ must either stay in $[-k^{d+\zeta},k^{d+\zeta}]\cap\mathbbm{Z}$, in which case its length is at least $k$ by our choice of $x,y$; or must cross from $x$ to $\mathbbm{Z}\setminus[-k^{d+\zeta},k^{d+\zeta}]$, in which case its length is also at least $k$ since the $\mathcal{M}$-graph distance from $\overline{\ell}$ to $\mathbbm{Z}\setminus[-k^{d+\zeta},k^{d+\zeta}]$ is at least $k$. Taking the infimum over all such paths gives our claim. Since $X$ is sampled uniformly from $[-k^{d+\zeta},k^{d+\zeta}]\cap\mathbbm{Z}$, independently from $(\mathcal{M},\Gamma)$, the preceding paragraph implies that $\mathbbm{P}\mathopen{}\mathclose{{}\left[X\in\overline{\ell}\,\middle|\,(\mathcal{M},\Gamma)}\right]\mathbbm{1}_{E_{k}}\geq\frac{1}{2}k^{-(d-1)-\zeta}\mathbbm{1}_{E_{k}}.$ (4.12) Since $\mathbbm{P}[E_{k}]\geq 1/10$ and the $\mathcal{M}$-graph-distance diameter of $\overline{\ell}$ is at least $k$ on $E_{k}$, taking the expectations on both sides of (4.12) gives $\mathbbm{P}\mathopen{}\mathclose{{}\left[\mathopen{}\mathclose{{}\left(\text{$\mathcal{M}$-graph- distance diameter of $\ell_{X}$}}\right)\geq k}\right]\geq\frac{1}{20}k^{-(d-1)-\zeta}.$ (4.13) By (4.11), (4.13) implies (1.6) upon sending $\zeta\rightarrow 0$. ∎ The following proposition is the analog of Theorem 1.5 for the UIMS. It is the main input in the proof of Theorem 1.5. ###### Proposition 4.5. For each $\zeta\in(0,1)$, there exists $\beta>0$ and $a_{0},a_{1}>0$, depending on $\zeta$, such that for each $n\in\mathbbm{N}$, it holds with probability at least $1-a_{0}e^{-a_{1}n^{\beta}}$ that there is a segment of a loop or an infinite path in $\Gamma$ which hits only vertices of $[1,2n]\cap\mathbbm{Z}$ and has $\mathcal{M}$-graph-distance diameter at least $n^{1/d-\zeta}$. ###### Proof. Let $\beta\in(0,1)$ to be chosen later, depending on $\zeta$. The idea of the proof is as follows. We will consider a collection of $\operatorname{const}\times n^{\beta}$ disjoint sub-intervals of $[1,2n]$ of length $2\lfloor n^{1-\beta}\rfloor$. We will then use independence to show that with high probability, the event of Proposition 4.4 (with $\lfloor n^{1-\beta}\rfloor$ instead of $n$) occurs for at least one of these intervals. For $n\in\mathbbm{N}$ and $x\in\mathbbm{Z}$, let $G_{n}(x)$ be the event that the following is true: 1. $(\ddagger)$ There is a segment $\overline{\ell}$ of a loop or infinite path in $\Gamma$ which is contained in $[x-n,x+n]\cap\mathbbm{Z}$ such that the $\mathcal{M}[x-n,x+n]$-graph-distance diameter of $\overline{\ell}$ and the $\mathcal{M}$-graph distance from $\overline{\ell}$ to $\mathbbm{Z}\setminus[x-n,x+n]$ are each at least $n^{1/d-\beta}$. By the translation invariance of the law of $\mathcal{M}$, Proposition 4.4 implies that for each large enough $n\in\mathbbm{N}$, $\mathbbm{P}\mathopen{}\mathclose{{}\left[G_{n}(x)}\right]\geq\frac{1}{10},\quad\forall x\in\mathbbm{Z}.$ (4.14) Furthermore, the event $G_{n}(x)$ depends only on the set of edges of $\mathcal{M}$ between vertices in $[x-n,x+n]\cap\mathbbm{Z}$ and the set of vertices in $[x-n,x+n]\cap\mathbbm{Z}$ which are joined by edges of $\mathcal{M}$ to vertices not in $[x-n,x+n]\cap\mathbbm{Z}$. This information is determined by the restricted, shifted walk $(\mathcal{L}-\mathcal{L}_{x},\mathcal{R}-\mathcal{R}_{x})|_{[x-n,x+n]\cap\mathbbm{Z}}$. Consequently, $G_{n}(x)$ and $G_{n}(y)$ are independent if $|x-y|\geq 2n$. (4.15) There is a constant $c=c(\beta)>0$ such that for each $n\in\mathbbm{N}$, there is a deterministic set $X_{n}\subset[1,2n]\cap\mathbbm{Z}$ of cardinality at least $cn^{\beta}$ such that $\text{the intervals $\mathopen{}\mathclose{{}\left[x-\lfloor n^{1-\beta}\rfloor,x+\lfloor n^{1-\beta}\rfloor}\right]$ for $x\in X_{n}$ are disjoint and contained in $[1,2n]$}.$ (4.16) By (4.14) and (4.15), $\mathbbm{P}\mathopen{}\mathclose{{}\left[\text{$G_{\lfloor n^{1-\beta}\rfloor}(x)$ occurs for at least one $x\in X_{n}$}}\right]\geq 1-\mathopen{}\mathclose{{}\left(\frac{9}{10}}\right)^{cn^{\beta}}.$ (4.17) On the other hand, if $G_{\lfloor n^{1-\beta}\rfloor}(x)$ occurs, then the segment $\overline{\ell}$ as in the definition of $G_{\lfloor n^{1-\beta}\rfloor}$ has $\mathcal{M}[x-\lfloor n^{1-\beta}\rfloor,x+\lfloor n^{1-\beta}\rfloor]$-graph-distance diameter at least $\lfloor n^{1-\beta}\rfloor^{1/d-\beta}$ and the $\mathcal{M}$-graph distance from $\overline{\ell}$ to $\mathbbm{Z}\setminus[x-\lfloor n^{1-\beta}\rfloor,x+\lfloor n^{1-\beta}\rfloor]$ is at least $\lfloor n^{1-\beta}\rfloor^{1/d-\beta}$. Hence, the $\mathcal{M}$-graph-distance diameter of $\overline{\ell}$ is at least $\lfloor n^{1-\beta}\rfloor^{1/d-\beta}$. We now choose $\beta$ to be small enough, depending on $\zeta$, so that $(1-\beta)(1/d-\beta)>1/d-\zeta.$ Then (4.17) and the preceding paragraph give that if $n$ is large enough, then with probability at least $1-(9/10)^{cn^{\beta}}$, there is a segment of a loop in $\Gamma$ which intersects $[1,2n]\cap\mathbbm{Z}$ and has $\mathcal{M}$-graph-distance diameter at least $n^{1/d-\zeta}$. This gives the proposition statement for an appropriate choice of $a_{0},a_{1}>0$. ∎ ###### Proof of Theorem 1.5. Let $(\mathcal{M},\Gamma)$ be the loop-decorated planar map associated with an infinite meandric system. For $n\in\mathbbm{N}$, let $F_{n}$ be the event that there is no arc of the upper or lower arc diagram associated with $\mathcal{M}$ which has one endpoint in $[1,2n]\cap\mathbbm{Z}$ and one endpoint not in $[1,2n]\cap\mathbbm{Z}$. Equivalently, by (1.5), $F_{n}$ is the event that the encoding walks satisfy $\mathcal{L}_{2n}=\mathcal{R}_{2n}=0$ and $\mathcal{L}_{x},\mathcal{R}_{x}\geq 0$ for each $x\in[0,2n]\cap\mathbbm{Z}$. By a standard random walk estimate, there is a universal constant $c>0$ such that $\mathbbm{P}[F_{n}]\sim cn^{-3}\quad\text{as $n\rightarrow\infty$}.$ (4.18) By the definition of $F_{n}$, if $F_{n}$ occurs, then no infinite path in $\Gamma$ can hit $[1,2n]\cap\mathbbm{Z}$ and the set $\Gamma_{n}$ of loops of $\Gamma$ which hit $[1,2n]\cap\mathbbm{Z}$ is the same as the set of loops in $\Gamma$ which do not hit any vertices in $\mathbbm{Z}\setminus[0,2n]$. Moreover, the conditional law of $(\mathcal{L},\mathcal{R})|_{[0,2n]\cap\mathbbm{Z}}$ given $F_{n}$ is that of a pair of independent uniform $2n$-step simple random walk excursions. By the discussion surrounding (1.1), this implies that the conditional law of the loop-decorated planar map $(M[1,2n],\Gamma_{n})$ given $F_{n}$ is that of the planar map associated with a uniform meandric system of size $n$. Hence, it suffices to show that if we condition on $F_{n}$, then except on an event of conditional probability decaying faster than any negative power of $n$, there is a loop in $\Gamma_{n}$ which has $\mathcal{M}[1,2n]$-graph-distance diameter at least $n^{1/d-\zeta}$. By Proposition 4.5 and (4.18), if $\beta,a_{0},a_{1}$ are as in Proposition 4.5, then it holds with conditional probability at least $1-a_{0}c^{-1}n^{3}e^{-a_{1}n^{\beta}}$ given $F_{n}$ that there is a segment of a loop or an infinite path in $\Gamma$ which hits only vertices of $[1,2n]\cap\mathbbm{Z}$ and has $\mathcal{M}$-graph-distance diameter (and hence also $\mathcal{M}[1,2n]$-graph-distance diameter) at least $n^{1/d-\zeta}$. By the first sentence of the preceding paragraph, on $F_{n}$ this segment is in fact a segment of a loop $\ell\in\Gamma_{n}$. This loop $\ell$ has $\mathcal{M}[1,2n]$-graph-distance diameter at least $n^{1/d-\zeta}$. Since $a_{0}c^{-1}n^{3}e^{-a_{1}n^{\beta}}=O(n^{-p})$ for every $p>0$, this concludes the proof. ∎ ###### Proof of Proposition 1.4. Theorem 1.5 immediately implies that except on an event of probability decaying faster than any negative power of $n$, the graph-distance diameter of $\mathcal{M}_{n}$ is at least $n^{1/d-\zeta}$. To prove an upper bound for the graph-distance diameter of $\mathcal{M}_{n}$, we first apply [GHS19, Theorem 1.15], which tells us that there exists an exponent $\chi>0$ such that for each $\zeta\in(0,1)$ and each $n\in\mathbbm{N}$, it holds except on an event of probability decaying faster than any negative power of $n$ that the graph-distance diameter of $\mathcal{G}[1,2n]$ is at most $n^{\chi+\zeta/2}$. It was shown in [GP21, Theorem 3.1] that $\chi=1/d$. By combining the preceding paragraph with Theorem 4.2, we get that for each $\alpha>0$, there exists $A=A(\alpha)>0$ such that with probability at least $1-O_{n}(n^{-\alpha})$, the graph-distance diameter of $\mathcal{M}[1,2n]$ is at most $A(\log n)^{3}n^{1/d+\zeta/2}$, which is at most $n^{1/d+\zeta}$ if $n$ is large enough. Since $\alpha$ can be made arbitrarily large, we get that except on an event of probability decaying faster than any negative power of $n$, the graph-distance diameter of $\mathcal{M}[1,2n]$ is at most $n^{1/d+\zeta}$. To transfer this from $\mathcal{M}[1,2n]$ to $\mathcal{M}_{n}$, we define the event $F_{n}$ as in the proof of Theorem 1.5, condition on $F_{n}$, and apply (4.18), exactly as in the proof of Theorem 1.5. ∎ ## 5 Estimate for the mated-CRT map via SLE and LQG To complete the proofs of our main results, it remains to prove Proposition 4.3. We will do this using SLE and LQG. ### 5.1 SLE/LQG description of the mated-CRT map Recall that we previously defined the mated-CRT map using Brownian motion in Section 4.1. In this subsection we will give the SLE/LQG description of the mated-CRT map, which comes from the results of [DMS21]. We will not need many properties of the SLE/LQG objects involved, so we will not give detailed definitions. Instead, we give precise references. Let $h$ be the random generalized function on $\mathbbm{C}$ associated with the $\sqrt{2}$-quantum cone. The generalized function $h$ is a minor variant of the whole-plane Gaussian free field; see [DMS21, Definition 4.10] for a precise definition. One can associate with $h$ a random locally finite measure $\mu_{h}$ on $\mathbbm{C}$, the $\sqrt{2}$-LQG measure, which is a limit of regularized versions of $e^{\sqrt{2}h}\,dx\,dy$, where $dx\,dy$ denotes Lebesgue measure on $\mathbbm{C}$ [Kah85, DS11]. The measure $\mu_{h}$ assigns positive mass to every open subset of $\mathbbm{C}$ and zero mass to every point but is mutually singular with respect to Lebesgue measure. See [BP, Chapter 2] for a detailed account of the construction and properties of $\mu_{h}$. One can similarly associate with $h$ a random metric (distance function) $D_{h}$ on $\mathbbm{C}$, the $\sqrt{2}$-LQG metric [DDDF20, GM21b]. The metric $D_{h}$ induces the same topology on $\mathbbm{C}$ as the Euclidean metric, but the Hausdorff dimension $d$ of the metric space $(\mathbbm{C},D_{h})$ is strictly larger than 2 (this is the same $d$ appearing in Proposition 1.4). See [DDG21] for a survey of results about $D_{h}$. The metric measure space $(\mathbbm{C},D_{h},\mu_{h})$ possesses a scale invariance property which will be important for our purposes: $\text{$(\mathbbm{C},D_{h},\mu_{h})\overset{d}{=}(\mathbbm{C},r^{1/d}D_{h},r\mu_{h})$ as metric measure spaces $\forall r>0$}.$ (5.1) In fact, one has the following slightly stronger property: for each $r>0$, there is a random $\rho_{r}>0$ such that $\mathopen{}\mathclose{{}\left(D_{h},\mu_{h}}\right)\overset{d}{=}\mathopen{}\mathclose{{}\left(r^{1/d}D_{h}(\rho_{r}\cdot,\rho_{r}\cdot),r\mu_{h}(\rho_{r}\cdot)}\right),\quad\forall r>0.$ (5.2) The scaling property (5.2) follows from the scaling property of $h$ [DMS21, Proposition 4.13(i)] together with the fact that adding the constant $\frac{1}{\sqrt{2}}\log r$ to $h$ results in scaling $\mu_{h}$ by $r$ and $D_{h}$ by $r^{-1/d}$, both of which are immediate from the constructions of $\mu_{h}$ and $D_{h}$ (see the proof of [GS22, Proposition 2.17] for a more detailed explanation). Whole-plane SLE8 from $\infty$ to $\infty$ is a random space-filling curve $\eta$ which travels from $\infty$ to $\infty$ in $\mathbbm{C}$. It can be thought of as a two-sided version of chordal SLE8 (see [DMS21, Footnote 4] for a precise version of this statement). For each $z\in\mathbbm{C}$, a.s. $z$ is hit exactly once by $\eta$, but there exist zero-Lebesgue measure sets of points which are hit twice or three times. Now suppose that we sample $\eta$ independently from the random generalized function $h$ above, then re-parametrize $\eta$ so that $\eta(0)=0\quad\text{and}\quad\mu_{h}(\eta([a,b]))=b-a,\quad\forall a,b\in\mathbbm{R}\>\text{with}\>a<b.$ (5.3) The law of $\eta$ is invariant under spatial scaling (this is immediate from the definition in [DMS21, Footnote 4]), so it follows from (5.2) that $\mathopen{}\mathclose{{}\left(D_{h},\mu_{h},\eta}\right)\overset{d}{=}\mathopen{}\mathclose{{}\left(r^{1/d}D_{h}(\rho_{r}\cdot,\rho_{r}\cdot),r\mu_{h}(\rho_{r}\cdot),\rho_{r}^{-1}\eta(\cdot/r)}\right),\quad\forall r>0.$ (5.4) The connection between the pair $(h,\eta)$ and the mated-CRT map $\mathcal{G}$ comes by way of the following theorem, which is a consequence of [DMS21, Theorems 1.9 and 8.18]. ###### Theorem 5.1. With $h$ and $\eta$ as above, let $\mathcal{G}$ be the graph with vertex set $\mathbbm{Z}$, with two distinct vertices $x,y\in\mathbbm{Z}$ joined by an edge if and only if $\eta([x-1,x])\cap\eta([y-1,y])\not=\emptyset.$ (5.5) Then $\mathcal{G}$ has the same law (as a graph) as the mated-CRT map as defined in (4.1). The mated-CRT map has some double edges, but we do not worry about such edge multiplicity in Theorem 5.1 since in this section we are only interested in graph distances. ###### Remark 5.2. The results and proofs in this section all carry over verbatim if we replace $\sqrt{2}$-LQG by $\gamma$-LQG for $\gamma\in(0,2)$ and SLE8 by space-filling SLEκ for $\kappa=16/\gamma^{2}$. In this setting, the corresponding mated-CRT map is constructed from a pair of correlated Brownian motions with correlation $-\cos(\pi\gamma^{2}/4)$, instead of a pair of independent Brownian motions as in Section 4.1; and the value of $d$ depends on $\gamma$. ### 5.2 Proof of lower bounds for mated-CRT graph distances Henceforth assume that we are in the setting of Theorem 5.1. Our goal is to prove Proposition 4.3. To this end, we first prove a lemma that allows us to compare $D_{h}$-distances and graph distances in $\mathcal{G}$. For $n\in\mathbbm{N}$ and $z,w\in\eta([-n,n])\subset\mathbb{C}$, we slightly abuse notation by writing $\operatorname{dist}_{\mathcal{G}[-n,n]}(z,w):=\min\mathopen{}\mathclose{{}\left\\{\operatorname{dist}_{\mathcal{G}[-n,n]}(x,y):x,y\in[-n,n]\cap\mathbbm{Z},z\in\eta([x-1,x]),w\in\eta([y-1,y])}\right\\}.$ (5.6) Note that a.s. for Lebesgue-a.e. point $z$, the cell $\eta([x-1,x])$ containing $z$ is unique (since a.s. $z$ is hit exactly once by $\eta$), and that $\operatorname{dist}_{\mathcal{G}[-n,n]}(z,w)=0$ if $z$ and $w$ belong to the same cell. ###### Lemma 5.3. Fix $\zeta\in(0,1)$. It holds with polynomially high probability as $n\rightarrow\infty$ that $\operatorname{dist}_{\mathcal{G}[-n,n]}(z,w)\geq n^{-\zeta}D_{h}\mathopen{}\mathclose{{}\left(z,w}\right)-1,\quad\forall z,w\in\eta([-n,n]).$ (5.7) ###### Proof. By [GS22, Proposition 3.13], for each $p>0$ and each $t\in\mathbbm{R}$, the $p$th moment of the random variable $\sup_{u,v\in\eta([t-1,t])}D_{h}(u,v)$ is bounded above by a finite constant which depends only on $p$ (not on $t$). Therefore, we can apply Chebyshev’s inequality (for large positive moments) followed by a union bound over all $x\in[-n,n]\cap\mathbbm{Z}$ to get that with superpolynomially high probability as $n\rightarrow\infty$, $\max_{x\in[-n,n]\cap\mathbbm{Z}}\sup_{u,v\in\eta([x-1,x])}D_{h}(u,v)\leq n^{\zeta}.$ (5.8) Henceforth assume that (5.8) holds. Let $z,w\in\eta([-n,n])$ and let $N:=\operatorname{dist}_{\mathcal{G}[-n,n]}(z,w)$. By the description of $\mathcal{G}$ in Theorem 5.1, there exists $x_{0},x_{1},\dots,x_{N}\in[-n,n]\cap\mathbbm{Z}$ such that the union of the SLE8 segments $\eta([x_{j}-1,x_{j}])$ for $j=0,\dots,N$ contains a path from $z$ to $w$. By (5.8), the $D_{h}$-diameter of each of these cells is at most $n^{\zeta}$. By the triangle inequality, $D_{h}(z,w)\leq n^{\zeta}(N+1)$ which is (5.7). ∎ The following lemma is the main LQG estimate needed for the proof of Proposition 4.3. ###### Lemma 5.4. For each $t>0$, it holds with probability tending to 1 as $C\rightarrow\infty$, uniformly over the choice of $t$, that $D_{h}\mathopen{}\mathclose{{}\left(\eta([-t,t]),\partial\eta([-Ct,Ct])}\right)\geq t^{1/d}.$ (5.9) Moreover, for each fixed $C>1$ it holds with probability tending to 1 as $\delta\rightarrow 0$, uniformly over the choice of $t$, that $D_{h}\mathopen{}\mathclose{{}\left((-\infty,0]\cap\overline{\eta\mathopen{}\mathclose{{}\left([-Ct,Ct]\setminus[-t,t]}\right)},[0,\infty)\cap\overline{\eta\mathopen{}\mathclose{{}\left([-Ct,Ct]\setminus[-t,t]}\right)}}\right)\geq\delta t^{1/d}$ (5.10) where here we identify $\mathbbm{R}$ with $\mathbbm{R}\times\\{0\\}\subset\mathbbm{C}$. ###### Proof. By the scale invariance property (5.4), it suffices to prove both (5.9) and (5.10) in the case when $t=1$. We start with (5.9). Since $\eta$ a.s. fills all of $\mathbbm{C}$ and LQG metric balls of finite radius are a.s. compact, a.s. there exists some $C>1$ such that $\mathopen{}\mathclose{{}\left\\{z\in\mathbbm{C}\,:\,D_{h}(z,\eta([-1,1]))\leq 1}\right\\}\subset\eta([-C,C]).$ (5.11) Hence, (5.11) holds with probability tending to 1 as $C\rightarrow\infty$. This shows that (5.9) with $t=1$ holds with probability tending to 1 as $C\rightarrow\infty$. Since a.s. 0 is contained in the interior of $\eta([-1,1])$, for any fixed $C>1$ the compact sets $(-\infty,0]\cap\overline{\eta\mathopen{}\mathclose{{}\left([-C,C]\setminus[-1,1]}\right)}\quad\text{and}\quad[0,\infty)\cap\overline{\eta\mathopen{}\mathclose{{}\left([-C,C]\setminus[-1,1]}\right)}$ lie at positive Euclidean distance from each other. Moreover, since $D_{h}$ induces the Euclidean topology on $\mathbbm{C}$, the two compact sets above lie at positive $D_{h}$ distance from each other. Hence (5.10) with $t=1$ holds with probability tending to 1 as $\delta\rightarrow 0$. ∎ Figure 13: Illustration of the SLE/LQG event used in the proof of Proposition 4.3. On the event, the LQG distance between the inner and outer boundaries of the green, light-blue, and gray annular regions, as well as the LQG distance between the blue and red sets, are all bounded below. We note that the blue and red sets are not connected, but each of these sets contains a connected set whose closure intersects each of $\eta([-n,n])$ and $\overline{\mathbbm{C}\setminus\eta([-C^{2}n,C^{2}n])}$. ###### Proof of Proposition 4.3. See Figure 13 for an illustration. By Lemma 5.4, we can find $C>1$ such that for each $t>0$, (5.9) holds with probability at least $1-(1-p)/5$. Applying this with $t=n$, $t=Cn$, and $t=C^{2}n$ shows that with probability at least $1-3(1-p)/5$, $\displaystyle D_{h}\mathopen{}\mathclose{{}\left(\eta([-n,n]),\partial\eta([-Cn,Cn])}\right)\geq n^{1/d},\quad D_{h}\mathopen{}\mathclose{{}\left(\eta([-Cn,Cn]),\partial\eta([-C^{2}n,C^{2}n])}\right)\geq(Cn)^{1/d},\quad\text{and}$ $\displaystyle D_{h}\mathopen{}\mathclose{{}\left(\eta([-C^{2}n,C^{2}n]),\partial\eta([-C^{3}n,C^{3}n])}\right)\geq(C^{2}n)^{1/d}.$ (5.12) By (5.10) of Lemma 5.4 (applied with $C^{2}$ instead of $C$ and $t=n$), we can find $\delta\in(0,1)$ (depending on $C$) such that with probability at least $1-(1-p)/5$, $D_{h}\mathopen{}\mathclose{{}\left((-\infty,0]\cap\overline{\eta\mathopen{}\mathclose{{}\left([-C^{2}n,C^{2}n]\setminus[-n,n]}\right)},[0,\infty)\cap\overline{\eta\mathopen{}\mathclose{{}\left([-C^{2}n,C^{2}n]\setminus[-n,n]}\right)}}\right)\geq\delta n^{1/d}.$ (5.13) By Lemma 5.3 (applied with $C^{3}n$ in place of $n$ and $\zeta/2$ in place of $\zeta$), for each large enough $n\in\mathbbm{N}$ it holds with probability at least $1-(1-p)/5$ that $\operatorname{dist}_{\mathcal{G}[-C^{3}n,C^{3}n]}(z,w)\geq n^{-\zeta/2}D_{h}\mathopen{}\mathclose{{}\left(z,w}\right)-1,\quad\forall z,w\in\eta([-C^{3}n,C^{3}n]).$ (5.14) Henceforth assume that (5.2), (5.13), and (5.14) all occur, which happens with probability at least $p$ if $n$ is large enough. Recall that two vertices $x,y\in\mathbbm{Z}$ are joined by an edge of $\mathcal{G}$ if and only if $\eta([x-1,x])\cap\eta([y-1,y])\not=\emptyset$. Hence each path in $\mathcal{G}$ from $[-n,n]\cap\mathbbm{Z}$ to $\mathbbm{Z}\setminus[-Cn,Cn]$ has a sub-path which is contained in $[-Cn,Cn]\cap\mathbbm{Z}$ and which goes from $[-n,n]\cap\mathbbm{Z}$ to a vertex $x\in\mathbbm{Z}$ whose corresponding cell $\eta([x-1,x])$ intersects $\partial\eta([-Cn,Cn])$. Therefore, the first inequality in (5.2) together with (5.14) implies (a) in the lemma statement (provided $n$ is large enough that $n^{1/d-\zeta/2}-1\geq n^{1/d-\zeta}$). We similarly obtain (b) and (c) from the second and third inequalities in (5.2). To get (d), let $\Pi_{1}^{0}$ (resp. $\Pi_{2}^{0}$) be the set of $x\in\mathbbm{Z}$ such that the cell $\eta([x-1,x])$ intersects $(-\infty,0]\cap\overline{\eta\mathopen{}\mathclose{{}\left([-C^{2}n,C^{2}n]\setminus[-n,n]}\right)}$ (resp. $[0,\infty)\cap\overline{\eta\mathopen{}\mathclose{{}\left([-C^{2}n,C^{2}n]\setminus[-n,n]}\right)}$). Then each of $\Pi_{1}^{0}$ and $\Pi_{2}^{0}$ contains a connected subset of $\mathbbm{Z}$ which intersects both $[-n,n]\cap\mathbbm{Z}$ and $\mathbbm{Z}\setminus[-C^{2}n,C^{2}]$. So, we can find a path $\Pi_{1}$ (resp. $\Pi_{2}$) in $\mathcal{G}$ from $[-n,n]\cap\mathbbm{Z}$ to $\mathbbm{Z}\setminus[-C^{2}n,C^{2}]$ which is contained in $\Pi_{1}^{0}$ (resp. $\Pi_{2}^{0}$) and which is contained in $[-C^{2}n,C^{2}n]\cap\mathbbm{Z}$ except for its terminal endpoint. By (5.13) and (5.14), the $\mathcal{G}[-C^{3}n,C^{3}n]$-graph distance between $\Pi_{1}$ and $\Pi_{2}$ is at least $\delta n^{1/d-\zeta/2}-1$, which is at least $n^{1/d-\zeta}$ if $n$ is large enough. Furthermore, by (c) from the proposition statement (which we have already proven), the $\mathcal{G}$-graph distance from each of $\Pi_{1}$ and $\Pi_{2}$ to $\mathbbm{Z}\setminus[-Cn^{3},Cn^{3}]$ is at least $n^{1/d-\zeta}$. Hence, any path in $\mathcal{G}$ from $\Pi_{1}$ to $\Pi_{2}$ has to have length at least $n^{1/d-\zeta}$. This gives (d). ∎ ## 6 Proofs for the UIHPMS This section is organized as follows. In Section 6.1, we prove the equivalence of our two definitions of the UIHPMS (Lemma 1.11). The rest of the section is devoted to the proofs of Theorem 1.12 and Proposition 1.14. We start out in Section 6.2 by introducing a special sequence of boundary points for the UIHPMS, $\\{H_{m}\\}_{m\in\mathbbm{Z}}\subset\\{J_{k}\\}_{k\in\mathbbm{Z}}$. The points $\\{H_{m}\\}_{m\in\mathbbm{Z}}$ may be viewed as marked points of a minor variant of the UIHPMS that satisfies an invariance property under translation by $H_{m}$ as in (1.10), but (crucially) without the constraint that $m$ is even (Lemma 6.1). This property will be important for our purposes since we will eventually need to use some parity arguments. Sections 6.3 and 6.4 constitute the core part of the argument. In Section 6.3, we prove that a.s. there are no semi-infinite paths in the UIHPMS that start at one of the special boundary points $H_{m}$ and never hit any other special boundary point (Lemma 6.3). This is done using a Burton-Keane style argument, combined with the existence of certain special times for the pair of encoding walks $(\mathcal{L},|\mathcal{R}|)$ (Lemma 6.5). The non-existence of such semi-infinite paths allows us to define a perfect matching on $\mathbbm{Z}$ by saying that $m\in\mathbbm{Z}$ is matched to $m^{\prime}\in\mathbbm{Z}$ if and only if the path $P_{m}$ of arcs started from $H_{m}$ hits $H_{m^{\prime}}$ before hitting any other special boundary point (see Definition 6.2 for a precise definition of this path). In Section 6.4, we prove a general lemma for ergodic perfect matchings on $\mathbbm{Z}$ (Lemma 6.9) which in our setting implies that there are infinitely many paths between good boundary points which disconnect 0 from $\infty$ in the UIHPMS (Lemma 6.6). As explained in Section 6.5, these paths act as “shields” which cannot be crossed by any infinite path in the UIHPMS. This allows us to prove Theorem 1.12 and Proposition 1.14. Throughout this section, we assume that the arcs of the UIMS and the UIHPMS are drawn in such a way that each arc joining $x<y$ is contained in $[x,y]\times\mathbbm{R}$. With this convention, it is easy to see from either of the two definitions of the UIHPMS from Section 1.5 that for each boundary point $J_{k}$ there is no arc which crosses the downward ray $\\{J_{k}\\}\times(-\infty,0]$. ### 6.1 Equivalence of the definitions of the UIHPMS Recall the two definitions of the UIHPMS from Section 1.5, one by cutting the UIMS and one in terms of a simple random walk and an independent reflected random walk. We now prove that the two definitions are equivalent. We use a discrete version of celebrated Lévy’s theorem [Sim83], which we now recall. Let $\mathcal{X}=(\mathcal{X}_{n})_{n\geq 0}$ be a (one-sided) simple random walk with $\mathcal{X}_{0}=0$ and $M^{\mathcal{X}}$ be the running minimum process of $\mathcal{X}$ from time $0$, that is $M^{\mathcal{X}}_{n}=\min_{k\in[0,n]}\mathcal{X}_{k},\qquad\text{for }n\geq 0.$ (6.1) Then $(\mathcal{X}-M^{\mathcal{X}},-M^{\mathcal{X}})\stackrel{{\scriptstyle d}}{{=}}(|\mathcal{X}|-\textbf{1}_{\\{\mathcal{X}>0\\}},\ell^{\mathcal{X}})$ (6.2) where $\ell^{\mathcal{X}}_{n}$ denotes the number of times that $\mathcal{X}$ crosses $1/2$ during the time interval $[0,n]$, for all $n\geq 0$. ###### Proof of Lemma 1.11. Let $(\mathcal{L},\mathcal{R})$ be the pair of independent two-sided simple random walks on $\mathbbm{Z}$ used to construct the UIMS as in Section 1.4. We assume that $\mathcal{L}_{0}=\mathcal{R}_{0}=0$. We first analyze the cutting description of the UIHPMS. For each integer $k\in\mathbbm{Z}_{>0}$, let $T_{k}$ be the smallest time $x\in\mathbbm{Z}_{>0}$ such that $\mathcal{R}_{x}=-k$. Also, let $T_{-k}$ be the largest time $x\in\mathbbm{Z}_{<0}$ such that $\mathcal{R}_{x}=-k$. By (1.5), for each $k\in\mathbbm{Z}_{>0}$ the point $T_{k}$ is joined to $T_{-k}+1$ by a lower arc of the UIMS . Furthermore, arcs of this type are the only ones which cross $\\{1/2\\}\times(-\infty,0]$. Therefore, the UIHPMS under the cutting definition is obtained from the UIMS by first removing the lower arcs joining $T_{k}$ and $T_{-k}+1$ for $k\in\mathbbm{Z}_{>0}$, then adding lower arcs joining $T_{2k-1}$ and $T_{2k}$ for each $k\in\mathbbm{Z}_{>0}$; and lastly joining $T_{2k+1}+1$ and $T_{2k}+1$ for each $k\in\mathbbm{Z}_{<0}$. Let $\widetilde{\mathcal{R}}$ be the random walk on $\mathbbm{Z}$ obtained from $\mathcal{R}$ by replacing each of the downward steps $\mathcal{R}_{T_{2k-1}}-\mathcal{R}_{T_{2k-1}-1}$ for $k\in\mathbbm{Z}_{>0}$ by an upward step, and replacing each of the upward steps $\mathcal{R}_{T_{2k+1}+1}-\mathcal{R}_{T_{2k+1}}$ for $k\in\mathbbm{Z}_{<0}$ by a downward step (and otherwise leaving the steps of $\mathcal{R}$ unchanged). Then the cutting description of the UIHPMS is equivalent to the random walk description (1.5) with $\widetilde{\mathcal{R}}$ in place of $\mathcal{R}$. Therefore, it suffices to show that $\widetilde{\mathcal{R}}\overset{d}{=}|\mathcal{R}|$. Let $M^{\mathcal{R}}$ and $\ell^{\mathcal{R}}$ be as in (6.1) and (6.2), with $\mathcal{X}=\mathcal{R}\mid_{[0,\infty)\cap\mathbbm{Z}}$. Then for $x\in\mathbbm{Z}_{\geq 0}$, we have $\widetilde{\mathcal{R}}_{x}=\mathcal{R}_{x}-M_{x}^{\mathcal{R}}+\textbf{1}_{\\{\text{$M_{x}^{\mathcal{R}}$ is odd}\\}}.$ (6.3) On the other hand, the number $\ell_{x}^{\mathcal{R}}$ of crossings of $1/2$ by $\mathcal{R}$ during the time interval $[0,x]$ is odd if and only if $\mathcal{R}_{x}>0$, so $|\mathcal{R}|_{x}=|\mathcal{R}|_{x}-\textbf{1}_{\\{\mathcal{R}_{x}>0\\}}+\textbf{1}_{\\{\text{$\ell_{x}^{\mathcal{R}}$ is odd}\\}}.$ (6.4) By combining the preceding two identities with (6.2), we get that $\widetilde{\mathcal{R}}\mid_{[0,\infty)\cap\mathbbm{Z}}\overset{d}{=}|\mathcal{R}|\mid_{[0,\infty)\cap\mathbbm{Z}}$. We similarly get the desired equality in law for negative times. ∎ ### 6.2 Aperiodicity via good boundary points In this subsection we address the following technical point. The law of the UIHPMS is only invariant under _even_ translations along the boundary, i.e., translations by $J_{2k}$ for $k\in\mathbbm{Z}$; see (1.10). In one step of our proof of Theorem 1.12 (Section 6.4) we will use an ergodicity argument which requires us to know that a certain event for boundary points of odd index has the same probability as the corresponding event for boundary points of even index. For this reason, we need to introduce a variant of the UIHPMS where we have translation invariance for _all_ boundary points, not just even boundary points. The idea is to replace the reflected random walk $|\mathcal{R}|$ by a walk which has some constant steps at height 0. This is similar to how one can make a Markov chain aperiodic by introducing constant steps. Figure 14: Illustration of the variant $\mathcal{M}^{\operatorname{cut}}$ of the UIHPMS $\mathcal{M}^{\prime}$ constructed using the walk $\mathcal{Y}$ of (6.5). Dashed arcs are the ones which are part of the original UIHPMS $\mathcal{M}^{\prime}$ but not part of $\mathcal{M}^{\operatorname{cut}}$. The graph of the reflected random walk $|\mathcal{R}|$ used in the construction of the UIHPMS is shown at the bottom of the figure. Below each excursion of $|\mathcal{R}|$ we have also shown the outcomes of the i.i.d. Bernoulli$(1/2)$ random variables $\\{\xi_{k}\\}_{k\in\mathbbm{Z}}$. Let $\mathcal{R}$ be the two-sided simple random walk on $\mathbbm{Z}$ such that $|\mathcal{R}|$ is used to construct the UIHPMS. We define the modified walk $\mathcal{Y}_{x}:=|\mathcal{R}|_{x}-\textbf{1}_{\\{\mathcal{R}_{x}>0\\}}.$ (6.5) For $m\in\mathbbm{Z}$, let $H_{m}:=\mathopen{}\mathclose{{}\left(\text{$m$th smallest $x\in\mathbbm{Z}$ such that $\mathcal{Y}_{x}=\mathcal{Y}_{x-1}=0$}}\right),$ (6.6) with the numbering chosen so that $0\in[H_{0},H_{1}-1]\cap\mathbbm{Z}$. Recall that we denote by $J_{k}$ (resp. $J_{-k+1}$) the $k$th positive (resp. non-positive) boundary point of the UIHPMS. In particular, for each $k\in\mathbbm{Z}$, $J_{2k}=J_{2k+1}-1$ and $\mathcal{R}_{J_{2k}}=\mathcal{R}_{J_{2k+1}-1}=0$; and these are all the zeros of $\mathcal{R}$. Conditional on $|\mathcal{R}|$, the signs of the excursions $\\{\mathcal{R}|_{[J_{2k-2},J_{2k}]}\\}_{k\in\mathbbm{Z}}$ of $|\mathcal{R}|$ away from 0 are i.i.d. Bernoulli$(1/2)$ random variables. Therefore, $\mathcal{Y}$ can equivalently be constructed from $|\mathcal{R}|$ as follows. Start with a collection $\\{\xi_{k}\\}_{k\in\mathbbm{Z}}$ of i.i.d. Bernoulli$(1/2)$ random variables independent from $|\mathcal{R}|$. Then, for each $k\in\mathbbm{Z}$ such that $\xi_{k}=1$, replace the steps $|\mathcal{R}|_{J_{2k}}-|\mathcal{R}|_{J_{2k}-1}$ and $|\mathcal{R}|_{J_{2k-1}}-|\mathcal{R}|_{J_{2k-1}-1}$ by constant steps. In particular, $\\{H_{m}\\}_{m\in\mathbbm{Z}}\subset\\{J_{k}\\}_{k\in\mathbbm{Z}}\quad\text{and}\quad\mathbbm{P}\mathopen{}\mathclose{{}\left[J_{k}\in\\{H_{m}\\}_{m\in\mathbbm{Z}}\,|\,|\mathcal{R}|}\right]=\frac{1}{2},\quad\forall k\in\mathbbm{Z}.$ (6.7) We call $H_{m}$ the $m$th good boundary point of the UIHPMS. See Figure 14 for an illustration. Let $(\mathcal{M}^{\operatorname{cut}},\Gamma^{\operatorname{cut}})$ be the infinite planar map decorated by a collection of loops and paths which is defined exactly as in (1.5) but with $\mathcal{Y}$ in place of $\mathcal{R}$, with the convention that if $\mathcal{Y}_{x}=\mathcal{Y}_{x-1}=0$, then there is no lower arc incident to $x\in\mathbbm{Z}$. The discussion just above implies that $(\mathcal{M}^{\operatorname{cut}},\Gamma^{\operatorname{cut}})$ can be obtained from the UIHPMS by removing each lower arc joining boundary points $J_{2k-1}$ and $J_{2k}$ for $k\in\mathbbm{Z}$ such that $\xi_{k}=1$. The good boundary points are exactly these points $J_{2k-1}$ and $J_{2k}$ for $k\in\mathbbm{Z}$ such that $\xi_{k}=1$. In particular, they are not determined by the UIHPMS. One can also see from the proof of the equivalence of the two definitions of the UIHPMS (Lemma 1.11) that $(\mathcal{M}^{\operatorname{cut}},\Gamma^{\operatorname{cut}})$ has the same law as the decorated planar map obtained from the UIMS by removing all of the lower arcs which intersect $\\{1/2\\}\times(-\infty,0]$. Moreover, in this interpretation, the good boundary points are just the endpoints of such removed arcs. The following lemma is our main reason for introducing the objects in this subsection. It will be used in the proof of Lemma 6.6 below. ###### Lemma 6.1. Let $\widehat{\mathcal{Y}}$ be sampled from the conditional law of $\mathcal{Y}$ given the event $\\{\mathcal{Y}_{0}=\mathcal{Y}_{-1}=0\\}=\\{H_{0}=0\\}.$ (6.8) Let $(\widehat{\mathcal{M}}^{\operatorname{cut}},\widehat{\Gamma}^{\operatorname{cut}})$ be defined in the same manner as $(\mathcal{M}^{\operatorname{cut}},\Gamma^{\operatorname{cut}})$ with $\widehat{\mathcal{Y}}$ in place of $\mathcal{Y}$ and let $\\{\widehat{H}_{m}\\}_{m\in\mathbbm{Z}}$ be as in (6.6) with $\widehat{\mathcal{Y}}$ in place of $\mathcal{Y}$. Then $\mathopen{}\mathclose{{}\left(\widehat{\mathcal{M}}^{\operatorname{cut}}-\widehat{H}_{m},\widehat{\Gamma}^{\operatorname{cut}}-\widehat{H}_{m}}\right)\overset{d}{=}\mathopen{}\mathclose{{}\left(\widehat{\mathcal{M}}^{\operatorname{cut}},\widehat{\Gamma}^{\operatorname{cut}}}\right),\quad\forall m\in\mathbbm{Z}.$ (6.9) Furthermore, for each positive integer $r\in\mathbbm{Z}_{>0}$, each event which is invariant under translations of the form (6.9) with $m$ restricted to lie in $r\mathbbm{Z}$ has probability zero or one. ###### Proof. Since $(\widehat{\mathcal{M}}^{\operatorname{cut}},\widehat{\Gamma}^{\operatorname{cut}})$ is constructed from $\widehat{\mathcal{Y}}$ and an independent two-sided simple random walk $\mathcal{L}$ in the manner of (1.5), it suffices to show that $\widehat{\mathcal{Y}}_{\widehat{H}_{m}+\,\cdot}\overset{d}{=}\widehat{\mathcal{Y}}_{\cdot}\,,\quad\forall m\in\mathbbm{Z},$ (6.10) and that any event which is invariant under translations of the form (6.10) with $m$ restricted to lie in $r\mathbbm{Z}$ has probability zero or one. To this end, we first consider the unconditioned random walk $\mathcal{Y}=|\mathcal{R}|-\textbf{1}_{\\{\mathcal{R}>0\\}}$ from (6.5). Let $\\{\tau_{j}\\}_{j\in\mathbbm{Z}}$ be the times such that $\mathcal{Y}_{x}=0$, numbered from left to right in such a way that $\tau_{0}=0$. Let $(\mathcal{X},M^{\mathcal{X}})$ be a one-sided simple random walk on $\mathbbm{Z}$ started from $0$ at time $0$ and its running minimum process, as in (6.1). Also let $\mu$ be the law of $\mathcal{X}-M^{\mathcal{X}}$ stopped at the first positive time when it reaches zero. A sample from $\mu$ can be produced as follows: Flip a fair coin. If the coin comes up heads, we run a simple random walk started from 0, conditioned on the event that its first step is upward, until the first positive time when it reaches zero. If the coin comes up tails, we instead take the path which takes one step from 0 to 0. From this description, we get that the law of a path sampled from $\mu$ is invariant under time reversal. By the discrete version of Lévy’s theorem (6.2), the process $\mathcal{Y}|_{[0,\infty)\cap\mathbbm{Z}}$ has the same law as $\mathcal{X}-M^{\mathcal{X}}$. By the strong Markov property of $\mathcal{X}$, we get that the excursions $\mathcal{Y}|_{[\tau_{j-1},\tau_{j}]\cap\mathbbm{Z}}$ for $j\geq 1$ are i.i.d. samples from $\mu$. The law of $\mathcal{Y}$ is invariant under time reversal, so the increments $\mathcal{Y}|_{[\tau_{j-1},\tau_{j}]\cap\mathbbm{Z}}$ for $j\leq 0$ are also i.i.d., and have the same law as the time reversal of a sample from $\mu$. From the time reversal symmetry of $\mu$, we get that the law of the whole collection of increments $\mathcal{Y}|_{[\tau_{j-1},\tau_{j}]\cap\mathbbm{Z}}$ for $j\in\mathbbm{Z}$ are i.i.d. samples from $\mu$. The points $H_{m}$ for $m\in\mathbbm{Z}$ coincide precisely with the times $\tau_{j}$ for which $\tau_{j}=\tau_{j-1}+1$. From this and the above description of the increments $\mathcal{Y}|_{[\tau_{j-1},\tau_{j}]\cap\mathbbm{Z}}$, we infer that the increments $\widehat{\mathcal{Y}}|_{[\widehat{H}_{m-1},\widehat{H}_{m}]\cap\mathbbm{Z}}$ for $m\in\mathbbm{Z}$ are i.i.d. samples from the law of $\mathcal{Y}|_{[0,\infty)\cap\mathbbm{Z}}$ stopped at the first positive time $x$ such that $\mathcal{Y}_{x}=\mathcal{Y}_{x-1}=0$. From this and the fact that $\widehat{\mathcal{Y}}$ is sampled from the conditional law of $\mathcal{Y}$ given the event $\\{\mathcal{Y}_{0}=\mathcal{Y}_{-1}=0\\}$, the translation invariance property (6.10) is immediate. Furthermore, the desired ergodicity property for $\widehat{\mathcal{Y}}$ follows from the zero-one law for translation invariant events depending on a sequence of i.i.d. random variables, applied to the i.i.d. random variables $\widehat{\mathcal{Y}}|_{[\widehat{H}_{rn},\widehat{H}_{r(n+1)}]\cap\mathbbm{Z}}$ for $n\in\mathbbm{Z}$. We remark that one could also prove (6.10) by showing that if $U_{m}$ is sampled uniformly from $[-m,m]\cap\mathbbm{Z}$, then the law of $\mathcal{Y}_{H_{U_{m}}+\cdot}$ converges as $m\to\infty$ to the law of $\widehat{\mathcal{Y}}$. We will not give the details. ∎ ### 6.3 No semi-infinite paths started from the boundary Define the UIHPMS and its left-to-right ordered sequence of boundary points $\\{J_{k}\\}_{k\in\mathbbm{Z}}\subset\mathbbm{Z}$ as in Section 1.5. Also define the left-to-right ordered sequence of good boundary points $\\{H_{m}\\}_{m\in\mathbbm{Z}}$ as in (6.6). ###### Definition 6.2. For each good boundary point $H_{m}$ with $m\in\mathbbm{Z}$, let $P_{m}$ be the unique directed path of arcs in the UIHPMS starting from $H_{m}$, following the arc in the _upper_ half-plane incident to $H_{m}$, and ending at the first good boundary point other than $H_{m}$ which is hit by the path, if it exists; otherwise let $P_{m}$ be the whole semi-infinite path started from $H_{m}$. We call $P_{m}$ the boundary path started from $H_{m}$. By definition, the boundary path $P_{m}$ hits good boundary points of the UIHPMS only at its endpoints. However, it is in principle possible that for some values of $m$, the path $P_{m}$ never hits another good boundary point other than $H_{m}$, in which case it is semi-infinite. The first step in the proof of Theorem 1.12 is to rule this event out. ###### Lemma 6.3. Almost surely, the boundary path $P_{m}$ is finite for all $m\in\mathbbm{Z}$. The proof of Lemma 6.3 proceeds as follows. By even translation invariance (1.10), the probability that $J_{k}$ is a good boundary point (i.e., $J_{k}=H_{m}$ for some $m\in\mathbbm{Z}$) and $P_{m}$ is semi-infinite depends only on the parity of $k$. If this probability is positive for even values of $k$, say, then by the ergodic theorem a.s. $P_{m}$ is semi-infinite for a positive fraction of the indices $m\in\mathbbm{Z}$. We will show that this cannot be the case by a Burton-Keane style argument. Roughly speaking, we will use certain special times for the encoding walks (Lemma 6.5) to argue that there is not enough “room” for there to be a positive density of values of $m$ for which $P_{m}$ is semi-infinite. We start with some bounds on the probability that a bridge has at least some number of crossings of $1/2$. ###### Lemma 6.4. For a simple random walk $\mathcal{X}$ on $\mathbbm{Z}$ with $\mathcal{X}_{0}=0$, recall that $\ell^{\mathcal{X}}_{n}$ denotes the number of times that $\mathcal{X}$ crosses $1/2$ during the time interval $[0,n]$. For $n,k>0$, we have $\frac{2^{2k-1}\binom{2n-2k}{n}}{\binom{2n}{n}}\leq\mathbbm{P}\mathopen{}\mathclose{{}\left[\ell^{\mathcal{X}}_{2n}\geq 2k\;\middle|\;\mathcal{X}_{2n}=0}\right]\leq\frac{2^{k}\binom{2n-k}{n}}{\binom{2n}{n}}.$ (6.11) As a particular consequence, for each $\varepsilon>0$ and each integer $n>0$, $C_{1}\leq\mathbbm{P}\mathopen{}\mathclose{{}\left[\ell^{\mathcal{X}}_{2n}\geq\varepsilon\sqrt{n}\;\middle|\;\mathcal{X}_{2n}=0}\right]\leq C_{2}\quad\text{and}\quad C_{1}\leq\mathbbm{P}\mathopen{}\mathclose{{}\left[\ell^{\mathcal{X}}_{2n}\geq\varepsilon\sqrt{n}\;\middle|\;(\mathcal{X}_{2n-1},\mathcal{X}_{2n})=(1,0)}\right]\leq C_{2},$ (6.12) for constants $C_{1},C_{2}\in(0,1)$ only depending on $\varepsilon$. ###### Proof. We call a simple random walk path starting and ending at $0$ an _excursion_ if it does not hit $0$ except at its two endpoints (we allow excursions to be positive or negative). If $\mathcal{X}_{2n}=0$, we can decompose the path of $\mathcal{X}\mid_{[0,2n]\cap\mathbbm{Z}}$ into several excursions. The number $\ell^{\mathcal{X}}_{2n}$ is exactly twice the number of positive excursions. Hence, if $\mathcal{X}_{2n}=0$ and $\ell^{\mathcal{X}}_{2n}\geq 2k$, then the path of $\mathcal{X}\mid_{[0,2n]\cap\mathbbm{Z}}$ can be decomposed into at least $k$ excursions. Therefore, we compute $\mathbbm{P}\mathopen{}\mathclose{{}\left[\mathcal{X}\mid_{[0,2n]\cap\mathbbm{Z}}\text{has at least $k$ excursions}\;\middle|\;\mathcal{X}_{2n}=0}\right]$ in the next paragraph to obtain an upper bound for $\mathbbm{P}\mathopen{}\mathclose{{}\left[\ell^{\mathcal{X}}_{2n}\geq 2k\;\middle|\;\mathcal{X}_{2n}=0}\right]$. Assume that the path of $\mathcal{X}\mid_{[0,2n]\cap\mathbbm{Z}}$ with $\mathcal{X}_{2n}=0$ can be decomposed into at least $k$ excursions. By making the last $k$ excursions all positive (by flipping if necessary) and removing the last downward step of each of these $k$ excursions, we obtain a simple walk from $0$ to $k$ with $2n-k$ steps, i.e. with $n$ upward steps and $n-k$ downward steps. The total number of such walks is $\binom{2n-k}{n}$. In fact, this defines a $2^{k}$-to-$1$ map, as the modified positive excursions can be recovered from the resulting walk from $0$ to $k$ (the last visit of $j\in[1,k]\cap\mathbbm{Z}$ corresponds to where a downward step was removed), and there are $2^{k}$ ways to assign a sign for each excursion to obtain $\mathcal{X}\mid_{[0,2n]}$. As there are $\binom{2n}{n}$ possible walks from $0$ to $0$ with $2n$ steps, the upper bound in (6.11) follows. The lower bound of (6.11) is obtained via a similar argument. It suffices to notice that a simple walk from $0$ to $2k$ with $2n-2k$ steps can be turned (as done above) in a walk from 0 to 0 with $2n$ steps and at least $k$ positive excursions among the last $2k$ ones in at least $2^{2k-1}$ ways (by symmetry). The first bound in (6.12) is obtained by setting $k=\lfloor\varepsilon\sqrt{n}/2\rfloor$ in (6.11) and applying Stirling’s formula. To add the extra condition $\mathcal{X}_{2n-1}=1$ in the second bound in (6.12), we just fix the last excursion to be positive in the excursion decomposition defined above, which gives a similar bound. ∎ Let $\mathcal{L}$ and $\mathcal{R}$ be the independent two-sided simple random walks on $\mathbb{Z}$ with $\mathcal{L}_{0}=\mathcal{R}_{0}=0$ used in the definition of the UIHPMS from (1.9). We call $x\in\mathbbm{Z}_{>0}$ an upper block if $x$ is linked to some point in $(-\infty,0]\cap\mathbbm{Z}$ by an arc above the real line. If $x$ is an upper block, no arc above the real line connects $(0,x)\cap\mathbbm{Z}$ and $(x,\infty)\cap\mathbbm{Z}$; otherwise such an arc would have to cross the upper arc incident to $x$. In terms of the random walk description, $x\in\mathbbm{Z}_{>0}$ is an upper block if and only if $\mathcal{L}\mid_{[0,x]\cap\mathbbm{Z}}$ attains its _unique_ minimum value at time $x$. Recall that for $k>0$ we denoted by $J_{k}$ the $k$th positive boundary point of the UIHPMS. We call $x\in 2\mathbbm{Z}_{>0}$ a block if $x$ is an upper block and $x=J_{2k}$ for some $k>0$. That is, $x\in 2\mathbbm{Z}_{>0}$ is a block if and only if $|\mathcal{R}|_{x}=0$ and $\mathcal{L}\mid_{[0,x]\cap\mathbbm{Z}}$ attains its unique minimum value at time $x$. See Figure 15 for an example. Our motivation for the definition of a block is that if $x$ is a block, then no path of arcs started at a point in $(0,x)\cap\mathbbm{Z}$ can cross the vertical line $\\{x\\}\times\mathbbm{R}$ without first hitting $(-\infty,0]\cap\mathbbm{Z}$. Figure 15: An illustration of a block $x=J_{2k}$ with some upper blocks and boundary points in a subset of the UIHPMS. The right endpoints of the purple arcs are upper blocks, while the red arcs are not incident to any upper blocks. Boundary points are the endpoints of the lower green arcs. Lemma 6.5 describes a special type of block such that the number of purple arcs is much fewer than the number of green arcs. ###### Lemma 6.5. Fix $\varepsilon>0$. Almost surely, there are infinitely many $k>0$ such that $J_{2k}$ is a block and there are at most $\varepsilon k$ upper blocks in $(0,J_{2k}]\cap\mathbbm{Z}$. ###### Proof. In light of Lévy’s theorem (6.2), consider another two-sided simple random walk $\widetilde{\mathcal{L}}\overset{d}{=}\mathcal{L}$ such that $n>0$ is an upper block of $\mathcal{L}$ if and only if $\widetilde{\mathcal{L}}$ crosses $1/2$ in between times $n-1$ and $n$. Hence, the number of upper blocks in $(0,2n]\cap\mathbbm{Z}$ is equal to $\ell_{2n}^{\widetilde{\mathcal{L}}}$, where $\ell$ is defined in (6.2). Also, the number of zeros of $\mathcal{R}$ in $(0,2n]\cap\mathbbm{Z}$ is at least $\ell^{\mathcal{R}}_{2n}/2$ because at most two crossings of $1/2$ can correspond to the same zero of $\mathcal{R}$. Therefore, it is enough to prove that a.s. there are infinitely many integers $n>0$ such that $(\widetilde{\mathcal{L}}_{2n-1},\widetilde{\mathcal{L}}_{2n})=(1,0),\quad\mathcal{R}_{2n}=0,\quad\ell^{\mathcal{R}}_{2n}\geq 2\sqrt{n}\quad\text{ and }\quad\ell^{\widetilde{\mathcal{L}}}_{2n}<\varepsilon\sqrt{n}.$ (6.13) Indeed, the first two conditions guarantees that $2n$ is a block and so $2n=J_{2k}$ for some $k>0$, the third condition guarantees that there are at least $\sqrt{n}$ boundary points in $(0,J_{2k}]$ and so $2k\geq\sqrt{n}$, and the fourth condition guarantees that there are at most $\varepsilon\sqrt{n}$ upper blocks in $(0,J_{2k}]$. Let $A_{n}$ be the event that (6.13) holds. Using that $\mathbbm{P}[(\widetilde{\mathcal{L}}_{2n-1},\widetilde{\mathcal{L}}_{2n},\mathcal{R}_{2n})=(1,0,0)]\sim\frac{1}{2\pi n}$ for large $n$, the estimates in (6.12) implies that $\mathbbm{P}\mathopen{}\mathclose{{}\left[A_{n}}\right]\geq\frac{C_{3}}{n}$ for some constant $C_{3}>0$ only depending on $\varepsilon$. In particular, $\sum_{n=1}^{\infty}\mathbbm{P}\mathopen{}\mathclose{{}\left[A_{n}}\right]=\infty$. We now want to apply the Kochen-Stone theorem [KS64], which asserts that if $\sum_{n=1}^{\infty}\mathbbm{P}\mathopen{}\mathclose{{}\left[A_{n}}\right]=\infty$ and $\liminf_{N\to\infty}\frac{\sum_{n_{1},n_{2}=1}^{N}\mathbbm{P}[A_{n_{1}}\cap A_{n_{2}}]}{(\sum_{n=1}^{N}\mathbbm{P}[A_{n}])^{2}}<\infty,$ (6.14) then $\mathbbm{P}[A_{n}\text{ infinitely often}]>0$. For $n_{2}>n_{1}>0$, we have $\mathbbm{P}[A_{n_{1}}\cap A_{n_{2}}]\leq\mathbbm{P}[(\widetilde{\mathcal{L}}_{2n_{1}},\widetilde{\mathcal{L}}_{2n_{2}},\mathcal{R}_{2n_{1}},\mathcal{R}_{2n_{2}})=(0,0,0,0)]\leq\frac{C_{4}}{n_{1}(n_{2}-n_{1})},$ and similarly $\mathbbm{P}[A_{n}]\leq\frac{C_{4}}{n}$ for $n>0$, with some constant $C_{4}$. It follows that $\sum_{n_{1},n_{2}=1}^{N}\mathbbm{P}[A_{n_{1}}\cap A_{n_{2}}]\leq 2\sum_{1\leq n_{1}\leq n_{2}\leq N}\mathbbm{P}[A_{n_{1}}\cap A_{n_{2}}]\leq 2C_{4}\mathopen{}\mathclose{{}\left(\sum_{n_{1}=1}^{N}\frac{1}{n_{1}}+\sum_{n_{1},n_{2}=1}^{N}\frac{1}{n_{1}n_{2}}}\right)\leq C_{5}\log^{2}(N),$ for another constant $C_{5}$. Combined with our above lower bound for $\mathbbm{P}[A_{n}]$, this implies (6.14). Therefore, $\mathbbm{P}[A_{n}\text{ infinitely often}]>0$, and Kolmogorov’s zero-one law assures that $A_{n}$ happens infinitely often, almost surely. ∎ ###### Proof of Lemma 6.3. We prove the lemma via a Burton-Keane type argument. Since the UIHPMS is invariant under even translations along the boundary (1.10), we first look at the case of good boundary points of the form $J_{2k}$ for $k\in\mathbbm{Z}$. The odd case will be treated analogously at the end of the proof. For $k\in\mathbbm{Z}$, let $E_{k}$ be the event that $J_{2k}$ is a good boundary point, i.e., $J_{2k}=H_{m}$ for some $m\in\mathbbm{Z}$, and the boundary path $P_{m}$ is semi-infinite, that is $P_{m}$ does not hit another good boundary point other than $H_{m}$. By the translation invariance of the UIHPMS (1.10) and the fact that the set of $k\in\mathbbm{Z}$ such that $J_{2k}$ is good is independent from the UIHPMS (see the discussion just below (6.7)), $p:=\mathbbm{P}[E_{k}]$ does not depend on $k$. We need to show that $p=0$. Assume for contradiction that $p>0$. By the Birkhoff ergodic theorem, $\lim_{k\to\infty}\frac{1}{k}\\#\mathopen{}\mathclose{{}\left\\{k^{\prime}\in[1,k]\cap\mathbbm{Z}:E_{k^{\prime}}\text{ occurs}}\right\\}=p.$ (6.15) By Lemma 6.5 with $\varepsilon=p/4$ and (6.15), almost surely, there exist arbitrarily large values of $k\in\mathbbm{Z}_{>0}$ such that: 1. (i) $J_{2k}$ is a block; 2. (ii) there are at most $pk/4$ upper blocks in $(0,J_{2k})\cap\mathbbm{Z}$; 3. (iii) $\\#\\{k^{\prime}\in[1,k]\cap\mathbbm{Z}:E_{k^{\prime}}\text{ occurs}\\}>pk/2$. Fix $k\in\mathbbm{Z}_{>0}$ with the above three properties. For each $k^{\prime}\in(0,k)\cap\mathbbm{Z}$ such that $E_{k^{\prime}}$ occurs, let $m(k^{\prime})\in\mathbbm{Z}$ such that $H_{m(k^{\prime})}=J_{2k^{\prime}}$. Each semi-infinite path $P_{m(k^{\prime})}$ for $k^{\prime}\in(0,k)\cap\mathbbm{Z}$ such that $E_{k^{\prime}}$ occurs does not intersect any other such semi-infinite path. Since $J_{2k}$ is a block, there is no arc connecting $(0,J_{2k})\cap\mathbbm{Z}$ and $(J_{2k},\infty)\cap\mathbbm{Z}$. Therefore, each semi-infinite path $P_{m(k^{\prime})}$ for $k^{\prime}\in(0,k)\cap\mathbbm{Z}$ must exit the interval $(0,J_{2k})\cap\mathbbm{Z}$ via some arc joining $(0,J_{2k})\cap\mathbbm{Z}$ to $(-\infty,0]\cap\mathbbm{Z}$. This arc must be above the real line, since by the construction of the UIHPMS there are no arcs below the real line that cross the ray $\\{1/2\\}\times(-\infty,0]$. In other words, the right endpoint of such an arc is an upper block. It follows that the number of upper blocks in $(0,J_{2k})\cap\mathbbm{Z}$ is at least the number of semi-infinite paths started from points in $(0,J_{2k})\cap\mathbbm{Z}$, which is at least $pk/2$. But, we have chosen $k$ so that the number of upper blocks in $(0,J_{2k})\cap\mathbbm{Z}$ is at most $pk/4$. This yields a contradiction, and thus $p=0$. Exactly the same argument shows that a.s. the event $E_{k}^{\prime}$ that $J_{2k+1}=H_{m}$ for some $m\in\mathbbm{Z}$ and $P_{m}$ is semi-infinite does not occur for any $k\in\mathbbm{Z}$. This finishes the proof. ∎ ### 6.4 Infinitely many paths separating the origin from $\infty$ Lemma 6.3 implies that a.s. there are no semi-infinite paths in the UIHPMS which start at a good boundary point and never hit another good boundary point. The following lemma will allow us in Section 6.5 to rule out more general types of infinite paths. ###### Lemma 6.6. Almost surely, there exist infinitely many $m\in\mathbbm{Z}_{>0}$ such that the boundary path $P_{m}$ ends at $H_{m^{\prime}}$ for some $m^{\prime}\in\mathbbm{Z}_{\leq 0}$. We will deduce Lemma 6.6 from a general result (Lemma 6.9) for random perfect matchings. ###### Definition 6.7. A perfect matching of $\mathbbm{Z}$ is a function $\phi:\mathbbm{Z}\rightarrow\mathbbm{Z}$ such that $\phi(k)\not=k$ and $\phi(\phi(k))=k$ for each $k\in\mathbbm{Z}$. We say that $j,k\in\mathbbm{Z}$ are matched if $\phi(j)=k$, equivalently, $\phi(k)=j$. We say that $\phi$ is non-crossing if there do not exist integers $j_{1},j_{2}$ such that $j_{1}<j_{2}<\phi(j_{1})<\phi(j_{2})$. ###### Definition 6.8. A random perfect matching $\phi$ is stationary if $\phi(\cdot+m)\overset{d}{=}\phi(\cdot)$ for each $m\in\mathbbm{Z}$. A stationary perfect matching is strongly ergodic if any event which is invariant under shifts of the form $\phi(\cdot)\mapsto\phi(\cdot+2m)$ for all $m\in\mathbbm{Z}$ has probability zero or one. ###### Lemma 6.9. Let $\phi$ be a strongly ergodic random perfect matching of $\mathbbm{Z}$. Almost surely, for each $j\in\mathbbm{Z}$ there exists $k\geq j$ such that $\phi(k)\leq j-1$. If $\phi$ is non-crossing, then a.s. for each $j\in\mathbbm{Z}$ there are infinitely many integers $k\geq j$ such that $\phi(k)\leq j-1$. ###### Proof of Lemma 6.9. We say that $j\in\mathbbm{Z}$ is exposed if there does not exist $k\geq j$ such that $\phi(k)\leq j-1$. If we represent the matching by an arc diagram (with each arc matching $x$ and $\phi(x)$ plotted in the region $[x\wedge\phi(x),x\vee\phi(x)]\times[0,\infty)$) then $j$ being exposed is equivalent to the condition that there are no arcs crossing $\\{j-1/2\\}\times[0,\infty)$. We claim that a.s. there are no exposed integers. By the stationarity of $\phi$, the probability that $j\in\mathbbm{Z}$ is exposed does not depend on $j$. By the Birkhoff ergodic theorem, if any integer has a positive probability to be exposed, then a.s. there are infinitely many exposed integers. Hence we just need to show that the probability that there are infinitely many exposed integers is zero. Suppose that $j_{1}\leq j_{2}$ are both exposed. By the definition of exposed, every integer in the set $\\{j_{1},\dots,j_{2}-1\\}$ must have its match in $\\{j_{1},\dots,j_{2}-1\\}$. Hence $\\#\\{j_{1},\dots,j_{2}-1\\}$ must be even, so $j_{1}$ and $j_{2}$ must have the same parity. Consequently, a.s. either every exposed integer is even or every exposed integer is odd. By stationarity, the probabilities of the events $\\{\text{$\exists$ infinitely many even exposed integers}\\}\quad\text{and}\quad\\{\text{$\exists$ infinitely many odd exposed integers}\\}$ (6.16) are the same. The two events in (6.16) are invariant under translations of the form $\phi\mapsto\phi(\cdot+2m)$ for $m\in\mathbbm{Z}$, so by strong ergodicity each of these events has probability zero or one. Since we have just seen that these two events are disjoint, they must each have probability zero. Hence a.s. there are no exposed integers. Now assume that $\phi$ is non-crossing and consider a $j\in\mathbbm{Z}$ with the property that there are only finitely many integers $k\geq j$ such that $\phi(k)\leq j-1$. We will show that there exists an exposed integer. This, combined with our earlier result, will then imply that no such $j$ exists. To construct an exposed integer, let $k_{*}$ be the largest $k\geq j$ such that $\phi(k)\leq j-1$. We claim that $k_{*}+1$ is exposed. Indeed, suppose for contradiction that there exists $k\geq k_{*}+1$ with $\phi(k)\leq k_{*}$. By the maximality of $k_{*}$, we must have $\phi(k)\geq j$. Since $\phi(k_{*})\leq j-1$, we have $\phi(k)\not=k_{*}$ and so $\phi(k)\leq k_{*}-1$. Therefore, $\phi(k_{*})\leq j-1<j\leq\phi(k)<k_{*}<k,$ which contradicts the non-crossing condition. ∎ ###### Proof of Lemma 6.6. We divide the proof in three main steps. Step 1: Reducing to the setting of Lemma 6.1. Let $(\mathcal{M}^{\operatorname{cut}},\Gamma^{\operatorname{cut}})$ be the variant of the UIHPMS where we remove the lower arcs joining the good boundary points $\\{H_{m}\\}_{m\in\mathbbm{Z}}$, as in Section 6.2. By Definition 6.2, the boundary path $P_{m}$ for $m\in\mathbbm{Z}$ does not traverse any lower arcs in the UIHPMS joining pairs of good boundary points. Hence, the definition of $P_{m}$ and the statement of the lemma are unaffected if we replace the UIHPMS by $(\mathcal{M}^{\operatorname{cut}},\Gamma^{\operatorname{cut}})$. Let $(\widehat{\mathcal{M}}^{\operatorname{cut}},\widehat{\Gamma}^{\operatorname{cut}})$ be sampled from the law of $(\mathcal{M}^{\operatorname{cut}},\Gamma^{\operatorname{cut}})$ conditioned on the event that $H_{0}=0$, as in Lemma 6.1. Define the good boundary points $\\{\widehat{H}_{m}\\}_{m\in\mathbbm{Z}}$ and the boundary paths $\\{\widehat{P}_{m}\\}_{m\in\mathbbm{Z}}$ with $(\widehat{\mathcal{M}}^{\operatorname{cut}},\widehat{\Gamma}^{\operatorname{cut}})$ in place of the original UIHPMS. We claim that it suffices to show that a.s. 1. $(\boxdot)$ $\forall m_{0}\in\mathbbm{Z}$, $\exists$ infinitely many $m>m_{0}$ s.t. the path $\widehat{P}_{m}$ in $\widehat{\mathcal{M}}^{\operatorname{cut}}$ ends at $\widehat{H}_{m^{\prime}}$ for some $m^{\prime}\leq m_{0}$. Indeed, the law of $(\widehat{\mathcal{M}}^{\operatorname{cut}},\widehat{\Gamma}^{\operatorname{cut}})$ is absolutely continuous with respect to the law of $({\mathcal{M}}^{\operatorname{cut}},\Gamma^{\operatorname{cut}})$. So ($\boxdot$) implies that with positive probability, the event ($\boxdot$) holds with $\mathcal{M}^{\operatorname{cut}}$, $P_{m}$ and $H_{m^{\prime}}$ in place of $\widehat{\mathcal{M}}^{\operatorname{cut}}$, $\widehat{P}_{m}$ and $\widehat{H}_{m^{\prime}}$. This event depends on the UIHPMS in a manner which is invariant by translations of the form (1.10). By the zero-one law for translation invariant events, we get that this event in fact has probability one. Step 2: Constructing a random perfect matching. Recall that $\widehat{P}_{m}$ is defined as in Definition 6.2 with $(\widehat{\mathcal{M}}^{\operatorname{cut}},\widehat{\Gamma}^{\operatorname{cut}})$ in place of the UIHPMS. Define $\phi:\mathbbm{Z}\to\mathbbm{Z}$ by the condition that $\phi(m)=m^{\prime}$ if and only if $\widehat{P}_{m}$ ends at $\widehat{H}_{m^{\prime}}$. By Lemma 6.3 and the absolutely continuity of $(\widehat{\mathcal{M}}^{\operatorname{cut}},\widehat{\Gamma}^{\operatorname{cut}})$ with respect to $({\mathcal{M}}^{\operatorname{cut}},\Gamma^{\operatorname{cut}})$, a.s. each $\widehat{P}_{m}$ is a finite path ending at a good boundary point. Hence a.s. $\phi$ is well-defined. We will prove ($\boxdot$) by applying Lemma 6.9 to $\phi$. We first check that $\phi$ is a perfect matching. By the definition of $(\widehat{\mathcal{M}}^{\operatorname{cut}},\widehat{\Gamma}^{\operatorname{cut}})$, none of the good boundary points $\widehat{H}_{m}$ for $m\in\mathbbm{Z}$ is incident to a lower arc of $\widehat{\mathcal{M}}^{\operatorname{cut}}$. Hence, $\widehat{P}_{m}$ traverses an upper arc immediately before hitting $\widehat{H}_{\phi(m)}$. Therefore, $\widehat{P}_{\phi(m)}$ is the time reversal of $\widehat{P}_{m}$ and so $\phi(\phi(m))=m$ and $\phi(m)\neq m$. Thus $\phi$ is a perfect matching of $\mathbbm{Z}$. Figure 16: The sets $A$ and $U$ used in Step 3 of the proof of Lemma 6.6. Good boundary points are shown in green, and arcs not traversed by $\widehat{P}_{m}\subset A$ are not shown. The key property of $A$ is that it is not crossed by any arc of $\widehat{\mathcal{M}}^{\operatorname{cut}}$. Step 3: Non-crossing, stationarity, and ergodicity. We next show that $\phi$ is non-crossing. Fix $m\in\mathbbm{Z}$. We are going to show that each point in $(m\wedge\phi(m),m\vee\phi(m))\cap\mathbbm{Z}$ is matched with a point in $(m\wedge\phi(m),m\vee\phi(m))\cap\mathbbm{Z}$. See Figure 16 for an illustration. Let $A:=\widehat{P}_{m}\cup\mathopen{}\mathclose{{}\left(\\{\widehat{H}_{m}\\}\times(-\infty,0]}\right)\cup\mathopen{}\mathclose{{}\left(\\{\widehat{H}_{\phi(m)}\\}\times(-\infty,0]}\right).$ (6.17) By the definition of boundary points, for each $m^{\prime}\in\mathbbm{Z}$ no arc of $\widehat{\mathcal{M}}^{\operatorname{cut}}$ can cross the infinite ray $\\{\widehat{H}_{m^{\prime}}\\}\times(-\infty,0]$. This implies in particular that $A$ is the trace of a bi-infinite simple path in $\mathbbm{C}$ from $\infty$ to $\infty$. By the Jordan curve theorem $A$ separates $\mathbbm{C}$ into exactly two open connected components. Let $U$ be the open connected component which contains an unbounded subset of the semi-infinite box $(\widehat{H}_{\phi(m)}\wedge\widehat{H}_{m},\widehat{H}_{\phi(m)}\vee\widehat{H}_{m})\times(-\infty,0)$. If $m^{\prime}\in\mathbbm{Z}\setminus\\{m,\phi(m)\\}$, then the infinite ray $\\{\widehat{H}_{m^{\prime}}\\}\times(-\infty,0]$ intersects $U$ (resp. $\mathbbm{C}\setminus\overline{U}$) provided $m^{\prime}\in(m\wedge\phi(m),m\vee\phi(m))\cap\mathbbm{Z}$ (resp. $m^{\prime}\in\mathbbm{Z}\setminus[m\wedge\phi(m),m\vee\phi(m)]$). This ray cannot cross $A$ (by definition of boundary points), so must be entirely contained in either $U$ or $\mathbbm{C}\setminus\overline{U}$. Hence $U\cap\\{\widehat{H}_{m^{\prime}}\\}_{m^{\prime}\in\mathbbm{Z}}=\\{\widehat{H}_{m^{\prime}}\\}_{m^{\prime}\in(m\wedge\phi(m),m\vee\phi(m))\cap\mathbbm{Z}}.$ Since the paths $\widehat{P}_{m^{\prime}}$ for $m^{\prime}\in\mathbbm{Z}\setminus\\{m,\phi(m)\\}$ cannot cross $A$, this implies the desired property. By Lemma 6.1, $\phi$ is stationary and strongly ergodic in the sense of Definition 6.8. Therefore, we can apply Lemma 6.9 to get that a.s. ($\boxdot$) holds. ∎ ### 6.5 Proofs of Theorem 1.12 and Proposition 1.14 We are now ready to prove our main theorem (Theorem 1.12) for the UIHPMS. The key idea is that the paths $P_{m}$ as in Lemma 6.6 act as “shields” which an infinite path in the UIHPMS cannot cross. ###### Proof of Theorem 1.12. Fix $n\in\mathbbm{N}$. We will show that a.s. there is no bi-infinite path of arcs in the UIHPMS which intersects $[-n,n]$. Sending $n\to\infty$ shows that a.s. there is no bi-infinite path of arcs in the UIHPMS. See Figure 17 for an illustration. Figure 17: Illustration of the proof of Theorem 1.12. The numbers $m\in\mathbbm{Z}_{>0}$ and $m^{\prime}\in\mathbbm{Z}_{<0}$ are chosen so that the boundary path $P_{m}$ ends at $H_{m^{\prime}}$ and is disjoint from $[-n,n]\times(-\infty,0]$. This implies that any bi-infinite path of arcs which hits a point of $[-n,n]\cap\mathbbm{Z}$ must enter the domain $U$. This is possible only through the lower arcs $l_{m^{\prime}}$ and $l_{m}$ of the UIHPMS incident to $H_{m^{\prime}}$ and $H_{m}$. We show that, regardless of the configuration of these arcs, the bi-infinite path cannot enter $U$ (the two possible configurations are shown in the upper and lower panels of the figure) concluding that there is no bi-infinite path. By, e.g., the definition of the UIHPMS by cutting, there are only finitely many arcs of the UIHPMS which intersect the semi-infinite box $[-n,n]\times(-\infty,0]$ (recall that we are assuming that each arc joining $x<y$ is draw in such a way that it is contained in $[x,y]\times\mathbbm{R}$). If $m,m^{\prime}\in\mathbbm{Z}$ are chosen so that $P_{m}$ does not end at $H_{m^{\prime}}$, then the paths $P_{m}$ and $P_{m^{\prime}}$ are disjoint. Therefore, there are only finitely many $m\in\mathbbm{Z}$ such that $P_{m}$ traverses an arc of the UIHPMS which intersects $[-n,n]\times(-\infty,0]$. From this and Lemma 6.6, we get that a.s. there exists a large enough $m\in\mathbbm{Z}_{>0}$ such that $P_{m}$ ends at $H_{m^{\prime}}$ for some $m^{\prime}\in\mathbbm{Z}_{<0}$ and $P_{m}$ is disjoint from $[-n,n]\times(-\infty,0]$. Define $A$ and $U$ exactly as in the proof of Lemma 6.6 with $P_{m}$ instead of $\widehat{P}_{m}$, see the paragraph including (6.17). The condition that $P_{m}$ is disjoint from $[-n,n]\times(-\infty,0]$ implies that $[-n,n]\subset U$. There are only finitely many arcs of the UIHPMS which intersect $U$, so a bi-infinite path of arcs $\gamma$, if it exists, can spend only a finite amount of time in $\overline{U}$. Since $\gamma$ is bi-infinite, it follows that if $\gamma$ hits a point of $[-n,n]\cap\mathbbm{Z}\subset U$, then there are at least two times when $\gamma$ traverses an arc which joins a point of $\overline{U}\cap\mathbbm{Z}$ and a point of $\mathbbm{Z}\setminus\overline{U}$. By the definition of $A$ and the fact that no lower arcs of the UIHPMS cross the infinite rays $\\{H_{m}\\}\times(-\infty,0]$ and $\\{H_{m^{\prime}}\\}\times(-\infty,0]$, the only arcs of the UIHPMS which intersect $A$ are the arcs traversed by $P_{m}$ and the lower arcs incident to $H_{m^{\prime}}$ and $H_{m}$. Call these two lower arcs $l_{m^{\prime}}$ and $l_{m}$, respectively. Since $P_{m}\subset A=\partial U$, these two lower arcs $l_{m^{\prime}}$ and $l_{m}$ are the only arcs of the UIHPMS which can possibly join a point of $\overline{U}\cap\mathbbm{Z}$ and a point of $\mathbbm{Z}\setminus\overline{U}$. Hence, if $\gamma$ intersects $[-n,n]\cap\mathbbm{Z}$ then each of $l_{m^{\prime}}$ and $l_{m}$ joins $\overline{U}\cap\mathbbm{Z}$ and a point of $\mathbbm{Z}\setminus\overline{U}$, i.e., $l_{m}$ joins $H_{m}$ to a point of $\mathbbm{Z}\setminus\overline{U}$ and similarly for $l_{m^{\prime}}$ (see the first panel in Figure 17). But, $l_{m^{\prime}}$ and $l_{m}$ are also the only two arcs of the UIHPMS which can possibly join a point of $U\cap\mathbbm{Z}$ to a point of $\mathbbm{Z}\setminus U$ (see the second panel in Figure 17). But we have just seen that if $\gamma$ intersects $[-n,n]\cap\mathbbm{Z}$, then neither $l_{m^{\prime}}$ nor $l_{m}$ has an endpoint in $U\cap\mathbbm{Z}$. Hence, in this case there are no arcs joining a point of $U\cap\mathbbm{Z}$ to a point of $\mathbbm{Z}\setminus U$, so a.s. there is no bi-infinite path of arcs in the UIHPMS which intersects $[-n,n]\cap\mathbbm{Z}\subset U$. ∎ Using the same idea, we prove that there is a unique infinite path in the PIHPMS. ###### Proof of Proposition 1.14. By the definition of the PIHPMS given in (1.11), we can couple the PIHPMS with the UIHPMS in such a way that the upper arcs for both infinite meandric systems are the same, the lower arcs joining non-boundary points for both meandric systems are the same, and the sequence of boundary points $\\{J_{k}\\}_{k\in\mathbbm{Z}}$ for both meandric systems is the same. Hence we can identify the good boundary points $\\{H_{m}\\}_{m\in\mathbbm{Z}}$ with boundary points of the PIHPMS. This implies that the boundary paths $P_{m}$ for $m\in\mathbbm{Z}$ (Definition 6.2) are unaffected if we replace the UIHPMS by the PIHPMS. Recall that the path of arcs $\gamma^{\circ}$ started from $0$ in the PIHPMS is always semi-infinite because $\gamma^{\circ}$ cannot finish forming a loop as no arc in the lower half plane is incident to $0$ by construction (note that $\gamma^{\circ}$ is different from $P_{0}$ since it does not stop when it hits a boundary point). By Lemma 6.6, a.s. there exist infinitely many $m\in\mathbbm{Z}_{>0}$ such that $P_{m}$ ends at $H_{m^{\prime}}$ for some $m^{\prime}\in\mathbbm{Z}_{\leq 0}$. Since $\gamma^{\circ}$ is semi-infinite, the same argument as in the proof of Theorem 1.12 shows that $\gamma^{\circ}$ must hit either $H_{m}$ or $H_{m^{\prime}}$ for each such $m$, and hence $\gamma^{\circ}$ must traverse $P_{m}$ or $P_{m^{\prime}}$ (the time reversal of $P_{m}$) for each such $m$. This shows that $\gamma^{\circ}$ hits infinitely many points in each of $\\{H_{m}\\}_{m<0}\subset\\{J_{k}\\}_{k<0}$ and $\\{H_{m}\\}_{m>0}\subset\\{J_{k}\\}_{k>0}$, almost surely. The same argument as in the proof of Theorem 1.12 also shows that any other infinite path of arcs in the PIHPMS must also traverse infinitely many of the paths $P_{m}$ for $m\in\mathbbm{Z}_{>0}$ as above. Therefore, any other infinite path must share a portion of arcs with $\gamma^{\circ}$. But this is possible only if such infinite path is a subpath of $\gamma^{\circ}$. So $\gamma^{\circ}$ is the unique maximal infinite path of arcs in the PIHMPS, i.e, the unique infinite path in $\Gamma^{\circ}$. ∎ ## 7 Justification for Conjectures 1.2 and 1.3 ### 7.1 SLE8 on $\sqrt{2}$-LQG via mating of trees Let $\mathfrak{S}_{n}$ be a uniform meandric system of size $n$ and let $(\mathcal{M}_{n},P_{n},\Gamma_{n})$ be the associated planar map decorated by a Hamiltonian path (corresponding to the real line) and a collection of loops, as in the discussion at the beginning of Section 1.2. We have already seen in Sections 4 and 5 that the infinite-volume analog $(\mathcal{M},P)$ of $(\mathcal{M}_{n},P_{n})$ is closely connected to $\sqrt{2}$-LQG decorated by SLE8. More precisely, due to the encoding of the UIMS by random walks (Section 1.4) and the convergence of random walk to Brownian motion, $(\mathcal{M},P)$ is connected to the mated-CRT map $\mathcal{G}$, equipped with the left-right ordering of its vertices (Section 4). Furthermore, the mated-CRT map $\mathcal{G}$ is closely connected to $\sqrt{2}$-LQG decorated by SLE8 due to the results of [DMS21] (Theorem 5.1). This strongly suggests that the scaling limit of $(\mathcal{M}_{n},P_{n})$ should be given by $\sqrt{2}$-LQG decorated by SLE8. In fact, the random walk excursions which encode the upper and lower arc diagrams of $(\mathcal{M}_{n},P_{n})$ converge under appropriate scaling to two independent Brownian excursions. This can already be interpreted as a convergence statement for $(\mathcal{M}_{n},P_{n})$ toward $\sqrt{2}$-LQG decorated by SLE8 in the so-called peanosphere sense. See [GHS23] for a survey of this type of convergence and the results that can be proven using it. To justify Conjecture 1.2, we still need to explain why $\Gamma_{n}$ should converge to CLE6. To do this, we will give a physics argument which also provides an alternative heuristic for the scaling limit of $(\mathcal{M}_{n},P_{n})$. ### 7.2 Physics argument We will now explain why Conjecture 1.2 can be viewed as a special case of Conjecture [BGS22, Conjecture 6.2], which in turn is derived from physics heuristics in [DFGG00], building on [JK98, JK99, DCN04, JZJ04]. Let $n,m>0$ and let $G$ be a finite 4-regular graph. The fully packed $O(n\times m)$ loop model on $G$ is the probability measure on pairs of collections of loops (simple cycles) $\Gamma_{1},\Gamma_{2}$ in $G$ with the following properties: * • Each vertex of $G$ is visited by exactly one loop in each of $\Gamma_{1}$ and $\Gamma_{2}$. * • Each edge of $G$ is visited by either exactly one loop in $\Gamma_{1}$ or exactly one loop in $\Gamma_{2}$ (but not both). The probability of each possible configuration $(\Gamma_{1},\Gamma_{2})$ is proportional to $n^{\\#\Gamma_{1}}m^{\\#\Gamma_{2}}$. Suppose now that $G$ is a planar map, instead of just a graph. This in particular means that we have a canonical cyclic ordering of the four edges incident to each vertex of $G$. The crossing fully packed $O(n\times m)$ loop model on $G$ is the variant of the $O(n\times m)$ loop model where we impose the additional constraint that a loop of $\Gamma_{1}$ and a loop of $\Gamma_{2}$ cross at each vertex of $G$. Equivalently, the edges incident to each vertex, in cyclic order, must alternate between edges of $\Gamma_{1}$ and edges of $\Gamma_{2}$. One can also define the (crossing) fully packed $O(n\times m)$ loop model when one or both of $n$ or $m$ is equal to zero. In the case when, e.g., $n=0$, one considers pairs $(\Gamma_{1},\Gamma_{2})$ as above subject to the constraint that $\Gamma_{1}$ consists of a single loop (necessarily a Hamiltonian cycle), and the probability of each configuration is proportional to $m^{\\#\Gamma_{2}}$. The relevance of the (crossing) fully packed $O(n\times m)$ loop model to the present paper is that meandric systems are equivalent to the crossing fully packed $O(0\times 1)$ loop model on a planar map, as we now explain. Let $\mathfrak{S}_{n}$ be a meandric system of size $n$ and consider the corresponding decorated planar map $(\mathcal{M}_{n},P_{n},\Gamma_{n})$. Then $\mathcal{M}_{n}$ is a 4-regular planar map with $2n$ vertices, $P_{n}$ is a Hamiltonian cycle on $\mathcal{M}_{n}$, $\Gamma_{n}$ is a collection of loops on $\mathcal{M}_{n}$ such that each vertex is visited by exactly one loop, and each edge is visited by either the Hamiltonian cycle or one of the loops (but not both). Moreover, at each vertex of $\mathcal{M}_{n}$, the Hamiltonian cycle $P_{n}$ is crossed by a loop of $\Gamma_{n}$ (this corresponds to the condition that loops of $\mathfrak{S}_{n}$ do not touch $\mathbbm{R}$ without crossing it in Definition 1.1). Conversely, every 4-regular planar map decorated by a Hamiltonian path and a collection of loops which satisfy the above conditions gives rise to a meandric system: just choose a planar embedding of the map into $\mathbbm{R}^{2}\cup\\{\infty\\}$ under which the Hamiltonian cycle is mapped to the real line. Hence, the set of possibilities for $(\mathcal{M}_{n},P_{n},\Gamma_{n})$ is exactly the same as the set of possibilities for a 4-regular planar map with $2n$ edges decorated by a realization of the crossing fully packed $O(0\times 1)$ loop model. In particular, if $\mathfrak{S}_{n}$ is sampled uniformly, then $\mathcal{M}_{n}$ is a sample from the set of 4-regular planar maps with $2n$ vertices, weighted by the number of possible crossing fully packed $O(0\times 1)$ loop model configurations on the map. Moreover, the conditional law of $(P_{n},\Gamma_{n})$ given $\mathcal{M}_{n}$ is that of the fully packed $O(0\times 1)$ loop model on $\mathcal{M}_{n}$. Jacobsen and Kondev [JK98, JK99] gave a prediction for the scaling limit of the fully packed $O(n\times m)$ loop model on $\mathbbm{Z}^{2}$ in terms of a conformal field theory whose central charge is an explicit function of $n$ and $m$, which was later verified (at a physics level of rigor) in [DCN04, JZJ04]. Building on this, Di Franceso, Golinelli, and Guitter [DFGG00, Section 3] gave a prediction for the scaling limit of the crossing fully packed $O(n\times m)$ loop model on a random planar map. Section 6 of [BGS22] reviews the arguments leading to these predictions and translates the predictions into the language of SLE and LQG. In particular, from [BGS22, Conjecture 6.2] for $n=0$ and $m=1$, we get that the scaling limit of random planar maps decorated by the crossing fully packed $O(0\times 1)$ loop model should be given by $\sqrt{2}$-LQG decorated by SLE8 and CLE6. Since this decorated random planar map model is equivalent to a uniform meandric system (as discussed just above), this leads to Conjecture 1.2. ### 7.3 Predictions for exponents via KPZ We now explain how Conjecture 1.2 leads to predictions for various exponents associated with meandric systems (including Conjecture 1.3). In this and the next section, we assume that the reader has some familiarity with SLE and LQG. Consider a $\gamma$-LQG sphere, represented by a metric and a measure on $\mathbbm{C}$. If $X\subset\mathbbm{C}$ is a random set sampled independently from this metric and measure, then we can define the Hausdorff dimensions $\Delta_{0}$ and $\Delta_{\gamma}$ of $X$ with respect to the Euclidean and $\gamma$-LQG metrics, respectively. We re-scale $\Delta_{\gamma}$ by the reciprocal of the dimension of the whole space, so that $\Delta_{\gamma}\in[0,1]$. The Knizhnik-Polyakov-Zamolodchikov (KPZ) formula [KPZ88] states that a.s. $\Delta_{0}=\mathopen{}\mathclose{{}\left(2+\frac{\gamma^{2}}{2}}\right)\Delta_{\gamma}-\frac{\gamma^{2}}{2}\Delta_{\gamma}^{2}.$ (7.1) See [BS09, DS11, RV11, BJRV13, DRSV14, Aru15, BGRV16, GHS19, GHM20, DMS21, GM21a, GP22] for various rigorous versions of the KPZ formula. Now consider $(\mathcal{M}_{n},P_{n},\Gamma_{n})$ as above and for $k\in\mathbbm{N}$ let $\ell_{n}^{k}$ be the $k$th largest loop in $\Gamma_{n}$, i.e., the one with the $k$th largest number of vertices. Conjecture 1.2 tells us that if $k$ is fixed, then as $n\rightarrow\infty$, the loop $\ell_{n}^{k}$ should converge to the $k$th largest loop in a CLE6 on an independent $\sqrt{2}$-LQG sphere, i.e., the one with the $k$th largest $\sqrt{2}$-LQG length. The Euclidean dimension of the $k$th largest loop in a CLE6 is $\Delta_{0}=7/4$, the same as the dimension of an SLE6 curve [Bef08]. By (7.1), the (re-scaled) dimension of the $k$th largest loop in a CLE6 with respect to the $\sqrt{2}$-LQG metric is $\alpha:=\Delta_{\sqrt{2}}=\frac{1}{2}\mathopen{}\mathclose{{}\left(3-\sqrt{2}}\right)\approx 0.7929.$ (7.2) This means that the number of $\sqrt{2}$-LQG balls of $\sqrt{2}$-LQG mass $\varepsilon$ needed to cover the $k$th largest loop in a CLE6 is about $\varepsilon^{-\alpha}$. The analog of this in the discrete setting says that $\mathbbm{E}\mathopen{}\mathclose{{}\left[\\#\\{\text{vertices in $\ell_{n}^{k}$}\\}}\right]\approx n^{\alpha},\quad\text{with $\alpha$ as in~{}\eqref{eqn-loop-exponent-kpz}}.$ (7.3) This gives Conjecture 1.3. Using similar techniques to the ones above, we can also derive predictions for other exponents associated with uniform meandric systems. For example, let $\ell_{n}^{1}$ be the largest loop in $\Gamma_{n}$ and let $\operatorname{Cross}_{n}$ be the number of times that $\ell_{n}^{1}$ crosses the vertical line $\\{n\\}\times\mathbbm{R}$, i.e., the number of arcs of $\ell_{n}^{1}$ which disconnect $n$ from $\infty$ on either of the two sides of the horizontal line. To estimate the growth rate of $\operatorname{Cross}_{n}$ as $n\rightarrow\infty$, we look at the continuum analog of the set of crossing points. Consider an SLE8 curve $\eta$ from $\infty$ to $\infty$ and an independent whole-plane CLE6. Assume that $\eta$ is parametrized by $\sqrt{2}$-LQG area with respect to an independent unit area quantum sphere, so that $\eta:[0,1]\rightarrow\mathbbm{R}$. According to Conjecture 1.2, the continuum analog of the set of crossing points in the definition of $\operatorname{Cross}_{n}$ is the intersection of $\eta([0,1/2])\cap\eta([1/2,1])$ with the largest (in the sense of $\sqrt{2}$-LQG length) loop in the CLE6. The set $\eta([0,1/2])\cap\eta([1/2,1])$ is the union of two SLE2-type curves [DMS21, Footnote 4]. Therefore, the Euclidean Hausdorff dimension of the intersection of this set with the largest loop in the CLE6 should be equal to the Euclidean Hausdorff dimension of an SLE2 curve intersected with an independent SLE6 curve. Using [Bef08], on the event that the intersection is non-empty, its Euclidean Hausdorff dimension should be $\Delta_{0}=2-(2-5/4)-(2-7/4)=1.$ (7.4) Plugging this into (7.1) (with $\gamma=\sqrt{2}$) as above, we arrive at the prediction $\mathbbm{E}\mathopen{}\mathclose{{}\left[\operatorname{Cross}_{n}}\right]\approx n^{\nu},\quad\text{where}\quad\nu=\frac{1}{2}(3-\sqrt{5})\approx 0.3820.$ (7.5) ### 7.4 Simulations We give details on our numerical simulations in this section. A sample of a (finite) uniform meandric system of size $n$ can be obtained from a sample of two independent simple random walks of size $2n$ started from $0$, conditioned to stay non-negative and to end at $0$ using (1.1). These conditioned walks $\mathcal{X}_{i}$ can be sampled quickly from the conditional probability $\mathbbm{P}\mathopen{}\mathclose{{}\left[\mathcal{X}_{i}-\mathcal{X}_{i-1}=1\,|\,\mathcal{X}_{1},\dots,\mathcal{X}_{i-1}}\right]$, which boils down to counting the number of simple walk bridges conditioned to be non-negative. Those numbers can be explicitly computed by the reflection property of simple random walks in any case. The corresponding meandric system of size $n$ is then constructed by (1.1) as a combinatorial graph and analyzed by JuliaGraphs[FBS+21]. Our first simulation computes the median sizes (numbers of vertices) of the 5 largest loops from 100 uniformly sampled meandric systems of size $n$, for increasing values of $n$. The result in Figure 18 (Left) suggests that the constant $\alpha\approx 0.7929$ in Conjecture 1.3 is correct. This is distinguishable from the value $0.8$ guessed in [Kar20] because the 95% confidence interval for the largest loop exponent is $0.7929\pm 0.0022$. The size of $k$th largest loop can be written as $C_{k}n^{\alpha}$ where $C_{k}$ is some random variable. In Figure 18 (Left), the $y$-intercept is related to $\log(C_{k})$ for each $k$. One implication of Conjecture 1.2 is that the law of $C_{k+1}/C_{k}$ is tight for each $k$. Our second simulation computes the medians of $\operatorname{Cross}_{n}$ (defined in Section 7.3) from 100 uniformly sampled meandric systems of size $n$, for increasing values of $n$. The result in Figure 18 (Right) suggests that the KPZ prediction (7.5) is correct. This is strong evidence that our conjectured limiting objects SLE8 and CLE6 are independent, which is not trivial from the definition of meandric systems. See the discussion after Conjecture 1.2. Figure 18: Left: The plot of $\log(|\ell^{k}_{n}|)$ versus $\log(2n)$, where $|\ell^{k}_{n}|$ is the number of vertices of the $k$th largest loop of $\mathfrak{S}_{n}$. The 95% confidence intervals for the slopes are $0.7929\pm 0.0022,0.7932\pm 0.0021,0.7939\pm 0.0021,0.7938\pm 0.0020,0.7948\pm 0.0019$, respectively, which all include our conjectured value $(3-\sqrt{2})/2\approx 0.7929.$ Right: The plot of $\log(|\operatorname{Cross}_{n}|)$ versus $\log(2n)$, where $\operatorname{Cross}_{n}$ is defined in Section 7.3. The 95% confidence interval for the slope is $0.3815\pm 0.0015$, which includes our conjectured value $\frac{1}{2}(3-\sqrt{5})\approx 0.3820$. We now give the details of the simulations in Figure 2. There, we also visualize how the arcs of arc diagrams (which are usually embedded as smooth circular arcs in $\mathbbm{R}^{2}$ throughout the paper) associated with a uniformly sampled meandric system may look like fractal loops of a CLE in a proper embedding. For this purpose, we consider a uniform meandric system _with boundary_ (as defined below; its infinite-volume version is defined in Section 1.5) and embed the underlying planar map via the Tutte embedding [GMS21], a.k.a. harmonic embedding, so that it conjecturally approximates the $\sqrt{2}$-LQG wedge on the disk. We use a finite meandric system with boundary because Tutte embeddings are easier to define for planar maps with boundary. The uniform meandric system with boundary can be constructed as in (1.1) from two independent conditioned simple random walks with $2n$ steps started from $0$, both conditioned to stay non-negative, but one ends at $0$ as before while the other ends at $2\lfloor n^{1/2}\rfloor$. These conditioned random walks $\mathcal{X}_{i}$ can be also sampled from the corresponding conditional probability $\mathbbm{P}\mathopen{}\mathclose{{}\left[\mathcal{X}_{i}-\mathcal{X}_{i-1}=1\,|\,\mathcal{X}_{1},\dots,\mathcal{X}_{i-1}}\right]$ exactly as before. Then we may use (1.1) to construct arc diagrams, but for the second sampled walk, there are $2\lfloor n^{1/2}\rfloor$ unmatched points called _(good) boundary points_ , c.f. Section 1.5 and Section 6.2. With these boundary points, we may apply the Tutte embedding. A path of arcs started from any of the boundary points ends at another boundary point because boundary points have degree 1, whereas non-boundary points have degree 2. Thus it defines a _boundary path_ , c.f. Definition 6.2. We also link each pair of boundary points successively to form loops, c.f. Figure 4 (Left) for the corresponding procedure in the UIHPMS. After coloring loops according to their sizes, we finally obtain the pictures shown in Figure 2. This construction is analogous to the random walk construction of the UIHPMS defined in Section 1.5.555Simple walks with $2n$ steps started from $0$ conditioned to stay positive and end at $2\lfloor n^{1/2}\rfloor$ are in bijection with reflected simple walks with $2n+2\lfloor n^{1/2}\rfloor$ steps started from $0$ conditioned to have exactly $2\lfloor n^{1/2}\rfloor$ zeros (after time zero) and to end at $0$, from the bijection in the proof of Lemma 6.4. With the latter random walk, an arc diagram can be constructed from (1.9), but no extra linking is necessary as there are no unmatched points. See also Figure 4 (Right). Instead of linking every successive pair of boundary points as above, we may exclude two marked points (specifically, we marked $0$ and another point close to $n$ in our simulations). As each point other than these two marked points is incident to two arcs, there is always a path from one marked point to the other. We expect this path approximates a chordal SLE6 curve between two marked points. See the discussion prior to Conjecture 1.15. The simulation in Figure 5 does not contradict any known qualitative behaviors of SLE6. We also note that there are other embeddings, e.g., circle packing, Riemann uniformization, or Smith embedding (square tiling) under which planar maps without boundary should converge to LQG. However, the representation of paths and loops on these embeddings may not be as straightforward as in the Tutte embedding. Every simulation in this section can also be done for random meandric systems encoded by correlated random walks. For example, for $\gamma=1$ (correlation $-\cos(\pi\gamma^{2}/4)=-1/\sqrt{2}$), the simulation suggests the size of the largest loop grows like $n^{\alpha}$ where $\alpha$ is in the 95% confidence interval $0.8379\pm 0.0100$. This is consistent with the prediction $\frac{1}{2}(5-\sqrt{11})\approx 0.8417$ coming from KPZ if we assume that the loops converge to CLE6. See Remark 1.16 for further discussion. ## References * [Aru15] J. Aru. KPZ relation does not hold for the level lines and SLEκ flow lines of the Gaussian free field. Probab. Theory Related Fields, 163(3-4):465–526, 2015, 1312.1324. MR3418748 * [Bef08] V. Beffara. The dimension of the SLE curves. Ann. Probab., 36(4):1421–1452, 2008, math/0211322. MR2435854 (2009e:60026) * [BGRV16] N. Berestycki, C. Garban, R. Rhodes, and V. Vargas. KPZ formula derived from Liouville heat kernel. J. Lond. Math. Soc. (2), 94(1):186–208, 2016, 1406.7280. MR3532169 * [BGS22] J. Borga, E. Gwynne, and X. Sun. Permutons, meanders, and SLE-decorated Liouville quantum gravity. ArXiv e-prints, July 2022, 2207.02319. * [BJRV13] J. Barral, X. Jin, R. Rhodes, and V. Vargas. Gaussian multiplicative chaos and KPZ duality. Comm. Math. Phys., 323(2):451–485, 2013, 1202.5296. MR3096527 * [BP] N. Berestycki and E. Powell. Gaussian free field, Liouville quantum gravity, and Gaussian multiplicative chaos. Available at https://homepage.univie.ac.at/nathanael.berestycki/Articles/master.pdf. * [BS01] I. Benjamini and O. Schramm. Recurrence of distributional limits of finite planar graphs. Electron. J. Probab., 6:no. 23, 13 pp. (electronic), 2001, 0011019. MR1873300 (2002m:82025) * [BS09] I. Benjamini and O. Schramm. KPZ in one dimensional random geometry of multiplicative cascades. Comm. Math. Phys., 289(2):653–662, 2009, 0806.1347. MR2506765 (2010c:60151) * [CGHP20] N. Crawford, A. Glazman, M. Harel, and R. Peled. Macroscopic loops in the loop $O(n)$ model via the XOR trick. ArXiv e-prints, January 2020, 2001.11977. * [CKST19] N. Curien, G. Kozma, V. Sidoravicius, and L. Tournier. Uniqueness of the infinite noodle. Ann. Inst. Henri Poincaré D, 6(2):221–238, 2019, 1701.01083. MR3950654 * [CN06] F. Camia and C. M. Newman. Two-dimensional critical percolation: the full scaling limit. Comm. Math. Phys., 268(1):1–38, 2006, math/0605035. MR2249794 * [CN08] F. Camia and C. M. Newman. ${\mathrm{S}LE}_{6}$ and ${\mathrm{C}LE}_{6}$ from critical percolation. In Probability, geometry and integrable systems, volume 55 of Math. Sci. Res. Inst. Publ., pages 103–130. Cambridge Univ. Press, Cambridge, 2008, math/0611116. MR2407594 (2009h:82032) * [DCGPS21] H. Duminil-Copin, A. Glazman, R. Peled, and Y. Spinka. Macroscopic loops in the loop $O(n)$ model at Nienhuis’ critical point. J. Eur. Math. Soc. (JEMS), 23(1):315–347, 2021, 1707.09335. MR4186469 * [DCN04] D. Dei Cont and B. Nienhuis. The packing of two species of polygons on the square lattice. Journal of Physics A: Mathematical and General, 37(9):3085, 2004, cond-mat/0311244. * [DDDF20] J. Ding, J. Dubédat, A. Dunlap, and H. Falconet. Tightness of Liouville first passage percolation for $\gamma\in(0,2)$. Publ. Math. Inst. Hautes Études Sci., 132:353–403, 2020, 1904.08021. MR4179836 * [DDG21] J. Ding, J. Dubedat, and E. Gwynne. Introduction to the Liouville quantum gravity metric. ArXiv e-prints, September 2021, 2109.01252. * [DDGG22] P. Di Francesco, B. Duplantier, O. Golinelli, and E. Guitter. Exponents for Hamiltonian paths on random bicubic maps and KPZ. ArXiv e-prints, October 2022, 2210.08887. * [DFGG00] P. Di Francesco, O. Golinelli, and E. Guitter. Meanders: exact asymptotics. Nuclear Phys. B, 570(3):699–712, 2000, cond-mat/9910453. MR1749396 * [DG18] J. Ding and E. Gwynne. The fractal dimension of Liouville quantum gravity: universality, monotonicity, and bounds. Communications in Mathematical Physics, 374:1877–1934, 2018, 1807.01072. * [DKRV16] F. David, A. Kupiainen, R. Rhodes, and V. Vargas. Liouville quantum gravity on the Riemann sphere. Comm. Math. Phys., 342(3):869–907, 2016, 1410.7318. MR3465434 * [DMS21] B. Duplantier, J. Miller, and S. Sheffield. Liouville quantum gravity as a mating of trees. Astérisque, (427):viii+257, 2021, 1409.7055. MR4340069 * [DRSV14] B. Duplantier, R. Rhodes, S. Sheffield, and V. Vargas. Renormalization of critical Gaussian multiplicative chaos and KPZ relation. Comm. Math. Phys., 330(1):283–330, 2014, 1212.0529. MR3215583 * [DS11] B. Duplantier and S. Sheffield. Liouville quantum gravity and KPZ. Invent. Math., 185(2):333–393, 2011, 1206.0212. MR2819163 (2012f:81251) * [FBS+21] J. Fairbanks, M. Besançon, S. Simon, J. Hoffiman, N. Eubank, and S. Karpinski. Juliagraphs/graphs.jl: an optimized graphs package for the julia programming language, 2021. * [FN19] M. Fukuda and I. Nechita. Enumerating meandric systems with large number of loops. Ann. Inst. Henri Poincaré D, 6(4):607–640, 2019, 1609.02756. MR4033682 * [FT22] V. Féray and P. Thévenin. Components in Meandric Systems and the Infinite Noodle. International Mathematics Research Notices, 07 2022, 2201.11572. rnac156. * [GHM20] E. Gwynne, N. Holden, and J. Miller. An almost sure KPZ relation for SLE and Brownian motion. Ann. Probab., 48(2):527–573, 2020, 1512.01223. MR4089487 * [GHS19] E. Gwynne, N. Holden, and X. Sun. A distance exponent for Liouville quantum gravity. Probability Theory and Related Fields, 173(3):931–997, 2019, 1606.01214. * [GHS20] E. Gwynne, N. Holden, and X. Sun. A mating-of-trees approach for graph distances in random planar maps. Probab. Theory Related Fields, 177(3-4):1043–1102, 2020, 1711.00723. MR4126936 * [GHS21] E. Gwynne, N. Holden, and X. Sun. Joint scaling limit of site percolation on random triangulations in the metric and peanosphere sense. Electron. J. Probab., 26:Paper No. 94, 58, 2021, 1905.06757. MR4278605 * [GHS23] E. Gwynne, N. Holden, and X. Sun. Mating of trees for random planar maps and Liouville quantum gravity: a survey. In Topics in statistical mechanics, volume 59 of Panor. Synthèses, pages 41–120. Soc. Math. France, Paris, 2023, 1910.04713. MR4619311 * [GM17] E. Gwynne and J. Miller. Scaling limit of the uniform infinite half-plane quadrangulation in the Gromov-Hausdorff-Prokhorov-uniform topology. Electron. J. Probab., 22:1–47, 2017, 1608.00954. * [GM21a] E. Gwynne and J. Miller. Characterizations of SLEκ for $\kappa\in(4,8)$ on Liouville quantum gravity. Astérisque, (429):vii+242, 2021, 1701.05174. * [GM21b] E. Gwynne and J. Miller. Existence and uniqueness of the Liouville quantum gravity metric for $\gamma\in(0,2)$. Invent. Math., 223(1):213–333, 2021, 1905.00383. MR4199443 * [GMS19] E. Gwynne, J. Miller, and S. Sheffield. Harmonic functions on mated-CRT maps. Electron. J. Probab., 24:no. 58, 55, 2019, 1807.07511. * [GMS21] E. Gwynne, J. Miller, and S. Sheffield. The Tutte embedding of the mated-CRT map converges to Liouville quantum gravity. Ann. Probab., 49(4):1677–1717, 2021, 1705.11161. * [GNP20] I. P. Goulden, A. Nica, and D. Puder. Asymptotics for a class of meandric systems, via the Hasse diagram of ${\rm NC}(n)$. Int. Math. Res. Not. IMRN, (4):983–1034, 2020, 1708.05188. MR4073935 * [GP19] E. Gwynne and J. Pfeffer. Bounds for distances and geodesic dimension in Liouville first passage percolation. Electronic Communications in Probability, 24:no. 56, 12, 2019, 1903.09561. * [GP21] E. Gwynne and J. Pfeffer. External diffusion-limited aggregation on a spanning-tree-weighted random planar map. Ann. Probab., 49(4):1633–1676, 2021, 1901.06860. MR4260464 * [GP22] E. Gwynne and J. Pfeffer. KPZ formulas for the Liouville quantum gravity metric. Trans. Amer. Math. Soc., 375(12):8297–8324, 2022, 1905.11790. MR4504639 * [GS22] E. Gwynne and J. Sung. The Minkowski content measure for the Liouville quantum gravity metric. ArXiv e-prints, November 2022, 2211.04701. * [Gwy20] E. Gwynne. Random surfaces and Liouville quantum gravity. Notices Amer. Math. Soc., 67(4):484–491, 2020, 1908.05573. MR4186266 * [JK98] J. L. Jacobsen and J. Kondev. Field theory of compact polymers on the square lattice. Nuclear Physics B, 532(3):635–688, 1998, cond-mat/9804048. * [JK99] J. L. Jacobsen and J. Kondev. Transition from the compact to the dense phase of two-dimensional polymers. Journal of statistical physics, 96(1):21–48, 1999, cond-mat/9811085. * [JZJ04] J. Jacobsen and P. Zinn-Justin. Algebraic Bethe ansatz for the FPL2 model. Journal of Physics A: Mathematical and General, 37(29):7213, 2004, math-ph/0402008. * [Kah85] J.-P. Kahane. Sur le chaos multiplicatif. Ann. Sci. Math. Québec, 9(2):105–150, 1985. MR829798 (88h:60099a) * [Kar20] V. Kargin. Cycles in random meander systems. J. Stat. Phys., 181(6):2322–2345, 2020, 2011.13449. MR4179808 * [KMT76] J. Komlós, P. Major, and G. Tusnády. An approximation of partial sums of independent RV’s, and the sample DF. II. Z. Wahrscheinlichkeitstheorie und Verw. Gebiete, 34(1):33–58, 1976\. MR0402883 * [KPZ88] V. Knizhnik, A. Polyakov, and A. Zamolodchikov. Fractal structure of 2D-quantum gravity. Modern Phys. Lett A, 3(8):819–826, 1988. * [KS64] S. Kochen and C. Stone. A note on the Borel-Cantelli lemma. Illinois Journal of Mathematics, 8(2):248–251, 1964. * [KW16] A. Kemppainen and W. Werner. The nested simple conformal loop ensembles in the Riemann sphere. Probab. Theory Related Fields, 165(3-4):835–866, 2016, 1402.2433. MR3520020 * [La 03] M. La Croix. Approaches to the enumerative theory of meanders. Available at https://math.mit.edu/~malacroi/Latex/Meanders.pdf, 2003. * [MWW16] J. Miller, S. S. Watson, and D. B. Wilson. Extreme nesting in the conformal loop ensemble. Ann. Probab., 44(2):1013–1052, 2016, 1401.0218. MR3474466 * [Poi12] H. Poincaré. Sur un téorème de géométrie. Rend. del Circ. Mat. Palermo, 33:375–407, 1912. * [RV11] R. Rhodes and V. Vargas. KPZ formula for log-infinitely divisible multifractal random measures. ESAIM Probab. Stat., 15:358–371, 2011, 0807.1036. MR2870520 * [Sch00] O. Schramm. Scaling limits of loop-erased random walks and uniform spanning trees. Israel J. Math., 118:221–288, 2000, math/9904022. MR1776084 (2001m:60227) * [She09] S. Sheffield. Exploration trees and conformal loop ensembles. Duke Math. J., 147(1):79–129, 2009, math/0609167. MR2494457 (2010g:60184) * [She22] S. Sheffield. What is a random surface? ArXiv e-prints, March 2022, 2203.02470. * [Sim83] G. Simons. A discrete analogue and elementary derivation of ‘Lévy’s equivalence’ for Brownian motion. Statistics & Probability Letters, 1(4):203–206, 1983. * [Smi01] S. Smirnov. Critical percolation in the plane: conformal invariance, Cardy’s formula, scaling limits. C. R. Acad. Sci. Paris Sér. I Math., 333(3):239–244, 2001, 0909.4499. MR1851632 (2002f:60193) * [Ste03] K. Stephenson. Circle packing: a mathematical tale. Notices of the AMS, 50(11):1376–1388, 2003. * [Zai98] A. Y. Zaitsev. Multidimensional version of the results of Komlós, Major and Tusnády for vectors with finite exponential moments. ESAIM Probab. Statist., 2:41–108, 1998. MR1616527 * [Zvo21] A. Zvonkin. Meanders: a Personal Perspective. Available at https://www.labri.fr/perso/zvonkin/Research/Meanders-for-Rosenstiehl.pdf, 2021\. Data availability statement: Data sharing is not applicable to this article as no new data were created or analyzed in this study. Conflict of interest: The authors have no competing interests to declare that are relevant to the content of this article.
# Super-resolution of positive near-colliding point sources ††thanks: This work was supported in part by the Swiss National Science Foundation grant number 200021–200307. Ping Liu Department of Mathematics, ETH Zürich, Rämistrasse 101, CH-8092 Zürich, Switzerland<EMAIL_ADDRESS>[email protected]). Habib Ammari22footnotemark: 2 ###### Abstract In this paper, we analyze the capacity of super-resolution of one-dimensional positive sources. In particular, we consider the same setting as in [2] and generalize the results there to the case of super-resolving positive sources. To be more specific, we consider resolving $d$ positive point sources with $p\leqslant d$ nodes closely-spaced and forming a cluster, while the rest of the nodes are well separated. Similarly to [2], our results show that when the noise level $\epsilon\lesssim\mathrm{SRF}^{-2p+1}$, where $\mathrm{SRF}=(\Omega\Delta)^{-1}$ with $\Omega$ being the cutoff frequency and $\Delta$ the minimal separation between the nodes, the minimax error rate for reconstructing the cluster nodes is of order $\frac{1}{\Omega}\mathrm{SRF}^{2p-2}\epsilon$, while for recovering the corresponding amplitudes $\left\\{a_{j}\right\\}$ the rate is of order $\mathrm{SRF}^{2p-1}\epsilon$. For the non-cluster nodes, the corresponding minimax rates for the recovery of nodes and amplitudes are of order $\frac{\epsilon}{\Omega}$ and $\epsilon$, respectively. Our numerical experiments show that the Matrix Pencil method achieves the above optimal bounds when resolving the positive sources. ## 1 Introduction In recent years, the problem of super-resolution (SR), which seeks to extract fine details of a signal from its noisy Fourier data in a bounded frequency domain, draws increasing interest in the field of applied mathematics. In particular, considerable progress has been made in the study of super- resolution of sparse signals, e.g. a host of algorithms [3, 9, 23, 26, 25, 22, 21, 7] were devised for resolving signals with a sparse prior. The sparse signals are frequently modeled as discrete measures $F(x)=\sum_{j=1}^{d}a_{j}\delta\left(x-x_{j}\right),\quad x_{j}\in\mathbb{R},$ where $\delta$ is the Dirac’s $\delta$-distribution. Let $\mathcal{F}[F]$ denote the Fourier transform of $F$ : $\mathcal{F}[F](s)=\int_{-\infty}^{\infty}F(x)e^{-2\pi isx}\mathrm{\leavevmode\nobreak\ d}x.$ The noisy spectral data of the signal $F$ is modeled as a function $\Phi$ satisfying, $|\Phi(s)-\mathcal{F}[F](s)|\leqslant\epsilon,\quad s\in[-\Omega,\Omega],$ (1.1) where $\epsilon>0$ represents the noise level and $\Omega>0$ is the cutoff frequency. The sparse SR problem considered in this paper reads: given $\Phi$ as above, estimate the unknown parameters of $F$, namely the amplitudes $\left\\{a_{j}\right\\}$ and the nodes $\left\\{x_{j}\right\\}$. The minimax error rate for recovering the nodes and amplitudes from the spectral data $\Phi(s)$ has been established in [2]. In the present paper, we aim at exploring the corresponding minimax error rate for resolving positive signals. Since our result is a generalization of the result in [2] which deals with complex signals, we utilize the same notation, concepts, and configurations as those in [2] for the sake of consistency of the two papers and the convenience of reading. ### 1.1 Main contribution The main contribution of this paper is the generalization of the estimates in [2] for the minimax error rate for complex signals in the off-the-grid setting to the case of resolving positive signals. We consider the case where the nodes $\left\\{x_{j}\right\\}$ can take arbitrary real values and the amplitudes $\left\\{a_{j}\right\\}$ are known to be positive. We consider the same distribution of nodes as in [2] where it is assumed that $p$ nodes (approximately uniformly distributed), $x_{\kappa},\ldots,x_{\kappa+p-1}$, form a small cluster and the rest of the nodes are away from all the other nodes (see Definition 2.4 below). We show in Theorem 2.3 that for $\epsilon\lesssim(\Omega\Delta)^{2p-1}$ with $\Delta$ being the minimum separation of the clustered nodes, in the worst-case scenario, the errors of recovered nodes $x_{j}^{\prime}$ and amplitudes $a_{j}^{\prime}>0$ by any minimax algorithm (see Definition 2.2 below), satisfy * • For the non-cluster nodes: $\displaystyle\max_{j\notin\\{\kappa,\ldots,\kappa+p-1\\}}\left|x_{j}-x_{j}^{\prime}\right|\asymp\frac{\epsilon}{\Omega},$ $\displaystyle\max_{j\notin\\{\kappa,\ldots,\kappa+p-1\\}}\left|a_{j}-a_{j}^{\prime}\right|\asymp\epsilon;$ * • For the cluster nodes: $\displaystyle\max_{j\in\\{\kappa,\ldots,\kappa+p-1\\}}\left|x_{j}-x_{j}^{\prime}\right|\asymp\frac{\epsilon}{\Omega}(\Omega\Delta)^{-2p+2},$ $\displaystyle\max_{j\in\\{\kappa,\ldots,\kappa+p-1\\}}\left|a_{j}-a_{j}^{\prime}\right|\asymp\epsilon(\Omega\Delta)^{-2p+1}.$ Our results reveal that the minimax error rates for recovering the nodes and amplitudes of positive signals in the super-resolution problem are the same as those for resolving general complex signals [2]. To be more specific, for $\epsilon\lesssim\mathrm{SRF}^{-2p+1}$ with $\mathrm{SRF}:=(\Omega\Delta)^{-1}$, the minimax error rate for reconstructing the cluster nodes of positive signals is of the order $(\mathrm{SRF})^{2p-2}\frac{\epsilon}{\Omega}$, while for recovering the corresponding amplitudes the rate is of the order $(\mathrm{SRF})^{2p-1}\epsilon$. On the other hand, the corresponding minimax rates for the recovery of the non-cluster nodes and amplitudes are of the order $\frac{\epsilon}{\Omega}$ and $\epsilon$, respectively. These also indicate that the non-cluster nodes $\left\\{x_{j}\right\\}_{j\notin\\{\kappa,\ldots,\kappa+p-1\\}}$ can be recovered with much better stability than the cluster nodes. The main novelty we rely in analyzing the case of positive signals lies in a crucial observation in estimating the lower bound of diameter of the error set (Definition 2.3). In particular, in Theorem 4.1, we observe and demonstrate that the recovered and the underlying signals in the example constructed in [2] can actually be positive signals at the same time. We also examine the performance limit of the Matrix Pencil method in resolving positive signals by numerical experiments. The oberved error amplification in the experiments exactly verifies our theory for the minimax error rate. This also indicates that the Matrix Pencil method has the optimal performance in super-resolving positive sources. ### 1.2 Related work and discussion In 1992, Donoho first studied the possibility and difficulties of super- resolving multiple sources from noisy measurements. In particular, he considered measures supported on a lattice $\\{k\Delta\\}_{k=-\infty}^{\infty}$ and regularized by a so-called “Rayleigh index”. The measurement is then the noisy Fourier transform of the discrete measure with cutoff frequency $\Omega$. He derived both the lower and upper bounds for the minimax error of the amplitude recovery in terms of the noise level, grid spacing, cutoff frequency, and Rayleigh index. His results emphasize the importance of the sparsity in the super-resolution. The results were improved in recent years for the case when resolving $n$-sparse on-the- grid sources [6]. Concretely, the authors of [6] showed that the minimax error rate for amplitudes recovery scales like $\mathrm{SRF}^{2n-1}\epsilon$, where $\epsilon$ is the noise level and $\mathrm{SRF}:=\frac{1}{\Delta\Omega}$ is the super-resolution factor. Similar results for multi-clumps cases were also derived in [1, 10]. A closely related work to the present paper is [2], in which the authors derived sharp minimax errors for the location and the amplitude recovery of off-the-grid sources. They showed that for complex sources satisfying the $(p,h,T,\tau,\eta)$-clustered configuration (Definition 2.4) and $\epsilon\lesssim\mathrm{SRF}^{-2p+1}$ with $p$ being the number of the cluster nodes, the minimax error rate for reconstructing of the cluster nodes is of the order $(\mathrm{SRF})^{2p-2}\frac{\epsilon}{\Omega}$, while for recovering the corresponding amplitudes the rate is of the order $(\mathrm{SRF})^{2p-1}\epsilon$. Moreover, the corresponding minimax rates for the recovery of the non-cluster nodes and amplitudes are of the order $\frac{\epsilon}{\Omega}$ and $\epsilon$, respectively. As mentioned above, in the present paper we have generalized these results to the case when resolving positive sources. Thus, the minimax error estimations for super-resolving both one-dimensional complex and positive sources are well established now. On the other hand, in order to characterize the exact resolution in the number and location recovery, in the earlier works [16, 15, 14, 12, 13, 11] the authors have defined the so-called "computational resolution limits", which characterize the minimum required distance between point sources so that their number and locations can be stably resolved under certain noise level. It was shown that the computational resolution limits for the number and location recoveries in the $k$-dimensional super-resolution problem should be bounded above by respectively $\frac{C_{num}(k,n)}{\Omega}\left(\frac{\sigma}{m_{\min}}\right)^{\frac{1}{2n-2}}$ and $\frac{C_{supp}(k,n)}{\Omega}\left(\frac{\sigma}{m_{\min}}\right)^{\frac{1}{2n-1}}$, where $C_{num}(k,n)$ and $C_{supp}(k,n)$ are certain constants depending only on the source number $n$ and the space dimensionality $k$. In particular, these results were generalized to the case when resolving positive sources in [13]. In this paper, a similar idea is used to generalize the miminax error estimate to the positive cases. For other works related to the limit of super-resolution, we refer the readers to [20, 4] for understanding the resolution limit from the perceptive of sample complexity and to [24, 5] for the resolving limit of some algorithms. For the super-resolution of positive sources, to the best of our knowledge, the theoretical possibility for the super-resolution of positive sources was first considered in [8], where the authors characterized the relation between the sparsity of the on-the-grid signal $\mathbf{x}$ and the possibility of super-resolution in certain sense. Their definition and results focused on the possibility of overcoming Rayleigh limit in the presence of sufficiently small noise, while our work analyzes the non-asymptotic behavior of the reconstructions. In recent years, some researchers analyzed the stability of some super- resolution algorithms in a non-asymptotic regime [22, 21, 7] and derived similar stability results to those proved in this paper, which exhibit the optimal performance of these algorithms. ### 1.3 Organization of the paper The paper is organized in the following way. Section 2 presents the main results of the minimax error rate and Section 3 exhibits the performance of Matrix Pencil method by numerical experiments. Section 4 proves the main results stated in Section 2. ## 2 Minimax bound for the location and amplitude recoveries In this section, we present minimax error estimates for the location and amplitude recoveries in the super-resolution of positive signals. ### 2.1 Notation and preliminaries We shall denote by $\mathcal{P}_{d},\mathcal{P}_{d}^{+}$ the parameter space of respectively general and positive signals $F$ with amplitudes $a_{j}$’s and real, distinct and ordered nodes $x_{j}$’s: $\displaystyle\mathcal{P}_{d}=\left\\{(\mathbf{a},\mathbf{x}):\mathbf{a}=\left(a_{1},\ldots,a_{d}\right)\in\mathbb{C}^{d},\mathbf{x}=\left(x_{1},\ldots,x_{d}\right)\in\mathbb{R}^{d},x_{1}<x_{2}<\ldots<x_{d}\right\\},$ $\displaystyle\mathcal{P}_{d}^{+}=\left\\{(\mathbf{a},\mathbf{x}):\mathbf{a}=\left(a_{1},\ldots,a_{d}\right)\in(\mathbb{R}^{+})^{d},\mathbf{x}=\left(x_{1},\ldots,x_{d}\right)\in\mathbb{R}^{d},x_{1}<x_{2}<\ldots<x_{d}\right\\},$ and identify the $F$’s with their parameters $(\mathbf{a},\mathbf{x})\in\mathcal{P}_{d}$ or $\mathcal{P}_{d}^{+}$. We denote $\|F\|_{\infty}=\max\left(\|\mathbf{a}\|_{\infty},\|\mathbf{x}\|_{\infty}\right).$ We shall also denote the orthogonal coordinate projections of a signal $F$ to the $j$-th node and $j$-th amplitude, respectively, by $P_{\mathbf{x},j}:\mathcal{P}_{d}(\mathcal{P}_{d}^{+})\rightarrow\mathbb{R}$ and $P_{\mathbf{a},j}:\mathcal{P}_{d}(\mathcal{P}_{d}^{+})\rightarrow\mathbb{C}(\mathbb{R}^{+})$. Let $L_{\infty}[-\Omega,\Omega]$ be the space of bounded complex-valued functions defined on $[-\Omega,\Omega]$ with the norm $\|e\|_{\infty}=\max_{|s|\leqslant\Omega}|e(s)|$. ###### Definition 2.1. Given $\Omega>0$ and $U\subseteq\mathcal{P}_{d}^{+}$, we denote by $\mathfrak{F}(\Omega,U)$ the class of all admissible reconstruction algorithms, i.e., $\mathfrak{F}(\Omega,U)=\left\\{f:L_{\infty}[-\Omega,\Omega]\rightarrow U\right\\}.$ ###### Definition 2.2. Let $U\subset\mathcal{P}_{d}^{+}$. We consider the minimax error rate in estimating a signal $F\in U$ from $\Omega$-bandlimited data as in (1.1), with a measurement error $\epsilon>0$: $\mathcal{E}^{+}(\epsilon,U,\Omega)=\inf_{\mathfrak{f}\in\mathfrak{F}(\Omega,U)}\sup_{F\in U}\sup_{\|e\|_{\infty}\leqslant\epsilon}\|F-\mathfrak{f}(\mathcal{F}[F]+e)\|_{\infty}.$ Note that in order to analyze how the minimax error rate relates to the separation of the nodes and the magnitudes of the amplitudes, we will consider $U\in\mathcal{P}_{d}^{+}$ with certain specific constraints in the following discussions. Similarly, the minimax errors of estimating the individual nodes and the amplitudes of $F\in U$ are defined respectively by $\displaystyle\mathcal{E}^{+,\mathbf{x},j}(\epsilon,U,\Omega)=\inf_{\mathfrak{f}\in\mathfrak{F}(\Omega,U)}\sup_{F\in U}\sup_{\|e\|\leqslant\epsilon}\left|P_{\mathbf{x},j}(F)-P_{\mathbf{x},j}(\mathfrak{f}(\mathcal{F}(F)+e))\right|,$ $\displaystyle\mathcal{E}^{+,\mathbf{a},j}(\epsilon,U,\Omega)=\inf_{\mathfrak{f}\in\mathfrak{F}(\Omega,U)}\sup_{F\in U}\sup_{\|e\|_{\infty}\leqslant\epsilon}\left|P_{\mathbf{a},j}(F)-P_{\mathbf{a},j}(\mathfrak{f}(\mathcal{F}(F)+e))\right|.$ For a fixed $F\in\mathcal{P}_{d}^{+}$, we define the positive and general $\epsilon$-error set as follows. ###### Definition 2.3. The error set of positive signals $E_{\epsilon,\Omega}^{+}(F)\subset\mathcal{P}_{d}^{+}$ is the set consisting of all the signals $\hat{F}\in\mathcal{P}_{d}^{+}$ with $\left|\mathcal{F}[\hat{F}](\omega)-\mathcal{F}[F](\omega)\right|\leqslant\epsilon,\quad\omega\in[-\Omega,\Omega].$ (2.1) Moreover, the error set of general signal $E_{\epsilon,\Omega}(F)\subset\mathcal{P}_{d}$ is the set consisting of all the $\hat{F}\in\mathcal{P}_{d}$ satisfying (2.1). We will denote by $E_{\epsilon}^{+,\mathbf{x},j}(F)=E_{\epsilon,\Omega}^{+,\mathbf{x},j}(F)$ and $E_{\epsilon}^{+,\mathbf{a},j}(F)=E_{\epsilon,\Omega}^{+,\mathbf{a},j}(F)$ the projections of the error set of positive signals onto the individual nodes and the amplitudes components, respectively: $\displaystyle E_{\epsilon,\Omega}^{+,\mathbf{x},j}(F)=\left\\{\mathbf{x}_{j}^{\prime}\in\mathbb{R}:\left(\mathbf{a}^{\prime},\mathbf{x}^{\prime}\right)\in E_{\epsilon,\Omega}^{+}(F)\right\\}\equiv P_{\mathbf{x},j}E_{\epsilon,\Omega}^{+}(F),$ $\displaystyle E_{\epsilon,\Omega}^{+,\mathbf{a},j}(F)=\left\\{\mathbf{a}_{j}^{\prime}\in\mathbb{R}^{+}:\left(\mathbf{a}^{\prime},\mathbf{x}^{\prime}\right)\in E_{\epsilon,\Omega}^{+}(F)\right\\}\equiv P_{\mathbf{a}_{,}j}E_{\epsilon,\Omega}^{+}(F).$ Furthermore, we denote by $E_{\epsilon}^{\mathbf{x},j}(F)=E_{\epsilon,\Omega}^{\mathbf{x},j}(F)$ and $E_{\epsilon}^{\mathbf{a},j}(F)=E_{\epsilon,\Omega}^{\mathbf{a},j}(F)$ the projections of the error set of general signals onto the individual nodes and the amplitudes components, respectively: $\displaystyle E_{\epsilon,\Omega}^{\mathbf{x},j}(F)=\left\\{\mathbf{x}_{j}^{\prime}\in\mathbb{R}:\left(\mathbf{a}^{\prime},\mathbf{x}^{\prime}\right)\in E_{\epsilon,\Omega}(F)\right\\}\equiv P_{\mathbf{x},j}E_{\epsilon,\Omega}(F),$ $\displaystyle E_{\epsilon,\Omega}^{\mathbf{a},j}(F)=\left\\{\mathbf{a}_{j}^{\prime}\in\mathbb{C}:\left(\mathbf{a}^{\prime},\mathbf{x}^{\prime}\right)\in E_{\epsilon,\Omega}(F)\right\\}\equiv P_{\mathbf{a}_{,}j}E_{\epsilon,\Omega}(F).$ For any subset $V$ of a normed vector space with norm $\|\cdot\|$, the diameter of $V$ is given by $\mathrm{diam}(V)=\sup_{\mathbf{v}^{\prime},\mathbf{v}^{\prime\prime}\in V}\left\|\mathbf{v}^{\prime}-\mathbf{v}^{\prime\prime}\right\|_{\infty}.$ By the theory of optimal recovery [19, 17, 18], the minimax errors are directly linked to the diameter of the corresponding projections of the error set. More specifically, we have the following proposition that is similar to Proposition 2.4 in [2]. ###### Proposition 2.1. For $U\subset\mathcal{P}_{d}^{+},\Omega>0,1\leqslant j\leqslant d$ and $\epsilon>0$, we have $\displaystyle\frac{1}{2}\sup_{F:E_{\frac{1}{2}\epsilon,\Omega}^{+}(F)\subseteq U}\mathrm{diam}\left(E_{\frac{1}{2}\epsilon,\Omega}^{+}(F)\right)\leqslant\mathcal{E}^{+}(\epsilon,U,\Omega)\leqslant\sup_{F\in U}\mathrm{diam}\left(E_{2\epsilon,\Omega}^{+}(F)\right),$ (2.2) $\displaystyle\frac{1}{2}\sup_{F:E_{\frac{1}{2}\epsilon,\Omega}^{+}(F)\subseteq U}\mathrm{diam}\left(E_{\frac{1}{2}\epsilon,\Omega}^{+,\mathbf{x},j}(F)\right)\leqslant\mathcal{E}^{+,\mathbf{x},j}(\epsilon,U,\Omega)\leqslant\sup_{F\in U}\mathrm{diam}\left(E_{2\epsilon,\Omega}^{+,\mathbf{x},j}(F)\right),$ $\displaystyle\frac{1}{2}\sup_{F:E_{\frac{1}{2}\epsilon,\Omega}^{+}(F)\subseteq U}\mathrm{diam}\left(E_{\frac{1}{2}\epsilon,\Omega}^{+,\mathbf{a},j}(F)\right)\leqslant\mathcal{E}^{+,\mathbf{a},j}(\epsilon,U,\Omega)\leqslant\sup_{F\in U}\mathrm{diam}\left(E_{2\epsilon,\Omega}^{+,\mathbf{a},j}(F)\right).$ ### 2.2 Uniform estimates of minimax error for clustered configurations Similarly to [2], the main goal of this paper is to estimate $\mathcal{E}^{+,\mathbf{x},j}(\epsilon,U,\Omega),\mathcal{E}^{+,\mathbf{a},j}(\epsilon,U,\Omega)$, where $U\subset\mathcal{P}_{d}^{+}$ are certain compact subsets of $\mathcal{P}_{d}^{+}$ consisting of signals with $p\leqslant d$ nodes that are nearly uniformly distributed, forming a cluster. To be more specific, the set $U$ is defined as follows; See also [2, Definition 2.5]. ###### Definition 2.4. (Uniform cluster configuration) Given $0<\tau,\eta\leqslant 1$ and $0<h\leqslant T$, a node vector $\mathbf{x}=\left(x_{1},\ldots,x_{d}\right)\in\mathbb{R}^{d}$ is said to form a $(p,h,T,\tau,\eta)$-clustered configuration, if there exists a subset of $p$ nodes $\mathbf{x}^{c}=\left\\{x_{\kappa},\ldots,x_{\kappa+p-1}\right\\}\subset\mathbf{x},p\geqslant 2$, which satisfies the following conditions: * (i) for each $x_{j},x_{k}\in\mathbf{x}^{c},j\neq k$, $\tau h\leqslant\left|x_{j}-x_{k}\right|\leqslant h;$ * (ii) for $x_{\ell}\in\mathbf{x}\backslash\mathbf{x}^{c}$ and $x_{j}\in\mathbf{x},\ell\neq j$, $\eta T\leqslant\left|x_{\ell}-x_{j}\right|\leqslant T.$ One of the main contributions of [2] is an upper bound on $\operatorname{diam}\left(E_{\epsilon,\Omega}(F)\right)$, and its coordinate projections, for any signal $F$ forming a clustered configuration as above. Here, we generalize the result to the positive signal cases, which is a direct consequence of [2, Theorem 2.6]. ###### Theorem 2.1. (Upper bound) Let the positive signal $F=(\mathbf{a},\mathbf{x})\in\mathcal{P}_{d}^{+}$, such that $\mathbf{x}$ forms a $(p,h,T,\tau,\eta)$-clustered configuration and $0<m\leqslant\|\mathbf{a}\|$. Then, there exist positive constants $C_{1},\ldots,C_{5}$, depending only on $d,p,m$, such that for each $\frac{C_{4}}{\eta T}\leqslant\Omega\leqslant\frac{C_{5}}{h}$ and $\epsilon\leqslant C_{3}(\Omega\tau h)^{2p-1}$, it holds that $\begin{gathered}\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{x},j}(F)\right)\leqslant\frac{C_{1}}{\Omega}\epsilon\times\begin{cases}(\Omega\tau h)^{-2p+2},&x_{j}\in\mathbf{x}^{c},\\\ 1,&x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c},\end{cases}\\\ \operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{a},j}(F)\right)\leqslant C_{2}\epsilon\times\begin{cases}(\Omega\tau h)^{-2p+1},&x_{j}\in\mathbf{x}^{c},\\\ 1,&x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c}.\end{cases}\end{gathered}$ ###### Proof. By Theorem 2.6 in [2], under the same condition we have $\begin{gathered}\operatorname{diam}\left(E_{\epsilon,\Omega}^{\mathbf{x},j}(F)\right)\leqslant\frac{C_{1}}{\Omega}\epsilon\times\begin{cases}(\Omega\tau h)^{-2p+2},&x_{j}\in\mathbf{x}^{c},\\\ 1,&x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c},\end{cases}\\\ \operatorname{diam}\left(E_{\epsilon,\Omega}^{\mathbf{a},j}(F)\right)\leqslant C_{2}\epsilon\times\begin{cases}(\Omega\tau h)^{-2p+1},&x_{j}\in\mathbf{x}^{c},\\\ 1,&x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c}.\end{cases}\end{gathered}$ On the other hand, note that for $\epsilon>0$, according to the definition of $\mathrm{diam}(\cdot)$ we have $\mathrm{diam}\left(E_{\epsilon,\Omega}^{+,\boldsymbol{\mathbf{x}},j}(F)\right)\leq\mathrm{diam}\left(E_{\epsilon,\Omega}^{\boldsymbol{\mathbf{x}},j}(F)\right),\quad\mathrm{diam}\left(E_{\epsilon,\Omega}^{+,\boldsymbol{\mathbf{a}},j}(F)\right)\leq\mathrm{diam}\left(E_{\epsilon,\Omega}^{\boldsymbol{\mathbf{a}},j}(F)\right).$ This proves the theorem. ∎ The above estimates are optimal, as shown by our next main theorem. This is the main contribution of our paper, by which we non-trivially generalize the results in [2, Theorem 2.7]. For simplicity and without loss of generality, we assume that the index $\kappa$ is fixed in the result below. ###### Theorem 2.2. (Lower bound) Let $m\leqslant M,2\leqslant p\leqslant d,\tau\leqslant\frac{1}{p-1},\eta<\frac{1}{d},T>0$ be fixed. There exist positive constants $C_{1}^{\prime}\ldots,C_{5}^{\prime}$, depending only on $d,p,m,M$, such that for every $\Omega,h$ satisfying $h\leqslant C_{4}^{\prime}T$ and $\Omega h\leqslant C_{5}^{\prime}$, there exists $F=(\mathbf{a},\mathbf{x})\in\mathcal{P}_{d}^{+}$, with $\mathbf{x}$ forming a $(p,h,T,\tau,\eta)$-clustered configuration, and with $0<m\leqslant\|\mathbf{a}\|\leq M<\infty$, such that for certain indices $j_{1},j_{2}\in\\{\kappa,\ldots,\kappa+p-1\\}$ and every $\epsilon\leqslant C_{3}^{\prime}(\Omega\tau h)^{2p-1}$, it holds that $\displaystyle\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{x},j}(F)\right)\geqslant\frac{C_{1}^{\prime}}{\Omega}\epsilon\times\begin{cases}(\Omega\tau h)^{-2p+2},&j=j_{1},\\\ 1,&\forall j\notin\\{\kappa,\ldots,\kappa+p-1\\},\end{cases}$ $\displaystyle\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{a},j}(F)\right)\geqslant C_{2}^{\prime}\epsilon\times\begin{cases}(\Omega\tau h)^{-2p+1},&j=j_{2},\\\ 1,&\forall j\notin\\{\kappa,\ldots,\kappa+p-1\\}.\end{cases}$ ###### Proof. See the proof in Section 4. ∎ Thus combining Theorems 2.1 and 2.2 with Proposition 2.1 now, we can obtain the following theorem for the optimal rates of the minimax errors $\mathcal{E}^{+,\boldsymbol{\mathbf{x}},j},\mathcal{E}^{+,\boldsymbol{\mathbf{a}},j}$. This generalizes Theorem 2.8 in [2]. ###### Theorem 2.3. Let $m>0,2\leqslant p\leqslant d,\tau<\frac{1}{2(p-1)},\eta<\frac{1}{2d},T>0$ be fixed. There exist constants $c_{1},c_{2},c_{3}$, depending only on $d,p,m$ such that for all $\frac{c_{1}}{\eta T}\leqslant\Omega\leqslant\frac{c_{2}}{h}$ and $\epsilon\leqslant c_{3}(\Omega\tau h)^{2p-1}$, the minimax error rates for the set $\displaystyle U:=$ $\displaystyle U(p,d,h,\tau,\eta,T,m)$ $\displaystyle=$ $\displaystyle\left\\{(\mathbf{a},\mathbf{x})\in\mathcal{P}_{d}^{+}:0<m\leqslant\|\mathbf{a}\|\leq M<\infty,\text{$\mathbf{x}$ forms a $(p,h,T,\tau,\eta)$-clustered configuration}\right\\}$ satisfy the following: * • For the non-cluster nodes: $\forall j\notin\\{\kappa,\ldots,\kappa+p-1\\}:\quad\left\\{\begin{array}[]{l}\mathcal{E}^{+,\mathbf{x},j}(\epsilon,U,\Omega)\asymp\frac{\epsilon}{\Omega},\\\ \mathcal{E}^{+,\mathbf{a},j}(\epsilon,U,\Omega)\asymp\epsilon.\end{array}\right.$ * • For the cluster nodes: $\displaystyle\max_{j=\kappa,\ldots,\kappa+p-1}\mathcal{E}^{+,\mathbf{x},j}(\epsilon,U,\Omega)$ $\displaystyle\asymp\frac{\epsilon}{\Omega}(\Omega\tau h)^{-2p+2},$ $\displaystyle\max_{j=\kappa,\ldots,\kappa+p-1}\mathcal{E}^{+,\mathbf{a},j}(\epsilon,U,\Omega)$ $\displaystyle\asymp\epsilon(\Omega\tau h)^{-2p+1}.$ The proportionality constants in the above statements depend only on $d,p,m,M$. ###### Proof. The proof is the same as the one for Theorem 2.8 in [2]. Here, we present the details for the convenience of reading and completeness. Let $C_{3},C_{3}^{\prime},C_{4},C_{4}^{\prime},C_{5},C_{5}^{\prime}$ be the constants from Theorems 2.1 and 2.2. Let $c_{1}=C_{4}$ and $c_{2}=\min\left(C_{5},C_{5}^{\prime},C_{4}C_{4}^{\prime}\right)$. Let $\frac{c_{1}}{\eta T}\leqslant\Omega\leqslant\frac{c_{2}}{h}$, and $\epsilon\leqslant c_{3}(\Omega\tau h)^{2p-1}$, where $c_{3}\leqslant\min\left(C_{3},C_{3}^{\prime}\right)$ will be determined below. It can be verified that $\Omega,h$ and $\epsilon$ as above satisfy the conditions of both Theorems 2.1 and 2.2. Upper bound Directly follows from the upper bounds in Theorem 2.1 and Proposition 2.1. Lower bound Denote $U_{\epsilon}=\left\\{F\in U:E_{\frac{1}{2}\epsilon,\Omega}^{+}(F)\subseteq U\right\\}$. In order to prove the lower bounds on $\mathcal{E}^{+,\mathbf{x},j}$ and $\mathcal{E}^{+,\mathbf{a},j}$, by Proposition 2.1 it suffices to show that there exists an $F\in U_{\epsilon}\neq\emptyset$ such that the conclusions of Theorem 2.2 are satisfied for this $F$. Note that the set $U$ has a non-empty interior. Furthermore, one can choose $m^{\prime},M^{\prime}$ satisfying $m<m^{\prime}<M^{\prime}<M$, and also $T^{\prime}=0.99T,\tau^{\prime}=2\tau$ and $\eta^{\prime}=2\eta$, such that $U^{\prime}=U\left(p,d,h,\tau^{\prime},\eta^{\prime},T^{\prime},m^{\prime},M^{\prime}\right)\subset U,\quad\partial U^{\prime}\cap\partial U=\emptyset.$ By the construction of $U^{\prime}$, there exist positive constants $\widetilde{C}_{1},\widetilde{C}_{2}$, independent of $\Omega,h$ and $\tau,\eta$, such that $\displaystyle\inf_{u\in\partial U,u^{\prime}\in\partial U^{\prime}}\left|P_{\mathbf{x},j}(u)-P_{\mathbf{x},j}\left(u^{\prime}\right)\right|\geqslant\tilde{C}_{1}\times\begin{cases}\tau h,&x_{j}\in\mathbf{x}^{c},\\\ \eta T,&x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c};\end{cases}$ (2.3) $\displaystyle\inf_{u\in\partial U,u^{\prime}\in\partial U^{\prime}}\left|P_{\mathbf{a},j}(u)-P_{\mathbf{a},j}\left(u^{\prime}\right)\right|\geqslant\tilde{C}_{2}.$ Now, we use the fact that $\epsilon<c_{3}(\Omega\tau h)^{2p-1}$. Applying Theorem 2.1 to an arbitrary positive signal $F^{\prime}\in U^{\prime}$ and using the conditions $\frac{1}{\Omega}\leqslant\frac{\eta T}{c_{1}}$ and $\Omega\tau h\leqslant\Omega h\leqslant c_{2}$, we obtain that $\displaystyle\operatorname{diam}\left(E_{\frac{1}{2}\epsilon}^{+,\mathbf{x},j}\left(F^{\prime}\right)\right)$ $\displaystyle\leqslant\begin{cases}\frac{C_{1}c_{3}}{2}\tau h,&x_{j}\in\mathbf{x}^{c},\\\ \frac{C_{1}c_{3}}{2\Omega}(\Omega\tau h)^{2p-1}\leqslant\frac{C_{1}c_{3}}{2c_{1}}c_{2}^{2p-1}\eta T,&x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c};\end{cases}$ (2.4) $\displaystyle\operatorname{diam}\left(E_{\frac{1}{2}\epsilon}^{+,\mathbf{a},j}\left(F^{\prime}\right)\right)$ $\displaystyle\leqslant\begin{cases}\frac{C_{2}c_{3}}{2},&x_{j}\in\mathbf{x}^{c},\\\ \frac{C_{2}c_{3}}{2}c_{2}^{2p-1},&x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c}.\end{cases}$ Next, we set $c_{3}=\min\left(C_{3},C_{3}^{\prime},C_{3}^{\prime\prime}\right)$, where $C_{3}^{\prime\prime}=\min\left(1,c_{1}\right)\times\min\left(1,c_{2}^{-2p+1}\right)\times\min\left(\frac{2\tilde{C}_{1}}{C_{1}},\frac{2\tilde{C}_{2}}{C_{2}}\right).$ Combining (2.3) and (2.4), we obtain that $E_{\frac{1}{2}\epsilon,\Omega}^{+}(F^{\prime})\subseteq U$, i.e. $F^{\prime}\in U_{\epsilon}$. Since $F^{\prime}\in U^{\prime}$ is arbitrary, we conclude that $U^{\prime}\subseteq U_{\epsilon}$. Since clearly $U^{\prime}\neq\emptyset$, applying Proposition 2.1 and Theorem 2.2 finishes the proof. ∎ ## 3 Numerical optimality of Matrix Pencil method (MP method) Theorem 2.3 establishes the optimal error rate for super-resolving the locations and amplitudes of positive sources. In this section, we demonstrate by numerical experiments the optimal performance of MP method in recovering the locations of positive sources. Note that the numerical experiments in [2] have already demonstrated the optimal performance of Matrix Pencil method for resolving general sources. Here we conduct experiments similar to those in [2] but focusing on the case of resolving positive sources. ### 3.1 Review of the MP method In this section, we review the MP method . We assume that the noisy Fourier data of the signal $F$ is given by $\boldsymbol{\mathbf{Y}}(\omega)=\sum_{j=1}^{d}a_{j}e^{-2\pi ix_{j}\omega}+\epsilon(\omega),\quad\omega\in[-\Omega,\Omega].$ The measurements are usually taken at $N$ evenly spaced points, $\omega_{1}=-\Omega,\omega_{2}=-\Omega+h,\cdots,\omega_{N}=\Omega$ with $h$ being the spacing. From the measurement $\mathbf{Y}=(\mathbf{Y}(\omega_{1}),\mathbf{Y}(\omega_{2}),\cdots,\mathbf{Y}(\omega_{N}))^{\top}$ (3.1) and $\hat{N}=\lfloor\frac{N-1}{2}\rfloor$, we assemble the $(\hat{N}+1)\times(\hat{N}+1)$ Hankel matrix $\boldsymbol{\mathbf{H}}=\begin{pmatrix}\mathbf{Y}(\omega_{1})&\mathbf{Y}(\omega_{2})&\cdots&\mathbf{Y}(\omega_{\hat{N}})\\\ \mathbf{Y}(\omega_{2})&\mathbf{Y}(\omega_{3})&\cdots&\mathbf{Y}(\omega_{\hat{N}+1})\\\ \cdots&\cdots&\ddots&\cdots\\\ \mathbf{Y}(\omega_{\hat{N}})&\mathbf{Y}(\omega_{\hat{N}+1})&\cdots&\mathbf{Y}(\omega_{2\hat{N}+1})\end{pmatrix}.$ (3.2) Let $\boldsymbol{\mathbf{H}}_{u}:=\boldsymbol{\mathbf{H}}[1:\hat{N},:]$ (and $H_{l}:=H[2:\hat{N}+1,:]$) be the $\hat{N}\times(\hat{N}+1)$ matrix obtained from the Hankel matrix $\boldsymbol{\mathbf{H}}$ given by (3.2) by selecting the first $\hat{N}$ rows (respectively, the second to the $(\hat{N}+1)$-th rows). It turns out that, in the noiseless case, $e^{-2\pi ix_{j}h},1\leq j\leq n$, are exactly the nonzero generalized eigenvalues of the pencil $\boldsymbol{\mathbf{H}}_{l}-z\boldsymbol{\mathbf{H}}_{u}$. In the noisy case, when the sources are well-separated, each of the first $n$ nonzero generalized eigenvalues of the pencil $\boldsymbol{\mathbf{H}}_{l}-z\boldsymbol{\mathbf{H}}_{u}$ is close to $e^{-2\pi ix_{j}h}$ for some $j$. We summarize the Matrix Pencil method in Algorithm 1. Input: Source number $d$, measurement: $\mathbf{Y}$ in (3.1); 1: Let $\hat{N}=\lfloor\frac{N-1}{2}\rfloor$. Formulate the $(\hat{N}+1)\times(\hat{N}+1)$ Hankel matrix $\mathbf{H}$ given by (3.2) from $\mathbf{Y}$, and the matrices $\boldsymbol{\mathbf{H}}_{u},\boldsymbol{\mathbf{H}}_{l}$; 2: Compute the truncated Singular Value Decomposition (SVD) of $\boldsymbol{\mathbf{H}}_{u}$, $\boldsymbol{\mathbf{H}}_{l}$ of order $d$: $\boldsymbol{\mathbf{H}}_{u}=U_{1}\Sigma_{1}V_{1}^{*},\quad\boldsymbol{\mathbf{H}}_{l}=U_{2}\Sigma_{2}V_{2}^{*},$ where $U_{1},U_{2},V_{1},V_{2}$ are $\hat{N}\times d$ matrices and $\Sigma_{1},\Sigma_{2}$ are $d\times d$ matrices; 3: Generate the reduced pencil $\mathbf{\hat{H}}_{u}=U_{2}^{*}U_{1}\Sigma_{1}V_{1}^{*}V_{2},\quad\mathbf{\hat{H}}_{l}=\Sigma_{2},$ where $\mathbf{\hat{H}}_{u}$, $\mathbf{\hat{H}}_{l}$ are $d\times d$ matrices; 4: Compute the generalized eigenvalues $\\{\hat{z}_{j}\\}$ of the reduced pencil $(\mathbf{\hat{H}}_{u},\mathbf{\hat{H}}_{l})$, and put $\\{\hat{x}_{j}\\}=\\{\arg(\hat{z}_{j})\\},j=1,\cdots,n$, where the $\arg(z)$ is the argument of $z$; 5: Solving the linear least squares problem $\mathbf{\hat{b}}=\arg\min_{\mathbf{b}\in\mathbb{C}^{d}}\left|\left|{\boldsymbol{\mathbf{Y}}-V\mathbf{b}}\right|\right|_{2},$ where $V$ is the Vandermonde matrix $V=\left(e^{-2\pi i\hat{x}_{j}\omega_{k}}\right)_{k=1,\cdots,N}^{j=1,\cdots,d}$; 6: Compute $\hat{a}_{j}$ by $|\mathbf{\hat{b}}_{j}|$; Return: The estimated $\hat{x}_{j}$’s and $\hat{a}_{j}$’s. Algorithm 1 The Matrix Pencil algorithm ### 3.2 Numerical experiments We conduct 1000 random experiments (the randomness was in the choice of $a_{j},x_{j},\epsilon$) to examine the error amplification in the recovery of the nodes and amplitudes. In particular, we consider recovering $p=2$ cluster nodes and $1$ non-cluster node and their corresponding amplitudes. Each single experiment is summarized in Algorithm 2. The results are shown in Figure 3.1 and we observe that the error amplification is consistent with what we have predicted, i.e., the error amplification factors for resolving nodes and amplitudes are respectively $\mathrm{SRF}^{2p-2}$ and $\mathrm{SRF}^{2p-1}$ for the cluster nodes with size $p$. Moreover, for resolving non-cluster nodes, both the corresponding error amplification factors are bounded by a small constant. Input: $p,d,N,\epsilon$; 1: Construct the signal $F$ with $p$ closely-spaced sources and $d-p$ non- clustered sources; 2: Generate the measurement $\boldsymbol{\mathbf{Y}}$ defined by (3.1) with $\epsilon$ being the noise level; 3: Execute the MP method (Algorithm 1) and obtain $F_{MP}=\left(\mathbf{a}^{MP},\mathbf{x}^{MP}\right)$. The nodes in $\mathbf{x}^{MP}$ are ordered in an increasing manner; 4: for _each $j$_ do Compute the error for node $j$ : $e_{j}=\left|\mathbf{x}_{j}^{MP}-\mathbf{x}_{j}\right|.$ The success for node $j$ is defined as $\operatorname{Succ}_{j}=\left(e_{j}<\frac{\min_{\ell\neq j}\left|\mathbf{x}_{\ell}-\mathbf{x}_{j}\right|}{3}\right).$ if _Succ ${}_{j}==$ true_ then Compute normalized node error amplification factor $\mathcal{K}_{\mathbf{x},j}=\frac{\left|\mathbf{x}_{j}-\mathbf{x}_{j}^{MP}\right|\cdot\Omega}{\epsilon};$ Compute normalized amplitude error amplification factor $\mathcal{K}_{\mathbf{a},j}=\frac{\left|\mathbf{a}_{j}-\mathbf{a}_{j}^{MP}\right|}{\epsilon};$ Return: $\left(\mathcal{K}_{\mathbf{x},j},\mathcal{K}_{\mathbf{a},j},\operatorname{Succ}_{j}\right)$ for each node $j=1,\ldots,d$. Algorithm 2 A single experiment (a) (b) Figure 3.1: The error amplification factors. For the cluster nodes, the error amplification factors $\mathcal{K}_{\boldsymbol{\mathbf{x}},j},\mathcal{K}_{\boldsymbol{\mathbf{a}},j}$ scale like $\mathrm{SRF}^{2p-2},\mathrm{SRF}^{2p-1}$, respectively. For the non-cluster nodes, both error amplification factors are bounded by a small constant. ## 4 Proof of Theorem 2.2 ### 4.1 Normalization Similarly to [2], for ease of exposition, we should normalize the cluster configuration in some of the following discussions. Let us first define the scale transformation on $\mathcal{P}_{d}^{+}$. ###### Definition 4.1. For $F=\sum_{j=1}^{d}a_{j}\delta\left(x-x_{j}\right)\in\mathcal{P}_{d}^{+}$ and $T>0$, we define $SC_{T}:\mathcal{P}_{d}^{+}\rightarrow\mathcal{P}_{d}^{+}$ as follows: $SC_{T}(F)(x)=\sum_{j=1}^{d}a_{j}\delta\left(x-\frac{x_{j}}{T}\right).$ By the scale property of the Fourier transform, we have that for any $\epsilon>0$, $SC_{T}\left(E_{\epsilon,\Omega}^{+}(F)\right)=E_{\epsilon,\Omega T}^{+}\left(SC_{T}(F)\right).$ Thus the following proposition holds. ###### Proposition 4.1. Let $F=(\mathbf{a},\mathbf{x})\in\mathcal{P}_{d}^{+}$ and $T>0$. Then, for any $\epsilon>0$ and $1\leqslant j\leqslant d$, we have $\displaystyle\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{x},j}(F)\right)=T\operatorname{diam}\left(E_{\epsilon,\Omega T}^{+,\mathbf{x},j}\left(SC_{T}\left(F\right)\right)\right),$ $\displaystyle\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{a},j}(F)\right)=\operatorname{diam}\left(E_{\epsilon,\Omega T}^{+,\mathbf{a},j}\left(SC_{T}\left(F\right)\right)\right).$ ### 4.2 Auxiliary lemmas In this subsection, we introduce some notation and lemmas that are used in the following proofs. Set $\phi_{s}(t)=\left(1,t,\cdots,t^{s}\right)^{\top}.$ (4.1) ###### Lemma 4.1. Let $t_{1},\cdots,t_{k}$ be $k$ different real numbers and let $t$ be a real number. We have $\left(D_{k}(k-1)^{-1}\phi_{k-1}(t)\right)_{j}=\Pi_{1\leq q\leq k,q\neq j}\frac{t-t_{q}}{t_{j}-t_{q}},$ where $D_{k}(k-1):=\big{(}\phi_{k-1}(t_{1}),\cdots,\phi_{k-1}(t_{k})\big{)}$ with $\phi_{k-1}(\cdot)$ defined by (4.1). ###### Proof. This is [15, Lemma 5]. ∎ The following proposition is the main result for proving Theorem 2.2 and we present its detailed proof in Section 4.4. ###### Proposition 4.2. Let $F=(\mathbf{a},\mathbf{x})\in\mathcal{P}_{d}^{+}$, such that $\mathbf{x}$ forms a $(p,h,1,\tau,\eta)$-clustered configuration, with cluster nodes $\mathbf{x}^{c}=\left(x_{1},\ldots,x_{p}\right)$ (according to Definition 2.4), and with $\mathbf{a}\in(\mathbb{R}^{+})^{d}$ satisfying $m\leqslant\|\mathbf{a}\|\leqslant M$. Then, there exist constants $c_{1},k_{1},k_{2},k_{3},k_{4}$, depending only on $(d,p,\tau,m,M)$, such that for all $\epsilon<c_{1}(\Omega h)^{2p-1}$ and $\Omega h\leqslant 2$, there exists a signal $F_{\epsilon}\in\mathcal{P}_{d}^{+}$ satisfying, for some $j_{1},j_{2}\in\\{1,\ldots,p\\},$ $\displaystyle\left|P_{\mathbf{x},j_{1}}\left(F_{\epsilon}\right)-P_{\mathbf{x},j_{1}}(F)\right|\geqslant\frac{k_{1}}{\Omega}(\Omega h)^{-2p+2}\epsilon,$ $\displaystyle\left|P_{\mathbf{a},j_{2}}\left(F_{\epsilon}\right)-P_{\mathbf{a},j_{2}}(F)\right|\geqslant k_{2}(\Omega h)^{-2p+1}\epsilon,$ $\displaystyle\left|P_{\mathbf{x},j}\left(F_{\epsilon}\right)-P_{\mathbf{x},j}(F)\right|\geqslant\frac{k_{3}}{\Omega}\epsilon,\quad x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c},$ $\displaystyle\left|P_{\mathbf{a},j}\left(F_{\epsilon}\right)-P_{\mathbf{a},j}(F)\right|\geqslant k_{4}\epsilon,\quad x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c},$ $\displaystyle\left|\mathcal{F}\left(F_{\epsilon}\right)(s)-\mathcal{F}(F)(s)\right|\leqslant\epsilon,\quad|s|\leqslant\Omega.$ ### 4.3 Proof of Theorem 2.2 ###### Proof. After employing Proposition 4.2, the arguments for proving Theorem 2.2 is just the same as those in [2]. We present the details as follows. Let $\mathbf{a}\in(\mathbb{R}^{+})^{d}$ be any positive amplitude vector satisfying $m\leqslant\|\mathbf{a}\|\leqslant M$. Let $\Omega$, $h$ satisfy $\Omega h\leqslant 2$, and choose nodes $\mathbf{x}$ satisfying that $\mathbf{x}^{c}=\left(x_{1}=0,x_{1}=\tau h,\ldots,x_{p}=(p-1)\tau h\right),$ and the rest of the non-cluster nodes are equally spaced in $((p-1)\tau h,1)$. Now, let $h^{\prime}=(p-1)\tau h$ and $\tau^{\prime}=\frac{1}{p-1}$. Clearly, $\mathbf{x}$ is a $\left(p,h^{\prime},1,\tau^{\prime},\eta\right)$-clustered configuration for all sufficiently small $h$ (for instance $h<\frac{1}{d}<1-\eta(d-p+1)$).We now can apply Proposition 4.2 to the signal $F=(\mathbf{a},\mathbf{x})$. It then follows that for $\epsilon<c_{1}(p-1)^{2p-1}(\Omega\tau h)^{2p-1}$ and $\Omega h<\frac{2}{(p-1)\tau}$, there exist $j_{1},j_{2}\in\\{1,\ldots,p\\}$ such that $\displaystyle\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{x},j_{1}}(F)\right)$ $\displaystyle\geqslant\frac{k_{1}}{\Omega}(p-1)^{-2p+2}\epsilon(\Omega\tau h)^{-2p+2},$ $\displaystyle\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{a},j_{2}}(F)\right)$ $\displaystyle\geqslant k_{2}\epsilon(p-1)^{-2p+1}(\Omega\tau h)^{-2p+1}.$ Moreover, $\displaystyle\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{x},j}(F)\right)$ $\displaystyle\geqslant\frac{k_{3}}{\Omega}\epsilon,\quad x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c},$ $\displaystyle\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{a},j}(F)\right)$ $\displaystyle\geqslant k_{4}\epsilon,\quad x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c}.$ For the general case that $F=(\boldsymbol{\mathbf{a}},\boldsymbol{\mathbf{x}})\in\mathcal{P}_{d}^{+}$ such that $\mathbf{x}$ forms a $(p,h,T,\tau,\eta)$-clustered configuration, we consider $SC_{T}(F)=(\boldsymbol{\mathbf{a}},\mathbf{\tilde{x}})$, $\mathbf{\tilde{x}}=(\tilde{x}_{1},\cdots,\tilde{x}_{d})$, where $\tilde{x}_{j}=\frac{x_{j}}{T},j=1,\cdots,d$. Now the node vector $\mathbf{\tilde{x}}$ forms a $(p,\frac{h}{T},1,\tau,\eta)$-clustered configuration. Applying Proposition 4.1 and the above results, we obtain that $\displaystyle\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{x},j_{1}}(F)\right)=T\operatorname{diam}\left(E_{\epsilon,\Omega T}^{+,\mathbf{x},j_{1}}(SC_{T}(F))\right)\geqslant\frac{k_{1}}{\Omega}(p-1)^{-2p+2}\epsilon(\Omega\tau h)^{-2p+2},$ $\displaystyle\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{a},j_{2}}(F)\right)=\operatorname{diam}\left(E_{\epsilon,\Omega T}^{+,\mathbf{a},j_{2}}(SC_{T}(F))\right)\geqslant k_{2}\epsilon(p-1)^{-2p+1}(\Omega\tau h)^{-2p+1},$ $\displaystyle\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{x},j}(F)\right)=T\operatorname{diam}\left(E_{\epsilon,\Omega T}^{+,\mathbf{x},j}(SC_{T}(F))\right)\geqslant\frac{k_{3}}{\Omega}\epsilon,\quad x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c},$ $\displaystyle\operatorname{diam}\left(E_{\epsilon,\Omega}^{+,\mathbf{a},j}(F)\right)=\operatorname{diam}\left(E_{\epsilon,\Omega T}^{+,\mathbf{a},j}(SC_{T}(F))\right)\geqslant k_{4}\epsilon,\quad x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c}.$ This finishes the proof of Theorem $2.7$ with $C_{1}^{\prime}=\max\left(\frac{k_{1}}{(p-1)^{2p-2}},k_{3}\right)$, $C_{2}^{\prime}=\max\left(k_{4},\frac{k_{2}}{(p-1)^{2p-1}}\right),C_{3}^{\prime}=c_{1}(p-1)^{2p-1},C_{4}^{\prime}=\frac{1}{d}$ and $C_{5}^{\prime}=2$. ### 4.4 Proof of Proposition 4.2 We separate the proof into three steps. Step 1. In this step we prove the following theorem. ###### Theorem 4.1. Given the parameters $0<h\leqslant 2,0<\tau\leqslant 1,0<m\leqslant M<\infty$, let the signal $F=(\mathbf{a},\mathbf{x})\in\mathcal{P}_{p}^{+}$ with $\mathbf{a}\in(\mathbb{R}^{+})^{p}$ form a single uniform cluster as follows: * • (centered) $x_{p}=-x_{1}$; * • (uniform) for $1\leqslant j<k\leqslant p$ we have $\tau h\leqslant\left|x_{j}-x_{k}\right|\leqslant h;$ * • $m\leqslant\left\|a_{j}\right\|\leqslant M$. Then, there exist constants $K_{1},\ldots,K_{5}$ depending only on $(d,\tau,m,M)$ such that for every $\epsilon<K_{5}h^{2d-1}$, there exists a signal $F_{\epsilon}=(\mathbf{b},\mathbf{y})\in\mathcal{P}_{p}^{+}$ satisfying the following conditions: 1. (i) $m_{k}(F)=m_{k}\left(F_{\epsilon}\right)$ for $k=0,1,\ldots,2p-2$, where $m_{k}(F):=\sum_{j=1}^{p}a_{j}x_{j}^{k};$ (4.2) 2. (ii) $m_{2p-1}\left(F_{\epsilon}\right)=m_{2p-1}(F)+\epsilon$; 3. (iii) $K_{1}h^{-2p+2}\epsilon\leqslant\|\mathbf{x}-\mathbf{y}\|\leqslant K_{2}h^{-2p+2}\epsilon$; 4. (iv) $K_{3}h^{-2p+1}\epsilon\leqslant\|\mathbf{a}-\mathbf{b}\|\leqslant K_{4}h^{-2p+1}\epsilon$. ###### Proof. The case for $F=(\boldsymbol{\mathbf{a}},\boldsymbol{\mathbf{x}})\in\mathcal{P}_{p}$ and $F_{\epsilon}=(\boldsymbol{\mathbf{b}},\boldsymbol{\mathbf{y}})\in\mathcal{P}_{p}$ is the Theorem 6.2 in [2]. Now we prove that for the case when $\boldsymbol{\mathbf{a}}\in(\mathbb{R}^{+})^{p}$, from condition (i) in the theorem we actually have $\boldsymbol{\mathbf{b}}\in(\mathbb{R}^{+})^{p}$. This is enough to prove the theorem. Let $b_{j}$’s be elements in $\boldsymbol{\mathbf{b}}$ and $y_{j}$’s be elements in $\boldsymbol{\mathbf{Y}}$. We first consider the case when $x_{j^{*}}=y_{q^{*}}$ for certain $j^{*},q^{*}$. If $b_{q^{*}}\neq a_{j^{*}}$, then condition (i) in the theorem yields that $B\beta=A\alpha,$ (4.3) where $\alpha=(a_{1},\cdots,\ a_{j^{*}-1},\ a_{j^{*}}-b_{q^{*}},\ a_{j^{*}+1},\cdots,a_{p})^{\top}$, $\beta=(b_{1},\cdots,b_{q^{*}-1},b_{q^{*}+1},\cdots,b_{p})$ and $A=\big{(}\phi_{2p-2}(x_{1}),\cdots,\phi_{2p-2}(x_{p})\big{)},\ B=\big{(}\phi_{2p-2}(y_{1}),\cdots,\phi_{2p-2}(y_{q^{*}-1}),\phi_{2p-2}(y_{q^{*}+1}),\cdots,\phi_{2p-2}(y_{p})\big{)},$ with $\phi_{2p-2}(\cdot)$ being defined by (4.1). Since all the elements in $\alpha$ are nonzero by $b_{q^{*}}\neq a_{j^{*}}$ and $a_{j}>0,1\leq j\leq p$, $A$ contains $p$ different Vandermonde vectors ($\phi_{2p-2}(\cdot)$), and $B$ contains at most $p-1$ Vandermonde vectors, it is impossible to have (4.3) by [16, Theorem 3.12]. Thus, we must have $b_{q^{*}}=a_{j^{*}}>0$ for $x_{j^{*}}=y_{q^{*}}$. Next, we prove that $b_{q}>0$ for those $y_{q}$ that are not equal to any of the $x_{j}$’s. Without loss of generality, we can actually assume that all the $y_{q}$’s are not equal to any of the $x_{j}$’s. Further, since $(\boldsymbol{\mathbf{b}},\boldsymbol{\mathbf{y}})\in\mathcal{P}_{p}$ and $(\boldsymbol{\mathbf{a}},\boldsymbol{\mathbf{x}})\in\mathcal{P}_{p}^{+}$, all the $y_{q}$’s and $x_{j}$’s are distinct from each other. We now claim that $x_{1}<y_{1}<x_{2}<y_{2}<\cdots<x_{p}<y_{p},\ \text{ or }\ y_{1}<x_{1}<y_{2}<x_{2}<\cdots<y_{p}<x_{p}.$ (4.4) We denote the $x_{j},y_{j}$’s from left to right by $t_{1}<t_{2}<\cdots<t_{2p}$ and the corresponding $a_{j},-b_{j}$’s by $\alpha_{1},\cdots,\alpha_{2p}$. By condition (i), it follows that $A\alpha=0,$ where $\alpha=(\alpha_{1},\cdots,\alpha_{2p})^{\top}$ and $A=\big{(}\phi_{2p-2}(t_{1}),\cdots,\phi_{2p-2}(t_{2p})\big{)}$ with $\phi_{2p-2}(\cdot)$ being defined by (4.1). Thus we can have $-\alpha_{2p}\phi_{2p-2}(t_{2p})=\big{(}\phi_{2p-2}(t_{1}),\cdots,\phi_{2p-2}(t_{2p-1})\big{)}(\alpha_{1},\cdots,\alpha_{2p-1})^{\top},$ and hence $-\alpha_{2p}\left(\phi_{2p-2}(t_{1}),\cdots,\phi_{2p-2}(t_{2p-1})\right)^{-1}\phi_{2p-2}(t_{2p})=(\alpha_{1},\cdots,\alpha_{2p-1})^{\top}.$ (4.5) If claim (4.4) does not hold, we have $t_{q}=x_{j_{q}}$ and $t_{q+1}=x_{j_{q}+1}$ for certain $q,j_{q}$. Applying Lemma 4.1 to (4.5) and considering the $q$-th and $(q+1)$-th elements in the vectors, we have $\displaystyle-\alpha_{2p}\Pi_{1\leq j\leq 2p-1,j\neq q}\frac{t_{2p}-t_{j}}{t_{q}-t_{j}}=\alpha_{q},$ (4.6) $\displaystyle-\alpha_{2p}\Pi_{1\leq j\leq 2p-1,j\neq q+1}\frac{t_{2p}-t_{j}}{t_{q+1}-t_{j}}=\alpha_{q+1}.$ (4.7) Observe first that, for $1\leq k\leq 2p-1$, $\Pi_{1\leq j\leq 2p-1,j\neq k}(t_{2p}-t_{j})$ is always positive. Moreover, it is obvious that $\Pi_{1\leq j\leq 2p-1,j\neq q}(t_{q}-t_{j})$ and $\Pi_{1\leq j\leq 2p-1,j\neq q+1}(t_{q+1}-t_{j})$ have different signs in (4.6) and (4.7), respectively. It follows that $\alpha_{q}$ and $\alpha_{q+1}$ are of different signs. But the $\alpha_{q}$ and $\alpha_{q+1}$ are amplitudes of positive sources located at respectively $x_{q}$ and $x_{q+1}$, which yields a contradiction. Thus the case that $t_{q}=x_{j_{q}}$ and $t_{q+1}=x_{j_{q}+1}$ for certain $q,j_{q}$ will not happen and the claim (4.4) is thus proved. Suppose $y_{1}<x_{1}<y_{2}<x_{2}<\cdots<y_{p}<x_{p},$ (4.8) we now prove that the $b_{j}$’s are all positive. The another case can be demonstrated in the same way as below. By this setting, $t_{2j-1}=y_{j},\quad t_{2j}=x_{j},\quad j=1,\cdots,p.$ Since for $j=1,\cdots,2p-1$, we have $\displaystyle-\alpha_{2p}\Pi_{1\leq k\leq 2p-1,k\neq j}\frac{t_{2p}-t_{k}}{t_{j}-t_{k}}=\alpha_{j}.$ (4.9) For $1\leq j\leq 2p-1$, $\Pi_{1\leq k\leq 2p-1,k\neq j}(t_{2p}-t_{k})$ is always positive. For $j=2p-1$, since $\alpha_{2p}=a_{p}>0$, $-\alpha_{2p}\Pi_{1\leq k\leq 2p-1,k\neq 2p-1}(t_{2p-1}-t_{k})$ is negative in (4.9). Thus we have $\alpha_{2p-1}<0$. In the same fashion, we see that $\alpha_{j}>0$ for even $j$ and $\alpha_{j}<0$ for odd $j$. Note that by the setting (4.8), $\alpha_{2j-1}=-b_{j},\ j=1,\cdots,p$. This proves that $F_{\epsilon}$ is actually a positive signal and completes the proof of Theorem 4.1. ∎ Step 2. Now we start to prove Proposition 4.2. It is similar to the proof in [2]. Define $F^{c}$ and $F^{nc}$ to be the cluster and the non-cluster parts of $F\in\mathcal{P}_{d}^{+}$, respectively, i.e., $\displaystyle F^{c}$ $\displaystyle=\sum_{x_{j}\in\mathbf{x}^{c}}a_{j}\delta\left(x-x_{j}\right),$ $\displaystyle F^{nc}$ $\displaystyle=\sum_{x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c}}a_{j}\delta\left(x-x_{j}\right).$ We first analyze the non-cluster nodes. We construct that $F_{\epsilon}^{nc}=\sum_{x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c}}a_{j}^{\prime}\delta\left(x-x_{j}^{\prime}\right)$ where $a_{j}^{\prime}=a_{j}+\frac{\epsilon}{4(d-p)}$ and $x_{j}^{\prime}=x_{j}+\frac{\epsilon}{8\pi\Omega M(d-p)}$. For $|s|\leqslant\Omega$, the difference between the Fourier transforms of $F_{\epsilon}^{nc}$ and $F^{nc}$ satisfies $\displaystyle\left|\mathcal{F}[F_{\epsilon}^{nc}](s)-\mathcal{F}[F^{nc}](s)\right|$ $\displaystyle\leq\sum_{x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c}}\left|a_{j}e^{-2\pi ix_{j}s}-a_{j}^{\prime}e^{-2\pi ix_{j}^{\prime}s}\right|$ (4.10) $\displaystyle\leqslant\sum_{x_{j}\in\mathbf{x}\backslash\mathbf{x}^{c}}\left(\left|a_{j}e^{-2\pi ix_{j}s}\left(1-e^{-2\pi i\frac{\epsilon}{8\pi\Omega M(d-p)}s}\right)\right|+\frac{\epsilon}{4(d-p)}\right)$ $\displaystyle\leqslant\frac{\epsilon}{4}+\frac{\epsilon}{4}=\frac{\epsilon}{2}.$ Note that $a_{j}^{\prime}>0$ since $a_{j}^{\prime}=a_{j}+\frac{\epsilon}{4(d-p)}$ and $F_{\epsilon}^{nc}$ is a positive signal. We next analyze the cluster nodes. We suppose that $x_{1}+x_{p}=0$. For the case when $x_{1}+x_{p}\neq 0$, utilizing decomposition $\sum_{j=1}^{p}a_{j}e^{-2\pi ix_{j}s}=\sum_{j=1}^{p}a_{j}e^{-2\pi i\frac{x_{1}+x_{p}}{2}s}e^{-2\pi i\left(x_{j}-\frac{x_{1}+x_{p}}{2}\right)s},$ we nearly transfer the problem to the case when $x_{1}+x_{p}=0$ and it is not difficult to see that the following arguments hold as well with only a small modification, which is enough to prove the proposition. Next, we define a blowup of $F^{c}$ by $\Omega$ as $F_{(\Omega)}^{c}=SC_{\frac{1}{\Omega}}\left(F^{c}\right)=\sum_{x_{j}\in\mathrm{x}^{c}}a_{j}\delta\left(x-\Omega x_{j}\right)$ where $SC$ is defined by Definition 4.1. Let $\tilde{h}=\Omega h$ and $c_{1}=K_{5}(p,\tau,m,M)$ as in Theorem 4.1. Let $\epsilon\leqslant c_{1}(\Omega h)^{2p-1}$. Now, we apply Theorem 4.1 with parameters $p,\tilde{h},\tau,m,M,\tilde{\epsilon}=c_{2}\epsilon$ and the signal $F_{(\Omega)}^{c}$, where $c_{2}\leqslant 1$ will be determined below. We can obtain a signal $F_{(\Omega),\epsilon}^{c}\in\mathcal{P}_{p}^{+}$ such that the following hold for the difference of signals $H=F_{(\Omega),\epsilon}^{c}-F_{(\Omega)}^{c}$ : $m_{k}(H)=0,\quad k=0,1,\ldots,2p-2,\quad m_{2p-1}(H)=\tilde{\epsilon};$ (4.11) while also, for some $j_{1},j_{2}\in\\{1,\ldots,p\\}$ $\displaystyle\left|P_{\mathbf{x},j_{1}}\left(F_{(\Omega),\epsilon}^{c}\right)-P_{\mathbf{x},j_{1}}\left(F_{(\Omega)}^{c}\right)\right|\geqslant K_{1}(\Omega h)^{-2p+2}\tilde{\epsilon},$ (4.12) $\displaystyle\left|P_{\mathbf{x},j}\left(F_{(\Omega),\epsilon}^{c}\right)-P_{\mathbf{x},j}\left(F_{(\Omega)}^{c}\right)\right|\leqslant K_{2}(\Omega h)^{-2p+2}\tilde{\epsilon},\quad j=1,\ldots,p,$ $\displaystyle\left|P_{\mathbf{a}j_{2}}\left(F_{(\Omega),\epsilon}^{c}\right)-P_{\mathbf{a},j_{2}}\left(F_{(\Omega)}^{c}\right)\right|\geqslant K_{3}(\Omega h)^{-2p+1}\tilde{\epsilon}.$ Now, considering $F_{\epsilon}^{c}=SC_{\Omega}\left(F_{(\Omega),\epsilon}^{c}\right),$ we obtain that $\displaystyle\left|P_{\mathbf{x},j_{1}}\left(F_{\epsilon}^{c}\right)-P_{\mathbf{x},j_{1}}\left(F^{c}\right)\right|\geqslant\frac{K_{1}}{\Omega}(\Omega h)^{-2p+2}\tilde{\epsilon},$ $\displaystyle\left|P_{\mathbf{a},j_{2}}\left(F_{\epsilon}^{c}\right)-P_{\mathbf{a},j_{2}}\left(F^{c}\right)\right|\geqslant K_{3}(\Omega h)^{-2p+1}\tilde{\epsilon}.$ From the above definitions, we have $H_{\Omega}=SC_{\Omega}(H)=F_{\epsilon}^{c}-F^{c}$. We next show that there is a choice of $c_{2}$ such that $\left|\mathcal{F}[H_{\Omega}](s)\right|\leqslant\frac{\epsilon}{2},\quad|s|\leqslant\Omega.$ (4.13) Put $\omega=s/\Omega$. Then, by $\mathcal{F}\left[H_{\Omega}\right](s)=\mathcal{F}[H](\omega)$, the above inequality is equivalent to $|\mathcal{F}[H](\omega)|\leq\frac{\epsilon}{2},\quad|\omega|\leq 1.$ (4.14) Step 3. In this step we prove that (4.14) holds for a choice of $c_{2}$. Note that we have the following Taylor expansion of $\mathcal{F}[H](\omega)$: $\mathcal{F}[H](\omega)=\sum_{k=0}^{\infty}\frac{1}{k!}m_{k}(H)(-2\pi i\omega)^{k}.$ (4.15) Next we apply the following Taylor domination property [2, Theorem 6.3]. ###### Theorem 4.2. Let $H=\sum_{j=1}^{2p}\beta_{j}\delta\left(x-t_{j}\right)$, and put $R=\min_{j=1,\ldots,2p}\left|t_{j}\right|^{-1}>0$. Then, for all $k\geqslant 2p$, we have the so-called Taylor domination property $\left|m_{k}(H)\right|R^{k}\leqslant\left(\frac{2ek}{2p}\right)^{2p}\max_{\ell=0,1,\ldots,2p-1}\left|m_{\ell}(H)\right|R^{\ell}.$ Recall that $H=F_{(\Omega),\epsilon}^{c}-F_{(\Omega)}^{c}$. By Definition 2.4, the nodes of $F_{(\Omega)}^{c}$ is inside the interval $\left[-\frac{\Omega h}{2},\frac{\Omega h}{2}\right]$. The nodes of $F_{(\Omega),\epsilon}^{c}$, by (4.12), satisfy $\displaystyle\left|P_{\mathbf{x},j}\left(F_{(\Omega),\epsilon}^{c}\right)\right|\leqslant\frac{\Omega h}{2}+K_{2}(\Omega h)^{-2p+2}\tilde{\epsilon}\leqslant\frac{\Omega h}{2}+K_{2}(\Omega h)^{-2p+2}c_{1}(\Omega h)^{2p-1}=(\Omega h)\left(c_{1}K_{2}+\frac{1}{2}\right).$ Since $\Omega h\leqslant 2$ by assumption, we can conclude that the factor $R$ in Theorem 4.2 is greater than $C_{4}=\frac{1}{2\left(c_{1}K_{2}+\frac{1}{2}\right)}$. Now, we continue the proof of (4.14). By Theorem 4.2 and (4.11), we have for $k\geqslant 2p$, $\displaystyle\left|m_{k}(H)\right|\leqslant\left(\frac{e}{p}\right)^{2p}k^{2p}R^{2p-1-k}\tilde{\epsilon}\leqslant C_{5}C_{4}^{2p-1-k}k^{2p}\tilde{\epsilon}.$ Plugging this into (4.15) we obtain that $|\mathcal{F}(H)(\omega)|\leqslant\frac{\tilde{\epsilon}|2\pi\omega|^{2p-1}}{(2p-1)!}+C_{5}C_{4}^{2p-1}\tilde{\epsilon}\sum_{k\geqslant 2p}\left(\frac{2\pi|\omega|}{C_{4}}\right)^{k}\frac{k^{2p}}{k!}.$ Let $\zeta=\frac{2\pi|\omega|}{C_{4}}$ and by $|\omega|\leq 1$, we further have $\displaystyle|\mathcal{F}(H)(\omega)|\leqslant C_{6}\tilde{\epsilon}\sum_{k\geqslant 2p-1}\zeta^{k}\frac{k^{2p}}{k!}\leqslant C_{7}\tilde{\epsilon}.$ Therefore, we can choose $c_{2}=\min\left(1,\frac{1}{2C_{7}}\right)$ to ensure that $|\mathcal{F}(H)(\omega)|\leqslant\frac{\epsilon}{2},\quad|\omega|\leqslant 1,$ which shows (4.13). Finally, we construct the positive signal $F_{\epsilon}=F_{\epsilon}^{nc}+F_{\epsilon}^{c}$. Thus we have $\left|F_{\epsilon}(s)-F(s)\right|\leq\left|F_{\epsilon}^{nc}(s)-F^{nc}(s)\right|+\left|F_{\epsilon}^{c}(s)-F^{c}(s)\right|\leq\epsilon,\ s\in[-\Omega,\Omega].$ This completes the proof of Proposition 4.2 with $k_{1}=K_{1},k_{2}=K_{3},k_{3}=\frac{1}{8\pi\Omega M(d-p)},k_{4}=\frac{1}{4(d-p)}$. ∎ ## References * [1] Dmitry Batenkov, Laurent Demanet, Gil Goldman, and Yosef Yomdin. Conditioning of partial nonuniform fourier matrices with clustered nodes. SIAM Journal on Matrix Analysis and Applications, 41(1):199–220, 2020. * [2] Dmitry Batenkov, Gil Goldman, and Yosef Yomdin. Super-resolution of near-colliding point sources. Information and Inference: A Journal of the IMA, 05 2020. iaaa005. * [3] Emmanuel J. Candès and Carlos Fernandez-Granda. Towards a mathematical theory of super-resolution. Communications on Pure and Applied Mathematics, 67(6):906–956, 2014\. * [4] Sitan Chen and Ankur Moitra. Algorithmic foundations for the diffraction limit. pages 490–503, 2021. * [5] Maxime Ferreira Da Costa and Yuejie Chi. On the stable resolution limit of total variation regularization for spike deconvolution. IEEE Transactions on Information Theory, 66(11):7237–7252, 2020\. * [6] Laurent Demanet and Nam Nguyen. The recoverability limit for superresolution via sparsity. arXiv preprint arXiv:1502.01385, 2015. * [7] Quentin Denoyelle, Vincent Duval, and Gabriel Peyré. Support recovery for sparse super-resolution of positive measures. Journal of Fourier Analysis and Applications, 23(5):1153–1194, 2017\. * [8] David L Donoho, Iain M Johnstone, Jeffrey C Hoch, and Alan S Stern. Maximum entropy and the nearly black object. Journal of the Royal Statistical Society: Series B (Methodological), 54(1):41–67, 1992. * [9] Vincent Duval and Gabriel Peyré. Exact support recovery for sparse spikes deconvolution. Foundations of Computational Mathematics, 15(5):1315–1355, 2015\. * [10] Weilin Li and Wenjing Liao. Stable super-resolution limit and smallest singular value of restricted fourier matrices. Applied and Computational Harmonic Analysis, 51:118–156, 2021. * [11] Ping Liu and Habib Ammari. A mathematical theory of super-resolution and diffraction limit. arXiv preprint arXiv:2211.15208, 2022. * [12] Ping Liu and Habib Ammari. Nearly optimal resolution estimate for the two-dimensional super-resolution and a new algorithm for direction of arrival estimation with uniform rectangular array. arXiv preprint arXiv:2205.07115, 2022. * [13] Ping Liu, Yanchen He, and Habib Ammari. A mathematical theory of resolution limits for super-resolution of positive sources. arXiv preprint arXiv:2211.13541, 2022. * [14] Ping Liu and Hai Zhang. A mathematical theory of computational resolution limit in multi-dimensional spaces. Inverse Problems, 37(10):104001, 2021. * [15] Ping Liu and Hai Zhang. A theory of computational resolution limit for line spectral estimation. IEEE Transactions on Information Theory, 67(7):4812–4827, 2021\. * [16] Ping Liu and Hai Zhang. A mathematical theory of computational resolution limit in one dimension. Applied and Computational Harmonic Analysis, 56:402–446, 2022. * [17] CA Micchelli and TJ Rivlin. Lectures on optimal recovery. In Numerical Analysis Lancaster 1984, pages 21–93. Springer, 1985\. * [18] Charles A Micchelli, Th J Rivlin, and Shmuel Winograd. The optimal recovery of smooth functions. Numerische Mathematik, 26(2):191–200, 1976. * [19] Charles A Micchelli and Theodore J Rivlin. A survey of optimal recovery. Optimal estimation in approximation theory, pages 1–54, 1977. * [20] Ankur Moitra. Super-resolution, extremal functions and the condition number of vandermonde matrices. In Proceedings of the Forty-seventh Annual ACM Symposium on Theory of Computing, STOC ’15, pages 821–830, 2015. * [21] Veniamin I Morgenshtern. Super-resolution of positive sources on an arbitrarily fine grid. J. Fourier Anal. Appl., 28(1):Paper No. 4., 2021. * [22] Veniamin I. Morgenshtern and Emmanuel J. Candès. Super-resolution of positive sources: The discrete setup. SIAM Journal on Imaging Sciences, 9(1):412–444, 2016. * [23] Clarice Poon and Gabriel Peyré. Multidimensional sparse super-resolution. SIAM Journal on Mathematical Analysis, 51(1):1–44, 2019. * [24] Gongguo Tang. Resolution limits for atomic decompositions via markov-bernstein type inequalities. In 2015 International Conference on Sampling Theory and Applications (SampTA), pages 548–552. IEEE, 2015. * [25] Gongguo Tang, Badri Narayan Bhaskar, and Benjamin Recht. Near minimax line spectral estimation. IEEE Transactions on Information Theory, 61(1):499–512, 2014. * [26] Gongguo Tang, Badri Narayan Bhaskar, Parikshit Shah, and Benjamin Recht. Compressed sensing off the grid. IEEE transactions on information theory, 59(11):7465–7490, 2013\.
# Feedback-controlled solute transport through chemo-responsive polymer membranes Sebastian Milster Applied Theoretical Physics – Computational Physics, Physikalisches Institut, Albert-Ludwigs-Universität Freiburg, Hermann-Herder Strasse 3, D-79104 Freiburg, Germany Won Kyu Kim Korea Institute for Advanced Study, Seoul 02455, Republic of Korea Joachim Dzubiella <EMAIL_ADDRESS>Applied Theoretical Physics – Computational Physics, Physikalisches Institut, Albert-Ludwigs-Universität Freiburg, Hermann-Herder Strasse 3, D-79104 Freiburg, Germany Research Group for Simulations of Energy Materials, Helmholtz-Zentrum Berlin für Materialien und Energie, Hahn-Meitner-Platz 1, D-14109 Berlin, Germany ###### Abstract Polymer membranes are typically assumed to be inert and nonresponsive to the flux and density of the permeating particles in transport processes. Here, we study theoretically the consequences of membrane responsiveness and feedback on the steady-state force–flux relations and membrane permeability using a nonlinear-feedback solution-diffusion model of transport through a slab-like membrane. Therein, the solute concentration inside the membrane depends on the bulk concentration, $c_{0}$, the driving force, $f$, and the polymer volume fraction, $\phi$. In our model, solute accumulation in the membrane causes a sigmoidal volume phase transition of the polymer, changing its permeability, which, in return, affects the membrane’s solute uptake. This feedback leads to nonlinear force–flux relations, $j(f)$, which we quantify in terms of the system’s differential permeability, $\mathcal{P}_{\text{\tiny sys}}^{\Delta}\propto{\mathrm{d}j}/{\mathrm{d}f}$. We find that the membrane feedback can increase or decrease the solute flux by orders of magnitude, triggered by a small change in the driving force, and largely tunable by attractive versus repulsive solute–membrane interactions. Moreover, controlling the input, $c_{0}$ and $f$, can lead to steady-state bistability of $\phi$ and hysteresis in the force–flux relations. This work advocates that the fine-tuning of the membrane’s chemo-responsiveness will enhance the nonlinear transport control features, providing great potential for future (self-)regulating membrane devices. ## I Introduction The precise and selective control of molecular transport through membranes is of fundamental importance for various applications in industry and medicine, such as water purification,[1, 2, 3] food-processing,[4, 5] nano-catalysis,[6, 7, 8, 9] drug-delivery,[10, 11, 12, 13] and tissue engineering.[14, 15] Modern membrane technology becomes increasingly inspired by responsive bio-membranes with nonlinear potential-, pressure- or flux-gated permeabilities, bistable behavior and memristive properties.[16, 17, 18, 19, 20, 21, 22, 23] Such features allow the design of highly selective membrane devices that efficiently control molecular transport, autonomously regulate the chemical milieu, and may act as logical operators, artificial synapses, or analogous filters for electrical or chemical signals. Moreover, the possible memristive properties create the foundation for information storage, adaptive responses to stimuli based upon past events, and neuromorphic systems. [24, 25, 26] In general, such self-regulation premises a feedback mechanism controlling the transport properties in a nonlinear fashion.[27, 28, 29, 30] In the scope of membrane applications this may arise from various system-dependent effects, such as autocatalysis, substrate or product inhibition,[31, 32] the interplay of voltage and hydrodynamic pressure,[33, 34] or, as highlighted in this work, the reciprocal impacts of molecular fluxes and membrane permeability.[35, 36, 37, 38, 39, 40, 41, 42] In this regard many polymeric compounds offer great potential as they are versatile in their response to various physico-chemical stimuli and environmental conditions, such as temperature, electric field, solvent quality, etc.[43, 44, 45] For example, the polymer responds with a volume phase transition, either from a swollen to a collapsed state, or vice versa, in which the polymer volume fraction, $\phi$, may change by orders of magnitude.[46, 47, 48, 49, 50, 51, 36] Such a drastic change of the polymer’s physical features, in turn, has substantial, nonlinear effects on the solute permeability of the membrane device. Figure 1: Essential feedback loop of chemo-responsive polymer membranes pointing out the nonlinear, reciprocal dependence of the polymer volume fraction, $\phi(c_{\text{\tiny in}})$, and the solute concentration inside the membrane, $c_{\text{\tiny in}}(\phi)$. A change in the solute bulk concentration, $c_{0}$, or the external force, $f$, acting on the solutes has nontrivial effects on $c_{\text{\tiny in}}$ and $\phi$, and thus on the transport properties of the membrane. Very illustrative examples are so-called _smart gating membranes_ ,[52, 53, 54, 55, 56, 57, 58] which are (rather solid) porous membranes with polymer- coated channels that can reversibly open and close, triggered by external stimuli or, through autonomous feedback, by molecular recognition. Moreover, literature on the solution-diffusion model[59, 60, 61, 62] suggests that the use of more flexible, responsive polymeric membranes enables feedback- controlled solute transport with further valuable features, such as multiple steady states and hysteresis transitions.[35, 63, 38, 37, 36] However, more research is needed here to understand the role of the membrane feedback, especially in the presence of external driving, and how hysteresis transitions can occur. For nonresponsive polymer membranes, we have previously shown that the Smoluchowski equation[64] well describes solute flux and concentration profiles under stationary nonequilibrium conditions.[65] Therein, we reported that the membrane’s solute uptake, $c_{\text{\tiny in}}$, is not only a function of the polymer volume fraction, $\phi$, the membrane permeability, $\mathcal{P}_{\text{\tiny mem}}(\phi)$, and bulk concentration, $c_{0}$, but also tuneable in nonequilibrium by a the external driving force, $f$. The latter leads to a nonlinear flux, $j(f)$, with significant differences between the low- and high-force regimes. The nonlinear intermediate crossover was quantified using the newly introduced system’s differential permeability, $\mathcal{P}_{\text{\tiny sys}}^{\Delta}\propto{\mathrm{d}j}/{\mathrm{d}f}$. Motivated by the above features and open questions, in this work we turn our attention to polymer membranes that are responsive to the penetrants, and highlight the key differences compared to nonresponsive membranes. Specifically, we include a mean-field model for the polymer response in the Smoluchowski framework, i.e., $\phi\to\phi(c_{\text{\tiny in}})$ is a sigmoidal function of the average solute uptake, which enters $\mathcal{P}_{\text{\tiny mem}}(\phi)$ and, in turn, controls $c_{\text{\tiny in}}$, leading to a membrane-intrinsic feedback mechanism [Fig. 1]. Eventually, we use empirical expressions for $\mathcal{P}_{\text{\tiny mem}}(\phi)$ to study the feedback effect on $j$ and $\mathcal{P}_{\text{\tiny sys}}^{\Delta}$ as function of $c_{0}$ and $f$. Compared to nonresponsive membranes, we find substantial enhancement of the nonlinear characteristics, such as an order of magnitude change in $j$ due to a very small change in $f$, and report the emergence of multiple steady sates, bifurcations, and hysteresis in the force–flux relations. ## II Theoretical framework ### II.1 Steady-state Smoluchowski equation and system setup We consider the solute transport across a polymer membrane as a one- dimensional drift-diffusion process (in $z$-direction) of ideal solutes [see the system sketch in Fig. 2(a)]. The membrane has the width $d$, and is located in the center of the system of length $L$, yielding interfaces at $(L\pm d)/2$. It is in contact with two solute reservoirs of equal concentration $c_{0}$ via boundary layers on the feed and permeate sides. [66] The steady-state flux in the overdamped limit derived from the Smoluchowski equation, reads [67] $\displaystyle j=-D(z)\mathopen{}\mathclose{{}\left[\frac{\partial c(z)}{\partial z}+\beta c(z)\mathopen{}\mathclose{{}\left(\frac{\partial G(z)}{\partial z}-f}\right)}\right],$ (1) with the inverse temperature, $\beta=1/(k_{\text{\tiny B}}T)$, and the (external) driving force, $f$, which may result from various sources. We assume that the diffusion and energy landscapes, $D(z)$ and $G(z)$, are piecewise homogeneous, cf. Fig. 2(b), precisely $\displaystyle D(z)=\mathopen{}\mathclose{{}\left\\{\begin{array}[]{l l}D_{\text{\tiny in}}&\frac{L-d}{2}\leq z\leq\frac{L+d}{2},\\\ D_{0}&\text{elsewhere},\\\ \end{array}}\right.$ (4) and $\displaystyle G(z)=\mathopen{}\mathclose{{}\left\\{\begin{array}[]{l l}G_{\text{\tiny in}}&\frac{L-d}{2}\leq z\leq\frac{L+d}{2},\\\ G_{0}&\text{elsewhere},\\\ \end{array}}\right.$ (7) where the subscripts ‘$0$’ and ‘in’ refer to the regions outside and inside the membrane, respectively. Figure 2: (a): System setup showing a membrane (red) of width $d$ in $z$-direction (periodic in $x$ and $y$) in the center of the system of size $L$, and an example solute concentration profile $c(z)$ (blue) in a steady state with external driving, $f>0$. The system is in contact with identical solute bulk reservoirs with constant concentration, $c(0)=c(L)=c_{0}$. The concentration in the boundary and membrane layers, described by the Smoluchowski framework, is determined by $f$, and the energy and diffusion landscapes, $G(z)$ and $D(z)$, depicted in (b). ### II.2 The membrane permeability The quantities $D_{\text{\tiny in}}$ and $\Delta G=G_{\text{\tiny in}}-G_{0}$ define the membrane permeability in the solution–diffusion picture,[59, 60, 61, 62] $\displaystyle\mathcal{P}_{\text{\tiny mem}}=D_{\text{\tiny in}}\mathcal{K},$ (8) where $\mathcal{K}=\exp(-\beta\Delta G)$ is the equilibrium partitioning. The membrane permeability is a function of the polymer volume fraction, $\phi$, and depends on the solute–polymer interactions. For not too attractive solute-polymer interactions, the solute diffusivity inside the membrane is well described by Yasuda’s free-volume theory,[60] $\displaystyle D_{\text{\tiny in}}(\phi)=D_{0}\exp\mathopen{}\mathclose{{}\left(-A\frac{\phi}{1-\phi}}\right),$ (9) with $A$ a positive parameter accounting for steric solute–polymer effects. For dilute solute systems, the partitioning of the ideal solutes can be well approximated by[68] $\displaystyle\mathcal{K}=\exp\mathopen{}\mathclose{{}\left(-B\phi}\right),$ (10) where $B=2B_{2}v_{0}^{-1}$ is related to the second virial coefficient, $B_{2}$, rescaled by effective monomer volume $v_{0}$. In fact, there exist many extended versions of scaling laws for $D_{\text{\tiny in}}$ and $\mathcal{K}$ which take into account further microscopic details, such as the chemistry, shape and size of the solutes, the solvent and the membrane types, and the architecture of the polymer network.[69, 70, 68, 71, 72, 61, 73, 74, 75] However, we use Eqs. 9 and 10 to explain the feedback effects of responsive polymers at the simplest level of detail. Further, the presented framework assumes that the equilibrium findings for $\mathcal{K}$ and $D_{\text{\tiny in}}$ are also valid under moderate nonequilibrium conditions, i.e., they are independent of the flux or the force. The validity of this assumption was demonstrated in our preceding work with nonequilibrium coarse-grained simulations of membrane-bulk systems.[65] ### II.3 Solute flux and concentration inside membrane With known $\mathcal{P}_{\text{\tiny mem}}$ and $c(0)=c_{0}$, we solve Eq. 1 to obtain the concentration profile, yielding $\displaystyle c(z)=\mathopen{}\mathclose{{}\left[c_{0}-j\mathcal{I}(0,z)}\right]e^{-\beta(G(z)-fz)},$ (11) with $\displaystyle\mathcal{I}(0,z)=\int\limits_{0}^{z}\mathrm{d}y\frac{e^{\beta(G(y)-fy)}}{D(y)}.$ (12) Using $c(L)=c_{0}$, the flux can be expressed as[65] $\displaystyle j=D_{0}c_{0}\beta f\mathopen{}\mathclose{{}\left[1+\mathopen{}\mathclose{{}\left(\frac{D_{0}}{\mathcal{P}_{\text{\tiny mem}}}-1}\right)S(f)}\right]^{-1},$ (13) with $S(f)=\sinh(\beta fd/2)/\sinh(\beta fL/2)$, which determines the impact of $\mathcal{P}_{\text{\tiny mem}}/D_{0}$ in the low- and high-force limits (see Appendix A for further details). By reinserting Eq. 13, one obtains the solute concentration profile throughout the system (see Appendix B for full expressions). We are particularly interested in the mean solute concentration inside the membrane, $c_{\text{\tiny in}}:=\langle c(z)\rangle_{\text{\tiny in}}=d^{-1}\int_{\text{\tiny in}}\mathrm{d}z\ c(z)$, with $z_{\text{\tiny in}}\in\mathopen{}\mathclose{{}\left[(L-d)/2,(L+d)/2}\right]$, because it is the key stimulus for our membrane response model. After integration, we obtain $\displaystyle c_{\text{\tiny in}}(c_{0},f,\phi)=$ $\displaystyle c_{0}\mathcal{K}\ \frac{2\mathopen{}\mathclose{{}\left({D_{0}}-\mathcal{P}_{\text{\tiny mem}}}\right)S(f)\sinh\mathopen{}\mathclose{{}\left({\beta f(d-L)}/{2}}\right)+\beta fd{D_{0}}}{\beta fd\mathopen{}\mathclose{{}\left[({D_{0}}-\mathcal{P}_{\text{\tiny mem}})S(f)+\mathcal{P}_{\text{\tiny mem}}}\right]}.$ (14) The detailed behavior of Eq. 14 is described in the results section with an appropriate choice of the model parameters. ### II.4 The polymer response to the solute concentration The above theoretical framework is straightforward for membranes with constant $\phi$. We now extend the model to membranes that are responsive (in $\phi$) to $c_{\text{\tiny in}}$. Motivated by experimental, theoretical, and computational findings, [46, 47, 48, 49, 50, 51, 36, 76, 77, 78] we consider that the polymer networks undergo a sigmoidal volume phase transition in the vicinity of a crossover concentration $c_{\text{\tiny c}}$. We assume the following form $\displaystyle\phi(c_{\text{\tiny in}})=\phi_{\text{\tiny c}}\pm\frac{\Delta\phi}{2}\tanh\mathopen{}\mathclose{{}\left(\frac{c_{\text{\tiny in}}-c_{\text{\tiny c}}}{\Delta c}}\right),$ (15) where $\Delta\phi=(\phi_{\text{\tiny max}}-\phi_{\text{\tiny min}})$ is the maximum change, and $\phi_{\text{\tiny c}}=\phi(c_{\text{\tiny c}})=(\phi_{\text{\tiny min}}+\phi_{\text{\tiny max}})/{2}$ the polymer volume fraction at $c_{\text{\tiny c}}$. Hence, we call ($c_{\text{\tiny c}},\phi_{\text{\tiny c}}$) the crossover point. The transition may occur from a swollen state ($\phi_{\text{\tiny min}}$) to a collapsed state ($\phi_{\text{\tiny max}}$) or vice versa as $c_{\text{\tiny in}}$ increases (denoted by the ‘$\pm$’-symbol in Eq. 15), depending on the interactions between the solutes, the solvent, and the polymer. Effectively attractive solute-membrane interactions ($B<0$) are expected to cause a swollen-to- collapsed transition ($+$), while a transition from the collapsed to the swollen phase ($-$) is expected for repulsive interactions ($B>0$). Further, the parameter $\Delta c$ determines the sharpness of the transition, ranging from almost irresponsive ($\Delta c\gg c_{\text{\tiny c}}$) to very sharp transitions ($\Delta c\ll c_{\text{\tiny c}}$). Note that Eq. 15 assumes continuous transitions although hysteresis has been reported in experimental studies.[49, 77, 79, 80] However, this work will demonstrate that hysteresis and bistability can result from the mutual dependencies of $\phi$ and $c_{\text{\tiny in}}$. Furthermore, Eq. 15 is a mean-field approach as it does not resolve spatial inhomogeneities of $\phi$ and $c_{\text{\tiny in}}$ [cf. the example concentration profile in Fig. 2(a)]. We assume that the system is small and that the thin membranes do not change the width in the direction of the solute flux. Despite the multiple assumptions, our simplified model enables the investigation of the effect of a responsive membrane permeability on the transport. Table 1: Summary of the model parameters and the corresponding values for $\mathcal{K},D_{\text{\tiny in}}$, and $\mathcal{P}_{\text{\tiny mem}}$ at $\phi_{\text{\tiny min}}$, $\phi_{\text{\tiny max}}$, and $\phi_{\text{\tiny c}}$. Length scales are given in units of $\sigma$, the radius of one monomer. The transition width is rescaled by the crossover concentration $c_{\text{\tiny c}}$ [cf. Eq. 15]. Permeabilities and diffusivities are expressed in units of $D_{0}$, the solute bulk diffusion. The arrow ($\Rightarrow$) indicates that the presented values are direct consequences of the parameter choice. The approximate Lennard-Jones energy $\varepsilon_{\text{\tiny LJ}}$ stems from a comparison with the second virial coefficient, $B=2B_{2}v_{0}^{-1}$. The symbols ‘$+$’ and ‘$-$’ correspond the swollen-to-collapsed and the collapsed-to-swollen transition, respectively [see also Eq. 15]. polymer response [Eq. 15] --- | $\phi_{\text{\tiny min}}$ | $0.05$ | swollen | $\phi_{\text{\tiny max}}$ | $0.35$ | collapsed $\Rightarrow$ | $\phi_{\text{\tiny c}}$ | $0.2$ | | $\Rightarrow$ | $\Delta\phi$ | $0.3$ | | | | sharp | gradual | weak | $\Delta c/c_{\text{\tiny c}}$ | $0.1$ | $1.0$ | $10.0$ lengths (Fig. 2) | $L/\sigma$ | $100$ | system size | $d/\sigma$ | $90$ | membrane width solute diffusion inside membrane [Eq. 9] | $A$ | $5$ | | $\Rightarrow$ | ${D_{\text{\tiny in}}(\phi_{\text{\tiny min}})}/{D_{0}}$ | $0.77$ | | $\Rightarrow$ | ${D_{\text{\tiny in}}(\phi_{\text{\tiny c}})}/{D_{0}}$ | $0.27$ | | $\Rightarrow$ | ${D_{\text{\tiny in}}(\phi_{\text{\tiny max}})}/{D_{0}}$ | $0.07$ | | solute–membrane interactions and partitioning [Eq. 10] | | repulsive | weakly attr. | attractive | $B$ | $5.26$ | $-6.25$ | $-17.8$ $\Rightarrow$ | $\beta\varepsilon_{\text{\tiny LJ}}$ (approx.) | $0.03$ | $0.55$ | $0.9$ $\Rightarrow$ | ${\mathcal{K}(\phi_{\text{\tiny min}})}$ | $0.77$ | $1.37$ | $2.4$ $\Rightarrow$ | ${\mathcal{K}(\phi_{\text{\tiny c}})}$ | $0.35$ | $3.49$ | $34.9$ $\Rightarrow$ | ${\mathcal{K}(\phi_{\text{\tiny max}})}$ | $0.16$ | $8.91$ | $501.2$ $\Rightarrow$ | transition [Eq. 15] | $-$ | $+$ | $+$ membrane permeability [Eq. 8] | | repulsive | neutral | attractive $\Rightarrow$ | ${\mathcal{P}_{\text{\tiny mem}}(\phi_{\text{\tiny min}})}/{D_{0}}$ | $0.59$ | $1.05$ | $1.9$ $\Rightarrow$ | ${\mathcal{P}_{\text{\tiny mem}}(\phi_{\text{\tiny c}})}/{D_{0}}$ | $0.10$ | $1.00$ | $10.0$ $\Rightarrow$ | ${\mathcal{P}_{\text{\tiny mem}}(\phi_{\text{\tiny max}})}/{D_{0}}$ | $0.01$ | $0.60$ | $33.9$ ### II.5 Model parameters All length scales are expressed in units of $\sigma$, the effective diameter of one monomer. We set the system size to $L=100\sigma$ and fix the membrane width to $d=90\sigma$, i.e., the boundary layers between the membrane and the two bulk reservoirs with concentration $c_{0}$ have the width $5\sigma$. The concentrations $c_{0}$, $c_{\text{\tiny in}}$ and the transition width $\Delta c$ are rescaled by the crossover concentration $c_{\text{\tiny c}}$ of the volume phase transition [Eq. 15]. We choose three different transition widths, $\Delta c/c_{\text{\tiny c}}\in\mathopen{}\mathclose{{}\left\\{0.1,1.0,10}\right\\}$. The force, $\beta f\sigma$, is rescaled by the thermal energy and the solute size. The solute diffusivity inside the membrane, $D_{\text{\tiny in}}$, and the permeability, $\mathcal{P}$, are expressed in units of the solute bulk diffusivity $D_{0}$. The parameters $A$ and $B$, which enter $D_{\text{\tiny in}}(\phi)$ [Eq. 9] and $\mathcal{K}(\phi)$ [Eq. 10], respectively, as well as the limits of the polymer volume fraction, $\phi_{\text{\tiny min}}=0.05$ and $\phi_{\text{\tiny max}}=0.35$, are based on our group’s coarse-grained simulations.[68, 70] We fix $A=5$, assuming that the diffusion is dominated by steric exclusion. The interaction parameter $B$ is chosen in a way to yield three different values for the equilibrium membrane permeability at $\phi_{\text{\tiny c}}=0.2$, and, hence, we denote the membranes as repulsive ($\mathcal{P}_{\text{\tiny mem}}(\phi_{\text{\tiny c}})/D_{0}=0.1$), neutral ($\mathcal{P}_{\text{\tiny mem}}(\phi_{\text{\tiny c}})/D_{0}=1.0$), and attractive ($\mathcal{P}_{\text{\tiny mem}}(\phi_{\text{\tiny c}})/D_{0}=10.0$). In fact, due to typical cancelling effects of $\mathcal{K}(\phi)$ and $D_{\text{\tiny in}}(\phi)$,[68, 70] we can safely assume that the permeability of our neutral membrane does not significantly deviate from unity throughout the range of $\phi$. In a system with Lennard-Jones (LJ) interactions between the solutes and the membrane monomers of equal size, the characteristic LJ interactions strengths would take the approximate values $\beta\varepsilon\approx 0.03$ (repulsive), $\beta\varepsilon\approx 0.55$ (weakly attractive), and $\beta\varepsilon\approx 0.9$ (attractive), respectively. All parameter values and relevant quantities are summarized in Table 1. ## III Results and Discussion Figure 3: Mean solute concentration inside the membrane [Eq. 14] as a function of $\phi$ for different values of the driving force, $f$ [color- coded, see colorbar in panel (b)], and different interaction strengths, $B\in\mathopen{}\mathclose{{}\left\\{5.26,-6.25,-17.8}\right\\}$, which correspond to a repulsive [(a): $\mathcal{P}_{\text{\tiny mem}}(\phi_{\text{\tiny c}})=0.1D_{0}$], neutral [(b): $\mathcal{P}_{\text{\tiny mem}}(\phi_{\text{\tiny c}})=1.0D_{0}$], and attractive membrane [(c): $\mathcal{P}_{\text{\tiny mem}}(\phi_{\text{\tiny c}})=10D_{0}$], respectively, as indicated above the panels. The blue and the red dotted line represent the zero and the infinite force limits, $\lim_{f\to 0}c_{\text{\tiny in}}/c_{0}=\mathcal{K}(\phi)$ [Eq. 10] and $\lim_{f\to\infty}c_{\text{\tiny in}}/c_{0}=D_{0}/D_{\text{\tiny in}}(\phi)$ [Eq. 9], respectively. While for the repulsive membrane [panel (a)] the $c_{\text{\tiny in}}$ increases with an increase in force, it decreases for the case of the attractive membranes [panel (c)]. The solute concentration in the neutral membrane [panel (b)] depends on $f$, yet is essentially a function of $\phi$. ### III.1 Force-controlled solute uptake From Eq. 14, the low- and high-force limits for the mean solute concentration inside the membrane, $c_{\text{\tiny in}}$, can be deduced, which has been discussed and substantiated with concentration profiles from theory and coarse-grained simulations in our previous work.[65] Here, we recapture the main findings and discuss the results for the parameters used in this work. In Fig. 3, we depict $c_{\text{\tiny in}}$ [Eq. 14] as a function of $\phi$ for different values of $\beta f\sigma\in\mathopen{}\mathclose{{}\left[0.01,10}\right]$ (color-coded) and for three different values of $\mathcal{P}_{\text{\tiny mem}}(\phi_{c})/D_{0}\in\mathopen{}\mathclose{{}\left\\{0.1,1.0,10.0}\right\\}$ (different panels). In the zero force limit, $c_{\text{\tiny in}}$ reduces to the expected equilibrium value, $\lim_{f\to 0}c_{\text{\tiny in}}=c_{0}\mathcal{K}(\phi)$, which monotonously decreases for the repulsive membrane [panel (a)] and increases otherwise [panels (b) and (c)]. The same limiting result is obtained, if $\mathcal{P}_{\text{\tiny mem}}(\phi)=D_{0}\ \forall\phi$, which applies approximately for the ‘neutral’ membrane [panel (b)]. In the high-force limit, the concentration profiles become piecewise constant with $\lim_{f\to\infty}c_{\text{\tiny in}}=c_{0}D_{0}D_{\text{\tiny in}}^{-1}(\phi)$, and one further finds $\lim_{f\to\infty}j=c_{0}\beta fD_{0}=c_{\text{\tiny in}}\beta fD_{\text{\tiny in}}$ for all membrane types, since it is independent of $\mathcal{K}$. The solute uptake of the repulsive membrane at fixed $\phi$ increases with $f$, while it decreases in the attractive membrane [see Figs. 3(a) and 3(c) ]. For ‘neutral’ membrane [panel (b)], $c_{\text{\tiny in}}$ shows no significant force dependence. Figure 4: Phase plane showing $\phi(c_{\text{\tiny in}})$ [Eq. 15] and $c_{\text{\tiny in}}(\phi,c_{0},f)$ [Eq. 14]. The color-coded lines [see colorbar in panel (c)] depict $c_{\text{\tiny in}}(\phi,c_{0},f)$ in the attractive membrane with $\mathcal{P}(\phi_{\text{\tiny c}})=10D_{0}$ (cf. Fig. 3(c)), for three different bulk concentrations $c_{0}$ as indicated above the panels. The black lines depict the (swollen-to-collapsed) transition function $\phi(c_{\text{\tiny in}})$ [Eq. 15] for three different values of the transition sharpness $\Delta c$ [see legend in panel (a)]. Each interception point of one colored line and one black line refers to a steady- state solution ($c_{\text{\tiny in}}^{*},\phi^{*}$) that depends on $c_{0}$, $\Delta c$ and $f$. The solutions, $\phi^{*}(f)$, are summarized in Fig. 5. In panel (a), we show $c_{0}=c_{\text{\tiny c}}{D_{\text{\tiny in}}(\phi_{\text{\tiny c}})}/{D_{0}}$, i.e., the high-force limit (red dotted line) intercepts with the crossover point. In panel (c), we show $c_{0}=c_{\text{\tiny c}}/{\mathcal{K}(\phi_{\text{\tiny c}})}$, and the zero force limit (blue dotted line) intercepts with the crossover point $(c_{\text{\tiny c}},\phi_{\text{\tiny c}})$. In panel (b) the geometric mean of the two limits is chosen, i.e., $c_{0}=c_{\text{\tiny c}}\sqrt{{D_{\text{\tiny in}}(\phi_{\text{\tiny c}})}/\mathopen{}\mathclose{{}\left({D_{0}\mathcal{K}(\phi_{\text{\tiny c}})}}\right)}$. In this phase plane, $c_{0}$ performs a horizontal shift of $c_{\text{\tiny in}}/c_{\text{\tiny c}}$. Figure 5: Steady-state solutions of the polymer volume fraction, $\phi^{*}$, for attractive membranes ($\mathcal{P}(\phi_{\text{\tiny c}})=10D_{0}$) as function of the external driving force, $f$. The columns differ in terms of the bulk concentration, i.e., $c_{0}=c_{\text{\tiny c}}{D_{\text{\tiny in}}(\phi_{\text{\tiny c}})}/{D_{0}}$ (panels in the left column), $c_{0}=c_{\text{\tiny c}}\sqrt{{D_{\text{\tiny in}}(\phi_{\text{\tiny c}})}\mathopen{}\mathclose{{}\left({D_{0}\mathcal{K}(\phi_{\text{\tiny c}})}}\right)}$ (central column), $c_{0}=c_{\text{\tiny c}}/{\mathcal{K}(\phi_{\text{\tiny c}})}$ (right column). Each row refers to one value of the transition sharpness, $\Delta c$ (see labels right of rows). In general, the force $f$ tunes $\phi^{*}$ from high ($\phi_{\text{\tiny max}}$) to low ($\phi_{\text{\tiny min}}$) values (since $c_{\text{\tiny in}}(f)$ decreases for attractive membranes). One observes regions of multiple steady states (with two stable branches and one unstable branch) which may occur in the entire force range [e.g., panels (g) and (h)]. ### III.2 Multiple steady-state solution With Eqs. 15 and 14 the feedback loop depicted in Fig. 1 is closed. We obtain numerically the steady-state solutions, $(c_{\text{\tiny in}}^{*},\phi^{*},)$, by finding the intersection points of $c_{\text{\tiny in}}(\phi,f)$ and $\phi(c_{\text{\tiny in}})$ in the phase plane. In this section we show the results with the attractive membrane only and demonstrate the general procedure. (For the repulsive membrane, we show a representative phase plane in Fig. 7 in Appendix C.) In Fig. 5, the black lines depict the polymer’s volume phase transition of $\phi(c_{\text{\tiny in}})$, Eq. 15, for three different values of $\Delta c/c_{\text{\tiny c}}\in\mathopen{}\mathclose{{}\left\\{0.1,1,10}\right\\}$. The colored lines, $c_{\text{\tiny in}}(c_{0},f,\phi)$, Eq. 14, are the inverted images of Fig. 3(c) and are plotted in Fig. 5 for three different bulk concentrations, $c_{0}$, from high [panel (a)] to low values [panel (c)]. As obvious, changing $c_{0}$ performs a shift of $c_{\text{\tiny in}}(c_{0},f,\phi)$ along the horizontal axis. In panel (a), we choose $c_{0}=c_{\text{\tiny c}}D_{\text{\tiny in}}(\phi_{c})/D_{0}\approx 0.29$ such that the high-force limit of $c_{\text{\tiny in}}$ intercepts with the crossover point $(c_{\text{\tiny c}},\phi_{\text{\tiny c}})$. In panel (c), we impose that the low-force limit intercepts with the crossover point, i.e., $c_{0}=c_{\text{\tiny c}}/(\mathcal{K}(\phi_{c})\approx 0.03$. In panel (b), the geometric mean, $c_{0}=c_{\text{\tiny c}}\sqrt{D_{\text{\tiny in}}(\phi_{c})/(D_{0}\mathcal{K}(\phi_{\text{\tiny c}}))}\approx 0.09$ is used as intermediate probe concentration. For fixed $f$, $c_{0}$ and $\Delta c$, we find one or three interception points of $c_{\text{\tiny in}}(c_{0}f,\phi)$ and $\phi(c_{\text{\tiny in}})$, yielding the steady-state solutions, $\mathopen{}\mathclose{{}\left(c_{\text{\tiny in}}^{*},\phi^{*}}\right)$. In Fig. 5(b), for instance, the low-force limit (blue dotted line) intercepts with the black solid line ($\Delta c/c_{\text{\tiny c}}=1$) only once at $\phi^{*}\approx\phi_{\infty}$, while it has three interceptions with the black dotted line ($\Delta c/c_{\text{\tiny c}}=0.1$) at $\phi_{1}^{*}\approx\phi_{\text{\tiny min}}$, $\phi_{2}^{*}\approx 0.15$, and $\phi_{3}^{*}\approx\phi_{\text{\tiny max}}$. In the case of triple solutions, the intermediate one is an unstable solution, while the other two are stable solutions. Precisely, the latter correspond to asymptotically stable solutions of the time-dependent Smoluchowski equation, $\dot{c}(z,t)=-\partial j(z)/\partial z$, i.e., the steady-state solution, $c^{*}(z)$, is restored after a small perturbation. A more detailed discussion on the stability of multiple solutions and consequences for the bistable domains is provided in a separate section below. We summarize the steady-state solutions for the attractive membrane by plotting $\phi^{*}(f)$ for different $c_{0}$, and $\Delta c$ in Fig. 5. We observe a swelling (decrease in $\phi$) with hysteresis due to an increase in $f$ [see panels (a), (b), and (e)]. In more detail, higher $c_{0}$ can shift the force-induced $\phi$-transition to higher force values [e.g., compare panels (a) and (b)] and whether transitions may occur at all. For instance, as in the case of low transition sharpness, $\Delta c/c_{\text{\tiny c}}=10$, and low solute concentration [panel (c)], there is no significant effect on $\phi^{*}$. Similarly for the moderate sharpness, $\Delta c/c_{\text{\tiny c}}=1$, no transition is induced if $c_{0}$ is too high [panel (d)]. Further, $\Delta c$, plays an important role as it tunes the width of the bistable domains, e.g., while only small force ranges with bistability are observed for weakly responsive membranes ($\Delta c/c_{\text{\tiny c}}=10$), see Figs. 5(a) and 5(b), it can exist in the entire force range for sufficiently sharp transitions ($\Delta c/c_{\text{\tiny c}}=0.1$), see Figs. 5(g) and 5(h). We already conclude that the membrane’s feedback response can lead to large bistable domains in $\phi$ tuned by $f$ and $c_{0}$, which is characterized by drastic switching of the membrane properties, such as the permeability, due to the bifurcations at the critical values. attractive membrane repulsive membrane Figure 6: Force-dependent permeability and flux of responsive membranes undergoing a very sharp volume phase transition with $\Delta c=0.1c_{\text{\tiny c}}$. All panels on the left (right) hand side show the results for the attractive (repulsive) membrane. Top panels (a) and (h) depict the system’s differential permeability, $\mathcal{P}_{\text{\tiny sys}}^{\Delta}/D_{0}$, as heatmaps in the $f$-$c_{0}$ plane. The heatmaps share the same color-code ranging from $10^{-2}$ to $10^{1}$ (see colorbar). The white lines labeled with roman numbers, I-VI, depict selected values of $c_{0}$, for which $j$ and $\mathcal{P}_{\text{\tiny sys}}^{\Delta}$ are presented in the panels below. The black dotted lines indicate the bifurcation at which the system changes from mono- to bi-stable (or vice versa), while the two solutions in the bistable domain are presented in a striped pattern. In examples I-VI, the solutions are distinguished by the membrane’s volume phase, i.e., blue corresponds to $\phi<\phi_{\text{\tiny min}}$ (swollen), and red to $\phi>\phi_{\text{\tiny max}}$ (collapsed). In fact, we find $\phi$ is either fully swollen or collapsed except for example IV, where a gradual crossover from $\phi(f=0)=\phi_{\text{\tiny max}}$ to $\phi(f\to\infty)\approx 0.15$ is observed [cf. panel (h), where the loosely dotted line indicates $\phi=\phi_{\text{\tiny c}}$]. The pale red and blue dotted lines in I-VI are the references for nonresponsive membranes in the fully collapsed and swollen case, respectively. The gray dashed lines in I-VI correspond to the bulk references, i.e., $j=D_{0}c_{0}\beta f$ and $\mathcal{P}_{\text{\tiny sys}}=D_{0}$, respectively, which yield the asymptotic values for $f\to\infty$ and $\phi\to 0$. More details are provided in the main text. ### III.3 Consequences for the transport properties The flux $j(f)$, given by Eq. 13, is a nonlinear function of $f$ determined by two contributions: The change in the membrane permeability $\mathcal{P}_{\text{\tiny mem}}(\phi)$ (due to the change in $\phi$) and the spatial setup (see Appendix A and our previous work[65]). The nonlinear characteristics of $j(f)$ are quantified by the _differential_ system permeability,[65] defined as $\displaystyle\mathcal{P}^{\text{\tiny$\Delta$}}_{\text{\tiny sys}}(f)=\frac{1}{\beta c_{0}}\frac{\mathrm{d}j}{\mathrm{d}f},$ (16) which describes the change in the steady-state flux induced by a change in the external driving force. We make use of $\mathcal{P}^{\text{\tiny$\Delta$}}_{\text{\tiny sys}}(f)$ to highlight the novel nonlinear effects on $j(f)$ caused by the membrane’s feedback response. We limit ourselves to the very sharp membrane response ($\Delta c/c_{\text{\tiny c}}=0.1$), and point out the significant difference between the fluxes in attractive and repulsive membranes. In Fig. 6, we present heatmaps of $\mathcal{P}^{\text{\tiny$\Delta$}}_{\text{\tiny sys}}$ in the $f$-$c_{0}$ plane for the attractive [panel (a)] and the repulsive membrane [panel (h)]. The white lines labeled with roman numerals depict selected values of $c_{0}$, and correspond to the panels below, showing $j$ and $\mathcal{P}^{\text{\tiny$\Delta$}}_{\text{\tiny sys}}$. The heatmaps share the same color-scale (see colorbars), allowing a direct comparison between the results of attractive and repulsive membranes. The attractive membrane exhibits, in general, larger $\mathcal{P}^{\text{\tiny$\Delta$}}_{\text{\tiny sys}}$ values than the repulsive one, particularly in the low-force and collapsed regime ($\phi=\phi_{\text{\tiny max}}$), in which the influence of $\mathcal{P}_{\text{\tiny mem}}$ is the greatest. For $f\to 0$, we find $\mathcal{P}^{\text{\tiny$\Delta$}}_{\text{\tiny sys}}\approx 7D_{0}$ and $\mathcal{P}^{\text{\tiny$\Delta$}}_{\text{\tiny sys}}\approx 0.01D_{0}$ for the attractive and repulsive membrane in the collapsed state, respectively. In the high-force limit, the system permeability converges to $D_{0}$ irrespective of the volume phase. If the membrane is swollen ($\phi=\phi_{\text{\tiny min}}$), the permeability of the repulsive and the attractive membrane are of the same order of magnitude, i.e., $\mathcal{P}_{\text{\tiny mem}}(\phi_{\text{\tiny min}})\approx 0.6D_{0}$ and $\mathcal{P}_{\text{\tiny mem}}(\phi_{\text{\tiny min}})\approx 2D_{0}$, respectively, and $\mathcal{P}^{\text{\tiny$\Delta$}}_{\text{\tiny sys}}$ does not deviate significantly from bulk diffusivity $D_{0}$, even for low forces (compare blue lines in lower panels of Fig. 6). Due to the sharp response with $\Delta c=0.1c_{c}$, the membrane is either fully swollen ($\phi_{\text{\tiny min}}$), or fully collapsed ($\phi_{\text{\tiny max}}$). Moreover, this also leads to large bistable domains in the $c_{0}$-$f$ plane, visualized as striped patterns in Figs. 6(a) and 6(h). Crossing the boundary of these domains leads to a discontinuous volume phase transitions accompanied with an order of magnitude change in the solute flux (examples I, III, V, and VI). In the case of the repulsive membrane, the flux can be switched even by two orders of magnitude, particularly for small $f$. In examples I-VI [Figs. 6(b)–6(g), and 6(i)–6(n)], we also depict the results for nonresponsive membranes in the fully swollen and collapsed case for comparison. In the nonresponsive case, $\mathcal{P}^{\text{\tiny$\Delta$}}_{\text{\tiny sys}}$ can also be tuned in the same range by only controlling $f$. Nonetheless, the membrane’s responsiveness brings about more dramatic effects, such as bistability (I, III, V, VI) and hysteresis (V), yielding new control mechanisms to switch between two flux states. While nonresponsive membranes require large forces to exhibit bulk-like properties ($D_{0}$), a transition to this neutral state can be achieved in responsive membranes in a much sharper fashion and for lower force values, e.g., see Fig. 6 I, V, and VI. In V, for example, the crossover from the low- to the high-permeability state occurs abruptly around $\beta f\sigma\approx 0.2$ (red solid line), whereas for the collapsed, nonresponsive membrane a gradual change is observed in the range $\beta f\sigma\approx 0.1-1.0$. Further, even if the polymer volume phase transition occurs without bifurcation, nonlinearities in the force–flux relations can be significantly enhanced. For instance, in panel (l) (example IV), we find a tenfold maximization of $\mathcal{P}^{\text{\tiny$\Delta$}}_{\text{\tiny sys}}\approx 12D_{0}$ at roughly $\beta f\sigma\approx 0.5$, which even exceeds the maximum differential system permeability measured for the attractive membrane. ### III.4 Discussion: nonequilibrium steady-state stability Our model results into well-defined force–flux relations in the domains with unique steady-state solutions. In the bistable domains, however, the question arises whether the states coexist or whether only one survives under real conditions. This far, the solutions were simply deduced from a deterministic interpretation of the macroscopic model equations, i.e., by evaluating the self-consistency equation $c_{\text{\tiny in}}=R(c_{\text{\tiny in}})$, with $R(c_{\text{\tiny in}})=d^{-1}\int_{\text{\tiny in}}c(z,\phi(c_{\text{\tiny in}}))\mathrm{d}z$. The steady state is asymptotically stable, if ${\mathrm{d}R(c_{\text{\tiny in}})}/{\mathrm{d}c_{\text{\tiny in}}}|_{c_{\text{\tiny in}}^{*}}<1$.[81] However, our approach neglects larger fluctuations in $\phi$ and $c_{\text{\tiny in}}$, and does not analyze further nonequilibrium extremum principles.[82] In the following, we first discuss the consequences of the deterministic perspective, and then briefly review alternative interpretations. In the case of negligible fluctuations the (deterministic) transition between states can be either _reversible_ or _irreversible_.[35] One _reversible_ transition is example V [Figs. 6(j) and 6(m)]. Here, the membrane is in the collapsed state (red line) for small $f$. With increasing $f$, the membrane is driven into the bistable regime, yet remains in the collapsed state. Only if $f$ exceeds the second bifurcation line, the membrane swells. In the same example V, if $f$ is decreased from high force values, the membrane stays swollen in the bistable domain and returns to the collapsed state until the first bifurcation line is passed. Hence, we find a _reversible_ transition with hysteresis between the two state in V. In contrast, consider example I or VI, and assume that the membrane is in the collapsed state at $f=0$, an increase in $f$ leads to a swelling when the bifurcation line is crossed. Decreasing the force again, however, does not induce a collapse, and, hence, this transition can be termed _irreversible_ in the deterministic interpretation. This is because only the swollen case survives once the threshold is surpassed. Analogously, see example III, a collapsed-to-swollen transition cannot be induced by increasing $f$. Although two stable states may coexist in the deterministic model, one of them could be metastable and practically unoccupied under experimental conditions. In literature one finds nonequilibrium principles, e.g., based on the maximization of entropy, the minimization of entropy production (least dissipation), or the minimization of power, providing various possible routes.[83, 84, 85, 86, 31, 87, 88, 89, 82, 90] Such extremum principles may lead to unique solutions in the _bistable_ regime, and to different values for $f$ and $c_{0}$, where the switching between the high and low flux states occurs. For example, it should be the minimum flux, if the least-dissipation principle applies. This has direct consequences on the flux–force relation and the critical transition values of $f$ and $c_{0}$, where the phase transition in V would occur always at the first bifurcation line, i.e., without bistability and hysteresis. Furthermore, the presented diffusion process can also be modeled with the stochastic Smoluchowski equation,[91] and possibly further coarse-grained to a stochastic differential equation for $\dot{c}_{\text{\tiny in}}$.[92, 93] Hence, given the fluctuations are large enough, a stochastic switching between the two steady states may be observed in the bistable domains, and the effective force–flux relations are determined by the averaged values of $\phi$ and $c_{\text{\tiny in}}$. Consequently, changing $f$ results in a continuous transition between the two states, implying that example V does not exhibit hysteresis behavior, but is rather similar to the transition in example IV. Ultimately, the appropriate stability interpretation remains to be verified, and is likely specific to the membrane material and the experimental nonequilibrium conditions. Nonetheless, a strong amplification of nonlinear characteristics and a critical switching in the force–flux relations can be expected due to the membrane’s responsiveness. ## IV Summary We have investigated the driven steady-state solute transport through polymeric membranes with a sigmoidal volume phase response to the penetrant uptake. The change in the polymer volume fraction is decisive for the membrane permeability, which we modeled with exponential functions. This, in turn, impacts on the solute uptake, leading to novel feedback-induced effects in force–flux relations that cannot be achieved by nonresponsive membranes. We quantified our findings in terms of the system’s differential permeability. The feedback effects of responsive membranes are most pronounced in the low- force regime, where the bulk concentration largely tunes the membrane density between the swollen and the collapsed state. Increasing the force can lead to a membrane swelling accompanied with a strong amplification of nonlinear characteristics and critical switching in the force–flux relations. For instance, the swelling of membranes with repulsive polymer-solute interactions can be caused by a small change in the driving force, for which we report an increase in the flux by two orders of magnitude, and a pronounced maximization of the differential permeability, i.e., a tenfold increase compared to the case of nonresponsive membranes. Moreover, of particular note is the feedback-induced coexistence of two stable steady states, while the size of the bistable domains increases with the sharpness of the sigmoidal polymer response. The bifurcations from mono- to bistability occur at critical values of the driving force, and the solute bulk concentration, leading to discontinuous changes in the flux of up to two orders of magnitude. Thus, the force-dependent switching between high and low flux states provides a valuable control mechanism for molecular transport. It can be fine-tuned also to control the appearance of hysteresis, enabling the presented feedback membranes to function as memristive devices. Moreover, the coupling of the permeability hysteresis to (non-oscillatory) chemical reactions may lead to biomimetic features, such as membrane excitability and autonomous oscillations, as first proposed by theory,[35, 63, 31, 32] and eventually validated by experiments.[36, 37, 38] Hysteresis transitions found in literature[36, 37, 38, 80, 49, 77, 79] are usually rationalized by a bistability in the polymer’s conformational free energy,[94, 95, 79] and attributed to the complex microscopic interactions or the competition between entropic and energetic contributions.[94, 95, 79] Nonetheless, many polymers exhibit a hysteresis-free response, for which the presented feedback mechanism provides a novel explanation of how hysteresis transitions can be generated and tuned in polymer membranes. We disclosed in this work how nonlinear solute transport through chemo- responsive polymer membranes is controlled by membrane feedback. It thus provides the theoretical basis for the rational design of self-regulating membranes with nonlinear control features for molecular transport. Adaptations employing more complex functions for partitioning, diffusivity, and the polymer response, to differing feed and permeate bulk concentrations as well as extensions to different spatial arrangements and geometries could be interesting for future studies. ## Acknowledgments The authors thank Matej Kanduč for fruitful discussions. This work was supported by the Deutsche Forschungsgemeinschaft (DFG) via the Research Unit FOR 5099 “Reducing complexity of nonequilibrium systems”. W.K.K. acknowledges the support by the KIAS Individual Grants (CG076001 and CG076002) at Korea Institute for Advanced Study. ## Author declarations ### Conflict of Interest The authors have no conflicts to disclose. ## Appendix A The flux in the low- and high-force limits The flux, $j$ [Eq. 13], in the small-force regime reads[65] $\displaystyle\lim_{f\to 0}j=D_{0}\beta fc_{0}\mathopen{}\mathclose{{}\left[1+\mathopen{}\mathclose{{}\left(\frac{D_{0}}{\mathcal{P}_{\text{\tiny mem}}}-1}\right)\frac{d}{L}}\right]^{-1},$ (17) which converges to $\lim_{f\to 0}j\to\mathcal{P}_{\text{\tiny mem}}c_{0}\beta f$ for $d\to L$. The membrane thickness, $d$, determines the crossover to the high-force regime, for which $\lim_{f\to\infty}j=D_{0}c_{0}\beta f$ results. For moderate to large forces, we find $S(f)\approx\exp(-\beta f(L-d)/2)$. With increasing $f$, the denominator in Eq. 13 converges to unity, governed by $(L-d)$. So, the larger $d$ with respect to $L$, the higher $f$ has to be in order to reach the high-force limit. Obviously, the onset of the high-force limit also depends on the membrane permeability, precisely, large values of $\mathcal{P}_{\text{\tiny mem}}$ will shift the crossover to smaller force values. ## Appendix B Concentration profiles The solute concentration profile in a system depicted in described in Sec. II reads[65] $\displaystyle c(z)=c_{0}\mathopen{}\mathclose{{}\left[1-\dfrac{D_{0}\beta f\mathcal{I}(0,z)}{1+\mathopen{}\mathclose{{}\left(\frac{D_{0}}{\mathcal{P}_{\text{\tiny mem}}}-1}\right)S(f)}}\right]e^{-\beta\mathopen{}\mathclose{{}\left(G(z)-fz}\right)},$ (18) We can split Eq. 18 into the piecewise homogeneous layers, precisely, the feed boundary, membrane (‘in’), and permeate boundary layer, and can write the respective full expressions as $\displaystyle c(z)|_{\text{\tiny feed}}=$ $\displaystyle c_{0}\mathcal{K}\ \dfrac{e^{\beta fz}\mathopen{}\mathclose{{}\left({D_{0}}-\mathcal{P}_{\text{\tiny mem}}}\right)S(f)+{P_{\text{\tiny mem}}}}{({D_{0}}-\mathcal{P}_{\text{\tiny mem}})S(f)+\mathcal{P}_{\text{\tiny mem}}},$ (19) $\displaystyle c(z)|_{\text{\tiny in}}=$ $\displaystyle c_{0}\mathcal{K}\ \dfrac{\frac{e^{\beta f\mathopen{}\mathclose{{}\left(z-L/2}\right)}}{\sinh(\beta fL/2)}\mathopen{}\mathclose{{}\left({D_{0}}-\mathcal{P}_{\text{\tiny mem}}}\right){\sinh\mathopen{}\mathclose{{}\left(\beta f\frac{d-L}{2}}\right)}+{D_{0}}}{({D_{0}}-\mathcal{P}_{\text{\tiny mem}})S(f)+\mathcal{P}_{\text{\tiny mem}}},$ (20) $\displaystyle c(z)|_{\text{\tiny permeate}}=$ $\displaystyle c_{0}\mathcal{K}\ \dfrac{e^{\beta f(z-L)}\mathopen{}\mathclose{{}\left({D_{0}}-\mathcal{P}_{\text{\tiny mem}}}\right)S(f)+{P_{\text{\tiny mem}}}}{({D_{0}}-\mathcal{P}_{\text{\tiny mem}})S(f)+\mathcal{P}_{\text{\tiny mem}}}.$ (21) ## Appendix C Phase plane for repulsive membranes Figure 7: Phase plane showing $\phi(c_{\text{\tiny in}})$ [Eq. 15] and $c_{\text{\tiny in}}(\phi,c_{0},f)$ [Eq. 14]. The color-coded lines (see colorbar) depict $c_{\text{\tiny in}}(\phi,c_{0},f)$ for the repulsive membrane ($\mathcal{P}(\phi_{\text{\tiny c}})=0.1D_{0}$, cf. Fig. 3(a)), with selected probe concentration $c_{0}$. The black lines depict the (collapsed- to-swollen) transition function $\phi(c_{\text{\tiny in}})$ [Eq. 15] for three different values of the transition sharpness $\Delta c$ (see legend). Each interception point of a colored line [Eq. 14] and a black line [Eq. 15] refers to a steady-state solution ($c_{\text{\tiny in}}^{*},\phi^{*}$) that depends on $c_{0}$, $\Delta c$ and $f$. In Fig. 7, $c_{\text{\tiny in}}(c_{0},f,\phi)$ [Eq. 14] and $\phi(c_{\text{\tiny in}})$ [Eq. 15] are presented in the $c_{0}$-$\phi$ plane. Interception points of $c_{\text{\tiny in}}$ and $\phi$ correspond to the force-dependent steady-state solution. The results were used to calculate $\phi^{*}(c_{0},f)$, which enter the flux and the differential permeability as depicted in Figs. 6(h)–(n). The bulk concentration, $c_{0}$, and the external force, $f$, yield, in general, an increase in the mean inside concentration, $c_{\text{\tiny in}}$, resulting in a swelling of the membrane. For very sharp transitions, e.g., $\Delta c=0.1c_{\text{\tiny c}}$, one can find multiple solutions for fixed $c_{0}$ and $f$, giving rise to bistability and hysteresis. ## References ## References * Shannon _et al._ [2008] M. A. Shannon, P. W. Bohn, M. Elimelech, J. G. Georgiadis, B. J. Mariñas, and A. M. Mayes, Nature 452, 301 (2008). * Geise _et al._ [2010] G. M. Geise, H.-S. Lee, D. J. Miller, B. D. Freeman, J. E. McGrath, and D. R. Paul, J. Polym. Sci. Part B Polym. Phys. 48, 1685 (2010). * Pendergast and Hoek [2011] M. M. Pendergast and E. M. Hoek, Energy Environ. Sci. 4, 1946 (2011). * Cuperus and Nijhuis [1993] F. P. Cuperus and H. H. Nijhuis, Trends Food Sci. Technol. 4, 277 (1993). * Kumar _et al._ [2013] P. Kumar, N. Sharma, R. Ranjan, S. Kumar, Z. F. Bhat, and D. K. Jeong, Asian-Australasian J. Anim. Sci. 26, 1347 (2013). * Renggli _et al._ [2011] K. Renggli, P. Baumann, K. Langowska, O. Onaca, N. Bruns, and W. Meier, Adv. Funct. Mater. 21, 1241 (2011). * Roa _et al._ [2017] R. Roa, W. K. Kim, M. Kanduč, J. Dzubiella, and S. Angioletti-Uberti, ACS Catal. 7, 5604 (2017). * Lu and Ballauff [2011] Y. Lu and M. Ballauff, Prog. Polym. Sci. 36, 767 (2011). * Kanduč _et al._ [2020] M. Kanduč, W. K. Kim, R. Roa, and J. Dzubiella, Mol. Syst. Des. Eng. 5, 602 (2020). * Li and Mooney [2016] J. Li and D. J. Mooney, Nat. Rev. Mater. 1, 1 (2016). * Brudno and Mooney [2015] Y. Brudno and D. J. Mooney, J. Control. Release 219, 8 (2015). * Moncho-Jordá _et al._ [2020] A. Moncho-Jordá, A. B. Jódar-Reyes, M. Kanduč, A. Germán-Bellod, J. M. López-Romero, R. Contreras-Cáceres, F. Sarabia, M. García-Castro, H. A. Pérez-Ramírez, and G. Odriozola, ACS Nano 14, 15227 (2020). * Mitchell _et al._ [2021] M. J. Mitchell, M. M. Billingsley, R. M. Haley, M. E. Wechsler, N. A. Peppas, and R. Langer, Nat. Rev. Drug Discov. 20, 101 (2021). * Stamatialis _et al._ [2008] D. F. Stamatialis, B. J. Papenburg, M. Girones, S. Saiful, S. N. M. Bettahalli, S. Schmitmeier, and M. Wessling, J. Memb. Sci. 308, 1 (2008). * Lee [2018] J. H. Lee, Biomater. Res. 22, 1 (2018). * Hodgkin and Huxley [1952] A. L. Hodgkin and A. F. Huxley, J. Physiol. 117, 500 (1952). * Henisch and Smith [1974] H. K. Henisch and W. R. Smith, Appl. Phys. Lett. 24, 589 (1974). * Yang _et al._ [2006] Y. Yang, J. Ouyang, L. Ma, R. J. H. Tseng, and C. W. Chu, Adv. Funct. Mater. 16, 1001 (2006). * Keener and Sneyd [2009] J. Keener and J. Sneyd, _Mathematical Physiology I: Cellular Physiology_ (Springer-Verlag New York, 2009) p. 547. * Lee Takhee and Chen Yong [2012] Lee Takhee and Chen Yong, Marerials Res. Soc. Bull. 37, 144 (2012). * Lei _et al._ [2014] Y. Lei, Y. Liu, Y. Xia, X. Gao, B. Xu, S. Wang, J. Yin, and Z. Liu, AIP Adv. 4, 077105 (2014). * Hasan _et al._ [2018a] M. S. Hasan, J. S. Najem, R. Weiss, C. D. Schuman, A. Belianinov, C. P. Collier, S. A. Sarles, and G. S. Rose, IEEE 13th Nanotechnol. Mater. Devices Conf. , 1 (2018a). * Minoura [1998] N. Minoura, Langmuir 14, 2145 (1998). * Hasan _et al._ [2018b] M. S. Hasan, C. D. Schuman, J. S. Najem, R. Weiss, N. D. Skuda, A. Belianinov, C. P. Collier, S. A. Sarles, and G. S. Rose, IEEE 13th Dallas Circuits Syst. Conf. , 1 (2018b). * Jo _et al._ [2010] S. H. Jo, T. Chang, I. Ebong, B. B. Bhadviya, P. Mazumder, and W. Lu, Nano Lett. 10, 1297 (2010). * Prodromakis and Toumazou [2010] T. Prodromakis and C. Toumazou, 2010 17th IEEE Int. Conf. Electron. Circuits Syst. , 934 (2010). * Frank [2005] T. Frank, _Nonlinear Fokker-Planck equations: fundamentals and applications_ , 1st ed. (Springer Science & Business Media, Heidelberg, 2005). * Giacomelli _et al._ [2012] G. Giacomelli, F. Marino, M. A. Zaks, and S. Yanchuk, Europhys. Lett. 99, 58005 (2012). * Gernert and Klapp [2015] R. Gernert and S. H. L. Klapp, Phys. Rev. E 92, 22132 (2015). * Holubec _et al._ [2022] V. Holubec, A. Ryabov, S. A. M. Loos, and K. Kroy, New J. Phys. 24, 023021 (2022). * Katchalsky and Spangler [1968] A. Katchalsky and R. Spangler, Q. Rev. Biophys. 1, 127 (1968). * Borckmans _et al._ [2007] P. Borckmans, P. De Kepper, A. R. Khokhlov, and S. Métens, _Chemomechanical Instabilities in Responsive Materials_ (Springer, 2007). * Teorell [1951] T. Teorell, Zeitschrift für Elektrochemie und Angew. Phys. Chemie 55, 460 (1951). * Teorell [1962] T. Teorell, Biophys. J. 2, 25 (1962). * Hahn, Ortoleva, and Ross [1973] H. S. Hahn, P. J. Ortoleva, and J. Ross, J. Theor. Biol. 41, 503 (1973). * Bell _et al._ [2021] D. J. Bell, D. Felder, W. G. von Westarp, and M. Wessling, Soft Matter 17, 592 (2021). * Zou and Siegel [1999] X. Zou and R. A. Siegel, J. Chem. Phys. 110, 2267 (1999). * Leroux and Siegel [1999] J. C. Leroux and R. A. Siegel, Chaos An Interdiscip. J. Nonlinear Sci. 9, 267 (1999). * Sparr _et al._ [2009] E. Sparr, C. Åberg, P. Nilsson, and W. Håkan, Soft Matter 5, 3225 (2009). * Costa-Balogh _et al._ [2005] F. O. Costa-Balogh, C. Åberg, J. J. Sousa, and E. Sparr, Langmuir 21, 10307 (2005). * Åberg and Wennerström [2009] C. Åberg and H. Wennerström, Phys. Chem. Chem. Phys. 11, 9075 (2009). * Tikhonov and Tokmachev [2011] N. A. Tikhonov and M. G. Tokmachev, J. Math. Chem. 49, 629 (2011). * Stuart _et al._ [2010] M. A. C. Stuart, W. T. S. Huck, J. Genzer, M. Müller, C. Ober, M. Stamm, G. B. Sukhorukov, I. Szleifer, V. V. Tsukruk, M. Urban, F. Winnik, S. Zauscher, I. Luzinov, and S. Minko, Nat. Mater. 9, 101 (2010). * Schattling, Jochum, and Theato [2014] P. Schattling, F. D. Jochum, and P. Theato, Polym. Chem. 5, 25 (2014). * Wandera, Wickramasinghe, and Husson [2010] D. Wandera, S. R. Wickramasinghe, and S. M. Husson, J. Memb. Sci. 357, 6 (2010). * Tanaka [1978] T. Tanaka, Phys. Rev. Lett. 40, 820 (1978). * Tanaka _et al._ [1980] T. Tanaka, D. Fillmore, S.-T. Sun, I. Nishio, G. Swislow, and A. Shah, Phys. Rev. Lett. 45, 1636 (1980). * Hirokawa and Tanaka [1984] Y. Hirokawa and T. Tanaka, J. Chem. Phys. 81, 6379 (1984). * Koga, Sasaki, and Maeda [2001] S. Koga, S. Sasaki, and H. Maeda, J. Phys. Chem. B 105, 4105 (2001). * Roiter and Minko [2005] Y. Roiter and S. Minko, J. Am. Chem. Soc. 127, 15688 (2005). * Kim _et al._ [2017] W. K. Kim, A. Moncho-Jorda, R. Roa, M. Kanduč, and J. Dzubiella, Macromolecules 50, 6227 (2017). * Liu _et al._ [2016] Z. Liu, W. Wang, R. Xie, X. J. Ju, and L. Y. Chu, Chem. Soc. Rev. 45, 460 (2016). * Ito _et al._ [2002] T. Ito, T. Hioki, T. Yamaguchi, T. Shinbo, S. I. Nakao, and S. Kimura, J. Am. Chem. Soc. 124, 7840 (2002). * Ito _et al._ [2004] T. Ito, Y. Sato, T. Yamaguchi, and S. I. Nakao, Macromolecules 37, 3407 (2004). * Bhattacharyya _et al._ [2013] D. Bhattacharyya, T. Schäfer, S. R. Wickramasinghe, and S. Daunert, _Responsive Membranes and Materials_ (John Wiley & Sons, 2013). * Li _et al._ [2020] Y. Li, C. Chen, E. R. Meshot, S. F. Buchsbaum, M. Herbert, R. Zhu, O. Kulikov, B. McDonald, N. T. Bui, M. L. Jue, S. J. Park, C. A. Valdez, S. Hok, Q. He, C. J. Doona, K. J. Wu, T. M. Swager, and F. Fornasiero, Adv. Funct. Mater. 30, 1 (2020). * Xie, Li, and Chu [2007] R. Xie, Y. Li, and L. Y. Chu, J. Memb. Sci. 289, 76 (2007). * Tokarev and Minko [2010] I. Tokarev and S. Minko, Adv. Mater. 22, 3446 (2010). * Wijmans and Baker [1995] J. G. Wijmans and R. W. Baker, J. Memb. Sci. 107, 1 (1995). * Yasuda _et al._ [1969] H. Yasuda, A. Peterlin, C. K. Colton, K. A. Smith, and E. W. Merrill, Die Makromol. Chemie 126, 177 (1969). * Gehrke _et al._ [1997] S. H. Gehrke, J. P. Fisher, M. Palasis, and M. E. Lund, Ann. N. Y. Acad. Sci. 831, 179 (1997). * Missner and Pohl [2009] A. Missner and P. Pohl, Chem. Phys. Chem 10, 1405 (2009). * Siegel and Pitt [1995] R. A. Siegel and C. G. Pitt, J. Control. Release 33, 173 (1995). * Marian von Smoluchowski [1916] Marian von Smoluchowski, Ann. Phys. 353, 1103 (1916). * Kim _et al._ [2022] W. K. Kim, S. Milster, R. Roa, M. Kanduč, and J. Dzubiella, Macromolecules 55, 7327 (2022). * Grassi and Colombo [1999] M. Grassi and I. Colombo, J. Control. Release 59, 343 (1999). * Risken [1996] H. Risken, in _Fokker-Planck Equ._ (Springer, 1996). * Kim _et al._ [2020] W. K. Kim, R. Chudoba, S. Milster, R. Roa, M. Kanduč, and J. Dzubiella, Soft Matter 16, 1 (2020). * Milster _et al._ [2019] S. Milster, R. Chudoba, M. Kanduč, and J. Dzubiella, Phys. Chem. Chem. Phys. 21, 6588 (2019). * Milster _et al._ [2021] S. Milster, W. K. Kim, M. Kanduč, and J. Dzubiella, J. Chem. Phys. 154, 154902 (2021). * Kanduč _et al._ [2018] M. Kanduč, W. K. Kim, R. Roa, and J. Dzubiella, Macromolecules 51, 4853 (2018). * Kanduč _et al._ [2021] M. Kanduč, W. K. Kim, R. Roa, and J. Dzubiella, ACS Nano 15, 614 (2021). * Quesada-Pérez and Martín-Molina [2021] M. Quesada-Pérez and A. Martín-Molina, Adv. Colloid Interface Sci. 287, 102320 (2021). * Amsden [1998] B. Amsden, Macromolecules 31, 8382 (1998). * Hansing and Netz [2018] J. Hansing and R. R. Netz, Biophys. J. 114, 2653 (2018). * Wolf and Willms [1978] B. A. Wolf and M. M. Willms, Die Makromol. Chemie 179, 2265 (1978). * Kamemaru _et al._ [2018] K. Kamemaru, S. Usui, Y. Hirashima, and A. Suzuki, Gels 2018, Vol. 4, Page 45 4, 45 (2018). * Mohammadyarloo and Sommer [2022] Z. Mohammadyarloo and J.-U. Sommer, Macromolecules 55, 8975 (2022). * Annaka and Tanaka [1992] M. Annaka and T. Tanaka, Nat. 1992 3556359 355, 430 (1992). * Katchalsky and Neumann [1972] A. Katchalsky and E. Neumann, Int. J. Neurosci. 3, 175 (1972). * Frank [2004] T. D. Frank, Phys. Lett. A 327, 146 (2004). * Donskoy [2022] I. Donskoy, Energies 15, 7152 (2022). * Martyushev and Seleznev [2006] L. M. Martyushev and V. D. Seleznev, Phys. Rep. 426, 1 (2006). * P. Glansdorff, G. Nicolis, and I. Prigogine [1974] P. Glansdorff, G. Nicolis, and I. Prigogine, Proc. Natl. Acad. Sci. 71, 197 (1974). * Kondepudi and Prigogine [1998] D. Kondepudi and I. Prigogine, _Modern Thermodynamics_ (John Wiley & Sons, Chichester, 1998) p. 506. * Attard [2009] P. Attard, Annu. Reports Prog. Chem. Sect. C Phys. Chem. 105, 63 (2009). * Tomé [2006] T. Tomé, Brazilian J. Phys. 36, 1285 (2006). * Graham and Tél [1985] R. Graham and T. Tél, Phys. Rev. A 31, 1109 (1985). * Seifert [2012] U. Seifert, Reports Prog. Phys. 75, 58 (2012). * Schmidt [2022] M. Schmidt, Rev. Mod. Phys. 94, 015007 (2022). * Chavanis [2008] P.-H. Chavanis, Phys. A Stat. Mech. its Appl. 387, 5716 (2008). * Archer and Rauscher [2004] A. J. Archer and M. Rauscher, J. Phys. A. Math. Gen. 37, 9325 (2004). * Gardiner [2004] C. W. Gardiner, _Chemistry (Easton)._ , 3rd ed. (Springer-Verlag Berlin Heidelberg, 2004). * Hiller and Rubner [2003] J. Hiller and M. F. Rubner, Macromolecules 36, 4078 (2003). * Baker and Siegel [1996] J. P. Baker and R. A. Siegel, Macromol. Rapid Commun. 17, 409 (1996).
TTK-22-42, P3H-22-115 August 28, 2024 NLO QCD corrections and off-shell effects for $t\overline{t}H$ production in the Higgs characterisation model Jonathan Hermann111Work supported by German Research Foundation (DFG) Collaborative Research Centre/Transregio project CRC/TRR 257:P3H - Particle Physics Phenomenology after the Higgs Discovery, by the Research Training Group GRK 2497: The physics of the heaviest particles at the Large Hadron Collider and by a grant of the Bundesministerium für Bildung und Forschung (BMBF). Institute for Theoretical Particle Physics and Cosmology, RWTH Aachen University, D-52056 Aachen, Germany > In these proceedings we discuss $t\overline{t}H$ production at the Large > Hadron Collider in the Higgs characterisation model. We demonstrate the > importance of off-shell effects and higher-order QCD corrections for this > process and highlight the importance of single-resonant contributions for > the production of a $\mathcal{CP}$-odd Higgs boson. > PRESENTED AT > > > > > $15^{\mathrm{th}}$ International Workshop on Top Quark Physics > Durham, UK, 4–9 September, 2022 ## 1 Introduction Despite its small cross-section compared to gluon-gluon and vector-boson fusion, $t\overline{t}H$ production is one of the most important Higgs production channels at the Large Hadron Collider (LHC). It is particularly interesting for probing the top-Higgs Yukawa interaction as it already appears at tree level in $t\overline{t}H$ production. The $\mathcal{CP}$-nature of this interaction has been of great interest since the discovery of the Higgs boson in 2012. Both CMS and ATLAS have performed measurements of the mixing angle $\alpha_{CP}$ between a $\mathcal{CP}$-even and a $\mathcal{CP}$-odd Higgs boson [1, 2]. In both cases the results are in agreement with the Standard Model (SM) prediction of a $\mathcal{CP}$-even Higgs boson and the measurements have allowed the exclusion of a pure $\mathcal{CP}$-odd state with $3.7\,\sigma$ by CMS and $3.9\,\sigma$ by ATLAS. Nonetheless, there still remains a lot of freedom in the mixing angle and CMS has actually found that their fit results favour a $\mathcal{CP}$-mixed coupling. With our work we aim to provide state-of-the-art predictions for $t\overline{t}H$ production in the Higgs characterisation model [3] which extends the SM Lagrangian by a $\mathcal{CP}$-odd top-Higgs Yukawa interaction term. To this end, we present results for this process in the dilepton decay channel at next-to-leading order (NLO) in QCD and including full off-shell effects, i.e. $pp\to b\overline{b}e^{+}\mu^{-}\nu_{e}\overline{\nu}_{\mu}H+X$ production at order $\mathcal{O}\left(\alpha_{s}^{3}\alpha^{5}\right)$ for the LHC at $\sqrt{s}=13$ TeV. Here, full off-shell effects mean that we describe the unstable intermediate particles in the complex mass scheme and include all double-, single- and non-resonant contributions. These results are based on the ones presented in Ref. [4] and were computed using the HELAC-NLO framework [5, 6]. Similar calculations have already been performed for this process in the SM case in Refs. [7, 8, 9] but not for $\mathcal{CP}$-mixed top-Higgs interactions. For the latter, the previous state-of-the-art predictions only included NLO QCD corrections to the $t\overline{t}H$ production and not to the top-quark decays. In addition, only double-resonant diagrams were considered [10]. ## 2 Phenomenological results Table 1: Integrated fiducial cross-sections calculated in the NWA, NWA with LO top-quark decays and full off-shell approach for $\alpha_{CP}=0,\pi/4$ and $\pi/2$. Table was taken from [4]. $\alpha_{CP}$ | | Off-shell | NWA | Off-shell effects ---|---|---|---|--- $0$ (SM) | $\sigma_{\text{LO}}$ [fb] | $2.0313(2)^{+0.6275\,(31\%)}_{-0.4471\,(22\%)}$ | $2.0388(2)^{+0.6290\,(31\%)}_{-0.4483\,(22\%)}$ | $-0.37\%$ $\sigma_{\text{NLO}}$ [fb] | $2.466(2)^{+0.027\,(1.1\%)}_{-0.112\,(4.5\%)}$ | $2.475(1)^{+0.027\,(1.1\%)}_{-0.113\,(4.6\%)}$ | $-0.36\%$ $\sigma_{\text{NLO}_{\text{LOdec}}}$ [fb] | $-$ | $2.592(1)^{+0.161\,(6.2\%)}_{-0.242\,(9.3\%)}$ | | ${\cal K}=\sigma_{\text{NLO}}/\sigma_{\text{LO}}$ | $1.21$ | $1.21$ (LOdec: $1.27$) | $\pi/4$ | $\sigma_{\text{LO}}$ [fb] | $1.1930(2)^{+0.3742\,(31\%)}_{-0.2656\,(22\%)}$ | $1.1851(1)^{+0.3707\,(31\%)}_{-0.2633\,(22\%)}$ | $0.66\%$ $\sigma_{\text{NLO}}$ [fb] | $1.465(2)^{+0.016\,(1.1\%)}_{-0.071\,(4.8\%)}$ | $1.452(1)^{+0.015\,(1.0\%)}_{-0.069\,(4.8\%)}$ | $0.89\%$ $\sigma_{\text{NLO}_{\text{LOdec}}}$ [fb] | $-$ | $1.517(1)^{+0.097\,(6.4\%)}_{-0.144\,(9.5\%)}$ | | ${\cal K}=\sigma_{\text{NLO}}/\sigma_{\text{LO}}$ | $1.23$ | $1.23$ (LOdec: $1.28$) | $\pi/2$ | $\sigma_{\text{LO}}$ [fb] | $0.38277(6)^{+0.13123\,(34\%)}_{-0.09121\,(24\%)}$ | $0.33148(3)^{+0.11240\,(34\%)}_{-0.07835\,(24\%)}$ | $13.4\%$ $\sigma_{\text{NLO}}$ [fb] | $0.5018(3)^{+0.0083\,(1.2\%)}_{-0.0337\,(6.7\%)}$ | $0.4301(2)^{+0.0035\,(0.8\%)}_{-0.0264\,(6.1\%)}$ | $14.3\%$ $\sigma_{\text{NLO}_{\text{LOdec}}}$ [fb] | $-$ | $0.4433(2)^{+0.0323\,(7.3\%)}_{-0.0470\,(11\%)}$ | | ${\cal K}=\sigma_{\text{NLO}}/\sigma_{\text{LO}}$ | $1.31$ | $1.30$ (LOdec: $1.34$) | First we compare the integrated fiducial cross-sections for the $\mathcal{CP}$-even, -mixed and -odd Higgs boson production in the narrow- width approximation (NWA) and the full off-shell treatment at LO and NLO in QCD. The results are listed in Table 1. We find that the cross-sections for the $\mathcal{CP}$-mixed and $\mathcal{CP}$-even cases are about $3$ and $5$ times larger than for the $\mathcal{CP}$-odd one. NLO corrections are around $20\%$ for the $\mathcal{CP}$-even and -mixed and about $30\%$ for the $\mathcal{CP}$-odd Higgs boson. In all cases these corrections are consistent between the NWA and the full off-shell treatment and within the LO uncertainties. Concerning the off-shell effects, we find that these are of the expected order of $\Gamma_{t}/m_{t}\sim 0.8\%$ for the $\mathcal{CP}$-even and -mixed cases and thus negligible at the level of integrated fiducial cross- sections when compared to the NLO scale uncertainties of around $5\%$. However, in the pure $\mathcal{CP}$-odd case, the off-shell effects are significantly larger than these uncertainties at $14\%$ for the NLO results and should thus be taken into account even for the integrated fiducial cross- section. Figure 1: Differential distributions for the observables $p_{T,\,H}$ and $\Delta\phi_{e^{+}\mu^{-}}$ at NLO in QCD. The lower panels show the differential $\mathcal{K}\textrm{-factors}$. Figures were taken from [4]. Figure 2: Left: Differential distributions at NLO in QCD for $p_{T,H}$ for the full off-shell case (solid lines) and the NWA (dashed lines). The ratio to $\alpha_{CP}=0$ of the normalised differential distributions for the full off- shell case is shown in the middle panel, the ratio NWA/off-shell is given in the lower one. Center: Differential distributions at NLO in QCD for $p_{T,H}$ for the full off-shell case and the NWA as well as the double- (DR), single- (SR) and non- resonant (NR) contributions for the SM case. Right: Same as central plot but for the $\mathcal{CP}$-odd case. Figures were taken from [4]. The general trend of larger corrections for the $\mathcal{CP}$-odd Higgs boson can also be observed at the differential level. In Figure 1, where we show the differential distributions for $p_{T,H}$ and $\Delta\phi_{e^{+},\mu^{-}}$, one can for example see that NLO corrections are generally larger in the $\mathcal{CP}$-odd case than for the others. For most observables the $\mathcal{K}$-factors follow the same trend as for $p_{T,H}$, i.e. the corrections increase towards the distribution tail but the behaviour of the three $\mathcal{K}$-factors is roughly the same. However, for observables which involve decay products of both top quarks, like $\Delta\phi_{e^{+},\mu^{-}}$, we observe that the $\mathcal{K}$-factor for the $\mathcal{CP}$-odd Higgs boson is flatter than for the other two cases. This happens due to the harder Higgs boson radiation for the former case which suppresses the real radiation corrections that are responsible for the large $\mathcal{K}$-factors for small opening angles and large transverse momenta. Off-shell effects are also enhanced at the differential level, as can be seen from the lower panel of the left plot in Figure 2. For the depicted $p_{T,H}$ distribution, they exceed $30\%$ in the high-$p_{T}$ region for the $\mathcal{CP}$-odd Higgs boson. On the other hand, the effects remain negligible for the other two considered $\mathcal{CP}$-states. The different size of the off-shell effects can be attributed to the single-resonant (SR) contributions included in the full off-shell computation. In the central and right plot in Figure 2 we show the splitting of the off-shell distribution into its double-, single- and non-resonant parts as defined in Eq. (5.6) to (5.9) Ref. [4]. In the SM case (central plot), the ratio between the three contributions remains virtually constant throughout the entire spectrum which mimics the behaviour of the off-shell effects in this case. In contrast, for the $\mathcal{CP}$-odd state (right plot), the DR contribution decreases significantly towards the distribution tail, in line with NWA distribution. At the same time, the SR contribution increases to account for the difference. Hence, the larger off-shell effects in the latter case can be explained by the different behaviour of the SR contribution. ## 3 Summary In these proceedings we have presented integrated and differential fiducial cross-sections for $t\overline{t}H$ production with leptonic top-quarks decays in the Higgs characterisation framework. We have underlined the importance of both off-shell effects and higher-order QCD corrections for these predictions, particularly for the production of a $\mathcal{CP}$-odd Higgs boson. In addition, we have demonstrated that the large off-shell effects in the $\mathcal{CP}$-odd case are a result of the missing single-resonant contributions in the NWA. ## References * [1] CMS Collaboration, arXiv:2208.02686. * [2] ATLAS Collaboration, _Phys. Rev. Lett._ 125 (2020) 061802. * [3] P. Artoisenet et al., _JHEP_ 11 (2013) 043. * [4] J. Hermann, D. Stremmer, M. Worek, _JHEP_ 09 (2022) 138. * [5] G. Bevilacqua et al., _Comput. Phys. Commun._ 184 (2013) 986. * [6] G. Bevilacqua, H. B. Hartanto, M. Kraus, T. Weber, M. Worek, _JHEP_ 03 (2020) 154. * [7] A. Denner and R. Feger, _JHEP_ 11 (2015) 209. * [8] A. Denner, J.-N. Lang, M. Pellen and S. Uccirati, _JHEP_ 02 (2017) 053 * [9] D. Stremmer, M. Worek, _JHEP_ 02 (2022) 196. * [10] F. Demartin, F. Maltoni, K. Mawatari, B. Page, M. Zaro, _Eur. Phys. J. C_ 74 (2014) 3065.
# Monitored Recurrence of a One-parameter Family of Three-state Quantum Walks Martin Štefaňák Department of Physics, Faculty of Nuclear Sciences and Physical Engineering, Czech Technical University in Prague, Břehová 7, 115 19 Praha 1 - Staré Město, Czech Republic<EMAIL_ADDRESS> ###### Abstract Monitored recurrence of a one-parameter set of three-state quantum walks on a line is investigated. The calculations are considerably simplified by choosing a suitable basis of the coin space. We show that the Polya number (i.e. the site recurrence probability) depends on the coin parameter and the probability that the walker is initially in a particular coin state for which the walk returns to the origin with certainty. Finally, we present a brief investigation of the exact quantum state recurrence. ## 1 Introduction Quantum walks [1, 2, 3] are generalizations of classical random walks to evolution of a quantum particle on a graph or a lattice. Similar ideas appeared already in 1960’s in the works of Feynman and Hibbs on discretization of the Dirac equation [4] and in the 1980’s in the work of Gudder on quantum graphic dynamics [5, 6]. Quantum walks became versatile tools in various field of physics and computation, for a recent review see [7]. One of the interesting questions in stochastic processes is the problem of recurrence. For classical random walks, this was studied in detail by Polya [8] who proved that in 1 and 2 spatial dimensions the balanced walk returns with certainty, i.e. is recurrent, while for $d\geq 3$ the recurrence probability (also called Polya number) is strictly less than unity, implying transience. Nevertheless, walks on infinite line and square lattice are null- recurrent, since the expected recurrence time is infinite. On the other hand, irreducible walks on finite graphs are positive-recurrent since the expected recurrence time is finite. In fact, for finite state space Kac shown [9] that it is an integer. Studying recurrence of quantum walks requires a detailed description of the measurement process since it has a nontrivial effect on the wavefunction of the quantum walker. To minimize the influence of measurement we have proposed a scheme [10, 11] where the quantum walk evolves freely for a certain number of steps, then a measurement detecting the presence of the walker at the origin is performed, after which the process is restarted. In this concept the recurrence or transience of a quantum walk can be decided from the behaviour of the probability of finding the walker at the origin after $t$ steps (without prior measurement) in the same way as for the classical random walk. In fact, for classical random walks this scheme is equivalent to the Polya approach, where we monitor the origin after each step, in the sense that the walk is recurrent in one approach if and only if it is recurrent in the other scheme. This stems from the fact that in the classical setting, the fundamental role is played by probabilities and there is a direct relation between the return probabilities without prior measurement and the first return probabilities of the monitored walk. However, this does not hold anymore in the quantum case. Due to the effect of measurement on the wavefunction the two probabilities cannot be accessed in the same experiment and they are not related to each other. Nevertheless, there exists a relation between the return amplitudes of the unitary quantum walk and the first return amplitudes of the monitored walk, as shown in [12, 13] for a much broader setting of iterated unitary evolutions. Note that some particular examples were investigated before the general theory was developed, e.g. the monitored recurrence of the Hadamard walk on a line [14] which was shown to be equivalent to the absorption of the walk on the half-line [15]. In the monitored recurrence approach, the expected return time to the exact initial state (state recurrence) of a finite system is an integer [12] as in the classical case. This holds even for iterated open quantum evolutions [16, 17]. Note that in the coined quantum walks return to the initial position can be understood as a subspace recurrence, since we are not interested in the exact state of the quantum coin. For finite systems the expected return time for a subspace recurrence is a rational number [13]. We point out that in the recent years the study of recurrence was considerably broadened to open quantum systems, quantum Markov chains and open quantum walks, see eg. [18, 19, 20, 21, 22, 23, 24]. In contrast to the classical case the two schemes for detecting recurrence (restart after measurement vs monitored after each step) are not equivalent for quantum walks. As an example, the Hadamard walk on a line is recurrent in the restart scheme [10] while transient in the monitored case [14]. This was also demonstrated in an optical experiment [25] where the quantum walk was simulated by a weak laser pulse cycling in a time-delay loop, utilizing time- multiplexing to encode the position and number of steps of the walk into the time of detection of a photon. It was a considerable experimental challenge to perform the local measurement at the origin without disturbing the rest of the pulse sequence. The deterministic out-coupling from the loop was achieved by a programmable fast-switching electro-optical modulator, which allowed to address specific time slots corresponding to the walker being at the origin. In the present paper we take the monitored approach to the recurrence problem and apply it to the three-state quantum walks on a line with a particular one- parameter family of coins [26]. We show that there is a unique initial coin state for which the walk stays at the origin in the first step, so in this case the walk is recurrent. We then choose this initial state as one of the basis vectors in the coin space and complement the basis with two vectors from orthogonal complement. This leads to significant reduction of the complexity of the follow-up analytical calculations. We show that the subspace recurrence probability is determined by the coin parameter and the probability that the walker is initially in the unique recurrent state. For the exact quantum state recurrence we provide a numerical investigation. In particular, we identify initial states for which the state recurrence is greater than the subspace recurrence, a paradoxical feature of monitored quantum evolution which was already discussed previously in the literature [13]. The paper is organized as follows. Section 2 reviews the basic concepts and tools for investigation of monitored recurrence. Section 3 is dedicated to the study of site recurrence for a one-parameter family of three-state quantum walks. In Section 4 we present a numerical investigation of the state recurrence. We conclude and present an outlook in Section 5. ## 2 Monitored site recurrence of a quantum walk We begin by a brief overview of the methods for studying recurrence developed in [12, 13], adopted for a coined quantum walks with constant coin. Consider a discrete time quantum walk on a lattice starting from the origin. The Hilbert space is a tensor product of the position and the coin spaces $\mathcal{H}=\mathcal{H}_{p}\otimes\mathcal{H}_{c}.$ We denote the evolution operator by $U$, it has the usual decomposition $U=S\cdot(I_{p}\otimes C),$ (1) where $S$ is the shift operator and $C$ is the coin. Return of the quantum walker to the original site can be understood as a subspace recurrence [13], since we are not interested in the internal state of the coin, only the position of the quantum walker. Let us denote the orthogonal projector onto the origin subspace as $\Pi_{0}=|0\rangle\langle 0|\otimes I_{c}.$ (2) The evolution of the monitored quantum walk is given by the operator $\tilde{U}=(I-\Pi_{0})U,$ i.e. the walk continues if we do not find the walker at the origin. Normalized state after $n$ steps of the monitored walk reads $|\psi_{n}\rangle=\frac{1}{\sqrt{s_{n}}}\tilde{U}^{n}|\psi\rangle,\quad|\psi\rangle=|0\rangle\otimes|\psi_{c}\rangle,$ where $s_{n}$ is the survival probability until the $n$-th step $s_{n}=\lVert\tilde{U}^{n}\psi\rVert^{2}.$ $s_{n}$ is the probability that the walker has not returned to the origin in the first $n$ steps. Hence, the complement of the limiting value of $s_{n}$ $P(\psi)=1-\lim\limits_{n\to\infty}s_{n},$ corresponds to the probability that the walker ever returns, i.e. the site recurrence probability (or Polya number). Alternatively, we can derive the Polya number from the first return probabilities. Let us denote the probability of first return to the origin after $n$ steps as $q_{n}$. Since these events are mutually exclusive, the overall site recurrence probability is given by the sum $P(\psi)=\sum_{n=1}^{\infty}q_{n}.$ In case of iterated unitary evolution the first return probability after $n$ steps is given by $q_{n}=\lVert\Pi_{0}U\tilde{U}^{n-1}\psi\rVert^{2}=\lVert a_{n}\psi\rVert^{2},$ where we have introduced the first return amplitude operator (note that $\Pi_{0}\psi=\psi$) $a_{n}=\Pi_{0}U\tilde{U}^{n-1}\Pi_{0}.$ The Polya number of a quantum walk for a given initial state is therefore given by $P(\psi)=\sum_{n=1}^{\infty}\lVert a_{n}\psi\rVert^{2}.$ Determining the operators $a_{n}$ directly is rather difficult, however, they can be related to the $n$-th step return amplitude operators without prior monitoring, which we denote as $\mu_{n}=\Pi_{0}U^{n}\Pi_{0}.$ We define the operator valued generating functions (for complex variable $z$ with $|z|<1$) $\mu(z)=\sum_{n=0}^{\infty}\mu_{n}z^{n},\quad a(z)=\sum_{n=1}^{\infty}a_{n}z^{n}.$ Note that $\mu(z)$ is also called the Stieltjes operator and the first return generating function $a(z)$ is related to the Schur function $f(z)$ by $a(z)=zf^{\dagger}(z),$ which is more extensively used in the literature [12, 13, 20, 23]. Introducing the resolvents $G(z)=\sum_{n=0}^{\infty}U^{n}z^{n}=(I-zU)^{-1},\quad\tilde{G}(z)=\sum_{n=0}^{\infty}\tilde{U}^{n}z^{n}=(I-z\tilde{U})^{-1},$ we see that the generating functions can be written in the form $\mu(z)=\Pi_{0}G(z)\Pi_{0},\quad a(z)=z\Pi_{0}U\tilde{G}(z)\Pi_{0}.$ (3) Using the resolvent identities $G(z)-\tilde{G}(z)=zG(z)\Pi_{0}U\tilde{G}(z)=z\tilde{G}(z)\Pi_{0}UG(z),$ we can derive the renewal equations [12, 13] $\mu(z)a(z)=a(z)\mu(z)=\mu(z)-\Pi_{0}.$ (4) In the above equation all operators act on the origin subspace (see (2) and (3)), i.e. they are of the form $\mu(z)=|0\rangle\langle 0|\otimes\mu_{c}(z),\quad a(z)=|0\rangle\langle 0|\otimes a_{c}(z),$ where $\mu_{c}(z)$ and $a_{c}(z)$ act on the coin space. Thus, we can rewrite (4) into operator equation on the coin space $a_{c}(z)=I_{c}-\mu_{c}(z)^{-1}.$ (5) Hence, we can express the first return generating function $a_{c}(z)$ from the return generating function $\mu_{c}(z)$, which is easier to obtain. All functions can be extended to the unit circle in the complex plane by the radial limit. The site recurrence probability is expressed in terms of the boundary values [12, 13] by $P(\psi_{c})=\int\limits_{0}^{2\pi}\lVert a_{c}(e^{it})\psi_{c}\rVert^{2}\frac{dt}{2\pi}=\langle\psi_{c}|R|\psi_{c}\rangle,$ (6) where we have denoted the recurrence probability operator $R=\int\limits_{0}^{2\pi}a_{c}^{\dagger}(e^{it})a_{c}(e^{it})\frac{dt}{2\pi}.$ (7) In summary, the recipe to obtain the Polya number is to find the Stieltjes operator $\mu_{c}(z)$, from (5) we find the first return generating operator, construct the recurrence probability operator (7) and find its average value for a given initial state of the coin $\psi_{c}$ (6). The site recurrence probability of a two-state quantum walk on a line, where the walker can move to the right or left in each step, was evaluated explicitly in [12]. The unitary coin operator was parameterized in the following way $C=\begin{pmatrix}\rho&-\gamma\\\ \overline{\gamma}&\rho\end{pmatrix},\quad\rho=\sqrt{1-|\gamma|^{2}}.$ (8) Since the considered walk is translationally invariant, one can use Fourier transformation to diagonalize the step operator and evaluate the Stieltjes operator with the formula $\mu_{c}(z)=\int\limits_{0}^{2\pi}\frac{dp}{2\pi}(I_{c}-zU(p))^{-1}.$ (9) Here $U(p)$ is the evolution operator (1) in the Fourier picture $U(p)=S(p)\cdot C=\begin{pmatrix}e^{ip}&0\\\ 0&e^{-ip}\end{pmatrix}\cdot\begin{pmatrix}\rho&-\gamma\\\ \overline{\gamma}&\rho\end{pmatrix},$ which is a multiplication operator. Using the substitution $e^{ip}=x$, $dp=\frac{dx}{ix}$ the RHS of (9) is turned into an integral over the unit circle in the complex plane, which can be evaluated with the residue theorem. After some calculations it is found that the recurrence operator (7) for a two-state walk is a multiple of identity, i.e. the Polya number does not depend on the initial coin state, only on the quantum coin. Explicitly, the site recurrence probability for a two-state quantum walk on a line with the coin (8) is given by [12, 27] $P=\frac{2\left(|\gamma|\sqrt{1-|\gamma|^{2}}-(1-2|\gamma|^{2})\arcsin|\gamma|\right)}{\pi|\gamma|^{2}}.$ ## 3 Site recurrence of a three-state walk on a line Let us now consider a three-state walk on a line, where the walker can move to the right, stay, or move to the left. This corresponds to the standard basis of the coin space $|R\rangle$, $|S\rangle$ and $|L\rangle$. We choose the following one-parameter set of coins [26, 28] $C=\begin{pmatrix}-\rho^{2}&\rho\sqrt{2(1-\rho^{2})}&1-\rho^{2}\\\ \rho\sqrt{2(1-\rho^{2})}&2\rho^{2}-1&\rho\sqrt{2(1-\rho^{2})}\\\ 1-\rho^{2}&\rho\sqrt{2(1-\rho^{2})}&-\rho^{2}\end{pmatrix},\quad 0<\rho<1,$ (10) which reduces to the 3x3 Grover matrix for $\rho=\frac{1}{\sqrt{3}}$. Before turning to the derivation of the Polya number, we make the following observation. Since the coin operator (10) is equal to its inverse, the initial coin state $|\alpha_{1}\rangle=C|S\rangle=\rho\sqrt{2(1-\rho^{2})}(|R\rangle+|L\rangle)+(2\rho^{2}-1)|S\rangle=\begin{pmatrix}\rho\sqrt{2(1-\rho^{2})}\\\ 2\rho^{2}-1\\\ \rho\sqrt{2(1-\rho^{2})}\end{pmatrix},$ is mapped to $|S\rangle$. Hence, if we take $|\alpha_{1}\rangle$ as the initial coin state, the walker is absorbed with certainty after one step, i.e. for this initial state the site recurrence probability is one. Numerical simulations indicate that for all coin states in the orthogonal complement to $|\alpha_{1}\rangle$ the Polya number has the same value, which is less than unity. Let us complete the orthonormal basis in the coin space by $\displaystyle|\alpha_{2}\rangle$ $\displaystyle=\frac{2\rho^{2}-1}{\sqrt{2}}(|R\rangle+|L\rangle)-2\rho\sqrt{1-\rho^{2}}|S\rangle=\begin{pmatrix}\frac{2\rho^{2}-1}{\sqrt{2}}\\\ -2\rho\sqrt{1-\rho^{2}}\\\ \frac{2\rho^{2}-1}{\sqrt{2}}\end{pmatrix},$ $\displaystyle|\alpha_{3}\rangle$ $\displaystyle=\frac{1}{\sqrt{2}}(|R\rangle-|L\rangle)=\frac{1}{\sqrt{2}}\begin{pmatrix}1\\\ 0\\\ -1\end{pmatrix}.$ The discussion above indicates that the recurrence probability operator should be diagonal when expressed in the basis formed by $\left\\{|\alpha_{1}\rangle,|\alpha_{2}\rangle,|\alpha_{3}\rangle\right\\}$. From now on all matrices will be expressed in this basis. For this we will utilize the transition matrices $T$ and $T^{\dagger}$, where $T=\begin{pmatrix}\rho\sqrt{2(1-\rho^{2})}&2\rho^{2}-1&\rho\sqrt{2(1-\rho^{2})}\\\ \frac{2\rho^{2}-1}{\sqrt{2}}&-2\rho\sqrt{2(1-\rho^{2})}&\frac{2\rho^{2}-1}{\sqrt{2}}\\\ \frac{1}{\sqrt{2}}&0&-\frac{1}{\sqrt{2}}\end{pmatrix}.$ We denote the matrix $M$ in the $\alpha$ basis as ${}^{\alpha}M$. Turning to the derivation of the Polya number, we begin by expressing the Stieltjes operator through the Fourier transformation (9), where ${}^{\alpha}U(p)$ is given by ${}^{\alpha}U(p)$ $\displaystyle=T\cdot\begin{pmatrix}e^{ip}&0&0\\\ 0&1&0\\\ 0&0&e^{-ip}\end{pmatrix}\cdot C\cdot T^{\dagger}$ $\displaystyle=\begin{pmatrix}2\rho^{2}-1&-2\rho\sqrt{1-\rho^{2}}\cos{p}&-2i\rho\sqrt{1-\rho^{2}}\sin{p}\\\ -2\rho\sqrt{1-\rho^{2}}&(1-2\rho^{2})\cos{p}&i(1-2\rho^{2})\sin{p}\\\ 0&-i\sin{p}&-\cos{p}\end{pmatrix}.$ The resolvent in the Fourier space is equal to ${}^{\alpha}(I_{c}-zU(p))^{-1}$ $\displaystyle=\ \frac{1}{(z-1)(1+z(2-2\rho^{2}+z)+2\rho^{2}z\cos{p})}\times$ $\displaystyle\begin{pmatrix}u(z)-2\rho^{2}z\cos p&2\rho\sqrt{1-\rho^{2}}z(\cos p+z)&2i\rho\sqrt{1-\rho^{2}}z\sin p\\\ 2\rho\sqrt{1-\rho^{2}}z(z\cos p+1)&w(z)(z\cos p+1)&-izv(z)\sin p\\\ -2i\rho\sqrt{1-\rho^{2}}z^{2}\sin p&i(u(z)+z+1)\sin p&zv(z)\cos p+w(z)\end{pmatrix},$ where to shorten the formula we have denoted $u(z)=z^{2}(1-2\rho^{2})-1,\quad v(z)=1-2\rho^{2}+z,\quad w(z)=z(2\rho^{2}-1)-1.$ The integral in (9) can be again evaluated with residues. We will utilize the following integrals (with $n=1,0,-1$) $\displaystyle{\cal I}(n)=\int\limits_{0}^{2\pi}\frac{dp}{2\pi}\frac{e^{inp}}{(z-1)(1+z(2-2\rho^{2}+z)+2\rho^{2}z\cos{p})}$ $\displaystyle=\frac{1}{2\pi i}\oint\frac{x^{n}dx}{b(x)},$ where we have denoted $b(x)=(z-1)(\rho^{2}z+(1+2(1-\rho^{2})z+z^{2})x+\rho^{2}zx^{2}).$ The roots of the equation $b(x)=0$ are $x_{\pm}=\frac{2(\rho^{2}-1)z-z^{2}-1\pm\sqrt{\left(1+2(1-\rho^{2})z+z^{2}\right)^{2}-4\rho^{4}z^{2}}}{2\rho^{2}z},$ with $|x_{-}|>1$ for $|z|\leq 1$. Hence, for $n=0,1$ there is only the residue at $x_{+}$ and we find $\displaystyle{\cal I}(0)$ $\displaystyle={\rm Res}\left(\frac{1}{b(x)},x_{+}\right)=\frac{1}{(z^{2}-1)g(z)},$ $\displaystyle{\cal I}(1)$ $\displaystyle={\rm Res}\left(\frac{x}{b(x)},x_{+}\right)=\frac{(z+1)g(z)-1-2(1-\rho^{2})z-z^{2}}{2\rho^{2}z(z^{2}-1)g(z)},$ where we have used the notation $g(z)=\sqrt{1+2(1-2\rho^{2})z+z^{2}}.$ For $n=-1$ there is an additional residue at $x=0$ and we obtain $\displaystyle{\cal I}(-1)$ $\displaystyle={\rm Res}\left(\frac{1}{xb(x)},x_{+}\right)+{\rm Res}\left(\frac{1}{xb(x)},0\right)$ $\displaystyle=\frac{2\rho^{2}z}{(z^{2}-1)g(z)[(z+1)g(z)-1-2(1-\rho^{2})z-z^{2}]}+\frac{1}{\rho^{2}z(z-1)}.$ After some algebra, we express the Stieltjes operator in the form $\displaystyle^{\alpha}\mu_{c}(z)$ $\displaystyle=\frac{1}{(z-1)g(z)}\begin{pmatrix}2(\rho^{2}-1)z-g(z)&\frac{\sqrt{1-\rho^{2}}(g(z)+w(z))}{\rho}&0\\\ \frac{\sqrt{1-\rho^{2}}z(g(z)-v(z))}{\rho}&\frac{w(z)(g(z)-v(z))}{2\rho^{2}}&0\\\ 0&0&\frac{v(z)g(z)-1-z(z+2-4\rho^{2})}{2\rho^{2}}\end{pmatrix}.$ (11) From (5) we find the first return generating operator ${}^{\alpha}a_{c}(z)=\begin{pmatrix}(2\rho^{2}-1)z&-2\rho\sqrt{1-\rho^{2}}-\frac{4\rho^{3}\sqrt{1-\rho^{2}}(z-1)}{g(z)-v(z)}&0\\\ -2\rho\sqrt{1-\rho^{2}}z&1-2\rho^{2}+\frac{2\rho^{2}(2\rho^{2}-1)(z-1)}{g(z)-v(z)}&0\\\ 0&0&-\frac{j(z)}{2(\rho^{2}-1)}\end{pmatrix}.$ Here we have denoted $j(z)=1-2\rho^{2}z+z^{2}+(z-1)g(z).$ For the product ${}^{\alpha}a^{\dagger}_{c}(z)\ ^{\alpha}a_{c}(z)$ we obtain a diagonal matrix as expected $^{\alpha}r(z)=^{\alpha}a^{\dagger}_{c}(z)\ ^{\alpha}a_{c}(z)=\begin{pmatrix}|z|^{2}&0&0\\\ 0&\frac{|j(z)|^{2}}{4(\rho^{2}-1)^{2}}&0\\\ 0&0&\frac{|j(z)|^{2}}{4(\rho^{2}-1)^{2}}\end{pmatrix}.$ (12) To determine the matrix of the recurrence probability operator in the $\alpha$ basis ${}^{\alpha}R$ we have to take $z=e^{it}$ and perform the integral (7). The first diagonal entry in (12) is 1, and so is the result of the integral, corresponding to the fact that for the $|\alpha_{1}\rangle$ state the walker returns to the original site with certainty. Then we are left with evaluating $Q=\int\limits_{0}^{2\pi}\frac{dt}{8\pi(\rho^{2}-1)^{2}}|j(e^{it})|^{2}.$ We have to carefully express $g(e^{it})$ and $j(e^{it})$. We find the following $g(e^{it})=\left\\{\begin{array}[]{cc}e^{\frac{it}{2}}\operatorname{sgn}{(\sin{t})}\sqrt{2(\cos{t}+1-2\rho^{2})},&\cos{t}\geq 2\rho^{2}-1\\\ \\\ -ie^{\frac{it}{2}}\sqrt{-2(\cos{t}+1-2\rho^{2})},&\cos{t}<2\rho^{2}-1\end{array}\right..$ Hence, for $t\in(0,\arccos(2\rho^{2}-1))\cup(2\pi-\arccos(2\rho^{2}-1),2\pi)$ , we obtain $j(e^{it})=2e^{it}\left(\cos{t}-\rho^{2}+i\operatorname{sgn}{(\sin{t})}\sin(t/2)\sqrt{2(\cos{t}+1-2\rho^{2})}\right),$ which has a constant modulus square $\displaystyle|j(e^{it})|^{2}=4(\rho^{2}-1)^{2}.$ The contribution to $Q$ is then proportional to the length of the interval and is found to be $Q_{1}=\frac{\arccos(2\rho^{2}-1)}{\pi}.$ Turning to the case $\cos{t}<-2\rho^{2}-1$, i.e. the interval $t\in(\arccos(2\rho^{2}-1),2\pi-\arccos(2\rho^{2}-1))$, we find $j(e^{it})=2e^{it}\left(\cos{t}-\rho^{2}+\sin(t/2)\sqrt{-2(\cos{t}+1-2\rho^{2})}\right).$ The resulting integral can be evaluated directly $\displaystyle Q_{2}=$ $\displaystyle\int\limits_{\arccos{(2\rho^{2}-1)}}^{2\pi-\arccos(2\rho^{2}-1)}\frac{dt}{8\pi(\rho^{2}-1)}|j(e^{it})|^{2}$ $\displaystyle=$ $\displaystyle\int\limits_{\arccos{(2\rho^{2}-1)}}^{2\pi-\arccos(2\rho^{2}-1)}\frac{dt}{2\pi(\rho^{2}-1)}\left(\cos{t}-1+\sin(t/2)\sqrt{-2(\cos{t}+1-2\rho^{2})}\right)^{2}$ $\displaystyle=\rho\frac{2(1+2\rho^{2})\sqrt{1-\rho^{2}}-\rho(2+\rho^{2})\arccos(2\rho^{2}-1)}{\pi}.$ In summary, we find $Q=Q_{1}+Q_{2}=\frac{2\rho\left(2\rho^{2}+1\right)\sqrt{1-\rho^{2}}+\left(1-4\rho^{2}\right)\arccos\left(2\rho^{2}-1\right)}{\pi\left(\rho^{2}-1\right)^{2}}.$ (13) The matrix of the recurrence probability operator in the $\alpha$ basis is then given by ${}^{\alpha}R=\begin{pmatrix}1&0&0\\\ 0&Q&0\\\ 0&0&Q\end{pmatrix}.$ This means that the Polya number for the three-state walk with the coin (10) depends on the probability $p_{1}$ to be initially in the $|\alpha_{1}\rangle$ state $p_{1}=|\langle\alpha_{1}|\psi_{c}\rangle|^{2},$ and the coin parameter $\rho$. The resulting formula is $P(\psi_{c})=p_{1}+Q(1-p_{1}).$ (14) The site recurrence probability thus ranges between 1 and $Q$ (13). In particular, for the Grover walk corresponding to $\rho=\frac{1}{\sqrt{3}}$, we find $P(\psi_{c})=p_{1}+\frac{10\sqrt{2}-3\arccos{(-1/3)}}{4\pi}(1-p_{1})\stackrel{{\scriptstyle.}}{{=}}0.67+0.33p_{1}.$ We illustrate our results in Figures 1 and 2. Figure 1 shows the site recurrence probability (14) in dependence of $\rho$ and $p_{1}$. In Figure 2 we display the value of $Q$ (13) as a function of the coin parameter $\rho$. Figure 1: Density plot of the Polya number (14) of the three-state quantum walk as a function of the probability $p_{1}$ and the coin parameter $\rho$. Dashed lines show the contours of $P=k/10$ for $k=1,\ldots 9$. Figure 2: $Q$ (13) as a function of the coin parameter $\rho$. $Q$ represents the minimal value of the recurrence probability for the walk with a given parameter $\rho$. ## 4 State recurrence of a three-state walk on a line Let us now briefly comment on the state recurrence [12]. In this case we are interested in the return to the exact initial state, so the orthogonal projector (2) has the form $\Pi_{0}=|0\rangle\langle 0|\otimes|\psi_{c}\rangle\langle\psi_{c}|.$ The recipe to determine the state recurrence probability $S(\psi_{c})$ is similar to the site recurrence, but instead of the operator valued generating functions we will deal with scalars. The matrix element of the Stieltjes operator (11) with the initial coin state $|\psi_{c}\rangle$ yields the generating function for return amplitudes without prior monitoring $\mu_{\psi_{c}}(z)=\langle\psi_{c}|\mu_{c}(z)|\psi_{c}\rangle.$ The generating function for the first arrival amplitudes is then given by [12] $a_{\psi_{c}}(z)=1-\frac{1}{\mu_{\psi_{c}}(z)}.$ Finally, the state recurrence probability is obtained by a formula analogous to (7) $S(\psi_{c})=\int\limits_{0}^{2\pi}\lvert a_{\psi_{c}}(e^{it})\rvert^{2}\frac{dt}{2\pi}.$ (15) In contrast to the site recurrence, which was tractable, the integrand in (15) is usually a rather complicated function and we rely on numerical integration. Below we present plots with numerical results for several initial states. For Figure 3 we consider the basis states $|\alpha_{i}\rangle$ and plot the state recurrence probability as a function of the coin parameter $\rho$. For $|\alpha_{2}\rangle$ and $|\alpha_{3}\rangle$ the integral in (15) can be evaluated analytically and we find that the state recurrence is equal to the site recurrence. In fact, numerical simulations reveal that for the initial state $|\alpha_{3}\rangle$ the walker always returns to the origin in the $|\alpha_{3}\rangle$ state. However, for $|\alpha_{2}\rangle$ the situation is more complicated - the walker returns to the original site in a superposition of $|\alpha_{1}\rangle$ and $|\alpha_{3}\rangle$ with different weights in every time step. Nevertheless, the state recurrence probability equals (13). For $|\alpha_{1}\rangle$ the state recurrence probability is smaller than site recurrence probability (which equals 1), except for the boundary cases of $\rho=0,1$. Figure 3: State recurrence probability for the state $|\alpha_{1}\rangle$ (blue dots) and $|\alpha_{2}\rangle$ and $|\alpha_{3}\rangle$ (orange dots) in dependence on $\rho$. Figures 4 and 5 focus on the three-state Grover walk ($\rho=\frac{1}{\sqrt{3}}$). In Figure 4 we consider the initial state $\frac{1}{\sqrt{2}}(|\alpha_{1}\rangle+e^{i\phi}|\alpha_{2}\rangle)$ and plot $S$ as a function of the angle $\phi$. Note that for superpositions $\frac{1}{\sqrt{2}}(|\alpha_{1}\rangle+e^{i\phi}|\alpha_{3}\rangle)$ and $\frac{1}{\sqrt{2}}(|\alpha_{2}\rangle+e^{i\phi}|\alpha_{3}\rangle)$ the angle $\phi$ does not affect the state recurrence probability due to the block- diagonal form of the Stieltjes operator (11). Figure 5 shows state recurrence probability for superposition states $a|\alpha_{1}\rangle+\sqrt{1-a^{2}}|\alpha_{2}\rangle$, $a|\alpha_{1}\rangle+\sqrt{1-a^{2}}|\alpha_{3}\rangle$ and $a|\alpha_{2}\rangle+\sqrt{1-a^{2}}|\alpha_{3}\rangle$ as a function of $a$. Figure 4: State recurrence probability of the Grover walk ($\rho=1/\sqrt{3}$) for the state $\frac{1}{\sqrt{2}}(|\alpha_{1}\rangle+e^{i\phi}|\alpha_{2}\rangle)$ as a function of the angle $\phi$. Figure 5: State recurrence probabilities of the Grover walk ($\rho=1/\sqrt{3}$) for the states $a|\alpha_{1}\rangle+\sqrt{1-a^{2}}|\alpha_{2}\rangle$ (blue dots), $a|\alpha_{1}\rangle+\sqrt{1-a^{2}}|\alpha_{3}\rangle$ (orange dots) and $a|\alpha_{2}\rangle+\sqrt{1-a^{2}}|\alpha_{3}\rangle$ (green dots) as a function of $a$. Finally, we point out that the three-state walk shows similar paradoxical behavior as reported in [13] for a walk on a half-line or some 2D quantum walks, where the state recurrence probability can be greater than the site recurrence probability. In Figure 6 we consider the initial coin state $|\psi_{c}\rangle=\sqrt{\frac{1-\rho^{2}}{2}}|R\rangle+\rho|S\rangle+\sqrt{\frac{1-\rho^{2}}{2}}|L\rangle=\rho|\alpha_{1}\rangle-\sqrt{1-\rho^{2}}|\alpha_{2}\rangle,$ (16) which is one of the eigenvectors used in the construction of coin operator [26]. The plot indicates that the site recurrence probability, depicted by the orange curve, is smaller than the state recurrence probability up to $\rho\approx 0.79$. Figure 6: State recurrence (blue dots) versus site recurrence (orange curve) for the initial coin state (16). ## 5 Conclusions The presented results demonstrate that in contrast to the simple two-state quantum walks the site recurrence of the three-state model depends on the initial state. For the selected one-parameter family of coins we were able to derive the site recurrence probability in closed form. Extension to an arbitrary 3x3 unitary coin seems to be intractable due to the complexity and large number of free parameters. Nevertheless, for all three-state walks there will be an initial coin state with Polya number equal to one, namely the state given by $C^{-1}|S\rangle$. Indeed, such state will remain on the initial position after the first step, and thus is absorbed with certainty. It is an open question if such behaviour of site recurrence is possible for models without the staying put option in the shift operator, e.g. in quantum walks driven by Wigner rotation matrices of order $2j+1$ studied in [29, 30] for the case of half-integer $j$. As discussed in Section 4, obtaining closed formulas for state recurrence probability was possible only for particular initial coin states due to the complexity of the involved integrals. The same applies to both site and state recurrence for quantum walks on higher dimensional lattices, since in such cases already the evaluation of the Stieltjes operator through multidimensional Fourier transform cannot be easily reduced to calculation of residues as in the one-dimensional case. Nevertheless, the formulas allow to obtain approximations through numerical evaluation. It would be interesting to find if there exists site recurrent initial conditions e.g. for 2D quantum walks with Grover coin and its extensions. Additional twist can come from considering coined quantum walks combined with some form of classical randomness resulting in iterated open quantum evolution, e.g. quantum walks on dynamically percolated lattices [31, 32, 33]. Recent extension of monitored recurrence to arbitrary iterated closed operators on Banach spaces [20] and quantum Markov chains [23] allows to investigate such scenarios. MŠ is grateful for financial support from RVO14000 and ”Centre for Advanced Applied Sciences”, Registry No. CZ.02.1.01/0.0/0.0/16 019/0000778, supported by the Operational Programme Research, Development and Education, co-financed by the European Structural and Investment Funds. On a personal note, I want to express my gratitude to prof. Igor Jex for years of stimulating discussions and mediation of a myriad of personal contacts with excellent researchers worldwide. Live long and prosper! ## References ## References * [1] Y. Aharonov, L. Davidovich, and N. Zagury. Quantum random walks. Phys. Rev. A, 48:1687–1690, 1993. * [2] D. A. Meyer. From quantum cellular automata to quantum lattice gases. J. Stat. Phys., 85:551–574, 1996. * [3] E Farhi and S Gutmann. Quantum computation and decision trees. Phys. Rev. A, 58:915–928, 1998. * [4] R. P. Feynman and A. R. Hibbs. Quantum mechanics and path integrals. International series in pure and applied physics. McGraw-Hill, New York, 1965. * [5] S. P. Gudder. Quantum graphic dynamics. Found. Phys., 18:751–776, 1988. * [6] S. P. Gudder. Quantum Probability. Academic Press, New York, 1988. * [7] K. Kadian, S. Garhwal, and A. Kumar. Quantum walk and its application domains: A systematic review. Comput. Sci. Rev., 41, 2021. * [8] G. Pólya. Uber eine aufgabe betreffend die irrfahrt im strassennetz. Math. Ann., 84:149–160, 1921. * [9] M. Kac. On the notion of recurrence in discrete stochastic processes. Bull. Amer. Math. Soc., 53:1002, 1947. * [10] M. Štefaňák, I. Jex, and T. Kiss. Recurrence and Pólya number of quantum walks. Phys. Rev. Lett., 100:020501, 2008. * [11] M. Štefaňák, T. Kiss, and I. Jex. Recurrence properties of unbiased coined quantum walks on infinite d-dimensional lattices. Phys. Rev. A, 78:032306, 2008. * [12] F. A. Grünbaum, L. Velázquez, A. H. Werner, and R. F. Werner. Recurrence for Discrete Time Unitary Evolutions. Communications in Mathematical Physics, 320:543, 2013. * [13] J. Bourgain, F. A. Grünbaum, L. Velazquez, and J. Wilkening. Quantum recurrence of a subspace and operator-valued schur functions. Commun. Math. Phys., 329:1031–1067, 2014. * [14] T. Kiss, L. Kecskés, M. Štefaňák, and I. Jex. Recurrence in coined quantum walks. Phys. Script., T135:014055, 2009. * [15] A. Ambainis, E. Bach, A. Nayak, A. Vishwanath, and J. Watrous. One-dimensional quantum walks. In STOC’01 Proceedings of the 33rd annual ACM symposium on Theory of computing, pages 37–49, 2001. * [16] P. Sinkovicz, Z. Kurucz, T. Kiss, and J. K. Asbóth. Quantized recurrence time in unital iterated open quantum dynamics. Phys. Rev. A, 91:042108, 2015. * [17] P. Sinkovicz, T. Kiss, and J. K. Asbóth. Generalized Kac lemma for recurrence time in iterated open quantum systems. Phys. Rev. A, 93:050101, 2016. * [18] C. F. Lardizabal and R. R. Souza. On a Class of Quantum Channels, Open Random Walks and Recurrence. J. Stat. Phys., 159:772, 2015. * [19] S. L. Carvalho, L. F. Guidi, and C. F. Lardizabal. Site recurrence of open and unitary quantum walks on the line. Quantum Inform. Process., 16:17, 2016. * [20] F.A. Grünbaum and L. Velázquez. A generalization of Schur functions: Applications to Nevanlinna functions, orthogonal polynomials, random walks and unitary and open quantum walks. Adv. Math., 326:352, 2018. * [21] I. Bardet, H. Bringuier, Y. Pautrat, and C. Pellegrini. Recurrence and Transience of Continuous-Time Open Quantum Walks. In C. DonatiMartin, A. Lejay, and A. Rouault, editors, Seminaire De Probabilites L, volume 2252, page 493. Springer International Publishing Ag, Cham, 2019. * [22] A. Dhahri and F. Mukhamedov. Open quantum random walks, quantum Markov chains and recurrence. Rev. Math. Phys., 31:1950020, 2019. * [23] F. A. Grunbaum, C. F. Lardizabal, and L. Velazquez. Quantum Markov Chains: Recurrence, Schur Functions and Splitting Rules. Ann. Henri Poincare, 21:189, 2020. * [24] T. S. Jacq and C. F. Lardizabal. Homogeneous Open Quantum Walks on the Line: Criteria for Site Recurrence and Absorption. Quantum Inform. Comput., 21:37, 2021. * [25] T. Nitsche, S. Barkhofen, R. Kruse, L. Sansoni, M. Štefaňák, A. Gábris, V. Potoček, T. Kiss, I. Jex, and Ch. Silberhorn. Probing measurement-induced effects in quantum walks via recurrence. Sci. Adv., 4:eaar6444, 2018. * [26] M. Štefaňák, I. Bezděková, and I. Jex. Continuous deformations of the grover walk preserving localization. Eur. Phys. J. D, 66:142, 2012. * [27] M. Sabri, E. Segawa, and M. Štefaňák. Conditional limit measure of a one-dimensional quantum walk with an absorbing sink. Phys. Rev. A, 98:012136, 2018. * [28] M. Štefaňák, I. Bezděková, and I. Jex. Limit distributions of three-state quantum walks: The role of coin eigenstates. Phys. Rev. A, 90:012342, 2014. * [29] T. Miyazaki, M. Katori, and N. Konno. Wigner formula of rotation matrices and quantum walks. Phys. Rev A, 76:012332, 2007. * [30] I. Bezděková, M. Štefaňák, and I. Jex. Suitable bases for quantum walks with wigner coins. Phys. Rev. A, 92:022347, 2015. * [31] B. Kollár, T. Kiss, J. Novotný, and I. Jex. Asymptotic dynamics of coined quantum walks on percolation graphs. Phys. Rev. Lett., 108:230505, 2012. * [32] B. Kollár, J. Novotný, T. Kiss, and I. Jex. Percolation induced effects in two-dimensional coined quantum walks: analytic asymptotic solutions. New J. Phys., 16:023002, 2014. * [33] B. Kollár, J. Novotný, T. Kiss, and I. Jex. Discrete time quantum walks on percolation graphs. Eur. Phys. J. Plus, 129:103, 2014.
001<EMAIL_ADDRESS><EMAIL_ADDRESS> # Predictive Sampling: Real-time Behaviour Synthesis with MuJoCo Taylor Howell Stanford University DeepMind Nimrod Gileadi DeepMind Saran Tunyasuvunakool DeepMind Kevin Zakka DeepMind University of California Berkeley Tom Erez DeepMind Yuval Tassa DeepMind We introduce MuJoCo MPC (MJPC), an open-source, interactive application and software framework for real-time predictive control, based on MuJoCo physics. MJPC allows the user to easily author and solve complex robotics tasks, and currently supports three shooting-based planners: derivative-based iLQG and Gradient Descent, and a simple derivative-free method we call Predictive Sampling. Predictive Sampling was designed as an elementary baseline, mostly for its pedagogical value, but turned out to be surprisingly competitive with the more established algorithms. This work does not present algorithmic advances, and instead, prioritises performant algorithms, simple code, and accessibility of model-based methods via intuitive and interactive software. MJPC is available at github.com/deepmind/mujoco_mpc, a video summary can be viewed at dpmd.ai/mjpc. ## 1 Introduction Model-based approaches form the foundation of classical control and robotics. Since Kalman’s seminal work [Kalman, 1960], the _state_ along with its dynamics and observation models has played a central role. The classical approach is being challenged by learning-based methods, which forgo the explicit description of the state and associated models, letting internal representations emerge from the learning process [Lillicrap et al., 2015, Schulman et al., 2017, Salimans et al., 2017, Smith et al., 2022, Rudin et al., 2022]. The flexibility afforded by learned representations makes these methods powerful and general, but the requirement for large amounts of data and computation makes them slow. In contrast, pure model-based methods, like the ones described below, can synthesise behaviour in real time [Tassa et al., 2014]. Since both approaches ultimately generate behaviour by optimising an objective, there is reason to believe they can be effectively combined. Indeed, well-known discrete-domain breakthroughs like AlphaGo [Silver et al., 2016] are predicated on combining model-based search and learning-based value and policy approximation. We believe the same could happen for robotics and control, and describe our thinking on how this might happen in the Discussion (Section 5). However, before the community can rise to this challenge, a core problem must be overcome: Model-based optimisation is difficult to implement, often depends on elaborate optimisation algorithms, and is generally inaccessible. To address this deficit, we present MJPC, an open-source interactive application and software framework for predictive control, based on MuJoCo physics [Todorov et al., 2012], which lets the user easily author and solve complex tasks using predictive control algorithms in real time. The tool offers implementations of standard derivative-based algorithms: iLQG (second- order planner) and Gradient Descent (first-order planner). Additionally, it introduces _Predictive Sampling_ , a simple zero-order, sampling-based algorithm that works surprisingly well and is easy to understand. Importantly, the interactive simulation can be slowed down asynchronously, speeding up the planner with respect to simulation time. This means that behaviours can be generated on older, slower machines, leading to a democratisation of predictive control tooling. MJPC and Predictive Sampling advance our central goal of lowering the barriers to entry for predictive control in robotics research. An important secondary goal is to accelerate research velocity. When tweaking a parameter, a researcher should not need to wait hours or minutes, but should receive instantaneous feedback – which will measurably enhance their own cognitive performance [Lu and Dosher, 2022]. We believe that flexible, interactive simulation with researcher-authored graphical user interface is not just a “nice to have”, but a prerequisite for advanced robotics research. ## 2 Background In this section we provide a general background on Optimal Control and Trajectory Optimisation, then focus on Predictive Control and derivative-free optimisation. #### Optimal control. Optimal Control means choosing actions in order to minimise future costs (equivalently, to maximise future returns). A dynamical system with _state_ ${x}\in\mathbf{R}^{n}$, which takes a user-chosen _control_ (or _action_) ${u}\in\mathbf{R}^{m}$, evolves according to the discrete-time111The continuous-time formulation is generally equivalent, we choose discrete time for notation simplicity. dynamics: ${y}=f({x},{u}),$ (1) returning a new state, ${y}\in\mathbf{R}^{n}$. The behaviour of the system is encoded via the running cost: ${c}({x},{u}),$ (2) a function of state and control, where explicit time-dependence can be realised by folding time into the state. Future costs (a.k.a _cost-to-go_ or _value_), can be defined in several ways. The summation can continue to infinity, leading to the _average-cost_ or _discounted-cost_ formulations, favoured in temporal-difference learning, which we discuss in Section 5. Here we focus on the _finite-horizon_ formulation, whereby the optimisation objective ${J}$ is given by: ${J}({x}_{0:T},{u}_{0:T})=\sum_{t=0}^{T}{c}({x}_{t},{u}_{t}),$ (3) where subscripts indicated discrete-time indices. #### Trajectory optimisation. Solving the finite-horizon optimal control problem (1, 3), i.e., optimising a fixed-length trajectory, is commonly known as _planning_ or _trajectory optimisation_. These algorithms [Von Stryk and Bulirsch, 1992, Betts, 1998] have a rich history reaching back to the Apollo Program [NASA, 1971, Smith and Yound, 1967]. An important distinction can be made between two classes of algorithms: * • _Direct_ or _simultaneous_ methods have both states and controls as decision variables and enforce the dynamics (1) as constraints. These methods (e.g., Von Stryk [1993]) specify a large, sparse optimisation problem, which is usually solved with general-purpose software [Wächter and Biegler, 2006, Gill et al., 2005]. They have the important benefit that non-physically-realisable trajectories can be represented, for example in order to clamp a final state, without initially knowing how to get to it. * • _Shooting_ methods like Differential Dynamic Programming [Jacobson and Mayne, 1970] use only the controls ${u}_{0:T}$ as decision variables and enforce the dynamics via forward simulation. In the shooting approach only physically- realisable trajectories can be considered, but they benefit from the reduced search space and from the optimiser not having to enforce the dynamics. The latter benefit is especially important for stiff systems like those with contact, where the difference between a physical and non-physical trajectory can be very small222For example, consider the physical scenario of a free rigid box lying flat on a plane under gravity, and then consider the non- physical scenario of the same box hovering above the plane or penetrating it by a few microns.. Unlike _direct_ methods which require dynamics derivatives, shooting methods can employ derivative-free optimisation, as discussed below. #### Predictive control. Algorithm 1 Predictive Control (asynchronous) Agent: (repeat) 1:Read the current action ${u}$ from the _nominal_ plan ${\mathbf{\Pi}}$, apply it to the controlled system. Planner: (repeat) 1:Measure the current state ${x}$. 2:Using the nominal ${\mathbf{\Pi}}$ to warm-start, optimise the finite- horizon objective ${J}$. 3:Update ${\mathbf{\Pi}}$. The key idea of Predictive Control, invented in the late 60s and first published in [Richalet et al., 1978], is to use trajectory optimisation in _real-time_ as the system dynamics are evolving. This class of algorithm has been successfully deployed in numerous real-world settings including: chemical and nuclear process control [Na et al., 2003, Lopez-Negrete et al., 2013], navigation for autonomous vehicles [Falcone et al., 2007], and whole-body control of humanoid robots [Kuindersma et al., 2016]. In the real-time setting, the current state ${x}$ needs to be estimated or measured, and the trajectory optimiser is required to return a set of optimal or near-optimal controls for the finite-horizon (here often called the _receding horizon_) problem, starting at ${x}$. We use ${\mathbf{\Pi}}$ to denote the _plan_ , the finite-horizon policy. In the context of shooting methods ${\mathbf{\Pi}}={u}_{0:T}$, though as we discuss later, in some cases it can be re-parameterised, rather than using the discrete-time control sequence directly. Predictive control is best thought of in terms of two asynchronous processes, the _agent_ and the _planner_ , see Algorithm 1. Predictive Control has the following notable properties: * • Faster computation improves performance. The reason for this is clear, the more optimisation steps the planner can take in one unit of time, the better the optimised control sequences will be. * • Warmstarting has a large beneficial effect. By reusing the plan from the previous planning step, the optimiser only needs to make small modifications in order to correct for the changes implied by the new state. Warm-starting also leads to an amortisation of the optimisation process across multiple planning steps. * • The optimisation is not required to converge, only to improve, see Section 5.1. * • Tasks with a behaviour timescale much longer than the planning horizon $T$ are often still solvable, though this property is task dependent. * • Predictive controllers can easily get stuck in local minima, especially those using derivatives. This is due to the myopic nature of the optimisation, and can be addressed by terminating the rollout with a value function approximation, see Section 5.3. #### Derivative-free optimisation. One of the main reasons for our initial focus on shooting methods is their ability to make use of derivative-free optimisation, also known as _sampling- based_ optimisation [Audet and Hare, 2017]. Despite not leveraging problem structure or gradient information, this class of algorithms can discover complex behaviours [Salimans et al., 2017, Mania et al., 2018]. Sampling-based methods maintain a search distribution over policy parameters and evaluate the objective at sampled points in order to find an improved solution. Popular algorithms include: random search [Matyas, 1965], genetic algorithms [Holland, 1992], and evolutionary strategies [Rechenberg, 1973], including CMA-ES [Hansen and Ostermeier, 2001]. This class of algorithms has a number of desirable properties. First, because derivative information is not required, they are well-suited for tasks with non-smooth and discontinuous dynamics. Second, these methods are trivially parallelisable. Sampling-based methods have been used in the predictive control context, but not very widely. Notable exceptions are Hämäläinen’s work [Hämäläinen et al., 2014, 2015], which is an early precursor to MJPC, and the oeuvre of Theodorou including [Williams et al., 2017] and related papers. While these methods are usually considered to be sample inefficient, we explain below why, specifically in the predictive control context, they can be surprisingly competitive. ## 3 MuJoCo MPC (MJPC) We introduce MJPC, an open-source interactive application and software framework for predictive control, that lets the user easily synthesise behaviours for complex systems using predictive control algorithms in real time. Behaviours are specified by simple, composable objectives that are risk- aware. The planners, including: Gradient Descent, Iterative Linear Quadratic Gaussian (iLQG), and Predictive Sampling are implemented in C++ and extensively utilise multi-threading for parallel rollouts. The framework is asynchronous, enabling simulation slow-down and emulation of a faster controller, allowing this tool to run on slow machines. An intuitive graphical user-interface enables real-time interactions with the environment and the ability to modify task parameters, planner and model settings, and to instantly see the effects of the modifications. The tool is available at: https://github.com/deepmind/mujoco_mpc. ### 3.1 Physics Simulation We build MJPC using the API of the open-source physics engine MuJoCo [Todorov et al., 2012]. MuJoCo is a good infrastructure for an interactive framework for robotics algorithms for two main reasons: first, MuJoCo supports simulating multiple candidate future trajectories in parallel by offering a thread-safe API, which maximises the utilisation of modern multi-core CPU architectures; second, MuJoCo affords faster-than-realtime simulation of high- dimensional systems with many contacts — for example, the humanoid (a 27-DoF system) can be simulated 4000 times faster than realtime on a single CPU thread. ### 3.2 Objective MJPC provides convenient utilities to easily design and compose costs in order to specify an objective (3); as well as automatically and efficiently compute derivatives. #### Costs. We use a “base cost” of the form: ${l}({x},{u})=\sum_{i=0}^{M}w_{i}\cdot\text{n}_{i}\big{(}{r}_{i}({x},{u})\big{)}.$ (4) This cost is a sum of $M$ terms, each comprising: * • A nonnegative weight $w\in\mathbf{R}_{+}$ determining the relative importance of this term. * • A twice-differentiable norm $\text{n}(\cdot):\mathbf{R}^{p}\rightarrow\mathbf{R}_{+}$, which takes its minimum at $0^{p}$. * • The residual ${r}\in\mathbf{R}^{p}$ is a vector of elements that are “small when the task is solved”. #### Risk sensitivity. Figure 1: Risk transformation $\rho(l;R)$. The function is evaluated between $0$ and $1$ for different values of the risk parameter $R$. We augment the base cost (4) with a risk-aware exponential scalar transformation, $\rho:\mathbf{R}_{+}\times\mathbf{R}\rightarrow\mathbf{R}$, corresponding to the classical risk-sensitive control framework [Jacobson, 1973, Whittle, 1981]. The final running cost ${c}$ is given by: ${c}({x},{u})=\rho({l}({x},{u});R)=\frac{e^{R\cdot{l}({x},{u})}-1}{R}$ (5) The scalar parameter $R\in\mathbf{R}$ denotes risk-sensitivity. $R=0$ (the default) is interpreted as risk-neutral, $R>0$ as risk-averse, and $R<0$ as risk-seeking. The mapping $\rho$ (see Figure 1) has the following properties: * • Defined and smooth for any $R$. * • If $R=0$, is the identity $\rho({l};0)={l}$ (in the limit). * • Zero is a fixed point: $\rho(0;R)=0$. * • The derivative at 0 is 1: $\partial\rho(0;R)/\partial{l}=1$. * • Non-negative: If ${l}\geq 0$ then $\rho({l};R)\geq 0$. * • Monotonic: $\rho({l};R)>\rho(z;R)$ if ${l}>z$. * • If $R<0$ then $\rho$ is bounded: $\rho({l};R)<-\frac{1}{R}$. * • $\rho({l};R)$ carries the same units as ${l}$. Note that for negative $R$, the transformation $\rho$ creates costs that are similar to the bounded rewards commonly used in reinforcement learning. For example, when using the quadratic norm $\text{n}(r)=r^{T}Wr$ for some SPD matrix $W=\Sigma^{-1}$ and a risk parameter $R=-1$, we get an inverted- Gaussian cost ${c}=1-e^{-r^{T}\Sigma^{-1}r}$, whose minimisation is equivalent to maximum-likelihood maximisation of the Gaussian. This leads to the interesting interpretation of the bounded rewards commonly used in RL as _risk-seeking_. We do not investigate this relationship further in this paper. #### Derivatives. MuJoCo provides a utility for computing finite-difference (FD) Jacobians of the dynamics, which is efficient in two ways. First, by avoiding re- computation where possible, for example when differencing w.r.t. controls, quantities that depend only on positions and velocities are not recomputed. Second, because FD computational costs scale with the dimension of the _input_ , outputs can be added cheaply. MuJoCo’s step function ${y},{r}=f({x},{u})$, computes both the next state ${y}$ and sensor values $r$, defined in the model. Because the FD approximation of the Jacobians, $\frac{\partial{y}}{\partial{x}},\frac{\partial{y}}{\partial{u}},\frac{\partial{r}}{\partial{x}},\frac{\partial{r}}{\partial{u}},$ (6) scales like the combined dimension of ${x}$ and ${u}$, adding more sensors ${r}$ is effectively “free”. MJPC automatically and efficiently computes cost derivatives as follows. #### Gradients. Cost gradients are computed with: $\displaystyle\frac{\partial{c}}{\partial{x}}$ $\displaystyle=e^{R{l}}\frac{\partial{l}}{\partial{x}}=e^{R{l}}\sum\limits_{i=0}^{M}w_{i}\frac{\partial\text{n}_{i}}{\partial r}\frac{\partial r_{i}}{\partial{x}},$ (7a) $\displaystyle\frac{\partial{c}}{\partial{u}}$ $\displaystyle=e^{R{l}}\frac{\partial{l}}{\partial{u}}=e^{R{l}}\sum\limits_{i=0}^{M}w_{i}\frac{\partial\text{n}_{i}}{\partial r}\frac{\partial r_{i}}{\partial{u}}.$ (7b) The norm gradients, $\partial\text{n}/\partial r$, are computed analytically. #### Hessians. Second-order derivatives use the Gauss-Newton approximation, ignoring second derivatives of ${r}$: $\displaystyle\frac{\partial^{2}{c}}{\partial{x}^{2}}\approx$ $\displaystyle\,e^{R{l}}\left[\sum\limits_{i=0}^{M}w_{i}{\frac{\partial r_{i}}{\partial{x}}}^{T}\frac{\partial^{2}\text{n}_{i}}{\partial r^{2}}\frac{\partial r_{i}}{\partial{x}}+R{\frac{\partial{l}}{\partial{x}}}^{T}\frac{\partial{l}}{\partial{x}}\right],$ (8a) $\displaystyle\frac{\partial^{2}{c}}{\partial{u}^{2}}\approx$ $\displaystyle\,e^{R{l}}\left[\sum\limits_{i=0}^{M}w_{i}{\frac{\partial r_{i}}{\partial{u}}}^{T}\frac{\partial^{2}\text{n}_{i}}{\partial r^{2}}\frac{\partial r_{i}}{\partial{u}}+R{\frac{\partial{l}}{\partial{u}}}^{T}\frac{\partial{l}}{\partial{u}}\right],$ (8b) $\displaystyle\frac{\partial^{2}{c}}{\partial{x}\partial{u}}\approx$ $\displaystyle\,e^{R{l}}\left[\sum\limits_{i=0}^{M}w_{i}{\frac{\partial r_{i}}{\partial{x}}}^{T}\frac{\partial^{2}\text{n}_{i}}{\partial r^{2}}\frac{\partial r_{i}}{\partial{u}}+R{\frac{\partial{l}}{\partial{x}}}^{T}\frac{\partial{l}}{\partial{u}}\right].$ (8c) The norm Hessians, $\partial^{2}\text{n}/\partial r^{2}$, are computed analytically. ### 3.3 Splines Figure 2: Time-indexed spline representation of the controls. Parameter points (black) utilised to construct: zero (magenta), linear (orange), and cubic (blue) interpolants. As we explain below, planners like iLQG require the _direct_ control-sequence representation ${u}_{0:T}$ due to the requirements of the Bellman Principle. Without this constraint, controls can be “compressed” into a lower-dimensional object. There are many ways to do this, we picked the simplest: splines. Action trajectories are represented as a time-indexed set of knots, or control-points, parameterised by a sequence of monotonic time points $\tau_{0:P}$ and parameter values $\theta_{0:P}$, where we use the shorthand $\theta=\theta_{0:P}$. Given a query point $\tau$, the evaluation of the spline is given by: ${u}=s\big{(}\tau;(\tau_{0:P},\theta)\big{)}.$ (9) We provide three spline implementations: traditional cubic Hermite splines, piecewise-linear interpolation, and zero-order hold. See (Fig. 2) for an illustration. The main benefit of compressed representations like splines is that they reduce the search space. They also smooth the control trajectory, which is often desirable. Spline functions belong to the class of linear bases, which includes the Fourier basis and orthogonal polynomials. These are useful because they allow easy propagation of gradients from the direct representation $\partial{\mathbf{\Pi}}$ back to the parameter values $\partial\theta$. In our case this amounts to computing: $\frac{\partial s}{\partial\theta},$ (10) which has a simple analytic formula (see code for details). Unlike other linear bases, splines have the convenient property that bounding the values $\theta$ also bounds the spline trajectory. This is exactly true for the zero and linear interpolations, and mostly-true for cubic splines. Bounding is important as most physical systems clamp controls to bounds, and there is no point searching outside of them. Expressions for cubic, linear and zero interpolations are provided in Appendix A. ### 3.4 Planners MJPC includes two derivative-based planners. #### iLQG. Algorithm 2 iLQG 1:initial state ${x}_{0}$, nominal plan ${\mathbf{\Pi}}={u}_{0:T}$ 2:Roll out nominal trajectory from ${x}_{0}$ using ${\mathbf{\Pi}}$ 3:Compute action improvements and feedback policy using Dynamic Programming. 4:Roll out parallel line search with feedback policy (11). 5:Best actions are new nominal actions. The iLQG333Equivalently, “iLQR”, since we don’t make use of the noise- sensitive term for which iLQG was originally developed [Li and Todorov, 2004]. We keep the name “iLQG” due to its provenance. planner [Tassa et al., 2012], a Gauss-Newton approximation of the DDP algorithm [Jacobson and Mayne, 1970], utilises first- and second-order derivative information to take an approximate Newton step over the open-loop control sequence ${u}_{0:T}$ via dynamic programming [Kalman, 1964], producing a time-varying linear feedback policy: ${u}_{t}=\bar{{u}}_{t}+K_{t}({x}_{t}-\bar{{x}}_{t})+\alpha k_{t}.$ (11) The _nominal_ , or current best trajectory, is denoted with overbars ($\bar{\phantom{x}}$), $K$ is a feedback gain matrix, and $k$ is an improvement to the current action trajectory. A parallel line search over the step size $\alpha\in[\alpha_{\textrm{min}},1]$ is performed to find the best improvement. Additional enhancements include a constrained backward pass [Tassa et al., 2014] that enforces action limits and adaptive regularisation. The details of iLQG are too involved to restate here, we refer the reader to the references above for details. #### Gradient descent. Algorithm 3 Gradient Descent 1:initial state ${x}_{0}$, nominal plan ${\mathbf{\Pi}}(\theta)$ 2:Roll out nominal from ${x}_{0}$ using ${\mathbf{\Pi}}(\theta)$ 3:Compute $\partial{J}/\partial{\mathbf{\Pi}}$ with (14) 4:Compute $\partial{J}/\partial\theta$ with (13) 5:Roll out parallel line-search with (12) 6:Pick the best one: $\theta\leftarrow\mbox{argmin}\bigl{(}{J}(\theta^{(i)})\bigr{)}$ This first-order planner, known as Pontryagin’s Maximum Principle [Mangasarian, 1966], utilises gradient information to improve action sequences, here represented as splines. The gradient of the total return is used to update the spline parameters, using a parallel line search over the step size $\alpha\in[\alpha_{\textrm{min}},\alpha_{\textrm{max}}]$: $\theta\leftarrow\theta-\alpha\frac{\partial J}{\partial\theta}.$ (12) The total gradient is given by: $\frac{\partial J}{\partial\theta}=\frac{\partial J}{\partial{\mathbf{\Pi}}}\frac{\partial{\mathbf{\Pi}}}{\partial\theta},$ (13) where the spline gradient $\partial{\mathbf{\Pi}}/\partial\theta$ is given by (10), while $\partial J/\partial{\mathbf{\Pi}}$ is computed with the Maximum Principle. Letting $\lambda$ denote the _co-state_ , the gradients with respect to ${u}$ are given by: $\displaystyle\lambda_{j}$ $\displaystyle=\frac{\partial c}{\partial{x}_{j}}+\Bigl{(}\frac{\partial f}{\partial{x}_{j}}\Bigr{)}^{T}\lambda_{j+1},$ (14a) $\displaystyle\frac{\partial{J}}{\partial{u}_{j}}$ $\displaystyle=\frac{\partial c}{\partial{u}_{j}}+\Bigl{(}\frac{\partial f}{\partial{u}_{j}}\Bigr{)}^{T}\lambda_{j+1}.$ (14b) The primary advantage of this first-order method compared to a computationally more expensive method like iLQG is that optimisation is performed over the smaller space of spline parameters, instead of the entire (non-parametric) sequences of actions. #### Predictive Sampling. Algorithm 4 Predictive Sampling Parameters: $N$ rollouts, noise scale $\sigma$ 1:initial state ${x}_{0}$, nominal plan ${\mathbf{\Pi}}(\theta)$ 2:Get $N\\!-\\!1$ samples $\theta_{i}\sim\mathcal{N}(\theta,\sigma^{2})$ 3:Including $\theta$, roll out all $N$ samples from ${x}_{0}$ 4:Pick the best one: $\theta\leftarrow\mbox{argmin}\bigl{(}{J}(\theta^{(i)})\bigr{)}$ This is a trivial, zero-order, sampling-based Predictive Control method that works well and is easy to understand. Designed as an elementary baseline, this algorithm turned out to be surprisingly competitive with the more elaborate derivative-based algorithms. #### Algorithm. A nominal sequence of actions, represented with spline parameters, is iteratively improved using random search [Matyas, 1965]. At each iteration, $N$ candidate splines are evaluated: the nominal itself and $N-1$ noisy samples from a Gaussian with the nominal as its mean and fixed standard deviation $\sigma$. After sampling, the actions are clamped to the control limits by clamping the spline parameters $\theta$. Each candidate’s total return is evaluated and the nominal is updated with the best candidate. See Algorithm 4 and pseudocode in Appendix C. Predictive Sampling is not innovative or performant, but is presented as a simple baseline, see Discussion below. ## 4 Results We provide a short textual description of our graphical user interface (GUI) for three example tasks. They are best understood by viewing the associated video at dpmd.ai/mjpc or better yet, by downloading the software and interacting with it. ### 4.1 Graphical User Interface Figure 3: Graphical User Interface. The left tab includes modules for Tasks and the Agent. In the Task module, models are selected from a drop-down menu and the risk value is set; cost weights and residual parameters are specified with interactive sliders. The Agent module provides access to a drop-down menu for planners, settings, and switches that toggle the planner and controller. The right tab includes simulation readings and predictions for the cost, including individual terms, and actions. Live, planner-specific information and various compute time results are shown below. Traces of the current policy and sampled trajectories are visualised and the user can interactively pause or slow down the simulation and apply external forces and moments to the system. The MJPC GUI, shown and described in Fig. 3, provides an interactive simulation environment, as well as modules containing live plots and parameters that can be set by the researcher. The intuitive interface makes policy design easy by enabling the researcher to interactively change cost parameters or planner settings and immediately see the results in both the simulation environment and live plots, allowing fast debugging and an enhanced understanding of the factors that influence behaviour. ### 4.2 Examples (a) Humanoid standing up off the floor. (b) Quadruped rolling off its back to stand up. (c) Hand manipulating a cube to a goal orientation. Figure 4: Behaviours generated with MuJoCo MPC. Time progresses left to right. In the following examples, we demonstrate the ability to synthesise complex locomotion and manipulation behaviours for a variety of high-dimensional systems in simulation on a single CPU. Further, we demonstrate that the behaviours are robust to disturbances and mismatch between the simulation and planning model, and can adapt extremely quickly in new scenarios. For all of the examples, the total planning time for a single update is between 1 and 20 milliseconds. We highlight three examples below and provide additional examples with the software. Experimental details for objectives and planner settings are found in the Appendix B. #### Humanoid. This 27-DOF human-like system, from DeepMind Control Suite [Tunyasuvunakool et al., 2020], has 21 actions and is tasked with standing. The system can be initialised on the floor and quickly stands in a manner that is robust to large disturbances. If a sufficiently large disturbance knocks the humanoid onto the floor, the system will stand back up (Fig. 4(a)). #### Quadruped. A Unitree A1 quadruped [Unitree, 2022], from MuJoCo Menagerie [MuJoCo Menagerie Contributors, 2022], exhibits agile behaviour to traverse uneven terrain which includes walking over a steep slope. On slower machines, the quadruped often struggles to ascend. In this scenario, the simulation slow down can be effectively utilised to provide the planner with addition simulation time to plan a successful climb. The system is also capable of rolling off its back and standing up (Fig. 4(b)). In order to perform long- horizon tasks like continuously navigating the terrain, a series of target poses are set. Once a goal is reached, an automatic transition occurs and the next target is set. #### Hand. A Shadow Hand [Tuffield and Elias, 2003], also from MuJoCo Menagerie, performs in-hand manipulation of a cube to a desired orientation (Fig. 4(c)), where this goal can be set by the researcher in real-time by interactively setting the target orientation. In-hand reorientation—a high-DoF system with complex contact dynamics—is considered difficult to solve [Chen et al., 2022] and to the best of our knowledge has not previously been solved from scratch, in real time. ## 5 Discussion The thrust of this paper is to make predictive control accessible via customisable, interactive, open-source tooling. We believe that responsive, GUI-based tools are a prerequisite for accelerated robotics research, and that due to their importance, these tools should be modifiable and the inner workings transparent to the researcher. We hope that our MJPC project will be embraced by the community, and look forward to improving and extending it together. ### 5.1 Predictive Sampling The effectiveness of this simple method suggests that fast, approximate optimisation can be competitive with more sophisticated methods which return better solutions but at a lower rate. Does the higher planning rate completely explain this surprising effectiveness? We believe there is another, subtler reason. Predictive Control is not well-described by the tenets of traditional optimisation. For example, it usually makes no sense to take more than one step of optimisation. Once a single iteration is complete, it is more important to measure a new value of the state and re-plan, than it is to continue to converge to the minimum of an already-outdated problem. The constant shifting of the optimisation landscape makes Predictive Control a _qualitatively different problem_ , more like surfing than mountain climbing. The goal is not to find the minimum, but to _remain in the basin-of-attraction of the minimum_. This is a different, weaker criterion, at which simple algorithms fair better than when measured by the traditional yardstick of convergence. To be clear, Predictive Sampling is not a novel algorithm; instead, it is a baseline. A corner-case of many existing methods, it can variously be described as “MPPI with infinite temperature”, “CEM with a non-adaptive distribution” or just “trivial random search”. Better algorithms exist, but none are so easy to describe or implement. We are introducing Predictive Sampling not because it is good, but because it is _not good_. It is the simplest possible sampling-based shooting method, and therefore establishes a _lower bound_ for performance baselines. ### 5.2 Use cases Before we discuss limitations and their possible resolutions, it is worth asking how can MJPC be used _now_ , as it is described above? 1. 1. Task design. MJPC makes it easy to add new tasks, expose task parameters to the GUI, and quickly generate the desired behaviour. The task can then be re- implemented in any other framework of choice. While we have not yet implemented time-dependent tasks, it is possible and easy; we expect MJPC to work especially well for motion-tracking tasks. 2. 2. Data generation. MJPC can be used to generate data for learning-based approaches, i.e., it can act like an “expert policy”. In this context, it is often the case that the model and task from which the data is generated do not have to exactly match the one used by the learner, and the data can likely be useful for a wide range of setups. 3. 3. Predictive Control research. For researchers interested in Predictive Control itself, MJPC provides an ideal playground. MJPC can switch planners on-the- fly, and its asynchronous design affords a fair comparison by correctly accounting for and rewarding faster planners. ### 5.3 Limitation and Future Work #### Can only control what MuJoCo can simulate. This is a general limitation of Predictive Control and is in fact stronger since one can only control what can be simulated _much faster than real-time_. For example, it is difficult to imagine any simulation of a very-high-DoF system, like fluid, cloth or soft bodies advancing so fast. One solution is improved simulation using a combination of traditional physics modeling and learning, e.g., [Ladicky et al., 2017]. Another possibility is to entirely learn the dynamics from observations. This approach, often termed Model Based Reinforcement Learning is showing great promise [Heess et al., 2015, Nagabandi et al., 2020, Wu et al., 2022]. We would recommend that where possible, when attempting Predictive Control using learned dynamics models, a traditional simulator be employed as a fallback as in [Schrittwieser et al., 2020], to disambiguate the effects of modeling errors. #### Myopic. The core limitation of Predictive Control is that it is _myopic_ and cannot see past the fixed horizon. This can be ameliorated in three conceptually straightforward ways: 1. 1. Learned policies. By adding a learned policy, information from past episodes can propagate to the present via policy generalisation [Byravan et al., 2021]. This approach is attractive since it can only _improve_ performance: when rolling out samples, one also rolls out the proposal policy. If the rollout is better, it becomes the new nominal. A learned policy is also expected to lead to more stereotypical, periodic behaviours, which are important in locomotion. 2. 2. Value functions. Terminating the rollout with a learned value function which estimates the remaining cost-to-go is the obvious way by which to increase the effective horizon. Combining learned policies and value functions with model- based search would amount to an “AlphaGo for control” [Silver et al., 2016, Springenberg et al., 2020]. 3. 3. High-level agent. A predictive controller could be used as the low-level module in a hierarchical control setup. In this scenario, the actions of the high-level agent have the semantics of setting the cost function of the predictive controller. The predictive controller remains myopic while the high-level agent contains the long-horizon “cognitive” aspects of the task. A benefit of this scenario is that the high-level actions have a much lower frequency than required for low-level control (e.g., torques). #### Hardware. MJPC is aimed at robotics research, which raises the question, can it be used to control hardware? 1. 1. Transfer learning. As mentioned in 5.2, using MJPC to generate data which can then be transferred to a real robot is already possible. 2. 2. Estimation. The most obvious yet difficult route to controlling hardware is to follow in the footsteps of classic control and couple MJPC to an estimator providing real-time state estimates. In the rare cases where estimation is easy, for example with fixed-base manipulators and static objects, controlling a robot directly with MJPC would be a straightforward exercise. The difficult and interesting case involves free-moving bodies and contacts, as in locomotion and manipulation. For certain specific cases, like locomotion on flat, uniform terrain, reasonable estimates should not be difficult to obtain. For the general case, we believe that contact-aware estimation is possible following the approach of [Lowrey et al., 2014], but that remains to be seen. Relatedly, we believe that high-quality estimators also require the same kind of interactive, GUI-driven interface used by MJPC. ## References * Audet and Hare [2017] Charles Audet and Warren Hare. _Derivative-Free and Blackbox Optimization_ , volume 2. Springer, 2017. * Betts [1998] John T. Betts. Survey of numerical methods for trajectory optimization. _Journal of Guidance, Control, and Dynamics_ , 21(2):193–207, 1998. * Byravan et al. [2021] Arunkumar Byravan, Leonard Hasenclever, Piotr Trochim, Mehdi Mirza, Alessandro Davide Ialongo, Yuval Tassa, Jost Tobias Springenberg, Abbas Abdolmaleki, Nicolas Heess, Josh Merel, et al. Evaluating model-based planning and planner amortization for continuous control. _arXiv:2110.03363_ , 2021. * Chen et al. [2022] Tao Chen, Jie Xu, and Pulkit Agrawal. A system for general in-hand object re-orientation. In _Conference on Robot Learning_ , pages 297–307, 2022. * Falcone et al. [2007] Paolo Falcone, Francesco Borrelli, Jahan Asgari, Hongtei Eric Tseng, and Davor Hrovat. Predictive active steering control for autonomous vehicle systems. _IEEE Transactions on Control Systems Technology_ , 15(3):566–580, 2007. * Gill et al. [2005] Philip E. Gill, Walter Murray, and Michael A. Saunders. SNOPT: An SQP algorithm for large-scale constrained optimization. _SIAM Review_ , 47(1):99–131, 2005. * Hämäläinen et al. [2014] Perttu Hämäläinen, Sebastian Eriksson, Esa Tanskanen, Ville Kyrki, and Jaakko Lehtinen. Online motion synthesis using sequential Monte Carlo. _ACM Transactions on Graphics_ , 2014. * Hämäläinen et al. [2015] Perttu Hämäläinen, Joose Rajamäki, and C. Karen Liu. Online control of simulated humanoids using particle belief propagation. _ACM Transactions on Graphics_ , 2015. * Hansen and Ostermeier [2001] Nikolaus Hansen and Andreas Ostermeier. Completely derandomized self-adaptation in evolution strategies. _Evolutionary Computation_ , 9(2):159–195, 2001\. * Heess et al. [2015] Nicolas Heess, Gregory Wayne, David Silver, Timothy Lillicrap, Tom Erez, and Yuval Tassa. Learning continuous control policies by stochastic value gradients. _Advances in Neural Information Processing Systems_ , 28, 2015. * Holland [1992] John H. Holland. Genetic algorithms. _Scientific American_ , 267(1):66–73, 1992. * Jacobson [1973] David Jacobson. Optimal stochastic linear systems with exponential performance criteria and their relation to deterministic differential games. _IEEE Transactions on Automatic Control_ , 18(2):124–131, 1973. * Jacobson and Mayne [1970] David H. Jacobson and David Q. Mayne. _Differential Dynamic Programming_. Number 24. Elsevier Publishing Company, 1970. * Kalman [1960] Rudolf E. Kalman. On the general theory of control systems. In _International Conference on Automatic Control_ , pages 481–492, 1960. * Kalman [1964] Rudolf E. Kalman. When Is a Linear Control System Optimal? _Journal of Basic Engineering_ , 86(1):51–60, 1964. * Kuindersma et al. [2016] Scott Kuindersma, Robin Deits, Maurice Fallon, Andrés Valenzuela, Hongkai Dai, Frank Permenter, Twan Koolen, Pat Marion, and Russ Tedrake. Optimization-based locomotion planning, estimation, and control design for the Atlas humanoid robot. _Autonomous Robots_ , 40(3):429–455, 2016. * Ladicky et al. [2017] Lubor Ladicky, SoHyeon Jeong, Nemanja Bartolovic, Marc Pollefeys, and Markus Gross. PhysicsForests: Real-time fluid simulation using machine learning. In _ACM SIGGRAPH Real Time Live!_ , page 22, 2017. ISBN 9781450350907. * Li and Todorov [2004] Weiwei Li and Emanuel Todorov. Iterative linear quadratic regulator design for nonlinear biological movement systems. In _International Conference on Informatics in Control, Automation and Robotics_ , pages 222–229, 2004. * Lillicrap et al. [2015] Timothy P. Lillicrap, Jonathan J. Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, David Silver, and Daan Wierstra. Continuous control with deep reinforcement learning. _arXiv:1509.02971_ , 2015. * Lopez-Negrete et al. [2013] Rodrigo Lopez-Negrete, Fernando J. D’Amato, Lorenz T. Biegler, and Aditya Kumar. Fast nonlinear model predictive control: Formulation and industrial process applications. _Computers & Chemical Engineering_, 51:55–64, 2013. * Lowrey et al. [2014] Kendall Lowrey, Svetoslav Kolev, Yuval Tassa, Tom Erez, and Emanuel Todorov. Physically-consistent sensor fusion in contact-rich behaviors. In _IEEE/RSJ International Conference on Intelligent Robots and Systems_ , pages 1656–1662, 2014. * Lu and Dosher [2022] Zhong-Lin Lu and Barbara Anne Dosher. Current directions in visual perceptual learning. _Nature Reviews Psychology_ , 1(11):654–668, 2022\. * Mangasarian [1966] O. L. Mangasarian. Sufficient conditions for the optimal control of nonlinear systems. _SIAM Journal on Control_ , 4(1):139–152, 1966\. * Mania et al. [2018] Horia Mania, Aurelia Guy, and Benjamin Recht. Simple random search of static linear policies is competitive for reinforcement learning. In _Advances in Neural Information Processing Systems_ , pages 1800–1809, 2018. * Matyas [1965] J. Matyas. Random optimization. _Automation and Remote Control_ , 26(2):246–253, 1965. * MuJoCo Menagerie Contributors [2022] MuJoCo Menagerie Contributors. MuJoCo Menagerie: A collection of high-quality simulation models for MuJoCo, 2022. URL http://github.com/deepmind/mujoco_menagerie. * Na et al. [2003] Man Gyun Na, Sun Ho Shin, and Whee Cheol Kim. A model predictive controller for nuclear reactor power. _Nuclear Engineering and Technology_ , 35(5):399–411, 2003. * Nagabandi et al. [2020] Anusha Nagabandi, Kurt Konolige, Sergey Levine, and Vikash Kumar. Deep dynamics models for learning dexterous manipulation. In _Conference on Robot Learning_ , pages 1101–1112, 2020. * NASA [1971] NASA. _Apollo 11 Mission Report_ , volume 238. Scientific and Technical Information Office, National Aeronautics and Space Administration, 1971. * Rechenberg [1973] Ingo Rechenberg. Evolutionsstrategie. _Optimierung technischer Systeme nach Prinzipien derbiologischen Evolution_ , 1973. * Richalet et al. [1978] Jacques Richalet, André Rault, J. L. Testud, and J. Papon. Model predictive heuristic control: Applications to industrial processes. _Automatica_ , 14(5):413–428, 1978. * Rudin et al. [2022] Nikita Rudin, David Hoeller, Philipp Reist, and Marco Hutter. Learning to walk in minutes using massively parallel deep reinforcement learning. In _Conference on Robot Learning_ , pages 91–100, 2022. * Salimans et al. [2017] Tim Salimans, Jonathan Ho, Xi Chen, Szymon Sidor, and Ilya Sutskever. Evolution strategies as a scalable alternative to reinforcement learning. _arXiv:1703.03864_ , 2017. * Schrittwieser et al. [2020] Julian Schrittwieser, Ioannis Antonoglou, Thomas Hubert, Karen Simonyan, Laurent Sifre, Simon Schmitt, Arthur Guez, Edward Lockhart, Demis Hassabis, Thore Graepel, Timothy Lillicrap, and David Silver. Mastering Atari, Go, chess and shogi by planning with a learned model. _Nature_ , 588(7839):604–609, 2020. * Schulman et al. [2017] John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. Proximal policy optimization algorithms. _arXiv:1707.06347_ , 2017. * Silver et al. [2016] David Silver, Aja Huang, Chris J. Maddison, Arthur Guez, Laurent Sifre, George van den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Veda Panneershelvam, Marc Lanctot, Sander Dieleman, Dominik Grewe, John Nham, Nal Kalchbrenner, Ilya Sutskever, Timothy Lillicrap, Madeleine Leach, Koray Kavukcuoglu, Thore Graepel, and Demis Hassabis. Mastering the game of Go with deep neural networks and tree search. _Nature_ , 529(7587):484–489, 2016. * Smith et al. [2022] Laura Smith, Ilya Kostrikov, and Sergey Levine. A walk in the park: Learning to walk in 20 minutes with model-free reinforcement learning. _arXiv:2208.07860_ , 2022. * Smith and Yound [1967] R. E. Smith and J. W. Yound. Trajectory optimization for an Apollo-type vehicle under entry conditions encountered during lunar return. Technical report, 1967. * Springenberg et al. [2020] Jost Tobias Springenberg, Nicolas Heess, Daniel Mankowitz, Josh Merel, Arunkumar Byravan, Abbas Abdolmaleki, Jackie Kay, Jonas Degrave, Julian Schrittwieser, Yuval Tassa, et al. Local search for policy iteration in continuous control. _arXiv:2010.05545_ , 2020. * Tassa et al. [2012] Yuval Tassa, Tom Erez, and Emanuel Todorov. Synthesis and stabilization of complex behaviors through online trajectory optimization. In _IEEE/RSJ International Conference on Intelligent Robots and Systems_ , pages 4906–4913, 2012. * Tassa et al. [2014] Yuval Tassa, Nicolas Mansard, and Emo Todorov. Control-limited differential dynamic programming. In _IEEE International Conference on Robotics and Automation_ , pages 1168–1175, 2014. * Todorov et al. [2012] Emanuel Todorov, Tom Erez, and Yuval Tassa. MuJoCo: A physics engine for model-based control. In _IEEE/RSJ International Conference on Intelligent Robots and Systems_ , pages 5026–5033, 2012. * Tuffield and Elias [2003] Paul Tuffield and Hugo Elias. The Shadow robot mimics human actions. _Industrial Robot: An International Journal_ , 2003. * Tunyasuvunakool et al. [2020] Saran Tunyasuvunakool, Alistair Muldal, Yotam Doron, Siqi Liu, Steven Bohez, Josh Merel, Tom Erez, Timothy Lillicrap, Nicolas Heess, and Yuval Tassa. dm_control: Software and tasks for continuous control. _Software Impacts_ , 6:100022, 2020. * Unitree [2022] Unitree. Unitree A1 Quadruped. https://m.unitree.com/products/a1, 2022. * Von Stryk [1993] Oskar Von Stryk. Numerical solution of optimal control problems by direct collocation. In _Optimal Control_ , pages 129–143. Springer, 1993. * Von Stryk and Bulirsch [1992] Oskar Von Stryk and Roland Bulirsch. Direct and indirect methods for trajectory optimization. _Annals of Operations Research_ , 37(1):357–373, 1992. * Wächter and Biegler [2006] Andreas Wächter and Lorenz T. Biegler. On the implementation of an interior-point filter line-search algorithm for large-scale nonlinear programming. _Mathematical Programming_ , 106(1):25–57, 2006\. * Whittle [1981] Peter Whittle. Risk-sensitive linear/quadratic/Gaussian control. _Advances in Applied Probability_ , 13(4):764–777, 1981. * Williams et al. [2017] Grady Williams, Andrew Aldrich, and Evangelos A. Theodorou. Model predictive path integral control: From theory to parallel computation. _Journal of Guidance, Control, and Dynamics_ , 40(2):344–357, 2017. * Wu et al. [2022] Philipp Wu, Alejandro Escontrela, Danijar Hafner, Ken Goldberg, and Pieter Abbeel. DayDreamer: World models for physical robot learning. _arXiv:2206.14176_ , 2022. ## Appendix A Interpolation This section provides analytical expressions for computing interpolations. The indices $j$ and $j+1$ correspond to the indices of the domain variables which contain the query point $\tau$. These values are efficiently found using binary search in $\mathbf{O}\left(\text{log}(n)\right)$ where $n$ is the dimension of the domain variable set. #### Zero. The zero-order interpolation simply returns the parameter values at the lower bound index: $\theta_{j}\leftarrow s.$ (15) #### Linear. The linear interpolation returns: $(1-q)\cdot\theta_{j}+q\cdot\theta_{j+1}\leftarrow s,$ (16) where $\quad q=(\tau-\tau_{j})/(\tau_{j+1}-\tau_{j})$. #### Cubic. The cubic interpolation leverages finite-difference approximations of the slope at the interval points: $\displaystyle\phi_{j}$ $\displaystyle=\frac{1}{2}\left(\frac{\theta_{j+1}-\theta_{j}}{\tau_{j+1}-\tau_{j}}+\frac{\theta_{j}-\theta_{j-1}}{\tau_{j}-\tau_{j-1}}\right),$ (17) $\displaystyle\phi_{j+1}$ $\displaystyle=\frac{1}{2}\left(\frac{\theta_{j+2}-\theta_{j+1}}{\tau_{j+2}-\tau_{j+1}}+\frac{\theta_{j+1}-\theta_{j}}{\tau_{j+1}-\tau_{j}}\right),$ (18) and returns: $a\cdot\theta_{j}+b\cdot\phi_{j}+c\cdot\theta_{j+1}+d\cdot\phi_{j+1}\leftarrow s,$ (19) where, $\displaystyle a$ $\displaystyle=2q^{3}-3q^{2}+1,$ (20) $\displaystyle b$ $\displaystyle=(q^{3}-2q^{2}+q)\cdot(\tau_{j+1}-\tau_{j}),$ (21) $\displaystyle c$ $\displaystyle=-2q^{3}+3q^{2},$ (22) $\displaystyle d$ $\displaystyle=(q^{3}-q^{2})\cdot(\tau_{j+1}-\tau_{j}).$ (23) ## Appendix B Tasks This section provides additional information about the examples provided in Section 4, including objective formulations and planner settings. #### Humanoid objective. This task comprises $M=6$ cost terms: * • Term $0$: * $\text{r}_{0}$: lateral center-of-mass position and average lateral feet position alignment * $\text{n}_{0}$: hyperbolic cosine * $w_{0}$: 100 * • Term $1$: * $\text{r}_{1}$: lateral torso position and lateral center-of-mass position alignment * $\text{n}_{1}$: smooth absolute value * $w_{1}$: 1 * • Term $2$: * – $\text{r}_{2}$: head height and feet height difference minus target height difference * – $\text{n}_{2}$: smooth absolute value * – $w_{2}$: 100 * • Term $3$: * $\text{r}_{3}$: lateral center-of-mass velocity * $\text{n}_{3}$: smooth absolute value * $w_{3}$: 10 * • Term $4$: * $\text{r}_{4}$: joint velocity * $\text{n}_{4}$: quadratic * $w_{4}$: 0.01 * • Term $5$: * $\text{r}_{5}$: control effort * $\text{n}_{5}$: quadratic * $w_{5}$: 0.025 #### Quadruped objective. This task comprises $M=4$ cost terms: * • Term $0$: * $\text{r}_{0}$: body height and average feet height difference minus target height * $\text{n}_{0}$: quadratic * $w_{0}$: 1 * • Term $1$: * $\text{r}_{1}$: body position minus goal position * $\text{n}_{1}$: quadratic * $w_{1}$: 5.0 * • Term $2$: * $\text{r}_{2}$: body orientation minus goal orientation * $\text{n}_{2}$: quadratic * $w_{2}$: 1.0 * • Term $3$: * $\text{r}_{3}$: control effort * $\text{n}_{3}$: quadratic * $w_{3}$: 0.25 #### Hand objective. This task comprises $M=3$ cost terms: * • Term $0$: * $\text{r}_{0}$: cube position minus hand palm position * $\text{n}_{0}$: quadratic * $w_{0}$: 20 * • Term $1$: * $\text{r}_{1}$: cube orientation minus goal orientation * $\text{n}_{1}$: quadratic * $w_{1}$: 3 * • Term $2$: * $\text{r}_{2}$: cube linear velocity * $\text{n}_{2}$: quadratic * $w_{2}$: 10 #### Planner settings. We provide the settings used for Predictive Sampling in Table 1. Table 1: Predictive Sampling settings Task | $P$ | $N$ | $T$ | $\sigma$ ---|---|---|---|--- Humanoid | 3 | 10 | 23 | 0.125 Quadruped | 3 | 10 | 35 | 0.25 Hand | 6 | 10 | 25 | 0.1 ## Appendix C Predictive Sampling Algorithm Algorithm 5 PredictiveSampling 1:procedure OptimizePolicy 2: task: $f$, $c$, $R$ 3: settings: $T$, $N$, $\sigma$, $s$ 4: initialise: $(\tau,\theta)$ 5: for $k=0,\dots,\infty$ 6: $(x,\tau)\leftarrow$ get state 7: $(\bar{\tau},\theta)\leftarrow\text{resample}$ 8: for $i=0,\dots,N$ (parallel) 9: $\tilde{\theta}^{(i)}=\theta+\begin{cases}0,\phantom{\mathcal{N}(0,\sigma^{2}\cdot I)}i=0,\\\ \mathcal{N}(0,\sigma^{2}\cdot I),\phantom{0}\text{else}\end{cases}$ 10: $x^{(i)}_{0}=x,\,\tau_{0}^{(i)}=\tau$, 11: for $t=0,\dots,T$ 12: $u^{(i)}_{t}=s(\tau^{(i)}_{t};(\bar{\tau},\tilde{\theta}^{(i)}))$ 13: $c^{(i)}_{t}=c(x^{(i)}_{t},u^{(i)}_{t};R)$ 14: $(x^{(i)}_{t+1},\tau^{(i)}_{t+1})=f(x^{(i)}_{t},u^{(i)}_{t})$ 15: end for 16: end for 17: $\theta\leftarrow\text{argmin}\left(J(\tilde{\theta}^{(i)})\right)$ 18: end for 19:end procedure 20:procedure ActionFromPolicy 21: $(x,\tau)\leftarrow$ get state 22: $u=s(\tau;(\bar{\tau},\theta))$ 23: return $u$ 24:end procedure ## Appendix D Compute Resources Experiments were performed on a Lenovo ThinkStation P920 with 48GB of memory and an Intel Xeon Gold 6154 72-core CPU. Additional experiments were performed on an Apple MacBook Pro (2021) with 16 GB of memory and an M1 Pro CPU.
# Towards Explainability in Modular Autonomous Vehicle Software Hongrui Zheng∗, Zirui Zang∗, Shuo Yang∗, Rahul Mangharam ∗Authors contributed equally. All authors are with University of Pennsylvania, Department of Electrical and Systems Engineering, 19104, Philadelphia, PA, USA. Emails: {hongruiz, zzang, yangs1<EMAIL_ADDRESS> ###### Abstract Safety-critical Autonomous Systems require trustworthy and transparent decision-making process to be deployable in the real world. The advancement of Machine Learning introduces high performance but largely through black-box algorithms. We focus the discussion of explainability specifically with Autonomous Vehicles (AVs). As a safety-critical system, AVs provide the unique opportunity to utilize cutting-edge Machine Learning techniques while requiring transparency in decision making. Interpretability in every action the AV takes becomes crucial in post-hoc analysis where blame assignment might be necessary. In this paper, we provide positioning on how researchers could consider incorporating explainability and interpretability into design and optimization of separate Autonomous Vehicle modules including Perception, Planning, and Control. ## I Introduction According to the Morning Consult and Politico poll [1], only 16% of respondents are “very likely” to ride as a passenger in an autonomous vehicle, while 28% of respondents state that they “not likely at all”. Moreover, only 22% of respondents believe self-driving cars are safer than the average human driver, while 35% of them believing self-driving cars are less safe than the average human driver. The public’s distrust in Autonomous Vehicles (AV) shows that improving explainability in AV software is a necessity. There exist many surveys on explainable AI (XAI) and robotics [2, 3, 4, 5]. Specifically, [6, 7, 8] surveys explainability in Autonomous Driving. Atakishiyev et al. [7] believes that AVs need to provide regulatory compliant operational safety and explainability in real-time decisions. It focuses on providing discussion through the cause-effect-solution perspective. Zablocki et al. [6] provides an in-depth overview of XAI methods in deep vision-based methods, but is limited to the scope of perception only. Omeiza et al. [8] also provides an overview of explanations in AVs in the full self-driving pipeline. Gilpin et al. [9] proposes explainability as the trade-off between interpretability and completeness. As described in [9], to be interpretable is to describe the internals of a system in such a way that is understandable to humans; to be complete is to describe the operation of a system in an accurate way. We position ourselves to provide insight in augmenting explanability in Autonomous Vehicle’s sense-plan-act software modules as a task of balancing interpretability and completeness. In this paper, we look at the explainability in existing works and in our recent contributions in localization, planning, and control. In each case, we want to be able to quantify the uncertainty at each step of the decision making and interpret the provenance of the outcome of the algorithm. ## II Explainability in Localization Robot localization is a problem of finding a robot’s pose using a map and sensor measurements, such as LiDAR scans. The map is pre-built and the environment is assumed to not change significantly after the map is captured. It is crucial for any moving robot to interact with the physical world correctly. However, the problem of finding the mappings between measurements and poses can be ambiguous, because sensor measurements from multiple distant poses can be similar. Therefore, to tightly integrate the localization module with other parts of the software stack and for the engineers implementing and tuning the algorithm, the explainability of localization algorithms using neural networks becomes important. We need to estimate the uncertainty of the localization results, and in the worst case, to know when and why the robot fails to localize on a certain map. Monte Carlo Localization (MCL)[10], the widely adopted method, uses random hypothesis sampling and sensor measurement updates to infer the pose. In MCL, the proposed particles are explicit poses on the map and we can interpret the distribution of the particles as the uncertainties. The random generation of the particles can be tuned with parameters that have physical meaning, providing an interface for humans to adjust the behavior of the algorithm. Many developments in localization seek to improve within the framework of MCL.[11, 12, 13] Although particle filter has been a popular localization framework for its robustness and reliable performance, it introduces random jitter into the localization results. Other common approaches are to use Bayesian filtering[14] or to find more distinguishable global descriptors on the map[15, 16]. In Bayesian filtering, the explainability lies in the conditional probability attached with the motion model and each measurement. The estimation of such probability is challenging. For the global descriptor approach, oftentimes manual selection of map features are needed, which increases the explainability of the system, but also increases the human workload and reduces robustness. Developments in localization research usually propose better measurement models or feature extractors within these frameworks. [17, 18]. Recent research in localization has also focused on the use of learning-based methods outside of the above frameworks [19]. Although learning-based methods may provide better localization precision with lower latency, the interpretability of the method decreases. While the traditional localization methods can be manually tuned according to the specific user scenarios, learning-based localization methods are usually not tunable once the network is trained. Uncertainty estimations of the neural networks also become a challenge for learning-based methods. There are efforts to approximate the uncertainty[20, 21, 22], but it hasn’t been widely applied. Our contribution: In our recent paper, Local_INN, we proposed a new approach to frame the localization problem as an ambiguous inverse problem and solve it with an invertible neural network (INN) [23]. It stores the map data implicitly inside the neural network. With the assumption that the environment doesn’t not change significantly from the map, by evaluating the reverse path of the neural network, we can get robot poses from LiDAR scans. It also provides uncertainty estimation from the neural network and is capable of learning and providing localization for complex environments. Localization is an inverse problem of finding a robot’s pose using a map and sensor measurements. This reverse process of inferring the pose from sensor measurements is ambiguous. Invertible neural networks such as normalizing flows[24] have been used to solve ambiguous inverse problems in various fields[25, 26, 27, 28]. The version of normalizing flows we used is called RealNVP[29], which uses a mathematical structure called coupling layers to ensure the invertibility while performing transformations with arbitrary neural network layers, such as MLPs. This framework of solving inverse problems with normalizing flows was introduced by Ardizonne et al.[25] and was later extended by [26, 30] to include a conditional input that is concatenated to the vectors inside the coupling layers. They proposed to use normalizing flows to learn a bijective mapping between two distributions and use a normal- distributed latent variable to encode the lost information in training due to the ambiguity of the problem. The network can be evaluated in both forward and reverse paths. During the evaluation, repeatedly sampling the latent variable can give the full posterior distribution given the input. In Local_INN, we use pose-scan data pairs to train such a bijective mapping. As shown in Fig. 1, The forward path is from pose to scan and the reverse path is from scan to pose. We use a conditional input calculated from the previous pose of the robot to reduce the ambiguity of the problem. Because INNs require the same input and output dimensions, we use a Variational Autoencoder[31] to reduce the dimension of the LiDAR scans and use Positional Encoding[32] to augment that of the poses. The network is trained with supervised loss functions on both sides. The data used for training the Local_INN can be simulated or real data recorded from LiDARs. In our experiments, we tested on both real and simulated data with 2D and 3D LiDARs. To collect training data, we uniformly sample $x,y$ position and heading $\theta$ on the drivable surface of each map, and use a LiDAR simulator to find the corresponding LiDAR ranges. This means the trained network will be able to localize everywhere on the map. For each different map, we need to train a separate network. Map files are compressed inside the neural network and are no longer needed during evaluation. TABLE I: Local_INN Experiments: Map Reconstruction and RMS Localization Errors with 2D LiDAR ($xy$[m], $\theta$[∘]) | Race Track (Simulation) | Outdoor (Real) ---|---|--- | Original Map --- Reconstruction Test Trajectory | Online PF | $0.168,2.107$ | $0.047,1.371$ Local_INN+EKF | $\mathbf{0.056},\mathbf{0.284}$ | $\mathbf{0.046},\mathbf{1.130}$ Figure 1: Top: Evaluation of Local_INN in forward direction gives compressed map information, and in the reverse direction gives accurate localization with fast runtime and uncertainty estimation. Bottom: Network structure of Local_INN. Solid arrows are from pose to lidar scan. Dashed arrows are from lidar scan to pose. Conditional input is calculated from the previous pose of the robot. We claim that INN is naturally suitable for the localization problem with improved explainability compared to other learning-based methods. In particular, uncertainty estimation and map representation are the two advantages that Local_INN provides in the context of explainability. Figure 2: Example of global localization finding the correct pose at the 2nd iterations (green arrow). Top: Narrowing down of the candidate poses in the first 3 iterations. We can see candidate poses on the map (orange dots), correct pose (red dot), selected pose (green dot). Bottom: Examples of LiDAR scan ranges at candidate poses at iteration 3 (orange boxes), and at the selected pose (green box). In the range plots, the horizontal axis is the angle of LiDAR scans and vertical axis is the measured distance. We can see a comparison of network expected ranges (blue curve) at various poses, and the actual LiDAR measurement (orange curve). The correct pose is selected where the measurement best matches the expected shape. The network is trained on simulated data and tested on real LiDAR data. ### II-A Explainability from Uncertainty Estimation When we use Local_INN to localize, the input to the reverse path of the INN consists of the LiDAR scans concatenated with a latent vector that is sampled from normal distribution. With this sampling of latent vector, the network can output not just a pose but a distribution of inferred poses. The covariance of this distribution can be used as the confidence of the neural network when fusing with other sensors. Uncertainty estimation improves explainability by providing information on the measurement quality of the prediction. Compared to learning methods that do not provide uncertainty estimates, it is much easier to determine whether the prediction of the neural network is lower in accuracy due to higher uncertainty, and improve the prediction results by augmentation. In our experiments, we used an EKF to fuse the localization result with the odometry information. The results show that this fusion significantly improved localization accuracy where the map geometry is ambiguous, which means this covariance is very effective in revealing the confidence of the network. As shown in Table I, The accuracy of Local_INN is at par with the current localization efforts. See [23] for a comparative analysis of Local_INN localization accuracy in 2D and 3D maps. ### II-B Explainability from Map Representation Local_INN provides an implicit map representation and a localization method within one neural network. The guaranteed invertibility of the neural network provides the use a direct way to check the neural network’s ’understanding’ of the map by reproducing part of the map with poses. That is, we can compare the reconstructed map to the original map to see how much detail is used by the neural network in localization. Again, this feature improves explainability in failure scenarios. When the localization fails, this comparison can help us explain the failure and guide us in improving the methods. For example, we can train with more data from a particular location on the map that was difficult to localize in. As an example of how the stored map information in the forward path of the neural network can help us explain the localization results, let us consider the algorithm for global localization. Global localization is needed when a robot starts with an unknown pose or when the robot encounters the ’kidnapping problem’. In this case, it is challenging to find the correct position on the map due to the ambiguity of problem. MCL algorithms usually do global localization by spreading the covariance all around the map and using iterations of control inputs and measurements to decrease the covariance, which gives an explicit visualization to see the progress of the global localization processes. For other learning-based method, this process is usually hard to explain as we rely on the neural network to output poses as a black box. With Local_INN, we can randomly initialize a set of random poses on the map as conditional inputs and use new lidar scans to narrow down the assumptions. In other words, we initially have multiple random assumptions of the robot’s location on the map and use them as the conditional inputs for the nerual network. As shown in figure 2 iteration 1, when we input a LiDAR scan to the network along with these assumptions, it will output multiple possible pose distributions. In our algorithm, for each possible pose distribution, we compare the sensor measurement with what the neural network expects at this location, and use the reciprocal of the error term to weight the assumptions differently. The weights for the assumptions is used to determine the amount of latent variable we use. This process repeats with each new LiDAR scan we get from the sensor. In our experiments, the convergence of candidate poses is fast and accurate. As shown in iteration 2 and 3 in figure 2, even if we still track multiple poses, in this example, the correct pose is determined at the 2nd iteration in a highly symmetrical map. The low part of figure 2 shows plots of the expected lidar scan from the neural network and the current LiDAR measurement. This reveals the black-box process of the neural network so that we can see why multiple poses are possible and how we should decide which one to pick. To summarize, the explainability of localization methods generally lies in the uncertainty estimation and the ability to explain and tune the methods when localization fails. Traditional localization methods usually offer higher interpretability than learning-based methods, whereas the learning-based methods can provide better empirical performance. The new method, Local_INN, we recently proposed uses an invertible network network architecture to solve the localization problem. It offers interpretability by giving uncertainty estimation through the covariance of the inferred pose distributions, and by ensuring the invertibility of the network so that we can reveal the what information the neural network is using during the localization. At the same time, it does sacrifice completeness by using a Variational Autoencoder (VAE) to model a latent space of the LiDAR scans. ## III Explainability in Planning Planning is the task of finding a viable motion plan for the robot to reach some predetermined goal given the current observation of the environment through various sensors. The planning step is usually the next step after perception, or localization. Traditionally, sampling-based and model- predictive methods are the most popular choices in Autonomous Vehicle motion planning. Planning algorithms provide explanability through human-designed objectives: e.g. maximizing the distance of a planned trajectory to obstacles and undrivable areas, maximizing velocity on the trajectories, minimizing the lateral acceleration the vehicle experiences on a planned trajectory. We propose a unique position in adding interpretability and completeness to planning algorithms: explainability through abstraction. Next, we show our approach based on our recent paper on Game-Theoretic objective space planning [33]. Figure 3: Overview of Game-theoretic Objective Space Planning In this case study, our primary context is two-player racing games in close proximity. An overall depiction of the pipeline is shown in Figure 3. We choose a game theoretic approach that models racing as a zero-sum extensive game with imperfect information and perfect recall. Extensive games model sequential decision-making of players and naturally form a game tree where each node is a decision point for a player. However, the planning problem presented in autonomous racing is continuous, and the state space of the agents, in turn, the game tree in the extensive game, will also be infinitely large if we model the game in the vehicle’s planning space. Since the decision made by a game-theoretic algorithm in the planning space cannot be explained in a way that a human can understand, we use a lower dimensional space for planning. We define the notion of Objective Space $\mathcal{O}$. For each short rollout in an adversarial environment, we can compute multiple metrics regarding this agent’s performance, such as safety and aggressiveness. These metrics also add to the interpretability of our planning algorithm while not losing the completeness. $\mathcal{O}$ models the average outcome of each agent against competent opponents. Using $\mathcal{O}$, our planner maps complex agent behaviors to a lower dimension where only the episodic outcome is recorded instead of the entire decision-making process in the planning space. We define an action in our game as movements in a dimension of $\mathcal{O}$. This action space is analogous to the planning space in a grid world with actions that move the agent to a neighboring cell. This means the planning problem is much simpler than the original problem. In our case study, we choose aggressiveness and restraint as the two dimensions of $\mathcal{O}$. Aggressiveness is scored on an agent’s lead over the other at the end of the rollout, and restraint is scored on an agent’s time to collision to the environment and the other agent. Two movements are available for each dimension: increasing or decreasing for a fixed distance along the axis. For example, four discrete actions are available at each turn when $\mathcal{O}\in\mathbb{R}^{2}$. Even with the formulation of agent action space, the possible objective values of an opponent agent or possible position in $\mathcal{O}$ is still infinite. We propose a predictive model for regret values within Counterfactual Regret Minimization (CFR) [34] to make the problem tractable. Finally with head-to-head racing experiments, we demonstrate that using the proposed planning pipeline above significantly improves the win rate that generalizes to unseen opponents in an unseen environment. ### III-A Explainability in Agent Actions Figure 4: Trajectories of ego moves in $\mathcal{O}$. Figure 5: Effect of making a move in $\mathcal{O}$ in motion planning space. In this section, we examine specific cases of agents moving in the lower dimension and investigate whether we achieve the task of instilling explainability in our algorithm. We choose a 2-D space that encodes aggressiveness and restraint. In Figure 4, we show four examples of races between two agents. The Ego (orange) uses our proposed game-theoretic approach in the Objective Space, and the Opponent (green) is static in the objective space. In the first two cases, the opponent is in the lower right quadrant, meaning that they’re more conservative than aggressive. Hence our planner chooses to increase in aggressiveness continuously to win the races. In the last two cases, the opponent is in the upper left quadrant, meaning that they’re more aggressive than conservative. Hence our planner chooses to first become more conservative, and once a chance presents itself, increase in aggressiveness and win the races. In Figure 5, we inspect a specific move to show the effect in the motion planning space. In the beginning of the rollout, both agents are side by side. At a decision point, our planner locates the opponent in the lower right quadrant as more conservative than aggressive. Then the ego decides to increase in aggressiveness to finally overtake the opponent. From these examples, it is clear that moving the planning problem in a lower dimension that encodes interpretable metrics for humans doesn’t decrease the capability of the planner, or the completeness in the algorithm. ## IV Explainability in Control With robust localization, the autonomous vehicle can plan its trajectory but requires a safe an interpretable controller to execute the plan. In this section, we show our recent progress on learning-based safety-critical control through control barrier functions (CBFs) and provide our positioning on explainable safe control. Specifically, we show that our proposed differentiable safety filter has more completeness than non-differentiable safety filter without sacrificing interpretability. ### IV-A Safety-critical control Learning-based control could provide high empirical performance thus it has become popular for controlling complex dynamical systems. However, learning- based controllers, such as neural network (NN), generally lack formal safety guarantees because of their black-box nature. This limits their deployments with complex safety-critical systems. To address safety risk, multiple methods have been proposed, such as model predictive control (MPC) [35], Hamilton- Jacobi reachability analysis [36], contraction theory [37], and control barrier functions (CBF) [38]. Among the many safety-critical control techniques, CBF is becoming a popular choice since it explicitly specifies a safe control set by using a Lyapunov- like condition and guards the system inside a safe invariant set. When a continuous-time control-affine system is considered, such projection reduces to a convex quadratic program (QP) which is referred to as CBF-QP. Due to its simplicity, flexibility, and formal safety guarantees, CBFs have been applied in safe learning control with many successful applications [39, 40, 41, 42, 43]. Compared with MPC, which needs to handle a possibly nonconvex optimization problem in the face of nonlinear dynamical systems, CBF-QP is computationally efficient to solve online. However, unlike MPC, the QP-based safety filter only operates in a minimally invasive manner, i.e., it generates the safe control input closest to the reference control input (in the Euclidean norm), as shown in Fig. 6, unaware of the long-term effects of its action. This indicates that the effects of the safety filter on the performance of the closed-loop system are hard to predict. Therefore, the application of the safety filter may give rise to myopic controllers [44] that induce subpar performance in the long term. Figure 6: Illustration of QP: If the NN output is not in the safe control set $Q$, it will be projected in minimally invasive way to $Q$; otherwise, the control keeps the same. Figure 7: Illustration of the gauge map from the $\ell_{\infty}$ ball $B_{\infty}$ to a polytopic set $Q$. The original point in $B_{\infty}$ and mapped point in $Q$ are in the same level set and with the same direction. To address the issue of myopic CBF-based safety filters, in our recent work [45], we propose to utilize CBF to construct safe-by-construction NN controllers that allow end-to-end learning. Incorporating safety layers in the NN controller allows the learning agent to take the effects of safety filters into account during training in order to maximize long-term performance. Figure 8: NN controller with CBF-QP safety filter. (a) Gauge map-based safe NN controller architecture. (b) CBF-QP-based safe NN controller architecture. Figure 9: Safe-by-construction NN controllers that utilize CBFs to construct differentiable safety layers (yellow blocks). We design a differentiable safety layer using the gauge map [46] (as shown in Fig. 7) which establishes a bijection mapping between the polytopic set of a NN output (e.g., an $\ell_{\infty}$ norm ball) and the CBF-based safe control set. The proposed architecture is denoted as NN-gauge (Fig. 9(a)). We compare NN-gauge with an alternative differentiable safe-by-construction NN controller called NN-diff-QP, which consists of a NN followed by a differentiable CBF-QP layer (Fig. 9(b)). Specifically, NN-diff-QP (Fig. 9(b)) concatenates a differentiable projection layer which can be implemented in a NN using toolboxes such as cvxpylayers [47], qpth [48]. NN-gauge involves finding an interior point of the safe control set since gauge map requires that $Q$ set in Fig. 7 is convex and cover the origin. It can be reduced to implicitly finding the Chebyshev center [49] of the safe control set. ###### Remark 1 In the online execution, NN-gauge requires closed-form evaluation or solving a linear program (LP) while NN-diff-QP solves a quadratic program. Both methods are significantly cheaper to run than MPC. As an example, let’s consider the adaptive cruise control (ACC) in which the ego car is expected to achieve the desired cruising speed while maintaining a safe distance from the leading car. As shown in Fig. 10(b), applying the CBF- QP safety filter directly to enforce safe control deteriorates the long-term closed-loop performance of the NN controller. However, both NN-gauge and NN- diff-QP achieve similar closed-loop performance, and they are comparable to MPC. (a) CBF values. (b) Velocity of the ego car. Figure 10: Results of adaptive cruise control. CBF values (left) and velocity of the ego car (right) under different controllers are evaluated in closed- loop for $20$s. A CBF value below zero indicates unsafety, and the optimal behavior of the ego car is expected to have a steady state velocity of $16m/s$, same as the leading car. ### IV-B Explainability of safety-critical control #### IV-B1 Non-differentiable QP filter Quadratic program (QP) is the classic minimally invasive safety filter to modify the nominal unsafe controller. Due to its simplicity and the nature of minimal modification of the original controller, it has always been equipped to CBF as the safety filter. As shown in Fig. 6, it is very understandable to humans, since one only needs to project the unsafe control input to the “nearest” point in the safe control set, which is specified by CBF. Thus, QP safety filter enjoys high interpretability. However, on the other hand, the QP operation is not necessarily the best choice from the perspective of optimizing a long-term system objective. For example, in the long run, sometimes it might be better to choose a “farther” point in the safe control set than the “nearest” point on the boundary when the nominal control is not safe. These modified actions are unaware of the long-term effects and may give rise to myopic controllers which deteriorate their performance. From the ACC case study, we can also observe in Fig. 10(b) that the modified control (i.e., NN-QP) is myopic with unpredictable long-term effects. Thus, vanilla QP filter lacks the completeness in the sense of controlling the system in the optimal way with the safety guarantee. #### IV-B2 Differentiable safety filters Our proposed two differentiable safety filters, i.e., differentiable QP and gauge map, enjoy both a high level of interpretability and completeness. * • Differentiable QP: the QP structure is the same with vanilla CBF-QP, so there is no interpretability loss in differentiable QP. On the other hand, since the NN controller considers the gradient of QP structure while training to minimize the long-term loss, so the trained NN adapts to QP much better than non-differentiable QP. Thus, even if the QP is applied to ensure safety when evaluating online, the modified controller could still have good performance. Therefore, from the explainability perspective, completeness is improved with no loss on interpretability. * • Gauge map: unlike QP, gauge map seeks to map the original control “proportianally” to the safe control set with the same direction. This mapping is also understandable to humans. Furthermore, since gauge map is sub- differentiable itself, so the end-to-end training holds naturally, which allows it to perform as good as differentiable QP after training. Thus, explanation could find the high completeness in this method as well. In the ACC example, it also has been shown that both NN-gauge and NN-diff-QP have comparable performance with MPC, which implies the high completeness from demonstration. Looking ahead, we are interested in exploring a more general parameterized gauge map method, which even not necessarily maps to the same level set. In that case, it will perhaps lose some interpretability but have better completeness as it is not limited to the same level set mapping. This way, there is more flexibility while choosing the safe control input. ## V Discussion and Conclusion In any autonomous vehicle which operates a sense-plan-act processing loop, the essential components are localization, planning and control. For humans to trust the system to make decisions and actions on their behalf in safety- critical situations, it is essential to have explainability across the stack. We view explainability as a trade-off between _interpretability_ of the machine’s actions and the _completeness_ in terms of describing the operation of the system in an accurate way. The explainability of localization methods generally lies in the uncertainty estimation and the ability to explain and tune the methods when localization fails. We introduce Local_INN, which utilizes an invertable neural network architecture that provides explainability for uncertainty estimation and explainability from map representation. Uncertainty estimation, through the covariance of the inferred pose distributions, improves explainability by providing information on the measurement quality of the prediction. Furthermore, the guaranteed invertibility of the neural network provides the use a direct way to check the neural network’s understanding of the map by analyzing the reconstructed map. This feedback from the internals of the localization engine allows the human to know where in the map the network has lower confidence and that they should augment it with more data. The expainability in planning is necessary to ensure the autonomous vehicle maintains a safe trajectory at all times in terms of maximizing the distance of a planned trajectory to obstacles and undrivable areas, maximizing velocity on the trajectories, minimizing the lateral acceleration the vehicle experiences on a planned trajectory, and so on. To achieve this, we introduced a new approach on Game-Theoretic Objective Space Planning where we map these complex planning objectives to a lower dimensional space of balancing aggressiveness and restraint. In a racing context, we show how our planner increases aggressiveness continuously to win the races. Similarly, it chooses to be more conservative to maintain safety. By moving the planning problem in a lower dimension that encodes interpretable metrics for humans we demonstrate how it doesn’t decrease the capability of the planner, or reduce the completeness in the algorithm. Finally, to ensure the plan is executed in a safe manner, we describe our efforts in explainable safe neural network controller design with Differentiable CBF-QP filters. The Control Barrier Function structure of the filter ensures the control output is always safe in an interpretable manner as there is a direct projection of the NN controller output into the safe control set. The performance of the proposed safe controllers is comparable to model- predictive controllers, which implies the high completeness from demonstrations. Through these three learning-based architectures for localization, planning and safe control, we have demonstrated the initial findings on explainability of decisions in the sense-plan-act process of autonomous systems. Further developments in explainability of autonomous systems will lead to more public trust and safer systems. ## References * [1] [Online]. Available: https://morningconsult.com/wp-content/uploads/2017/09/National-Tracking-Poll-170904.pdf * [2] F. K. Došilović, M. Brčić, and N. Hlupić, “Explainable artificial intelligence: A survey,” in _2018 41st International convention on information and communication technology, electronics and microelectronics (MIPRO)_. IEEE, 2018, pp. 0210–0215. * [3] S. R. Islam, W. Eberle, S. K. Ghafoor, and M. Ahmed, “Explainable artificial intelligence approaches: A survey,” _arXiv preprint arXiv:2101.09429_ , 2021\. * [4] E. Tjoa and C. Guan, “A survey on explainable artificial intelligence (xai): Toward medical xai,” _IEEE transactions on neural networks and learning systems_ , vol. 32, no. 11, pp. 4793–4813, 2020. * [5] T. Sakai and T. Nagai, “Explainable autonomous robots: a survey and perspective,” _Advanced Robotics_ , vol. 36, no. 5-6, pp. 219–238, 2022\. * [6] É. Zablocki, H. Ben-Younes, P. Pérez, and M. Cord, “Explainability of deep vision-based autonomous driving systems: Review and challenges,” _International Journal of Computer Vision_ , pp. 1–28, 2022. * [7] S. Atakishiyev, M. Salameh, H. Yao, and R. Goebel, “Explainable artificial intelligence for autonomous driving: A comprehensive overview and field guide for future research directions,” _arXiv preprint arXiv:2112.11561_ , 2021\. * [8] D. Omeiza, H. Webb, M. Jirotka, and L. Kunze, “Explanations in autonomous driving: A survey,” _IEEE Transactions on Intelligent Transportation Systems_ , 2021. * [9] L. H. Gilpin, D. Bau, B. Z. Yuan, A. Bajwa, M. Specter, and L. Kagal, “Explaining explanations: An overview of interpretability of machine learning,” in _2018 IEEE 5th International Conference on data science and advanced analytics (DSAA)_. IEEE, 2018, pp. 80–89. * [10] F. Dellaert, D. Fox, W. Burgard, and S. Thrun, “Monte carlo localization for mobile robots,” in _Proceedings 1999 IEEE international conference on robotics and automation (Cat. No. 99CH36288C)_ , vol. 2. IEEE, 1999, pp. 1322–1328. * [11] C. Zhang, M. H. Ang, and D. Rus, “Robust lidar localization for autonomous driving in rain,” in _2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)_. IEEE, 2018, pp. 3409–3415. * [12] X. Chen, I. Vizzo, T. Läbe, J. Behley, and C. Stachniss, “Range Image-based LiDAR Localization for Autonomous Vehicles,” in _Proc. of the IEEE Intl. Conf. on Robotics & Automation (ICRA)_, 2021. * [13] X. Chen, T. Läbe, L. Nardi, J. Behley, and C. Stachniss, “Learning an overlap-based observation model for 3d lidar localization,” in _2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)_. IEEE, 2020, pp. 4602–4608. * [14] V. Fox, J. Hightower, L. Liao, D. Schulz, and G. Borriello, “Bayesian filtering for location estimation,” _IEEE pervasive computing_ , vol. 2, no. 3, pp. 24–33, 2003. * [15] R. Dube, A. Cramariuc, D. Dugas, H. Sommer, M. Dymczyk, J. Nieto, R. Siegwart, and C. Cadena, “Segmap: Segment-based mapping and localization using data-driven descriptors,” _The International Journal of Robotics Research_ , vol. 39, no. 2-3, pp. 339–355, 2020. * [16] P.-E. Sarlin, C. Cadena, R. Siegwart, and M. Dymczyk, “From coarse to fine: Robust hierarchical localization at large scale,” in _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 2019, pp. 12 716–12 725. * [17] M. A. Uy and G. H. Lee, “Pointnetvlad: Deep point cloud based retrieval for large-scale place recognition,” in _Proceedings of the IEEE conference on computer vision and pattern recognition_ , 2018, pp. 4470–4479. * [18] L. Sun, D. Adolfsson, M. Magnusson, H. Andreasson, I. Posner, and T. Duckett, “Localising faster: Efficient and precise lidar-based robot localisation in large-scale environments,” in _2020 IEEE International Conference on Robotics and Automation (ICRA)_. IEEE, 2020, pp. 4386–4392. * [19] W. Lu, Y. Zhou, G. Wan, S. Hou, and S. Song, “L3-net: Towards learning based lidar localization for autonomous driving,” in _Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition_ , 2019, pp. 6389–6398. * [20] M. Cai, C. Shen, and I. Reid, “A hybrid probabilistic model for camera relocalization,” 2019. * [21] A. Kendall and R. Cipolla, “Modelling uncertainty in deep learning for camera relocalization,” in _2016 IEEE international conference on Robotics and Automation (ICRA)_. IEEE, 2016, pp. 4762–4769. * [22] H. Deng, M. Bui, N. Navab, L. Guibas, S. Ilic, and T. Birdal, “Deep bingham networks: Dealing with uncertainty and ambiguity in pose estimation,” _International Journal of Computer Vision_ , pp. 1–28, 2022. * [23] Z. Zang, H. Zheng, J. Betz, and R. Mangharam, “Local_inn: Implicit map representation and localization with invertible neural networks,” _arXiv preprint arXiv:2209.11925_ , 2022. * [24] G. Papamakarios, E. T. Nalisnick, D. J. Rezende, S. Mohamed, and B. Lakshminarayanan, “Normalizing flows for probabilistic modeling and inference.” _J. Mach. Learn. Res._ , vol. 22, no. 57, pp. 1–64, 2021. * [25] L. Ardizzone, J. Kruse, S. Wirkert, D. Rahner, E. W. Pellegrini, R. S. Klessen, L. Maier-Hein, C. Rother, and U. Köthe, “Analyzing inverse problems with invertible neural networks,” _arXiv preprint arXiv:1808.04730_ , 2018. * [26] L. Ardizzone, C. Lüth, J. Kruse, C. Rother, and U. Köthe, “Guided image generation with conditional invertible neural networks,” _arXiv preprint arXiv:1907.02392_ , 2019. * [27] T. J. Adler, L. Ardizzone, A. Vemuri, L. Ayala, J. Gröhl, T. Kirchner, S. Wirkert, J. Kruse, C. Rother, U. Köthe _et al._ , “Uncertainty-aware performance assessment of optical imaging modalities with invertible neural networks,” _International journal of computer assisted radiology and surgery_ , vol. 14, no. 6, pp. 997–1007, 2019. * [28] T. Wehrbein, M. Rudolph, B. Rosenhahn, and B. Wandt, “Probabilistic monocular 3d human pose estimation with normalizing flows,” in _International Conference on Computer Vision (ICCV)_ , Oct. 2021. * [29] L. Dinh, J. Sohl-Dickstein, and S. Bengio, “Density estimation using real nvp,” _arXiv preprint arXiv:1605.08803_ , 2016. * [30] C. Winkler, D. Worrall, E. Hoogeboom, and M. Welling, “Learning likelihoods with conditional normalizing flows,” _arXiv preprint arXiv:1912.00042_ , 2019\. * [31] D. P. Kingma and M. Welling, “Auto-encoding variational bayes,” _arXiv preprint arXiv:1312.6114_ , 2013. * [32] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” _Advances in neural information processing systems_ , vol. 30, 2017. * [33] H. Zheng, Z. Zhuang, J. Betz, and R. Mangharam, “Game-theoretic objective space planning,” _arXiv preprint arXiv:2209.07758_ , 2022. * [34] M. Zinkevich, M. Johanson, M. Bowling, and C. Piccione, “Regret Minimization in Games with Incomplete Information,” in _Advances in Neural Information Processing Systems_ , vol. 20. Curran Associates, Inc., 2007. * [35] E. F. Camacho and C. B. Alba, _Model predictive control_. Springer science & business media, 2013. * [36] S. Bansal, M. Chen, S. Herbert, and C. J. Tomlin, “Hamilton-jacobi reachability: A brief overview and recent advances,” in _2017 IEEE 56th Annual Conference on Decision and Control (CDC)_. IEEE, 2017, pp. 2242–2253. * [37] G. Chou, N. Ozay, and D. Berenson, “Safe output feedback motion planning from images via learned perception modules and contraction theory,” _arXiv preprint arXiv:2206.06553_ , 2022. * [38] A. D. Ames, X. Xu, J. W. Grizzle, and P. Tabuada, “Control barrier function based quadratic programs for safety critical systems,” _IEEE Transactions on Automatic Control_ , vol. 62, no. 8, pp. 3861–3876, 2016. * [39] A. Anand, K. Seel, V. Gjærum, A. Håkansson, H. Robinson, and A. Saad, “Safe learning for control using control lyapunov functions and control barrier functions: A review,” _Procedia Computer Science_ , vol. 192, pp. 3987–3997, 2021. * [40] C. Dawson, Z. Qin, S. Gao, and C. Fan, “Safe nonlinear control using robust neural lyapunov-barrier functions,” in _Conference on Robot Learning_. PMLR, 2022, pp. 1724–1735. * [41] A. Taylor, A. Singletary, Y. Yue, and A. Ames, “Learning for safety-critical control with control barrier functions,” in _Learning for Dynamics and Control_. PMLR, 2020, pp. 708–717. * [42] S. Dean, A. J. Taylor, R. K. Cosner, B. Recht, and A. D. Ames, “Guaranteeing safety of learned perception modules via measurement-robust control barrier functions,” _arXiv preprint arXiv:2010.16001_ , 2020. * [43] R. Cheng, G. Orosz, R. M. Murray, and J. W. Burdick, “End-to-end safe reinforcement learning through barrier functions for safety-critical continuous control tasks,” in _Proceedings of the AAAI Conference on Artificial Intelligence_ , vol. 33, no. 01, 2019, pp. 3387–3395. * [44] M. H. Cohen and C. Belta, “Approximate optimal control for safety-critical systems with control barrier functions,” in _2020 59th IEEE Conference on Decision and Control (CDC)_. IEEE, 2020, pp. 2062–2067. * [45] S. Yang, S. Chen, V. M. Preciado, and R. Mangharam, “Differentiable safe controller design through control barrier functions,” _arXiv preprint arXiv:2209.10034_ , 2022. * [46] D. Tabas and B. Zhang, “Computationally efficient safe reinforcement learning for power systems,” in _2022 American Control Conference (ACC)_. IEEE, 2022, pp. 3303–3310. * [47] A. Agrawal, B. Amos, S. Barratt, S. Boyd, S. Diamond, and J. Z. Kolter, “Differentiable convex optimization layers,” _Advances in neural information processing systems_ , vol. 32, 2019. * [48] B. Amos and J. Z. Kolter, “Optnet: Differentiable optimization as a layer in neural networks,” in _International Conference on Machine Learning_. PMLR, 2017, pp. 136–145. * [49] S. Boyd, S. P. Boyd, and L. Vandenberghe, _Convex optimization_. Cambridge university press, 2004.
# Quantum walk based state transfer algorithms on the complete $M$-partite graph S. Skoupý and M. Štefaňák Department of Physics, Faculty of Nuclear Sciences and Physical Engineering, Czech Technical University in Prague, Břehová 7, 115 19 Praha 1 - Staré Město, Czech Republic ###### Abstract We investigate coined quantum walk search and state transfer algorithms, focusing on the complete $M$-partite graph with $N$ vertices in each partition. First, it is shown that by adding a loop to each vertex the search algorithm finds the marked vertex with unit probability in the limit of a large graph. Next, we employ the evolution operator of the search with two marked vertices to perform a state transfer between the sender and the receiver. We show that when the sender and the receiver are in different partitions the algorithm succeeds with fidelity approaching unity for a large graph. However, when the sender and the receiver are in the same partition the fidelity does not reach exactly one. To amend this problem we propose a state transfer algorithm with an active switch, whose fidelity can be estimated based on the single vertex search alone. ## I Introduction Quantum walks [1] are quantum mechanical analogues of random walks. Their dynamics can be formulated either in discrete time steps [2], as we consider in the present paper, or in continuous time [3]. Both have found promising applications in quantum information processing [4], notably in quantum spatial search where an unsorted database is represented by a graph. The solution to the search problem corresponds to a marked vertex, where the local dynamics is different from the non-marked vertices. Usually, the initial state of the walk is taken as the equal weight superposition of all basis states. The walk is evolved coherently for $T$ steps, after which we perform a measurement which collapses the state of superposition and the walker is found on a single vertex. On various graphs quantum walk is capable of finding the marked vertex with sufficiently high probability in a number of steps that grows with the square root of the number of vertices $n$, i.e., the complexity is the same as for the abstract Grover search algorithm [5] which is known to be optimal. Initially, the investigation was mostly focused on graphs with some degree of symmetry or regularity. Continuous time quantum walks were shown to be optimal [6] for a complete graph, hypercube and lattices of dimensions greater than 4. Discrete time quantum walks with coins [7] are also optimal on these graphs [8, 9, 10, 11], in addition, they are optimal for lattices of dimensions greater than 2 [9]. Scattering quantum walk [12, 13, 14], which represents an alternative equivalent formulation of the coined quantum walk [15, 16], can perform optimal search e.g., on a star graph or a complete $M$-partite graph [17]. However, high symmetry is not required for the optimal performance of the quantum walk search [18, 19, 20]. In fact, it was shown [21] that continuous the time quantum walk is optimal on Erdös-Renyi random graphs as long as the probability of an edge existing between any pair of vertices is greater than $\left(\log^{\frac{3}{2}}n\right)/n$. However, on scale-free networks [22] the application of quantum walk search appears to be limited since the optimal run-time depends on the centrality of the marked node [23]. More recently [24] several sufficient and necessary conditions for continuous time quantum walk search to be optimal were derived. For the search algorithm (SA) it is not required that we find the marked vertex with unit probability. As long as the success probability is constant, we can repeat the SA several times depending on our error tolerance to find the marked vertex with high probability without changing the overall complexity of the algorithm. Even if the success probability is of the order of $1/\log n$, as for the discrete time quantum walk on the 2D lattice [9], we can use amplitude amplification [25] which increases the run-time of the SA by a factor of $\sqrt{\log n}$. There are several graphs where the quantum walk SA is exactly equivalent to the Grover search, e.g., the star graph [17], which means that the success probability is unity. It is interesting that for the discrete time quantum walks the success probability of the SA can be often increased close to unity by adding loops of appropriate weights at each vertex. This was found originally for the complete graph [9] and the hypercube [10]. Later investigations [26, 27, 28, 29, 30] found that this result is much more generic and the optimal weight of the loop depending on the size of the graph and degree of the vertex was identified. Recently, it was proven that adding loops improves the success probability of the SA on all regular locally arc-transitive graphs [31]. Quantum walks were also applied to the task of state transfer [32] between two vertices of a graph. In this context the initial state of the walk is localized on the sender vertex and we want to transfer it with high probability to the receiver vertex. Provided that the location of the sender and the receiver vertices are known, we can globally design the dynamics such that the walker is transferred from one to the other. This approach was investigated on different graphs such as circle [33, 34], 2D lattice [35], regular graphs [36] or more general networks [37]. When the sender and the receiver don’t know each other’s position they can perform state transfer by modifying the local coins at their own vertices, i.e., by implementing the evolution operator of the SA for two marked vertices. This approach was proposed for state transfer on lattices [38] and further analyzed on various types of finite graphs, e. g. on cycles and their variants [39, 40], star and complete graph with loops [41], complete bipartite graph [42], circulant graphs [43] or butterfly network [44]. Similar approach can be also applied for finding a path in a maze formed by star graphs [45] or trees [46]. We consider the state transfer algorithm (STA) following the second approach, i.e., based on the evolution operator of the SA with two marked vertices. For the STA it is desirable that we succeed in the first attempt, i.e., the fidelity of state transfer should be ideally one. Natural candidates for graphs where STA works with unit fidelity are those where also the SA succeeds with certainty. Indeed, the graphs considered in [41, 42] were chosen exactly based on this idea. However, in this paper we show an example of a graph where the SA works with certainty yet in some instance the STA does not have unit fidelity. We investigate search and state transfer on the complete $M$-partite graph with $N$ vertices in each partition, i.e., the graph has $n=NM$ vertices. Search on the complete $M$-partite graph was already investigated in [17] in the framework of the scattering quantum walk. In the coined walk the success probability of SA reaches $\frac{1}{2}$. We show that by adding a loop to each vertex the success probability tends to one for a large graph. Our approach is based on dimensional reduction [47, 48, 49, 19]. First, we find an exact invariant subspace $\cal I$ where the state of the algorithm evolves. Next, we investigate the eigenvectors of the evolution operator in the limit of a large graph $M\to\infty$ and $N\to\infty$ and determine those which have non- vanishing overlap with the initial state. These eigenvectors form an orthonormal basis of the relevant part of the invariant subspace. For the SA, we find an exact invariant subspace with dimension 8, and in the asymptotic limit the relevant part is three-dimensional. We then investigate the evolution operator of the search with two marked vertices for the sake of state transfer. There are two possible configurations - either sender and receiver are in the same partition or not. In the first case, we find an exact invariant subspace with dimension 11, while in the second case it has dimension 22. To simplify the calculations, we employ the symmetry of the graph which allows us to exchange the sender and the receiver vertex, or the whole partitions containing them in the latter case. This symmetry splits the invariant subspace $\cal I$ further into two closed subspaces - $\cal I_{+}$ in which the search with two marked vertices evolves, and the complementary subspace $\cal I_{-}$ needed for the state transfer. In the configuration where the sender and the receiver are in the same partition, the subspace $\cal I_{+}$ has dimension 8 and the complementary subspace $\cal I_{-}$ is three-dimensional. For the second configuration the subspaces have dimensions 12 and 10, respectively. Nevertheless, in the limit of a large graph only five eigenvectors of the evolution operator remain relevant in both configurations - three in the subspace $\cal I_{+}$ and two in $\cal I_{-}$. The corresponding eigenvalues can be also determined analytically. We show that when the sender and the receiver are in different partitions the phases of the relevant eigenvalues are harmonic. Hence, state transfer is achieved with unit fidelity. However, when the sender and the receiver are in the same partition, the phases are not harmonic, and the fidelity of state transfer is less than one. To fix this issue we propose an STA with an active switch, where initially only the sender vertex is marked, and after some number of steps the marking is switched to the receiver vertex. The fidelity reachable by this STA can be estimated based on the properties of the search for a single vertex. We show that STA with an active switch achieves perfect state transfer on the complete $M$-partite graph in the limit of large $N$ and $M$ for both configurations of the sender and the receiver vertex, and discuss its applicability on other graphs. The paper is organized as follows: In Section II we describe discrete time quantum walks with coin on finite graphs and introduce the quantum walk search and state transfer algorithms. In Section III search on the complete $M$-partite graph with one marked vertex is investigated in detail. Section IV is devoted to the STA. The cases of sender and receiver vertices being in the same or different partitions are analyzed in Section IV.1 and Section IV.2, respectively. In Section V we consider a STA with an active switch. We conclude and provide an outlook in Section VI. ## II Preliminaries In this section we overview the general design of the search and the state transfer algorithms based on the discrete-time quantum walks with coins [8], [9] and [38]. Before we turn to the algorithms, we describe the discrete time quantum walks with coins. Let us start with Hilbert space of the walk. Having a graph $G=\left(V,E\right)$ the corresponding Hilbert space ${\cal H}_{G}$ can be decomposed as a direct sum ${\cal H}_{G}=\bigoplus_{v}{\cal H}_{v},$ of local Hilbert spaces at each vertex $v\in V$. The orthonormal basis in ${\cal H}_{v}$ is given by vectors $\mathinner{|{v,w}\rangle}$ such that there is an edge between vertex $v$ and $w$ ${\cal H}_{v}={\rm Span}\left\\{|v,w\rangle|w\in V,\left\\{v,w\right\\}\in E\right\\}.$ In the basis state $\mathinner{|{v,w}\rangle}$ the first index $v$ describes the actual position of the walker, while the second index $w$ describes the direction of propagation of the walker. Movement of the walker is achieved by application of the flip-flop shift operator $\hat{S}$, which is defined in the following way $\displaystyle\hat{S}|v,w\rangle=|w,v\rangle.$ (1) To generate a nontrivial evolution a coin operator $\hat{C}$ is applied at every step before the shift takes place. The coin operator can be decomposed into a direct sum $\hat{C}=\bigoplus_{v}\hat{C}_{v}^{(l)},$ where $\hat{C}_{v}^{(l)}$ acts locally at a vertex $v$, i.e., it is a unitary operator on ${\cal H}_{v}$. The evolution operator $\hat{U}$ of one step of the walk is then given by $\displaystyle\hat{U}=\hat{S}\hat{C}.$ (2) The main idea of search and state transfer algorithms is to apply one local coin operator at the marked vertices and different local coin operator at the other vertices of the graph. Marked vertices are those that we want to find or between which we want transfer the walker. Usually, the local coin that is used at non-marked vertices is the Grover operator [5] $\displaystyle\hat{G}_{v}^{(l)}=2|\Omega_{v}\rangle\langle\Omega_{v}|-\hat{I}_{v}^{(l)},$ (3) where $\hat{I}_{v}^{(l)}$ is an identity operator at the subspace ${\cal H}_{v}$ and $|\Omega_{v}\rangle$ is an equal weight superposition of all basis states at the vertex $v$ given by $\displaystyle|\Omega_{v}\rangle=\frac{1}{\sqrt{d\left(v\right)}}\sum_{\begin{subarray}{c}w\\\ \left\\{v,w\right\\}\in E\end{subarray}}|v,w\rangle.$ (4) Here $d(v)$ denotes the degree of the vertex $v$, which is also the dimension of the subspace ${\cal H}_{v}$. The coin operator of the SA with one marked vertex $m$ reads $\displaystyle\hat{C}_{m}=\bigoplus_{\begin{subarray}{c}v\in V\\\ v\neq m\end{subarray}}\hat{G}_{v}^{(l)}\oplus\hat{C}_{m}^{(l)}$ (5) where $\hat{C}_{m}^{(l)}$ is the local coin operator at the marked vertex. A usual choice for the marked coin is either a simple phase shift by $\pi$ (i.e., $\hat{C}_{m}^{(l)}=-\hat{I}_{m}^{(l)}$), or the Grover operator followed by a phase shift by $\pi$ (i.e., $\hat{C}_{m}^{(l)}=-\hat{G}_{m}^{(l)}$). In the present paper we consider the latter case. Using the coin operator (5) we obtain the evolution operator of the SA $\hat{U}_{m}=\hat{S}\hat{C}_{m}.$ The steps of the SA are: 1. 1. Initialize the walk in the equal weight superposition of all basis states $\displaystyle|\Omega\rangle=\frac{1}{\sqrt{\sum\limits_{v\in V}d(v)}}\sum_{v\in V}\sqrt{d(v)}|\Omega_{v}\rangle.$ (6) 2. 2. Apply the evolution operator $\hat{U}_{m}$ $t$-times. The state of the walk after $t$ steps is given by $|\phi\left(t\right)\rangle=\hat{U}_{m}^{t}\mathinner{|{\Omega}\rangle}.$ 3. 3. Measure the walk. The probability to find the walker at the marked vertex is given by the summation over all basis states in the subspace ${\cal H}_{m}$ $\displaystyle P_{m}(t)=\sum_{\begin{subarray}{c}w\\\ \left\\{m,w\right\\}\in E\end{subarray}}|\langle m,w|\phi\left(t\right)\rangle|^{2}.$ (7) The optimal number of steps $T$ providing high success probability depends on the structure of the graph. In the case of STA we consider $2$ parties, sender and receiver sitting at vertices $s$ and $r$, respectively, which want to establish communication between each other. The sender and the receiver have access only to their local Hilbert spaces ${\cal H}_{s}$ and ${\cal H}_{r}$, respectively. Typical STA [38, 41, 42] uses the evolution operator of the search for two marked vertices $\hat{U}_{s,r}=S\hat{C}_{s,r},$ where we apply the same local coins at both marked vertices at the same time $\displaystyle\hat{C}_{s,r}=\bigoplus_{\begin{subarray}{c}v\in V\\\ v\neq s,r\end{subarray}}\hat{G}_{v}^{(l)}\oplus\hat{C}_{s}^{(l)}\oplus\hat{C}_{r}^{(l)}.$ (8) However, the initial state of STA is different — it starts localized at the sender vertex in some state $\mathinner{|{s}\rangle}$. Standard choice is the equal weight superposition of all basis states at the vertex $s$, i.e., $\mathinner{|{s}\rangle}=\mathinner{|{\Omega_{s}}\rangle}$. The steps of the STA are the following: 1. 1. Sender initializes the walk at its vertex in the state $\mathinner{|{s}\rangle}$. 2. 2. The evolution operator $\hat{U}_{s,r}$ is applied $t$-times. The state of the walk after $t$ steps is given by $|\phi\left(t\right)\rangle=\hat{U}_{s,r}^{t}\mathinner{|{s}\rangle}.$ 3. 3. Receiver measures the walk at its vertex. The fidelity of the STA, i.e., the probability that the receiver finds the walker at its vertex, is given by $\displaystyle{\cal F}(t)=\sum_{\begin{subarray}{c}w\\\ \left\\{r,w\right\\}\in E\end{subarray}}|\langle r,w|\phi\left(t\right)\rangle|^{2}.$ (9) The number of steps $T^{(st)}$ required to achieve state transfer with high fidelity depends again on the size and the structure of the graph. In contrast to the SA it is desirable that the STA performs with high fidelity in a single run. In such a case we talk about perfect state transfer. ## III Search on the complete M-partite graph Consider the complete $M$-partite graph (with $M>2$). Complete $M$-partite graph is a graph that has the set of vertices $V$ divided into $M$ subsets, where vertices have no edges between them, but are connected to all vertices in other subsets. We label the vertices of the graph as $v_{\alpha}$, where $\alpha=1,\ldots,M$ denotes the partition. The basis states of the quantum walk are therefore given by $\mathinner{|{v_{\alpha},w_{\beta}}\rangle}$, $\alpha\neq\beta$. We also limit ourselves to the case where all parts have the same size $N$, thus the whole graph has $n=MN$ vertices. This choice greatly simplifies the construction of the invariant subspace. The graph is $d$-regular with the vertex degree $d=N(M-1).$ Without loss of generality we assume that the marked vertex is in the first partition. Search on the complete $M$-partite graph was investigated in [17] in the framework of the scattering quantum walk. The difference between the two formulations is that in the coined walk the walker lives on the vertices, while in the scattering walk it lives on the edges. The results of [17] can be adopted for the coined quantum walk with a single modification. Namely, for the evaluation of the success probability (7) we consider only the overlap with the states where the walker is at the marked vertex, i.e., of the form $\mathinner{|{m_{1},k_{\alpha}}\rangle}$, $k=1,\ldots,N$, $\alpha=2,\ldots,M$, which form the basis of the local Hilbert space at marked vertex ${\cal H}_{m}$. The equal weight superposition of such states corresponds to the state $\mathinner{|{w_{2}}\rangle}$ in [17]. This gives us the success probability of $\frac{1}{2}$ for a large graph. Note that with probability close to $\frac{1}{2}$ we would find the walker in some state $\mathinner{|{k_{\alpha},m_{1}}\rangle}$, $k=1,\ldots,N$, $\alpha=2,\ldots,M$, i.e., where the walker is at the vertex $k_{\alpha}$ and would move to the marked vertex after the application of the shift $\hat{S}$. In the scattering walk framework these states are also considered in the evaluation of the success probability. Nevertheless, we focus on the coined formulation. In the end our goal is to investigate the state transfer between vertices where the coined formulation is more natural, since sender and receiver are considered to be restricted to their local Hilbert spaces ${\cal H}_{s}$ and ${\cal H}_{r}$, respectively. We show that one can improve the success probability of the coined walk search by adding one loop at each vertex. This is done by adding basis states $\mathinner{|{v_{\alpha},v_{\alpha}}\rangle}$ corresponding to the loop at each vertex, i.e. the dimension of the local Hilbert spaces ${\cal H}_{v_{\alpha}}$ increases by one. Moreover, we modify the local coin operator (3) by replacing the state $\mathinner{|{\Omega_{v_{\alpha}}}\rangle}$ (4) with the state $\mathinner{|{\Omega_{v_{\alpha}}(l)}\rangle}$ given by $\mathinner{|{\Omega_{v_{\alpha}}(l)}\rangle}=\frac{1}{\sqrt{d+l}}\left(\sum_{\begin{subarray}{c}\beta=1\\\ \beta\neq\alpha\end{subarray}}^{M}\sum_{k=1}^{N}\mathinner{|{{v_{\alpha}},k_{\beta}}\rangle}+\sqrt{l}\mathinner{|{{v_{\alpha}},{v_{\alpha}}}\rangle}\right),$ where $l$ is the weight of the loop. Note, however, that the initial state of the SA remains the same as before, i.e., we prepare the system in the equal weight superposition (6) of all basis states excluding the states corresponding to the loops. According to [29, 31] the optimal weight $l$ of the loops is given by $l=\frac{d}{NM}={1-\frac{1}{M}}.$ (10) Since we focus on the limit of a large graph we put $l=1$. This choice corresponds to the local coin operator being the Grover operator (3) of dimensions $d+1$, except for the marked vertex where we include an additional phase shift by $\pi$. First, we show that the SA evolves in an 8-dimensional invariant subspace ${\cal I}$. We note that in principle the dimension of the exact invariant subspace can be reduced further, since the evolution operator restricted on $\cal I$ still has some degenerate eigenvalues. However, the construction of the basis would not be as intuitive and the ensuing calculations will not simplify. Second, we consider a limit of a large graph which effectively reduces the dimension of the invariant subspace to three. Let us begin with the construction of the basis of the invariant subspace ${\cal I}$. The numerical simulations indicate that the state of the walk $\mathinner{|{\phi(t)}\rangle}$ evolves periodically close to a state corresponding to the loop on the marked vertex $\mathinner{|{m_{1},m_{1}}\rangle}$. Hence, we consider this desired target state of the SA as the first basis vector of the invariant subspace $\mathinner{|{\nu_{1}}\rangle}=\mathinner{|{m_{1},m_{1}}\rangle}.$ (11) Next, we add an equal weight superposition of all edges leaving the marked vertex $|\nu_{2}\rangle=\frac{1}{\sqrt{d}}\sum_{\alpha=2}^{M}\sum_{k=1}^{N}|m_{1},k_{\alpha}\rangle.$ (12) Concerning the non-marked vertices in the first partition, we add two states corresponding to a superposition of all loops and a superposition of all edges leaving the first partition $\displaystyle|\nu_{3}\rangle$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{(N-1)}}\sum_{j\neq m}^{N}|j_{1},j_{1}\rangle,$ $\displaystyle|\nu_{4}\rangle$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d(N-1)}}\sum_{j\neq m}^{N}\sum_{\alpha=2}^{M}\sum_{k=1}^{N}|j_{1},k_{\alpha}\rangle.$ Next, we consider the edges leading to the first partition from the outside and ending either on the marked or non-marked vertex, and construct the following two basis states $\displaystyle|\nu_{5}\rangle$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d}}\sum_{\alpha=2}^{M}\sum_{k=1}^{N}|k_{\alpha},m_{1}\rangle,$ (13) $\displaystyle|\nu_{6}\rangle$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d(N-1)}}\sum_{j\neq m}^{N}\sum_{\alpha=2}^{M}\sum_{k=1}^{N}|k_{\alpha},j_{1}\rangle.$ These states can be obtained by applying the shift operator on $\mathinner{|{\nu_{2}}\rangle}$ and $\mathinner{|{\nu_{4}}\rangle}$. To complete the basis we consider the states corresponding to the superposition of all edges between the vertices outside of the first partition $|\nu_{7}\rangle=\frac{1}{\sqrt{d(d-N)}}\sum_{\begin{subarray}{c}\alpha,\beta=2\\\ \beta\neq\alpha\end{subarray}}^{M}\sum_{j,k=1}^{N}|j_{\alpha},k_{\beta}\rangle$ and the superposition of all remaining loops $|\nu_{8}\rangle=\frac{1}{\sqrt{d}}\sum_{\alpha=2}^{M}\sum_{k=1}^{N}|k_{\alpha},k_{\alpha}\rangle.$ Clearly, the initial state of the SA (6) lies in ${\cal I}$ and has the following form $\displaystyle\mathinner{|{\Omega}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{MN}}\left[|\nu_{2}\rangle+|\nu_{5}\rangle+\sqrt{d-N}|\nu_{7}\rangle+\right.$ (14) $\displaystyle\left.+\sqrt{N-1}(\mathinner{|{\nu_{4}}\rangle}+\mathinner{|{\nu_{6}}\rangle})\right].$ By direct calculation one can show that the evolution operator $\hat{U}_{m}$ of the SA acts on the basis states according to $\displaystyle\hat{U}_{m}\mathinner{|{\nu_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left((d-1)\mathinner{|{\nu_{1}}\rangle}-2\sqrt{d}\mathinner{|{\nu_{5}}\rangle}\right),$ $\displaystyle\hat{U}_{m}\mathinner{|{\nu_{2}}\rangle}$ $\displaystyle=$ $\displaystyle-\frac{1}{d+1}\left(2\sqrt{d}\mathinner{|{\nu_{1}}\rangle}+(d-1)\mathinner{|{\nu_{5}}\rangle}\right),$ $\displaystyle\hat{U}_{m}\mathinner{|{\nu_{3}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left((1-d)\mathinner{|{\nu_{3}}\rangle}+2\sqrt{d}\mathinner{|{\nu_{6}}\rangle}\right),$ $\displaystyle\hat{U}_{m}\mathinner{|{\nu_{4}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\sqrt{d}\mathinner{|{\nu_{3}}\rangle}+(d-1)\mathinner{|{\nu_{6}}\rangle}\right),$ $\displaystyle\hat{U}_{m}\mathinner{|{\nu_{5}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left((1-d)\mathinner{|{\nu_{2}}\rangle}+2\sqrt{N-1}\mathinner{|{\nu_{4}}\rangle}+\right.$ $\displaystyle\left.+2\sqrt{d-N}\mathinner{|{\nu_{7}}\rangle}+2\mathinner{|{\nu_{8}}\rangle}\right),$ $\displaystyle\hat{U}_{m}\mathinner{|{\nu_{6}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\sqrt{N-1}\mathinner{|{\nu_{2}}\rangle}-(d+3-2N)\mathinner{|{\nu_{4}}\rangle}+\right.$ $\displaystyle+2\sqrt{(d-N)(N-1)}\mathinner{|{\nu_{7}}\rangle}+$ $\displaystyle\left.+2\sqrt{N-1}\mathinner{|{\nu_{8}}\rangle}\right),$ $\displaystyle\hat{U}_{m}\mathinner{|{\nu_{7}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\sqrt{d-N}\mathinner{|{\nu_{2}}\rangle}+\right.$ $\displaystyle\left.+2\sqrt{(d-N)(N-1)}\mathinner{|{\nu_{4}}\rangle}+\right.$ $\displaystyle\left.+(d-1-2N)\mathinner{|{\nu_{7}}\rangle}+2\sqrt{d-N}\mathinner{|{\nu_{8}}\rangle}\right),$ $\displaystyle\hat{U}_{m}\mathinner{|{\nu_{8}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\mathinner{|{\nu_{2}}\rangle}+2\sqrt{N-1}\mathinner{|{\nu_{4}}\rangle}+\right.$ (15) $\displaystyle\left.+2\sqrt{d-N}\mathinner{|{\nu_{7}}\rangle}+(1-d)\mathinner{|{\nu_{8}}\rangle}\right).$ Hence, $\cal I$ is indeed an invariant subspace of the SA and the state of the walk $\mathinner{|{\phi(t)}\rangle}$ remains in $\cal I$ for all $t$. Its evolution is determined by the eigenvalues and eigenvectors of $\hat{U}_{m}$, which is in the invariant subspace represented by an $8\times 8$ unitary matrix with matrix elements given by (15). While it is possible to diagonalize this matrix analytically, the procedure is rather onerous and the resulting expressions are quite lengthy. Nevertheless, the analysis can be considerably simplified in the limit of a large graph, i.e., when $N\to\infty$ and $M\to\infty$. As we can see from the expansion (14) for large $N$ and $M$ the initial state of the algorithm tends to $\mathinner{|{\nu_{7}}\rangle}$. Hence, the only eigenvectors of the evolution operator $\hat{U}_{m}$, which remain relevant in the asymptotic limit, are those which have non-vanishing overlap with $\mathinner{|{\nu_{7}}\rangle}$. It turns out that there are only three such states and their asymptotic form is given by $\displaystyle\mathinner{|{\psi_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}(\mathinner{|{\nu_{7}}\rangle}-\mathinner{|{\nu_{1}}\rangle}),$ $\displaystyle\mathinner{|{\psi_{2}^{(\pm)}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{2}(\mathinner{|{\nu_{1}}\rangle}+\mathinner{|{\nu_{7}}\rangle}\pm i(\mathinner{|{\nu_{5}}\rangle}-\mathinner{|{\nu_{2}}\rangle})).$ (16) It can be shown that for the other eigenvectors of $\hat{U}_{m}$ the overlap with the initial state decreases at least as $O(1/\sqrt{NM})$. Let us turn to the eigenvalues. The eigenvector $\mathinner{|{\psi_{1}}\rangle}$ corresponds to $\lambda_{1}=1$. For $\mathinner{|{\psi_{2}^{(\pm)}}\rangle}$ the eigenvalues have the form $\lambda_{2}^{(\pm)}=e^{\pm i\omega_{2}}.$ From the characteristic polynomial of $\hat{U}_{m}$ we find that $\cos\omega_{2}$ is given by the largest root of the quadratic equation $\displaystyle x^{2}-\left(1-\frac{N}{d+1}\right)x-\frac{(d+1)(N-2)+N-3}{{(d+1)}^{2}}=0.$ This leads us to $\displaystyle\omega_{2}$ $\displaystyle=$ $\displaystyle\arccos\left(1-\frac{1+NM-\sqrt{N^{2}M^{2}-6NM+4N+5}}{2{(d+1)}}\right)$ (17) $\displaystyle\approx$ $\displaystyle\frac{2}{\sqrt{NM}}.$ From the relations (16) we express the initial and the target state of the SA in terms of the eigenvectors of the evolution operator as $\displaystyle\mathinner{|{\Omega}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\mathinner{|{\psi_{1}}\rangle}+\frac{1}{2}\left(\mathinner{|{\psi_{2}^{(+)}}\rangle}+\mathinner{|{\psi_{2}^{(-)}}\rangle}\right),$ $\displaystyle\mathinner{|{\nu_{1}}\rangle}$ $\displaystyle=$ $\displaystyle-\frac{1}{\sqrt{2}}\mathinner{|{\psi_{1}}\rangle}+\frac{1}{2}\left(\mathinner{|{\psi_{2}^{(+)}}\rangle}+\mathinner{|{\psi_{2}^{(-)}}\rangle}\right).$ (18) The state after $t$ iterations of the SA reads $\mathinner{|{\phi(t)}\rangle}=\frac{1}{\sqrt{2}}\mathinner{|{\psi_{1}}\rangle}+\frac{1}{2}\left(e^{i\omega_{2}t}\mathinner{|{\psi_{2}^{(+)}}\rangle}+e^{-i\omega_{2}t}\mathinner{|{\psi_{2}^{(-)}}\rangle}\right).$ (19) The success probability (7) of the SA after $t$ steps can be expressed in the form $P_{m}(t)=|\langle\nu_{1}|\phi(t)\rangle|^{2}+|\langle\nu_{2}|\phi(t)\rangle|^{2}.$ (20) From (19) and (18) we see that the probability to find the walker in the target state $\mathinner{|{\nu_{1}}\rangle}$ is given by $|\langle\nu_{1}|\phi(t)\rangle|^{2}=\sin^{4}\left(\frac{\omega_{2}t}{2}\right).$ (21) The probability to find the walker in the state $\mathinner{|{\nu_{2}}\rangle}$ reads $|\langle\nu_{2}|\phi(t)\rangle|^{2}=\frac{1}{4}\sin^{2}\left(\omega_{2}t\right).$ (22) Put together we obtain the overall success probability of the SA $P_{m}(t)=\sin^{2}\left(\frac{\omega_{2}t}{2}\right).$ (23) We see that for $t=\frac{\pi}{\omega_{2}}$ the state of the SA is very close to the target state $\mathinner{|{\nu_{1}}\rangle}$. Hence, the number of steps needed to find the marked vertex with probability close to one is given by $T=\frac{\pi}{\omega_{2}}\approx\frac{\pi\sqrt{NM}}{2}+O\left(\frac{1}{\sqrt{NM}}\right).$ (24) For illustration we plot in Figure 1 the probability to find the marked vertex (23) as a function of the number of steps for a graph with $N=40$ and $M=100$. Figure 1: Overall success probability of SA (black dots) and the probability to find the walker in the target state $\mathinner{|{\nu_{1}}\rangle}$ (blue squares) as a function of the number of steps $t$ for $N=40$ and $M=100$. Red curves correspond to the analytical results (dashed line to eq. (21) and full line to eq. (23) ). Green diamonds denote the probability that the walker is on the marked vertex but not in the loop, which follows the curve (22). The success probability is close to one after $T\approx 100$ steps, in accordance with (24). We note that the result (23) holds in the limit of large $N$ and $M$. To investigate how quickly does the success probability at the optimal time (24) approaches unity we performed numerical simulations for various values of $N$ and $M$. The simulations indicate that the success probability is essentially independent of $N$ and with $M$ it scales according to $P_{m}(T)=1-O\left(\frac{1}{M}\right).$ The results are illustrated in Figure 2. Figure 2: Overall success probability of SA as a function of the number of partitions $M$ for $N=10$ (gray circles), $N=50$ (blue triangles) and $N=100$ (brown diamonds). For a given $N$ and $M$ we evaluate numerically the evolution of SA for the optimal number of steps $T$ given by (24) and determine $P_{m}(T)$ from the formula (20). To unravel the scaling of the success probability we plot $1-P_{m}(T)$ on the log-log scale. Independent of the value of $N$, the sets of data-points fit well onto the $1/M$ slope indicated by the red line. ## IV State transfer We now turn to the analysis of the STA. There are two possible configurations — the sender and the receiver are in the same partition or not. Numerical simulations indicate that for the graph without loops the STA does not work well. When the sender and the receiver are in the same partition the fidelity does not surpass 0.25. For the second configuration the first maximum of fidelity tends to 0.8. Hence, we turn to the graph with loops. Choosing the initial state of the sender as the equal weight superposition (4), i.e., $\mathinner{|{s}\rangle}=\mathinner{|{\Omega_{s}}\rangle}$, the numerical simulation reveals that the fidelity tends to 1 for the second configuration. However, when the sender and the receiver are in the same partition the fidelity is still limited. As we show in Figure 3 it does not surpass 0.35. A careful analysis would reveal that the culprit are two orthogonal eigenvectors of the evolution operator of the STA corresponding to the eigenvalue -1, one having a large overlap with $\mathinner{|{\Omega_{s}}\rangle}$ and the other one with $\mathinner{|{\Omega_{r}}\rangle}$. In the second configuration this does not happen, as both $\mathinner{|{\Omega_{s}}\rangle}$ and $\mathinner{|{\Omega_{r}}\rangle}$ have overlaps with the same eigenvectors of the evolution operator of the STA. Hence, the absence of an edge between the sender and the receiver vertex in the first configuration significantly limits the achievable fidelity when we use the equal weight superposition state $\mathinner{|{\Omega_{s}}\rangle}$ as the initial state of STA. Figure 3: The evolution of the fidelity ${\cal F}$ of the state transfer during $1000$ steps for $N=40$ and $M=100$. Sender and receiver are in the same part. The initial state $\mathinner{|{s}\rangle}$ is the equal weight superposition on the sender vertex $\mathinner{|{\Omega_{s}}\rangle}$. We see that the fidelity does not grow over $0.35$. We show that the fidelity of STA in the first configuration can be improved considerably by choosing the initial state $\mathinner{|{s}\rangle}$ as the loop on the sender vertex. Moreover, we prove that this initial state works well also in the second configuration. In both configurations the walker will be with high probability transferred to the loop at the receiver vertex. We denote this receiver state as $\mathinner{|{r}\rangle}$. ### IV.1 Sender and receiver in the same partition Let us first consider the case when the sender and the receiver are in the same partition of the graph , i.e., they are not connected directly by an edge. Without loss of generality we consider that they are in the first one. We begin by constructing the basis of the exact invariant subspace. The procedure is similar to the one for the SA, but we have to consider two marked vertices corresponding to the sender $s$ and the receiver $r$. Hence, the basis states of the form (11), (12), (13) will appear twice - once for $m=s$ and once for $m=r$. In the end we find the following 11 basis vectors $\displaystyle\mathinner{|{\nu_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{s_{1},s_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{2}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d}}\sum_{\alpha=2}^{M}\sum_{j=1}^{N}\mathinner{|{s_{1},j_{\alpha}}\rangle},$ $\displaystyle\mathinner{|{\nu_{3}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{r_{1},r_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{4}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d}}\sum_{\alpha=2}^{M}\sum_{j=1}^{N}\mathinner{|{r_{1},j_{\alpha}}\rangle},$ $\displaystyle\mathinner{|{\nu_{5}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{N-2}}\sum_{j\neq s,r}^{N}\mathinner{|{j_{1},j_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{6}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d(N-2)}}\sum_{j\neq s,r}^{N}\sum_{\alpha=2}^{M}\sum_{k=1}^{N}\mathinner{|{j_{1},k_{\alpha}}\rangle},$ $\displaystyle\mathinner{|{\nu_{7}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d}}\sum_{\alpha=2}^{M}\sum_{j=1}^{N}\mathinner{|{j_{\alpha},s_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{8}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d}}\sum_{\alpha=2}^{M}\sum_{j=1}^{N}\mathinner{|{j_{\alpha},r_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{9}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d(N-2)}}\sum_{j\neq s,r}^{N}\sum_{\alpha=2}^{M}\sum_{k=1}^{N}\mathinner{|{k_{\alpha},j_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{10}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d(d-N)}}\sum_{\alpha=2}^{M}\sum_{\beta=2,\beta\neq\alpha}^{M}\sum_{j,k=1}^{N}\mathinner{|{j_{\alpha},k_{\beta}}\rangle},$ $\displaystyle\mathinner{|{\nu_{11}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d}}\sum_{\alpha=2}^{M}\sum_{j}^{N}\mathinner{|{j_{\alpha},j_{\alpha}}\rangle}$ (25) Let us denote the subspace spanned by these vectors as $\cal I$. Clearly, it contains the initial and the desired target state of the STA ($\mathinner{|{s}\rangle}=\mathinner{|{\nu_{1}}\rangle}$ and $\mathinner{|{r}\rangle}=\mathinner{|{\nu_{3}}\rangle}$). It can be shown by direct calculation that it is closed under the action of $\hat{U}_{s,r}$. However, we will not provide the expression of $\hat{U}_{s,r}$ in this basis because $\cal I$ can be divided further into two invariant subspaces. This comes from a fact that the evolution of the STA is invariant with respect to the exchange of the sender and the receiver vertex. Let us denote the operator corresponding to this symmetry as $\hat{P}$. Clearly, it holds that $\hat{P}^{2}=\hat{I},\quad[\hat{P},\hat{U}_{s,r}]=0.$ $\hat{P}$ acts on the basis states (25) as $\displaystyle\hat{P}\mathinner{|{\nu_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{\nu_{3}}\rangle},\quad\hat{P}\mathinner{|{\nu_{2}}\rangle}=\mathinner{|{\nu_{4}}\rangle},\quad\hat{P}\mathinner{|{\nu_{7}}\rangle}=\mathinner{|{\nu_{8}}\rangle},$ $\displaystyle\hat{P}\mathinner{|{\nu_{j}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{\nu_{j}}\rangle},\quad j=5,6,9,10,11.$ Since $\hat{P}$ commutes with $\hat{U}_{s,r}$, they have common eigenvectors. From $\hat{P}^{2}=\hat{I}$ we see that the spectrum of $\hat{P}$ consists of two eigenvalues $1$ and $-1$. Hence, the invariant subspace ${\cal I}$ can be split into two subspaces ${\cal I}_{+}$ and ${\cal I}_{-}$ which correspond to eigenvalues $1$ and $-1$ of the operator $\hat{P}$. Basis of the invariant subspace ${\cal I}_{+}$ is spanned by eigenstates denoted as $\mathinner{|{\sigma_{i}}\rangle}$, $i=1,\ldots,8$ and it has the following form $\displaystyle\mathinner{|{\sigma_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{1}}\rangle}+\mathinner{|{\nu_{3}}\rangle}\right),$ $\displaystyle\mathinner{|{\sigma_{2}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{2}}\rangle}+\mathinner{|{\nu_{4}}\rangle}\right),$ $\displaystyle\mathinner{|{\sigma_{3}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{\nu_{5}}\rangle},$ $\displaystyle\mathinner{|{\sigma_{4}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{\nu_{6}}\rangle},$ $\displaystyle\mathinner{|{\sigma_{5}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{7}}\rangle}+\mathinner{|{\nu_{8}}\rangle}\right),$ $\displaystyle\mathinner{|{\sigma_{6}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{\nu_{9}}\rangle},$ $\displaystyle\mathinner{|{\sigma_{7}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{\nu_{10}}\rangle},$ $\displaystyle\mathinner{|{\sigma_{8}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{\nu_{11}}\rangle}$ Note that if we perform SA for two marked vertices instead of STA, the subspace ${\cal I}_{+}$ would be invariant with respect to the search. This is due to the fact that the initial state of SA algorithm lies within this subspace. Basis of the invariant subspace ${\cal I}_{-}$ is spanned by eigenstates denoted as $\mathinner{|{\tau_{i}}\rangle}$, $i=1,\ldots,3$ and it has the following form $\displaystyle\mathinner{|{\tau_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{1}}\rangle}-\mathinner{|{\nu_{3}}\rangle}\right),$ $\displaystyle\mathinner{|{\tau_{2}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{2}}\rangle}-\mathinner{|{\nu_{4}}\rangle}\right),$ $\displaystyle\mathinner{|{\tau_{3}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{7}}\rangle}-\mathinner{|{\nu_{8}}\rangle}\right)$ This subspace is needed only in STA since the initial state of SA is orthogonal to this subspace. The sender and the receiver states in the new basis read $\displaystyle\mathinner{|{s}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\sigma_{1}}\rangle}+\mathinner{|{\tau_{1}}\rangle}\right),$ $\displaystyle\mathinner{|{r}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\sigma_{1}}\rangle}-\mathinner{|{\tau_{1}}\rangle}\right).$ (26) The evolution operator in the new basis is block diagonal, i.e., ${\cal I}_{i}$ are the invariant subspaces of $\hat{U}_{s,r}$. We find the following relations for the basis states of ${\cal I}_{+}$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left((d-1)\mathinner{|{\sigma_{1}}\rangle}-2\sqrt{d}\mathinner{|{\sigma_{5}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{2}}\rangle}$ $\displaystyle=$ $\displaystyle-\frac{1}{d+1}\left(2\sqrt{d}\mathinner{|{\sigma_{1}}\rangle}+(d-1)\mathinner{|{\sigma_{5}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{3}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left((1-d)\mathinner{|{\sigma_{3}}\rangle}+2\sqrt{d}\mathinner{|{\sigma_{6}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{4}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\sqrt{d}\mathinner{|{\sigma_{3}}\rangle}+(d-1)\mathinner{|{\sigma_{6}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{5}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left((3-d)\mathinner{|{\sigma_{2}}\rangle}+2\sqrt{2(N-2)}\mathinner{|{\sigma_{4}}\rangle}+2\sqrt{2(d-N)}\mathinner{|{\sigma_{7}}\rangle}+2\sqrt{2}\mathinner{|{\sigma_{8}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{6}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\sqrt{2(N-2)}\mathinner{|{\sigma_{2}}\rangle}-\left(d-2N+5\right)\mathinner{|{\sigma_{4}}\rangle}+2\sqrt{(d-N)(N-2)}\mathinner{|{\sigma_{7}}\rangle}+2\sqrt{N-2}\mathinner{|{\sigma_{8}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{7}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\sqrt{2(d-N)}\mathinner{|{\sigma_{2}}\rangle}+2\sqrt{(d-N)(N-2)}\mathinner{|{\sigma_{4}}\rangle}+\left(d-2N-1\right)\mathinner{|{\sigma_{7}}\rangle}+2\sqrt{d-N}\mathinner{|{\sigma_{8}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{8}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\sqrt{2}\mathinner{|{\sigma_{2}}\rangle}+2\sqrt{N-2}\mathinner{|{\sigma_{4}}\rangle}+2\sqrt{d-N}\mathinner{|{\sigma_{7}}\rangle}-\left(d-1\right)\mathinner{|{\sigma_{8}}\rangle}\right).$ (27) In the second subspace ${\cal I}_{-}$ the evolution operator acts according to $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(\left(d-1\right)\mathinner{|{\tau_{1}}\rangle}-2\sqrt{d}\mathinner{|{\tau_{3}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{2}}\rangle}$ $\displaystyle=$ $\displaystyle-\frac{1}{d+1}\left(2\sqrt{d}\mathinner{|{\tau_{1}}\rangle}+\left(d-1\right)\mathinner{|{\tau_{3}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{3}}\rangle}$ $\displaystyle=$ $\displaystyle-\mathinner{|{\tau_{2}}\rangle}.$ (28) Let us now investigate the dynamics of the STA in the limit of a large graph. We denote by $\hat{U}_{\pm}$ the restriction of $\hat{U}_{s,r}$ on ${\cal I}_{\pm}$, and determine the spectrum and eigenvectors of these operators. For $\hat{U}_{+}$ the results are similar to those for the SA. In the subspace ${\cal I}_{+}$ three relevant eigenvectors remain, as for the others the overlap with $\mathinner{|{\sigma_{1}}\rangle}$ tends to zero at least as $O(1/\sqrt{NM})$. The limit form of the relevant eigenvectors is given by $\displaystyle\mathinner{|{\psi_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\sqrt{\frac{2}{3}}\mathinner{|{\sigma_{1}}\rangle}-\frac{1}{\sqrt{3}}\mathinner{|{\sigma_{7}}\rangle},$ $\displaystyle\mathinner{|{\psi_{2}^{(\pm)}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{6}}\mathinner{|{\sigma_{1}}\rangle}+\frac{1}{\sqrt{3}}\mathinner{|{\sigma_{7}}\rangle}\pm\frac{i}{2}(\mathinner{|{\sigma_{5}}\rangle}-\mathinner{|{\sigma_{2}}\rangle}).$ (29) The eigenvector $\mathinner{|{\psi_{1}}\rangle}$ corresponds to $\lambda_{1}=1$. In the case of $\mathinner{|{\psi_{2}^{(\pm)}}\rangle}$ the eigenvalues have the form $\lambda_{2}^{(\pm)}=e^{\pm i\omega_{2}}.$ From the characteristic polynomial of $\hat{U}_{+}$ we find that $\cos\omega_{2}$ is the largest root of the quadratic equation $x^{2}-\left(1-\frac{N}{d+1}\right)x-\frac{{(d+1)}(N-3)+N-5}{{(d+1)}^{2}}=0,$ which leads us to $\displaystyle\omega_{2}$ $\displaystyle=$ $\displaystyle\arccos\left(1-\frac{1+NM-\sqrt{N^{2}M^{2}-10NM+8N+9}}{2{(d+1)}}\right)$ (30) $\displaystyle\approx$ $\displaystyle\sqrt{\frac{6}{NM}}.$ In the subspace $\cal I_{-}$ there are two additional relevant eigenvectors in the asymptotic limit $\mathinner{|{\psi_{3}^{(\pm)}}\rangle}=\frac{1}{\sqrt{2}}\mathinner{|{\tau_{1}}\rangle}\pm\frac{i}{2}(\mathinner{|{\tau_{2}}\rangle}-\mathinner{|{\tau_{3}}\rangle}).$ (31) For the last eigenvector of $\hat{U}_{-}$ the overlap with the state $\mathinner{|{\tau_{1}}\rangle}$ behaves like $O(1/\sqrt{NM})$. The relevant eigenvalues have the form $\displaystyle\lambda_{3}^{(\pm)}$ $\displaystyle=$ $\displaystyle e^{\pm i\omega_{3}},$ with $\omega_{3}=\arccos\left(1-\frac{1}{{d+1}}\right)\approx\sqrt{\frac{2}{NM}}.$ (32) Using the results (29), (31) we see that for a large graph the sender and the receiver states (26) can be decomposed into the eigenvectors of the evolution operator $\hat{U}_{s,r}$ according to $\displaystyle\mathinner{|{s}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{3}}\mathinner{|{\psi_{1}}\rangle}+\frac{1}{\sqrt{12}}\left(\mathinner{|{\psi_{2}^{(+)}}\rangle}+\mathinner{|{\psi_{2}^{(-)}}\rangle}\right)+$ $\displaystyle+\frac{1}{2}\left(\mathinner{|{\psi_{3}^{(+)}}\rangle}+\mathinner{|{\psi_{3}^{(-)}}\rangle}\right),$ $\displaystyle\mathinner{|{r}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{3}}\mathinner{|{\psi_{1}}\rangle}+\frac{1}{\sqrt{12}}\left(\mathinner{|{\psi_{2}^{(+)}}\rangle}+\mathinner{|{\psi_{2}^{(-)}}\rangle}\right)-$ $\displaystyle-\frac{1}{2}\left(\mathinner{|{\psi_{3}^{(+)}}\rangle}+\mathinner{|{\psi_{3}^{(-)}}\rangle}\right).$ Hence, the evolution of STA takes place in a five dimensional subspace $\displaystyle\mathinner{|{\phi(t)}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{3}}\mathinner{|{\psi_{1}}\rangle}+$ (33) $\displaystyle+\frac{1}{\sqrt{12}}\left(e^{i\omega_{2}t}\mathinner{|{\psi_{2}^{(+)}}\rangle}+e^{-i\omega_{2}t}\mathinner{|{\psi_{2}^{(-)}}\rangle}\right)+$ $\displaystyle+\frac{1}{2}\left(e^{i\omega_{3}t}\mathinner{|{\psi_{3}^{(+)}}\rangle}+e^{-i\omega_{3}t}\mathinner{|{\psi_{3}^{(-)}}\rangle}\right).$ The fidelity of STA can be written as a sum ${\cal F}(t)=|\langle r|\phi(t)\rangle|^{2}+|\langle\nu_{4}|\phi(t)\rangle|^{2},$ (34) of probabilities that the walker is in the receiver state $\mathinner{|{r}\rangle}=\mathinner{|{\nu_{3}}\rangle}$ corresponding to the loop, or at the receiver vertex but not in the loop, i.e., the state $\mathinner{|{\nu_{4}}\rangle}$. From the relations (29), (31) and (33) we find that these probabilities are given by $\displaystyle|\langle r|\phi(t)\rangle|^{2}$ $\displaystyle=$ $\displaystyle\frac{1}{36}\left(2+\cos\left(\omega_{2}t\right)-3\cos\left(\omega_{3}t\right)\right)^{2},$ (35) $\displaystyle|\langle\nu_{4}|\phi(t)\rangle|^{2}$ $\displaystyle=$ $\displaystyle\frac{1}{24}\left(\sin(\omega_{2}t)-\sqrt{3}\sin(\omega_{3}t)\right)^{2}.$ (36) From the relations (30), (32) we see that the frequencies are not harmonic, since for a large graph $\omega_{2}=\sqrt{3}\omega_{3}.$ The overall fidelity of STA is then given by $\displaystyle{\cal F}(t)$ $\displaystyle=$ $\displaystyle\frac{1}{36}\left(2+\cos\left(\sqrt{3}\omega_{3}t\right)-3\cos\left(\omega_{3}t\right)\right)^{2}+$ (37) $\displaystyle+\frac{1}{24}\left(\sin(\sqrt{3}\omega_{3}t)-\sqrt{3}\sin(\omega_{3}t)\right)^{2}.$ The fidelity of STA will not reach one exactly. However, for a large graph the first maximum of fidelity reaches the value ${\cal F}_{1}\approx 0.94.$ (38) The number of steps required to reach the first maximum is approximately given by $T^{(st)}\approx 2.39\sqrt{NM}.$ (39) At this time the walker is with high probability in the receiver state $\mathinner{|{r}\rangle}$. For illustration we show in Figure 4 the evolution of fidelity for a graph with $N=40$ and $M=100$. Figure 4: Fidelity of the state transfer as a function of the number of steps $t$ for $N=40$ and $M=100$. The sender and the receiver vertices are in the same partition. Black dots are obtained from the numerical simulation, the full red curve corresponds to (37). Since the frequencies (30), (32) are not integer multiples the fidelity behaves an-harmonically. At the time of the first maximum (39) the walker is with high probability in the receiver state $\mathinner{|{r}\rangle}$, which is depicted by the blue squares. The probability to be in the receiver state follows the curve (35) represented by the red dashed curve. The green diamonds correspond to the probability that the walker is at the marked vertex but not in the loop, which follows the curve (36). The fidelity (38) in the first maximum is reached in the limit of large $N$ and $M$. We have performed numerical simulations to investigate how quickly does the fidelity at the optimal time (39) approaches the asymptotic value (38). Similarly to the results for the SA, the simulations indicate that the fidelity is essentially independent of $N$ and with $M$ it scales according to ${\cal F}(T^{(st)})={\cal F}_{1}-O\left(\frac{1}{M}\right).$ The results are illustrated in Figure 5. Figure 5: Overall fidelity of STA as a function of the number of partitions $M$ for $N=10$ (gray circles), $N=50$ (blue triangles) and $N=100$ (brown diamonds). The sender and the receiver vertices are in the same partition. For a given $N$ and $M$ we evaluate numerically the evolution of STA for the number of steps $T^{(st)}$ needed to reach the first maximum (39) and determine ${\cal F}(T^{(st)})$ according to (34). To unravel the scaling of the fidelity we plot ${\cal F}_{1}-{\cal F}(T^{(st)})$ on the log-log scale. The data-points follow the $1/M$ slope indicated by the red line, with almost no dependence on $N$. ### IV.2 Sender and receiver in different partitions Let us now turn to the case when the sender and the receiver are in different parts of the complete M-partite graph. Without loss of generality we label the partition containing the sender as 1 and the partition with the receiver as 2. The construction of the basis of the invariant subspace $\cal I$ is more involved, since we have to consider the second partition with the receiver vertex separately from the rest of the graph. We begin with the states starting at the sender vertex $\displaystyle\mathinner{|{\nu_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{s_{1},s_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{2}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{s_{1},r_{2}}\rangle},$ $\displaystyle\mathinner{|{\nu_{3}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{N-1}}\sum_{j\neq r}^{N}\mathinner{|{s_{1},j_{2}}\rangle},$ $\displaystyle\mathinner{|{\nu_{4}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d-N}}\sum_{\alpha=3}^{M}\sum_{j=1}^{N}\mathinner{|{s_{1},j_{\alpha}}\rangle},$ (40) which correspond to the loop, edge from the sender to the receiver, equal weight superposition of all edges from the sender to the remaining vertices in the second partition, and edges to all remaining vertices. We repeat the same for the receiver vertex $\displaystyle\mathinner{|{\nu_{5}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{r_{2},r_{2}}\rangle},$ $\displaystyle\mathinner{|{\nu_{6}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{r_{2},s_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{7}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{N-1}}\sum_{j\neq s}^{N}\mathinner{|{r_{2},j_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{8}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d-N}}\sum_{\alpha=3}^{M}\sum_{j=1}^{N}\mathinner{|{r_{2},j_{\alpha}}\rangle}$ (41) Next, we consider the same edges but starting at the non-marked vertices in the first partition and prepare the following superpositions $\displaystyle\mathinner{|{\nu_{9}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{N-1}}\sum_{j\neq s}^{N}\mathinner{|{j_{1},j_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{10}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{N-1}}\sum_{j\neq s}^{N}\mathinner{|{j_{1},r_{2}}\rangle},$ $\displaystyle\mathinner{|{\nu_{11}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{N-1}\sum_{j\neq s}^{N}\sum_{k\neq r}^{N}\mathinner{|{j_{1},k_{2}}\rangle},$ $\displaystyle\mathinner{|{\nu_{12}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{(d-N)(N-1)}}\sum_{j\neq s}^{N}\sum_{\alpha=3}^{M}\sum_{k=1}^{N}\mathinner{|{j_{1},k_{\alpha}}\rangle}.$ (42) The same procedure is repeated in the second partition $\displaystyle\mathinner{|{\nu_{13}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{N-1}}\sum_{j\neq r}^{N}\mathinner{|{j_{2},j_{2}}\rangle},$ $\displaystyle\mathinner{|{\nu_{14}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{N-1}}\sum_{j\neq r}^{N}\mathinner{|{j_{2},s_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{15}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{N-1}\sum_{j\neq r}^{N}\sum_{k\neq s}^{N}\mathinner{|{j_{2},k_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{16}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{(d-N)(N-1)}}\sum_{j\neq r}^{N}\sum_{\alpha=3}^{M}\sum_{k=1}^{N}\mathinner{|{j_{2},k_{\alpha}}\rangle}.$ (43) Next, we consider the edges leading to the sender or the receiver vertex from the outside of the first two partitions $\displaystyle\mathinner{|{\nu_{17}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d-N}}\sum_{\alpha=3}^{M}\sum_{j=1}^{N}\mathinner{|{j_{\alpha},s_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{18}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d-N}}\sum_{\alpha=3}^{M}\sum_{j=1}^{N}\mathinner{|{j_{\alpha},r_{2}}\rangle}.$ (44) Then we add states corresponding to edges leading to the non-marked vertices in the first or the second partition from the outside $\displaystyle\mathinner{|{\nu_{19}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{(d-N)(N-1)}}\sum_{j\neq s}^{N}\sum_{\alpha=3}^{M}\sum_{k=1}^{N}\mathinner{|{k_{\alpha},j_{1}}\rangle},$ $\displaystyle\mathinner{|{\nu_{20}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{(d-N)(N-1)}}\sum_{j\neq r}^{N}\sum_{\alpha=3}^{M}\sum_{k=1}^{N}\mathinner{|{k_{\alpha},j_{2}}\rangle}.$ (45) Finally, we consider all edges between vertices in the rest of the graph, and all remaining loops $\displaystyle\mathinner{|{\nu_{21}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{N(d-N)(M-3)}}\sum_{\alpha=3}^{M}\sum_{\beta=3,\beta\neq\alpha}^{M}\sum_{j,k=1}^{N}\mathinner{|{j_{\alpha},k_{\beta}}\rangle},$ $\displaystyle\mathinner{|{\nu_{22}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{d-N}}\sum_{\alpha=3}^{M}\sum_{j=1}^{N}\mathinner{|{j_{\alpha},j_{\alpha}}\rangle}$ (46) It can be shown by direct calculation that the 22 vectors (40)-(46) constitute an invariant subspace of the STA. To proceed further we employ the symmetry $\hat{P}$ which switches the sender and the receiver partitions. Its action on the basis states $\mathinner{|{\nu_{j}}\rangle}$ is given by $\displaystyle\hat{P}\mathinner{|{\nu_{j}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{\nu_{j+4}}\rangle},\quad j=1,2,3,4,9,10,11,12,$ $\displaystyle\hat{P}\mathinner{|{\nu_{i}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{\nu_{i+1}}\rangle},\quad i=17,19,$ $\displaystyle\hat{P}\mathinner{|{\nu_{k}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{\nu_{k}}\rangle},\quad k=21,22.$ Since $\hat{P}$ commutes with the evolution operator of the STA we can split $\cal I$ into subspaces ${\cal I}_{\pm}$ corresponding to eigenvalues $\pm 1$. Subspace ${\cal I}_{+}$ has dimension $12$ and it is spanned by the following eigenvectors of $\hat{P}$ $\displaystyle\mathinner{|{\sigma_{i}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{i}}\rangle}+\mathinner{|{\nu_{i+4}}\rangle}\right),\quad i=1,2,3,4$ $\displaystyle\mathinner{|{\sigma_{j}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{j+4}}\rangle}+\mathinner{|{\nu_{j+8}}\rangle}\right),\quad j=5,6,7,8$ $\displaystyle\mathinner{|{\sigma_{9}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{17}}\rangle}+\mathinner{|{\nu_{18}}\rangle}\right),$ $\displaystyle\mathinner{|{\sigma_{10}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{19}}\rangle}+\mathinner{|{\nu_{20}}\rangle}\right),$ $\displaystyle\mathinner{|{\sigma_{11}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{\nu_{21}}\rangle},$ $\displaystyle\mathinner{|{\sigma_{12}}\rangle}$ $\displaystyle=$ $\displaystyle\mathinner{|{\nu_{22}}\rangle}$ Subspace ${\cal I}_{-}$ has dimension $10$ and it is spanned by the following eigenvectors of $\hat{P}$ corresponding to the eigenvalue $-1$ $\displaystyle\mathinner{|{\tau_{i}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{i}}\rangle}-\mathinner{|{\nu_{i+4}}\rangle}\right),\quad i=1,2,3,4$ $\displaystyle\mathinner{|{\tau_{j}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{j+4}}\rangle}-\mathinner{|{\nu_{j+8}}\rangle}\right),\quad j=5,6,7,8$ $\displaystyle\mathinner{|{\tau_{9}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{17}}\rangle}-\mathinner{|{\nu_{18}}\rangle}\right),$ $\displaystyle\mathinner{|{\tau_{10}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\nu_{19}}\rangle}-\mathinner{|{\nu_{20}}\rangle}\right)$ In the new basis the sender and the receiver states have the following form $\displaystyle\mathinner{|{s}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\sigma_{1}}\rangle}+\mathinner{|{\tau_{1}}\rangle}\right),$ $\displaystyle\mathinner{|{r}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{2}}\left(\mathinner{|{\sigma_{1}}\rangle}-\mathinner{|{\tau_{1}}\rangle}\right).$ (47) The evolution operator $\hat{U}_{s,r}$ is block diagonal. We find the following relations for the basis vectors of ${\cal I}_{+}$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(\left(d-1\right)\mathinner{|{\sigma_{1}}\rangle}-2\mathinner{|{\sigma_{2}}\rangle}-2\sqrt{N-1}\mathinner{|{\sigma_{6}}\rangle}-2\sqrt{d-N}\mathinner{|{\sigma_{9}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{2}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(-2\mathinner{|{\sigma_{1}}\rangle}+\left(d-1\right)\mathinner{|{\sigma_{2}}\rangle}-2\sqrt{N-1}\mathinner{|{\sigma_{6}}\rangle}-2\sqrt{d-N}\mathinner{|{\sigma_{9}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{3}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(-2\sqrt{N-1}\mathinner{|{\sigma_{1}}\rangle}-2\sqrt{N-1}\mathinner{|{\sigma_{2}}\rangle}+\left(d-2N+3\right)\mathinner{|{\sigma_{6}}\rangle}-2\sqrt{(d-N)(N-1)}\mathinner{|{\sigma_{9}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{4}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(-2\sqrt{d-N}\mathinner{|{\sigma_{1}}\rangle}-2\sqrt{d-N}\mathinner{|{\sigma_{2}}\rangle}-2\sqrt{(d-N)(N-1)}\mathinner{|{\sigma_{6}}\rangle}-\left(d-2N-1\right)\mathinner{|{\sigma_{9}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{5}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\mathinner{|{\sigma_{3}}\rangle}-\left(d-1\right)\mathinner{|{\sigma_{5}}\rangle}+2\sqrt{N-1}\mathinner{|{\sigma_{7}}\rangle}+2\sqrt{d-N}\mathinner{|{\sigma_{10}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{6}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(-\left(d-1\right)\mathinner{|{\sigma_{3}}\rangle}+2\mathinner{|{\sigma_{5}}\rangle}+2\sqrt{N-1}\mathinner{|{\sigma_{7}}\rangle}+2\sqrt{d-N}\mathinner{|{\sigma_{10}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{7}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\sqrt{N-1}\mathinner{|{\sigma_{3}}\rangle}+2\sqrt{N-1}\mathinner{|{\sigma_{5}}\rangle}-\left(d-2N+3\right)\mathinner{|{\sigma_{7}}\rangle}+2\sqrt{(d-N)(N-1)}\mathinner{|{\sigma_{10}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{8}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\sqrt{d-N}\mathinner{|{\sigma_{3}}\rangle}+2\sqrt{d-N}\mathinner{|{\sigma_{5}}\rangle}+2\sqrt{(d-N)(N-1)}\mathinner{|{\sigma_{7}}\rangle}+\left(d-2N-1\right)\mathinner{|{\sigma_{10}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{9}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(-\left(d-3\right)\mathinner{|{\sigma_{4}}\rangle}+4\sqrt{N-1}\mathinner{|{\sigma_{8}}\rangle}+2\sqrt{2N(M-3)}\mathinner{|{\sigma_{11}}\rangle}+2\sqrt{2}\mathinner{|{\sigma_{12}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{10}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(4\sqrt{N-1}\mathinner{|{\sigma_{4}}\rangle}-\left(d-4N+5\right)\mathinner{|{\sigma_{8}}\rangle}+2\sqrt{2N(N-1)(M-3)}\mathinner{|{\sigma_{11}}\rangle}+2\sqrt{2(N-1)}\mathinner{|{\sigma_{12}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{11}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\sqrt{2N(M-3)}\mathinner{|{\sigma_{4}}\rangle}+2\sqrt{2N(N-1)(M-3)}\mathinner{|{\sigma_{8}}\rangle}+\left(d-4N-1\right)\mathinner{|{\sigma_{11}}\rangle}+2\sqrt{N(M-3)}\mathinner{|{\sigma_{12}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\sigma_{12}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(2\sqrt{2}\mathinner{|{\sigma_{4}}\rangle}+2\sqrt{2(N-1)}\mathinner{|{\sigma_{8}}\rangle}+2\sqrt{N(M-3)}\mathinner{|{\sigma_{11}}\rangle}-\left(d-1\right)\mathinner{|{\sigma_{12}}\rangle}\right).$ The action of the evolution operator on the basis vectors of ${\cal I}_{-}$ reads $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(\left(d-1\right)\mathinner{|{\tau_{1}}\rangle}+2\mathinner{|{\tau_{2}}\rangle}+2\sqrt{N-1}\mathinner{|{\tau_{6}}\rangle}-2\sqrt{d-N}\mathinner{|{\tau_{9}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{2}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(-2\mathinner{|{\tau_{1}}\rangle}-\left(d-1\right)\mathinner{|{\tau_{2}}\rangle}+2\sqrt{N-1}\mathinner{|{\tau_{6}}\rangle}-2\sqrt{d-N}\mathinner{|{\tau_{9}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{3}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(-2\sqrt{N-1}\mathinner{|{\tau_{1}}\rangle}+2\sqrt{N-1}\mathinner{|{\tau_{2}}\rangle}-\left(d-2N+3\right)\mathinner{|{\tau_{6}}\rangle}-2\sqrt{(d-N)(N-1)}\mathinner{|{\tau_{9}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{4}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(-2\sqrt{d-N}\mathinner{|{\tau_{1}}\rangle}+2\sqrt{d-N}\mathinner{|{\tau_{2}}\rangle}+2\sqrt{(d-N)(N-1)}\mathinner{|{\tau_{6}}\rangle}-\left(d-2N-1\right)\mathinner{|{\tau_{9}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{5}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(-2\mathinner{|{\tau_{3}}\rangle}-\left(d-1\right)\mathinner{|{\tau_{5}}\rangle}-2\sqrt{N-1}\mathinner{|{\tau_{7}}\rangle}+2\sqrt{d-N}\mathinner{|{\tau_{10}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{6}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(\left(d-1\right)\mathinner{|{\tau_{3}}\rangle}+2\mathinner{|{\tau_{5}}\rangle}-2\sqrt{N-1}\mathinner{|{\tau_{7}}\rangle}+2\sqrt{d-N}\mathinner{|{\tau_{10}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{7}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(-2\sqrt{N-1}\mathinner{|{\tau_{3}}\rangle}+2\sqrt{N-1}\mathinner{|{\tau_{5}}\rangle}+\left(d-2N+3\right)\mathinner{|{\tau_{7}}\rangle}+2\sqrt{(d-N)(N-1)}\mathinner{|{\tau_{10}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{8}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{d+1}\left(-2\sqrt{d-N}\mathinner{|{\tau_{3}}\rangle}+2\sqrt{d-N}\mathinner{|{\tau_{5}}\rangle}-2\sqrt{(d-N)(N-1)}\mathinner{|{\tau_{7}}\rangle}+\left(d-2N-1\right)\mathinner{|{\tau_{10}}\rangle}\right),$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{9}}\rangle}$ $\displaystyle=$ $\displaystyle-\mathinner{|{\tau_{4}}\rangle},$ $\displaystyle\hat{U}_{s,r}\mathinner{|{\tau_{10}}\rangle}$ $\displaystyle=$ $\displaystyle-\mathinner{|{\tau_{8}}\rangle}.$ (48) To investigate the dynamics of STA in more detail we again turn to the limit of a large graph. We denote by $\hat{U}_{\pm}$ the restriction of $\hat{U}_{s,r}$ on ${\cal I}_{\pm}$. For $\hat{U}_{+}$ there are three eigenstates which have non-vanishing overlap with the state $\mathinner{|{\sigma_{1}}\rangle}$, namely $\displaystyle\mathinner{|{\psi_{1}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{\sqrt{3}}{2}\mathinner{|{\sigma_{1}}\rangle}-\frac{1}{2\sqrt{3}}\mathinner{|{\sigma_{2}}\rangle}-\frac{1}{\sqrt{6}}\mathinner{|{\sigma_{11}}\rangle},$ (49) $\displaystyle\mathinner{|{\psi_{2}^{(\pm)}}\rangle}$ $\displaystyle=$ $\displaystyle\frac{1}{\sqrt{8}}(\mathinner{|{\sigma_{1}}\rangle}+\mathinner{|{\sigma_{2}}\rangle})\pm\frac{i}{2}(\mathinner{|{\sigma_{9}}\rangle}-\mathinner{|{\sigma_{4}}\rangle})+\frac{1}{2}\mathinner{|{\sigma_{11}}\rangle}.$ Note that for the other eigenstates the overlap with $\mathinner{|{\sigma_{1}}\rangle}$ decreases at least as $O(1/\sqrt{NM})$. Turning to the eigenvalues, we find that the eigenvector $\mathinner{|{\psi_{1}}\rangle}$ has eigenvalue $\lambda_{1}=1$. From the characteristic polynomial of $\hat{U}_{+}$ we find that the eigenvalues of $\mathinner{|{\psi_{2}^{(\pm)}}\rangle}$ have the form $\lambda_{2}^{(\pm)}=e^{\pm i\omega_{2}}$, where $\cos\omega_{2}$ is the largest root of the cubic equation $\displaystyle 0$ $\displaystyle=$ $\displaystyle x^{3}-\left(1-\frac{N+2}{d+1}\right)x^{2}-\frac{{(d+1)}(N-1)-N+5}{{(d+1)}^{2}}x+$ $\displaystyle+\frac{NM-4}{{(d+1)}^{2}}.$ We find that it has the following asymptotic form $\omega_{2}\approx\arccos\left(1-\frac{4}{NM}\right)\approx 2\sqrt{\frac{2}{NM}}.$ (50) Considering the subspace ${\cal I}_{-}$, there are two eigenvectors which remain relevant in the asymptotic limit (for the others the overlap with $\mathinner{|{\tau_{1}}\rangle}$ vanishes at least as $O(1/\sqrt{NM})$), namely $\mathinner{|{\psi_{3}^{(\pm)}}\rangle}=\frac{1}{\sqrt{2}}\mathinner{|{\tau_{1}}\rangle}\pm\frac{i}{2}(\mathinner{|{\tau_{9}}\rangle}-\mathinner{|{\tau_{4}}\rangle}).$ (51) The eigenvalues are $\lambda_{3}^{(\pm)}=e^{\pm i\omega_{3}}$, where $\cos\omega_{3}$ is the largest root of the quartic equation $\displaystyle 0$ $\displaystyle=$ $\displaystyle x^{4}+\frac{N-2}{d+1}x^{3}-\left(1-\frac{NM}{{(d+1)}^{2}}\right)x^{2}-$ $\displaystyle-\frac{(N-2)(d-1)}{{(d+1)}^{2}}x+\frac{N(M-2)}{{(d+1)}^{2}}.$ Its asymptotic form is given by $\omega_{3}\approx\arccos\left(1-\frac{1}{NM}\right)\approx\sqrt{\frac{2}{NM}}.$ (52) From (49), (51) we see that for a large graph the sender and the receiver states can be decomposed into the eigenvectors of the evolution operator $\hat{U}_{s,r}$ according to $\displaystyle\mathinner{|{s}\rangle}$ $\displaystyle=$ $\displaystyle\sqrt{\frac{3}{8}}\mathinner{|{\psi_{1}}\rangle}+\frac{1}{4}\left(\mathinner{|{\psi_{2}^{(+)}}\rangle}+\mathinner{|{\psi_{2}^{(-)}}\rangle}\right)+$ $\displaystyle+\frac{1}{2}\left(\mathinner{|{\psi_{3}^{(+)}}\rangle}+\mathinner{|{\psi_{3}^{(-)}}\rangle}\right),$ $\displaystyle\mathinner{|{r}\rangle}$ $\displaystyle=$ $\displaystyle\sqrt{\frac{3}{8}}\mathinner{|{\psi_{1}}\rangle}+\frac{1}{4}\left(\mathinner{|{\psi_{2}^{(+)}}\rangle}+\mathinner{|{\psi_{2}^{(-)}}\rangle}\right)-$ $\displaystyle-\frac{1}{2}\left(\mathinner{|{\psi_{3}^{(+)}}\rangle}+\mathinner{|{\psi_{3}^{(-)}}\rangle}\right).$ The evolution of STA takes place in a five dimensional subspace $\displaystyle\mathinner{|{\phi(t)}\rangle}$ $\displaystyle=$ $\displaystyle\sqrt{\frac{3}{8}}\mathinner{|{\psi_{1}}\rangle}+$ (53) $\displaystyle+\frac{1}{4}\left(e^{i\omega_{2}t}\mathinner{|{\psi_{2}^{(+)}}\rangle}+e^{-i\omega_{2}t}\mathinner{|{\psi_{2}^{(-)}}\rangle}\right)+$ $\displaystyle+\frac{1}{2}\left(e^{i\omega_{3}t}\mathinner{|{\psi_{3}^{(+)}}\rangle}+e^{-i\omega_{3}t}\mathinner{|{\psi_{3}^{(-)}}\rangle}\right).$ The fidelity of STA after $t$ steps (9) can be expressed as a sum ${\cal F}(t)=\sum_{j=5}^{8}|\langle\nu_{j}|\phi(t)\rangle|^{2}.$ (54) From (49), (51) and (53) we find that the transfer probabilities to individual states $\mathinner{|{\nu_{j}}\rangle}$ are given by $\displaystyle|\langle\nu_{5}|\phi(t)\rangle|^{2}$ $\displaystyle=$ $\displaystyle\frac{1}{64}\left(3+\cos\left(\omega_{2}t\right)-4\cos\left(\omega_{3}t\right)\right)^{2},$ $\displaystyle|\langle\nu_{6}|\phi(t)\rangle|^{2}$ $\displaystyle=$ $\displaystyle\frac{1}{16}\sin^{4}\left(\frac{\omega_{2}t}{2}\right),$ $\displaystyle|\langle\nu_{7}|\phi(t)\rangle|^{2}$ $\displaystyle=$ $\displaystyle 0,$ $\displaystyle|\langle\nu_{8}|\phi(t)\rangle|^{2}$ $\displaystyle=$ $\displaystyle\frac{1}{32}\left(\sin(\omega_{2}t)-2\sin(\omega_{3}t)\right)^{2}.$ (55) From the asymptotic expansions (50), (52) we see that for a large graph the frequencies are harmonic $\omega_{2}=2\omega_{3}.$ Hence, we find that with probability $|\langle r|\phi(t)\rangle|^{2}=\sin^{8}{\left(\frac{\omega_{3}t}{2}\right)},$ (56) the walker is in the receiver state $\mathinner{|{r}\rangle}=\mathinner{|{\nu_{5}}\rangle}$, i.e., at the receiver vertex in the loop, and with probability $\displaystyle|\langle\nu_{6}|\phi(t)\rangle|^{2}+|\langle\nu_{8}|\phi(t)\rangle|^{2}$ $\displaystyle=$ $\displaystyle\frac{1}{2}\sin^{2}\left(\omega_{3}t\right)\sin^{4}\left(\frac{\omega_{3}t}{2}\right)+$ (57) $\displaystyle+\frac{1}{16}\sin^{4}\left(\omega_{3}t\right),$ it is at the receiver vertex but not in the loop. Overall, the fidelity of STA for a large graph is given by ${\cal F}(t)=\sin^{4}\left(\frac{\omega_{3}t}{2}\right).$ (58) We conclude that the state transfer is achieved after $T^{(st)}$ steps, where $T^{(st)}\approx\pi\sqrt{\frac{NM}{2}}.$ (59) At this time the walker is with high probability in the receiver state $\mathinner{|{r}\rangle}$. For illustration we show in Figure 6 the evolution of fidelity for a graph with $N=40$ and $M=100$. Figure 6: Overall fidelity of the STA as a function of the number of steps $t$ for $N=40$ and $M=100$. The sender and the receiver vertices are in different partitions. Black dots are obtained from the numerical simulation, the full red curve corresponds to (58). The walker is transferred to the receiver vertex with fidelity close to one after $T^{(st)}\approx 140$ steps, in accordance with (59). At this time the walker is found with high probability in the loop (blue squares), as follows from the analytical prediction (56) depicted by the red dashed curve. Green diamonds represent the probability that the walker is at the marked vertex but not in the loop, which follows the curve (57). The result (58) holds in the limit of large $N$ and $M$. To investigate how quickly does the fidelity at the optimal time (59) approaches unity we performed numerical simulations for various values of $N$ and $M$. The simulations indicate that the fidelity can be again estimated by ${\cal F}(T^{(st)})=1-O\left(\frac{1}{M}\right),$ however, the dependence on $N$ is more complex than for search and STA with the sender and the receiver in the same partition. The results are illustrated in Figure 7. Figure 7: Overall fidelity of STA as a function of the number of partitions $M$ for $N=10$ (gray circles), $N=50$ (blue triangles) and $N=100$ (brown diamonds). The sender and the receiver vertices are in different partitions. For a given $N$ and $M$ we evaluate numerically the evolution of STA for the optimal number of steps $T^{(st)}$ given by (59) and determine ${\cal F}(T^{(st)})$ from the formula (54). To unravel the scaling of the fidelity we plot $1-{\cal F}(T^{(st)})$ on the log-log scale. The full red line has the $1/M$ slope, while the dashed red line follows $1/M^{2}$. The plot indicates that $O(1/M)>1-{\cal F}(T^{(st)})>O(1/M^{2})$, and that the fluctuations decrease with increasing $N$. ## V State transfer algorithm with an active switch As we have shown in the previous section, the STA does not perform with unit fidelity on the complete $M$-partite graph with loops when the sender and the receiver are in the same partition. Note that if the receiver does not know the position of the sender, the measurement should be made at the optimal time (59) corresponding to the more likely configuration, i.e., when the sender and the receiver are in different partition. This reduces the fidelity of STA further to approximately 0.91. To fix this issue we introduce an STA where the sender and the receiver will actively switch the local coins at their vertices. We use that for $M\to\infty$ and $N\to\infty$ the state of the SA (19) on the complete $M$-partite graph with loops evolves periodically from the initial state $\mathinner{|{\Omega}\rangle}$ to the target state $\mathinner{|{\nu_{1}}\rangle}$ and back with a period of $2T$, where $T$ is the run-time of the SA given by (24). Hence, we can perform the state transfer in the following way. The sender initializes the walk on its vertex in the target state of the search algorithm $\mathinner{|{s}\rangle}=\mathinner{|{\nu_{1}}\rangle}$ which corresponds to the loop at the sender vertex. For the first $T$ steps only the sender will use the marked coin, i.e., the walk will evolve according to the operator $\hat{U}_{s}$ of the SA with the marked vertex $s$. The sender state $\mathinner{|{s}\rangle}$ will evolve close to the equal weight superposition $\mathinner{|{\Omega}\rangle}$, i.e., the initial state of the SA (6). Afterwards, the sender switches off the marked coin, and the receiver switches it on, i.e., the walk evolves according to the operator $\hat{U}_{r}$ of the SA with the marked vertex $r$. After another $T$ steps the walk will evolve close to the state $\mathinner{|{r}\rangle}$ corresponding to the loop at the receiver vertex, and the receiver will detect it with high probability. In this way we can achieve state transfer with high fidelity on the complete $M$-partite graph with loops irrespective of the relative position of the sender and the receiver. For illustration, we show in Figure 8 the comparison of the fidelities of state transfer of the original STA and the STA with an active switch when the sender and the receiver are in the same partition. We see that the STA with an active switch takes more steps, however, the fidelity reaches one. Figure 8: Comparison of fidelity of the original STA (37) (red diamonds) with fidelity of the STA with an active switch (black dots) as a function of the number of steps $t$ for $N=40$ and $M=100$. The sender and the receiver vertices are in the same partition. The switch between $\hat{U}_{s}$ and $\hat{U}_{r}$ is done after $T\approx 100$ steps, corresponding to the run- time of the SA. Let us now formalize the STA with an active switch on more general graphs. Namely, we consider graphs where the optimal number of steps $T$ of the SA does not depend on the position of the marked vertex $m$. The steps of the STA with an active switch can be formulated as follows: 1. 1. Sender initializes the walk at its vertex in the state $\mathinner{|{s}\rangle}$ corresponding to the target state of the SA with the marked vertex $s$. 2. 2. Sender uses marked coin on his vertex for $T$ steps, i.e., the evolution operator $\hat{U}_{s}$ is applied $T$-times. 3. 3. Receiver uses marked coin on his vertex for $T$ steps, i.e., the evolution operator $\hat{U}_{r}$ is applied $T$-times. 4. 4. Receiver measures the walk at its vertex. We show that the fidelity of the STA with an active switch can be lower bounded using only the results from the SA with one marked vertex, which is not true for the original STA. As we have seen in the case of the complete $M$-partite graph with loops, the SA does not tell us anything about the evolution of the STA in the subspace ${\cal I}_{-}$. To derive the lower bound of fidelity of the STA with an active switch we first introduce two conditions on the SA. The first condition is related to target state of the SA. We suppose that after $T$ steps the state of the SA can be expressed in the form $\displaystyle\mathinner{|{\phi(T)}\rangle}=\hat{U}_{m}^{T}\mathinner{|{\Omega}\rangle}=\alpha_{m}\mathinner{|{m}\rangle}+\epsilon_{m}\mathinner{|{\eta_{m}}\rangle}$ (60) for every marked vertex $m$ in the graph. Here $\mathinner{|{m}\rangle}\in{\cal H}_{m}$ is the target state of the SA, i.e., if the walk is in this state the success probability of finding the marked vertex $m$ is exactly $1$, and $\mathinner{|{\eta_{m}}\rangle}$ is a unit vector orthogonal to $\mathinner{|{m}\rangle}$. Complex numbers $\alpha_{m}$ and $\epsilon_{m}$ are such that $|\alpha_{m}|$ is close to one and $|\epsilon_{m}|\ll 1$. $|\alpha_{m}|^{2}$ is closely related to the success probability of the SA, since $\displaystyle P_{m}(T)$ $\displaystyle=$ $\displaystyle|\langle m|\phi(T)\rangle|^{2}+\sum_{\begin{subarray}{c}\mathinner{|{j}\rangle}\in{\cal H}_{m}\\\ \langle j|m\rangle=0\end{subarray}}|\langle j|\phi(T)\rangle|^{2}$ (61) $\displaystyle=$ $\displaystyle|\alpha_{m}|^{2}+|\epsilon_{m}|^{2}\sum_{\begin{subarray}{c}\mathinner{|{j}\rangle}\in{\cal H}_{m}\\\ \langle j|m\rangle=0\end{subarray}}|\langle j|\eta_{m}\rangle|^{2}.$ Note that if the vector $\mathinner{|{\eta_{m}}\rangle}$ does not have a support at the marked vertex then the success probability is exactly $|\alpha_{m}|^{2}$. From the relation (60) for $m=s$ we express the initial state of the STA with an active switch as $\mathinner{|{s}\rangle}=\frac{1}{\alpha_{s}}\left(\hat{U}_{s}^{T}\mathinner{|{\Omega}\rangle}-\epsilon_{s}\mathinner{|{\eta_{s}}\rangle}\right).$ (62) The second condition describes the periodicity of the SA. Namely, we suppose that after $2T$ steps the state of the SA can be written in the form $\displaystyle\hat{U}_{m}^{2T}\mathinner{|{\Omega}\rangle}=\beta_{m}\mathinner{|{\Omega}\rangle}+\delta_{m}\mathinner{|{\rho_{m}}\rangle},$ (63) where $\mathinner{|{\rho_{m}}\rangle}$ is a unit vector orthogonal to the initial state $\mathinner{|{\Omega}\rangle}$. $\beta_{m}$ and $\delta_{m}$ are again complex numbers where $|\beta_{m}|^{2}$ is the return probability. We assume that it is close to one and that $|\delta_{m}|\ll 1$. In other words, this condition says that if we double the number of steps the SA returns close to its initial state. Assuming that the SA satisfies the conditions (60) and (63) we write the final state of the STA with an active switch in the following manner $\displaystyle\hat{U}_{r}^{T}\hat{U}_{s}^{T}|s\rangle$ $\displaystyle=$ $\displaystyle\frac{1}{\alpha_{s}}\hat{U}_{r}^{T}\hat{U}_{s}^{T}\left(\hat{U}_{s}^{T}|\Omega\rangle-\epsilon_{s}|\eta_{s}\rangle\right)$ (64) $\displaystyle=$ $\displaystyle\frac{1}{\alpha_{s}}\hat{U}_{r}^{T}\left(\hat{U}_{s}^{2T}|\Omega\rangle-\epsilon_{s}\hat{U}_{s}^{T}|\eta_{s}\rangle\right)$ $\displaystyle=$ $\displaystyle\frac{1}{\alpha_{s}}\hat{U}_{r}^{T}\left(\beta_{s}|\Omega\rangle+\delta_{s}|\rho_{s}\rangle-\epsilon_{s}\hat{U}_{s}^{T}|\eta_{s}\rangle\right)$ $\displaystyle=$ $\displaystyle\frac{\alpha_{r}}{\alpha_{s}}\beta_{s}|r\rangle+\frac{\beta_{s}}{\alpha_{s}}\epsilon_{r}|\eta_{r}\rangle+\frac{\delta_{s}}{\alpha_{s}}\hat{U}_{r}^{T}|\rho_{s}\rangle-$ $\displaystyle-\frac{\epsilon_{s}}{\alpha_{s}}\hat{U}_{r}^{T}\hat{U}_{s}^{T}|\eta_{s}\rangle$ where we have first used (62), then (63) and finally we again use (62) but for the state $|r\rangle$. The fidelity of STA with an active switch can be expressed as ${\cal F}=\left|\left\langle r\left|\hat{U}_{r}^{T}\hat{U}_{s}^{T}\right|s\right\rangle\right|^{2}+\sum_{\begin{subarray}{c}\mathinner{|{j}\rangle}\in{\cal H}_{r}\\\ \langle j|r\rangle=0\end{subarray}}\left|\left\langle j\left|\hat{U}_{r}^{T}\hat{U}_{s}^{T}\right|s\right\rangle\right|^{2}.$ Hence, the square root of the fidelity can be bounded from below by $\sqrt{\cal F}\geq\left|\left\langle r\left|\hat{U}_{r}^{T}\hat{U}_{s}^{T}\right|s\right\rangle\right|.$ To approximate $\left|\left\langle r\left|\hat{U}_{r}^{T}\hat{U}_{s}^{T}\right|s\right\rangle\right|$ we use the following estimates $\displaystyle\left|\left\langle r\left|\hat{U}_{r}^{T}\right|\rho_{s}\right\rangle\right|$ $\displaystyle\leq$ $\displaystyle|||r\rangle||\left|\left|\hat{U}_{r}^{T}\left|\rho_{s}\right\rangle\right|\right|=|||r\rangle||\left|\left|\left|\rho_{s}\right\rangle\right|\right|=1,$ $\displaystyle\left|\left\langle r\left|\hat{U}_{r}^{T}\hat{U}_{s}^{T}\right|\eta_{s}\right\rangle\right|$ $\displaystyle\leq$ $\displaystyle|||r\rangle||\left|\left|\hat{U}_{r}^{T}\hat{U}_{s}^{T}\left|\eta_{s}\right\rangle\right|\right|=1,$ (65) which follow from the Cauchy–Schwarz inequality and the unitarity of evolution operator $\hat{U}_{m}$. Combining (64) with (65) we find the lower bound for the square root of the fidelity which reads $\displaystyle\sqrt{\cal F}$ $\displaystyle\geq$ $\displaystyle\left|\frac{\alpha_{r}}{\alpha_{s}}\beta_{s}\langle r|r\rangle+\frac{\beta_{s}}{\alpha_{s}}\epsilon_{r}\langle r|\eta_{r}\rangle+\frac{\delta_{s}}{\alpha_{s}}\left\langle r\left|\hat{U}_{r}^{T}\right|\rho_{s}\right\rangle-\frac{\epsilon_{s}}{\alpha_{s}}\left\langle r\left|\hat{U}_{r}^{T}\hat{U}_{s}^{T}\right|\eta_{s}\right\rangle\right|$ (66) $\displaystyle\geq$ $\displaystyle\frac{|\alpha_{r}|}{|\alpha_{s}|}|\beta_{s}|-\frac{|\delta_{s}|}{|\alpha_{s}|}\left|\left\langle r\left|\hat{U}_{r}^{T}\right|\rho_{s}\right\rangle\right|-\frac{|\epsilon_{s}|}{|\alpha_{s}|}\left|\left\langle r\left|\hat{U}_{r}^{T}\hat{U}_{s}^{T}\right|\eta_{s}\right\rangle\right|$ $\displaystyle\geq$ $\displaystyle\frac{|\alpha_{r}|}{|\alpha_{s}|}|\beta_{s}|-\frac{|\delta_{s}|}{|\alpha_{s}|}-\frac{|\epsilon_{s}|}{|\alpha_{s}|}.$ It is easy to see from (66) that if $|\alpha_{s}|$, $|\alpha_{r}|$ and $|\beta_{s}|$ are close to one and if $|\epsilon_{s}|\ll 1$ and $|\delta_{s}|\ll 1$ then the fidelity of the state transfer is close to one. The result derived above guarantees that if the SA succeeds with unit probability in the limit of a large graph, then the STA with an active switch achieves perfect state transfer. Moreover, even for small graphs the STA with an active switch can actually achieve very good fidelity. For illustration we have investigated numerically the STA with an active switch on the complete $M$-partite graph for various values of $N$ and $M$. The results are similar to those presented in Figure 7. Figure 9: Overall fidelity of STA with an active switch as a function of the number of partitions $M$ for $N=10$ (gray circles), $N=50$ (blue triangles) and $N=100$ (brown diamonds). The walk is initialized at the sender vertex in the loop. For a given $N$ and $M$ we evaluate numerically the evolution operator $\hat{U}_{s}$ of SA with marked vertex $s$, apply it for the optimal number of steps $T$ given by (24). Then we repeat the same with $\hat{U}_{r}$. Finally, we make a measurement at the receiver vertex $r$ and determine the fidelity ${\cal F}(2T)$. To unravel the scaling of the fidelity we plot $1-{\cal F}(2T)$ on the log-log scale. The full red line has the $1/M$ slope, while the dashed red line follows $1/M^{2}$. The plot indicates that $O(1/M)>1-{\cal F}(2T)>O(1/M^{2})$. ## VI Conclusion We have investigated search and state transfer algorithms based on the coined quantum walks, focusing on the complete $M$-partite graph. It was shown that adding loops to all vertices increases the success probability of the SA close to one for a large graph. This is by now a standard method [9, 10, 26, 27, 28, 29, 30] which has a potential to significantly improve the success probability on a much broader class of graphs [31]. However, the analysis of the SA does not provide the necessary insight for the investigation of the STA, as the latter requires larger invariant subspace. As we have seen on the example of the complete $M$-partite graph with loops, success probability of the SA close to one does not guarantee STA with unit fidelity. The reason is that the phases of the relevant eigenvalues of the evolution operator are not harmonic when the sender and the receiver are in the same partition. Although the modification of the initial state has improved the fidelity considerably, the absence of an edge between the sender and the receiver vertex on the complete $M$-partite graph does not allow for perfect state transfer. It would be interesting to find out if this occurs for different graphs as well. In the present paper we have limited our investigations to the case where all partitions have the same number of vertices $N$. This enabled us to find exact invariant subspaces for SA and STA which have dimensions independent of $N$ and $M$. Allowing the partitions with different number of vertices appears to break this feature, and the dimension of the invariant subspace is likely to depend on $M$. We plan to investigate this behaviour in the near future for small values of $M$. To improve the STA on the complete $M$-partite graph we have introduced the STA with an active switch. This approach allows for perfect state transfer in the limit of a large graph in both configurations of the sender and the receiver vertex. The trade-off is that the STA with an active switch requires more steps than the original STA. Indeed, when the sender and the receiver are in different partitions, the number of steps for the original STA to reach unit fidelity is twice the number of steps of the search for two vertices. On the other hand, STA with an active switch takes twice the number of steps of the SA for one vertex. Since the search for two vertices is $\sqrt{2}$ faster than the search for one vertex, the STA with an active switch is slower by the same factor. The main advantages of the STA with an active switch are that it can be applied to other graphs, and that its fidelity can be estimated based on the analysis of the SA for one marked vertex alone. In this way we can achieve state transfer with high fidelity on graphs, where the SA for one marked vertex has success probability close to one and evolves almost periodically. For many symmetric graphs these conditions are well satisfied, at least in the limit of a large graph. Moreover, exact periodicity of the Grover walk (i.e., without the marked vertex) was recently investigated for various graphs [50, 51, 52]. It would be of interest to determine if SA works on these graphs as well. The STA with an active switch also has some disadvantages. As we have already mentioned, it will have a longer run-time in comparison with the original STA. Moreover, the sender and the receiver have to actively switch off or on their marked coins. However, this is only a local operations, and since we consider that the run-time of the SA is independent of the location of the marked vertex, the time of the switching is determined solely by the global properties of the graph, e.g., the number of vertices. Hence, the sender and the receiver still do not need to know each other’s position. Finally, we have to determine the target state of SA, which serves as the initial state for the STA with an active switch. Nevertheless, for highly symmetric graphs this target state is usually either the equal weight superposition of all direction or the state corresponding to a loop. ###### Acknowledgements. Both authors received support from the Czech Grant Agency (GAČR) under grant No. 17-00844S and from MSMT RVO 14000. This publication was funded by the project “Centre for Advanced Applied Sciences”, Registry No. CZ.$02.1.01/0.0/0.0/16\\_019/0000778$, supported by the Operational Programme Research, Development and Education, co-financed by the European Structural and Investment Funds and the state budget of the Czech Republic. ## References * Aharonov _et al._ [1993] Y. Aharonov, L. Davidovich, and N. Zagury, Phys. Rev. A 48, 1687 (1993). * Meyer [1996] D. A. Meyer, J. Stat. Phys. 85, 551 (1996). * Farhi and Gutmann [1998] E. Farhi and S. Gutmann, Phys. Rev. A 58, 915 (1998). * Aharonov _et al._ [2001] D. Aharonov, A. Ambainis, J. Kempe, and U. Vazirani, in Proceedings of the 33rd STOC (ACM Press, New York, 2001) , 50 (2001). * Grover [1997] L. K. Grover, Phys. Rev. Lett. 79, 325 (1997). * Childs and Goldstone [2004] A. M. Childs and J. Goldstone, Phy. Rev. A 70, 022314 (2004). * Ambainis _et al._ [2001] A. Ambainis, E. Bach, A. Nayak, A. Vishwanath, and J. Watrous, in _STOC’01: Proc. of the 33rd annual ACM Symposium on Theory of Computing_ (2001) pp. 37–49. * Shenvi _et al._ [2003] N. Shenvi, J. Kempe, and K. B. Whaley, Phys. Rev. A 67, 052307 (2003). * Ambainis _et al._ [2005] A. Ambainis, J. Kempe, and A. Rivosh, in _Proc. 16th ACM-SIAM symposium on Discrete algorithms_ (Society for Industrial and Applied Mathematics, 2005) pp. 1099–1108. * Potoček _et al._ [2009] V. Potoček, A. Gábris, T. Kiss, and I. Jex, Phys. Rev. A 79, 012325 (2009). * Hein and Tanner [2009a] B. Hein and G. Tanner, J. Phys. A 42, 085303 (2009a). * Hillery _et al._ [2003] M. Hillery, J. Bergou, and E. Feldman, Phys. Rev. A 68, 032314 (2003). * Feldman and Hillery [2004] E. Feldman and M. Hillery, Phys. Lett. A 324, 277 (2004). * Feldman and Hillery [2007] E. Feldman and M. Hillery, J. Phys. A: Math. Theor. 40, 11343 (2007). * Andrade and da Luz [2009] F. M. Andrade and M. G. E. da Luz, Phys. Rev. A 80, 052301 (2009). * Venancio _et al._ [2013] B. F. Venancio, F. M. Andrade, and M. G. E. da Luz, J. Phys. A 46, 165302 (2013). * Reitzner _et al._ [2009] D. Reitzner, M. Hillery, E. Feldman, and V. Bužek, Phys. Rev. A 79, 012323 (2009). * Janmark _et al._ [2014] J. Janmark, D. A. Meyer, and T. G. Wong, Phys. Rev. Lett. 112, 210502 (2014). * Novo _et al._ [2015] L. Novo, S. Chakraborty, M. Mohseni, H. Neven, and Y. Omar, Sci. Rep. 5, 13304 (2015). * Meyer and Wong [2015] D. A. Meyer and T. G. Wong, Phys. Rev. Lett. 114, 110503 (2015). * Chakraborty _et al._ [2016] S. Chakraborty, L. Novo, A. Ambainis, and Y. Omar, Phys. Rev. Lett. 116, 100501 (2016). * Albert and Barabási [2002] R. Albert and A.-L. Barabási, Rev. Mod. Phys. 74, 47 (2002). * Osada _et al._ [2020] T. Osada, B. Coutinho, Y. Omar, K. Sanaka, W. J. Munro, and K. Nemoto, Phys. Rev. A 101, 022310 (2020). * Chakraborty _et al._ [2020] S. Chakraborty, L. Novo, and J. Roland, Phys. Rev. A 102, 032214 (2020). * Brassard _et al._ [2002] G. Brassard, P. Høyer, M. Mosca, and A. Tapp, Contemp. Math. 350, 53 (2002). * Wong [2015] T. G. Wong, J. Phys. A 48, 435304 (2015). * Wong [2018] T. G. Wong, Quantum Inf. Process. 17, 68 (2018). * Rhodes and Wong [2019] M. L. Rhodes and T. G. Wong, Phys. Rev. A 100, 042303 (2019). * Rhodes and Wong [2020] M. L. Rhodes and T. G. Wong, Quantum Inf. Process. 19, 334 (2020). * Chiang [2020] C. Chiang, Quantum Inf. Process. 19, 364 (2020). * Høyer and Zhan [2020] P. Høyer and Y. Zhan, arXiv:2002.11234 (2020). * Bose [2003] S. Bose, Phys. Rev. Lett. 91, 207901 (2003). * Kurzynski and Wojcik [2011] P. Kurzynski and A. Wojcik, Phys. Rev. A 83, 062315 (2011). * Yalcinkaya and Gedik [2015] I. Yalcinkaya and Z. Gedik, J. Phys. A 48, 225302 (2015). * Zhan _et al._ [2014] X. Zhan, H. Qin, Z. H. Bian, J. Li, and P. Xue, Phys. Rev. A 90, 012331 (2014). * Shang _et al._ [2018] Y. Shang, Y. Wang, M. Li, and R. Q. Lu, EPL 124, 60009 (2018). * Chen _et al._ [2019] X. B. Chen, Y. L. Wang, G. Xu, and Y. X. Yang, IEEE ACCESS 7, 13634 (2019). * Hein and Tanner [2009b] B. Hein and G. Tanner, Phys. Rev. Lett. 103, 260501 (2009b). * Kendon and Tamon [2011] V. M. Kendon and C. Tamon, J. Comput. Theor. Nanosci. 8, 422 (2011). * Barr _et al._ [2014] K. Barr, T. Proctor, D. Allen, and V. M. Kendon, Quantum Inf. Comput. 14, 417 (2014). * Štefaňák and Skoupý [2016] M. Štefaňák and S. Skoupý, Phys. Rev. A 94, 022301 (2016). * Štefaňák and Skoupý [2017] M. Štefaňák and S. Skoupý, Quantum Inf. Process. 16, 72 (2017). * Zhan [2019] H. Zhan, Quantum Inf. Process. 18, 369 (2019). * Cao _et al._ [2019] W. F. Cao, Y. G. Yang, D. Li, J. R. Dong, Y. H. Zhou, and W. M. Shi, Mod. Phys. Lett. A 34, 1950317 (2019). * Reitzner _et al._ [2017] D. Reitzner, M. Hillery, and D. Koch, Phys. Rev. A 96, 032323 (2017). * Koch and Hillery [2018] D. Koch and M. Hillery, Phys. Rev. A 97, 012308 (2018). * Krovi and Brun [2007] H. Krovi and T. A. Brun, Phys. Rev. A 75, 062332 (2007). * Feldman _et al._ [2010] E. Feldman, M. Hillery, H.-W. Lee, D. Reitzner, H. Zheng, and V. Bužek, Phys. Rev. A 82, 040301 (2010). * Hillery _et al._ [2012] M. Hillery, H. Zheng, E. Feldman, D. Reitzner, and V. Bužek, Phys. Rev. A 85, 062325 (2012). * Kubota _et al._ [2018] S. Kubota, E. Segawa, T. Taniguchi, and Y. Yoshie, Linear Algebra Appl. 554, 371 (2018). * Kubota _et al._ [2019] S. Kubota, E. Segawa, T. Taniguchi, and Y. Yoshie, Linear Algebra Appl. 579, 217 (2019). * Yoshie [2019] Y. Yoshie, Graphs Combin. 35, 1305 (2019).
# Reflections on reflections Kuan-Nan Lin<EMAIL_ADDRESS>Department of Physics and Center for Theoretical Sciences, National Taiwan University, Taipei 10617, Taiwan, R.O.C. LeCosPA, National Taiwan University, Taipei 10617, Taiwan, R.O.C. Pisin Chen <EMAIL_ADDRESS>Department of Physics and Center for Theoretical Sciences, National Taiwan University, Taipei 10617, Taiwan, R.O.C. LeCosPA, National Taiwan University, Taipei 10617, Taiwan, R.O.C. Kavli Institute for Particle Astrophysics and Cosmology, SLAC National Accelerator Laboratory, Stanford University, Stanford, California 94305, U.S.A. ###### Abstract Analog Hawking radiation emitted by a perfectly reflecting mirror in (1+3)-dimensional flat spacetime is investigated. This is accomplished by studying the reflected frequency and momentum based on Einstein’s mirror, instead of the canonical way of solving, if possible, wave equations subjected to a dynamical Dirichlet boundary condition. In the case of a finite-size mirror, diffraction pattern appears in the radiation spectrum. Based on the relevant parameters in the proposed AnaBHEL experiment, where the Hawking temperature $T_{H}=0.03$ eV and the mirror area $\mathcal{A}\sim 0.1$ mm2, the Hawking photon yield is estimated to be $N\sim 16$ per laser shot (assuming a high reflectivity mirror can be generated in the proposed AnaBHEL experiment). ## I Introduction Since the discovery of Hawking radiation [1] in curved spacetime in the 1970s, most of the works of mimicking Hawking radiation via mirror-induced radiation (MIR) in flat spacetime have been devoted to (1+1) dimensions [2, 3, 4, 5, 6, 7, 8, 9, 10, 7, 8, 11]. This is because Hawking radiation only propagates radially in the spacetime with a black hole. However, what is experienced in a laboratory is a (1+3)-dimensional flat spacetime, and these additional degrees of freedom complicate the situation of how the vacuum fluctuations interact with a non-point-like mirror. The intent of this paper is to fill in this gap in the literature. The essence of Hawking radiation in curved spacetime is the gravitational red shift of the wave mode that propagates to the future infinity. In the flying mirror case, red shift of wave mode is accomplished by the Doppler effect, which is induced upon reflection off the mirror, instead. Thus, it is essential to study how wave modes are reflected by a flying mirror. The canonical way of dealing with the moving mirror model is to impose, say, a dynamical Dirichlet boundary condition on a scalar field, whose excitation leads to mirror-induced radiation. This can be exactly solved in (1+1) dimensions since everything is conformally flat in (1+1)D. However, to our awareness, there is no general exact way to solve the problem in spacetime dimension other than (1+1) when the mirror is relativistic. In (1+3)D, however, the case of a non-relativistic perfectly reflecting plane mirror [12, 13, 14] or a relativistic semitransparent plane mirror [15, 16] have been worked out. The necessity of extensions beyond the standard mirror-black hole correspondence is required by laboratory considerations. In particular, the recent AnaBHEL (Analog Black Hole Evaporation via Lasers) Collaboration [17] proposed to generate a flying mirror based on the Chen-Mourou [18, 19] proposal, in which the mirror would have low reflectivity and finite area and thickness, to investigate analog Hawking radiation and the quantum entanglement with its partner. In this paper, instead of working with the conventional approach, we begin by dealing with the reflected frequency and momentum, which can be derived according to Einstein’s special theory of relativity [20], and, from which, we are able to obtain the corresponding phase of the reflected wave mode, and finally obtain the quantum particle spectrum in higher spacetime dimensions by taking the reflected wave mode’s Fourier component. In fact, this approach to study particle creation is quite universal, i.e., it also applies to particle creation in other contexts, e.g., typical QFT in flat/curved spacetime, Unruh effect, MIR for perfect/semitransparent mirror, etc. This paper is organized as follows. In Sec.II, we begin by reviewing reflections from Einstein’s inertial mirror in (1+1)D and, from which, generalization to an accelerated mirror is made. In Sec.III, we review several crucial aspects, which do not appear in (1+1)D, of reflection by a mirror in (1+3)D. In Sec.IV, we first make connection of a general wave mode’s Fourier components to the Bogoliubov coefficients in various cases, e.g., cosmological particle creation, Unruh effect, etc., and later focus on the mirror-induced radiation. Particularly, the radiation spectrum emitted by a (in)finite-size plane mirror, which shows diffraction patterns, in (1+3)D is worked out. In addition, estimation of event yield based on proposed experiment is made. Finally, conclusion is given in Sec.V. Notation: throughout the paper, we use the mostly plus metric convention, $G=\hbar=c=k_{B}=1,\;\mathbb{R}^{1}=(-\infty,+\infty)$, and $\mathbb{R}^{-}=(-\infty,0)$. ## II Reflection in (1+1)D Suppose, according to a static observer in the lab frame $(t,x)$, there is a plane wave $\phi_{inc}=\text{exp}\left(-i\Omega t+iPx\right)$, where $\Omega=|P|>0$ is frequency and $P$ is momentum, that incidents upon an inertial moving mirror with trajectory $x_{r}(t_{r})=\beta t_{r}$, where $\beta=\dot{x}_{r}(t_{r})=$ const. is its velocity. According to Einstein’s theory of special relativity [20], upon reflection, the incident plane wave will undergo Doppler effect and this reflected wave, as seen by the static observer in the lab frame, can be described by $\phi_{ref}=\text{exp}\left(-i\omega t+ipx\right)$, where $\omega=\Omega\left(\frac{1\pm\beta}{1\mp\beta}\right),\quad p=-P\left(\frac{1\pm\beta}{1\mp\beta}\right),$ (1) are, respectively, the Doppler shifted frequency and momentum of the reflected wave, and the upper sign is valid for $P<0$, while the lower sign is valid for $P>0$. A first attempt to generalize the above discussions to an accelerated mirror can be made by giving the above mirror-related constant quantities time dependencies. That is, we shall have $\displaystyle\phi_{inc}$ $\displaystyle=\text{exp}\left(-i\Omega t+iPx\right),$ $\displaystyle\phi_{ref}$ $\displaystyle=\text{exp}\left(-i\int dt\;\omega(t,x)+i\int dx\;p(t,x)\right),$ (2) where $\Omega$ and $P$ are constants, while $\omega(t,x)$ and $p(t,x)$ are, respectively, the time-dependent reflected frequency and momentum given by $\displaystyle\omega(t,x)$ $\displaystyle=\Omega\left(\frac{1\pm\beta(t,x)}{1\mp\beta(t,x)}\right),$ (3) $\displaystyle p(t,x)$ $\displaystyle=-P\left(\frac{1\pm\beta(t,x)}{1\mp\beta(t,x)}\right).$ The observation point $(t,x)$ dependency in $\beta(t,x)$ is due to the fact that the retarded time $t_{r}$ at which the reflection occurs is a function of the observation point, i.e., $t_{r}=t_{r}(t,x)$. For a right/left-moving reflected wave, $p=\pm\omega$, the reflected waves become $\phi_{ref}=\text{exp}\left(-i\int dx_{\mp}\;\omega(t,x)\right),$ (4) where $x_{\mp}(t,x)=t\mp x$ are the light cone coordinates. On the worldline of a given reflected light ray, $t\mp x=t_{r}\mp x_{r}=$ const., where $(t_{r},x_{r}(t_{r}))$ is the spacetime point where the reflection occurs. From this, we also obtain $dx_{\mp}(t,x)=\left(1\mp\beta(t_{r})\right)dt_{r}$, which leads to $\phi_{ref}=\text{exp}\left(-i\Omega\int dt_{r}\left(1\pm\beta(t_{r})\right)\right).$ (5) In terms of the mirror’s proper time $\tau$, which is related to the retarded time by $d\tau=\gamma^{-2}(t_{r})dt_{r}$, one obtains $\phi_{ref}=\text{exp}\left(-i\Omega\int d\tau\gamma^{2}(t_{r})\left(1\pm\beta(t_{r})\right)\right).$ (6) These are the direct generalizations to an accelerated mirror in terms of observation point $(t,x)$, light cone coordinates $(x_{+},x_{-})$, retarded time $t_{r}$, and mirror’s proper time $\tau$ based on observations from the results of an inertial mirror. Alternatively, one may also begin by solving the (1+1)-dimensional wave equation: $\Box\phi(t,x)=0$, where $\Box=\partial^{\mu}\partial_{\mu}$ is the d’Alembertian operator, with the dynamical Dirichlet boundary condition: $\phi(t_{r},x_{r}(t_{r}))=0$, where $x_{r}(t_{r})$ is the mirror’s arbitrary trajectory [2, 3, 4, 5]. Suppose $\phi=\phi_{inc}-\phi_{ref}$, then for an incident plane wave $\phi_{inc}=\text{exp}\left(-i\Omega t+iPx\right)$, the reflected wave is solved to be $\phi_{ref}=\text{exp}\left(-i\Omega\eta_{\mp}(x_{\mp})\right)$, where $\eta_{\mp}(x_{\mp})=t_{r}\pm x_{r}(t_{r})$ is the ray tracing function, for $P<0$ (upper sign) and $P>0$ (lower sign). Noting that $\eta_{\mp}(x_{\mp})=\int dx_{\mp}\partial_{\mp}\eta_{\mp}(x_{\mp})$, the reflected wave can also be expressed as $\phi_{ref}=\text{exp}\left(-i\Omega\int dx_{\mp}\partial_{\mp}\eta(x_{\mp})\right),$ (7) which is identical to the previous generalization, Eq. (4), by identifying the time derivative of the ray tracing function as the reflected frequency, i.e., $\omega(t,x)=\Omega\partial_{\mp}\eta(x_{\mp})$. This justifies the validity of our previous generalizations to an accelerated mirror. In the case of an accelerated mirror, the integrations physically mean summing over distinct reflected waves with each wave labelled by the integration variables. Then, for distinct reflected waves that are close to each other, e.g., $x_{\mp}\approx\bar{x}_{\mp}+\varepsilon$, where $\bar{x}_{\mp}$ is a fixed reference point and $\varepsilon\rightarrow 0$, the wave in the vicinity of $\bar{x}_{\mp}$ is approximately, using Eq. (4), $\phi_{ref}\approx A(\bar{t},\bar{x})\cdot\text{exp}\left(-i\omega(\bar{t},\bar{x})\varepsilon\right),$ (8) where $A(\bar{t},\bar{x})$ is some function of the reference point. If one expands $\phi_{ref}=\text{exp}\left(-i\Omega\eta_{\mp}(x_{\mp})\right)$ around $\bar{x}_{\mp}$ instead, one obtains $\phi_{ref}\approx\text{exp}(-i\Omega\eta_{\mp}(\bar{x}_{\mp}))\cdot\text{exp}(-i\omega(\bar{t},\bar{x})\varepsilon).$ (9) From which we may identify $A(\bar{t},\bar{x})=\text{exp}(-i\Omega\eta_{\mp}(\bar{x}_{\mp}))$. Since $\varepsilon$ is small, the mirror has approximately a constant velocity in this small interval and thus the reflected wave has a similar expression, aside from the factor of $A(\bar{t},\bar{x})$ due to the velocity variation, as that reflected by an inertial mirror. From the above discussions, Eq. (3), one may observe that (I) whether the mirror is accelerated or not, the reflected frequencies are (I.a) only related to the mirror’s velocity at the retarded time and (I.b) the relations are exactly the same. (II) While the Doppler shifted frequencies are related to the velocities in the same manner for both inertial and accelerated mirrors, the expressions for the phase of the reflected waves are in fact different111The treatment in Ref.[21] to extend the discussion to an accelerated mirror is equivalent to treating the phase of the reflected wave as $\omega(t,x)x_{\mp}$ instead of reserving $\omega(t,x)$ inside the integral. However, in the case of constant four-acceleration, the resulting expression for the reflected wave happens to coincide with the result based on Eq. (4).. For waves reflected by an accelerated mirror, only those that are close to each other do they share similar expressions of phase as that reflected by an inertial mirror. ## III Reflection in (1+3)D In higher dimensional spacetimes, it is known, if possible, to be difficult to solve the wave equation: $\Box\phi(t,\mathbf{x})=0$ with the boundary condition: $\phi(t_{r},x_{r}(t_{r}),\mathbf{x}_{\perp})=0$ since conformal symmetry no longer holds. Nevertheless, from the previous section, we have seen the alternative route to obtain the expression of the reflected wave from its reflected frequency. We shall pursue this manner in this section. Suppose a plane wave $\phi_{inc}=\text{exp}(-i\Omega t+i\mathbf{P}\cdot\mathbf{x})$ incidents on a perfectly reflecting plane mirror moving at a constant velocity $\beta$, the reflected wave can be derived using Lorentz transformation222See, e.g., Ref.[22] for an alternative derivation based on purely geometric considerations without using Lorentz transformation. and the result is $\phi_{ref}=\text{exp}\left(-i\omega t+i\mathbf{p}\cdot\mathbf{x}\right)$, where $\displaystyle\omega$ $\displaystyle=\Omega\left(\frac{1\pm 2\beta\cos\theta_{i}+\beta^{2}}{1-\beta^{2}}\right),$ (10) $\displaystyle p_{x}$ $\displaystyle=-P_{x}\left(\frac{1\pm 2\beta\sec\theta_{i}+\beta^{2}}{1-\beta^{2}}\right),\quad\mathbf{p}_{\perp}=\mathbf{P}_{\perp},$ (11) where $\theta_{i}$ is the incident angle observed in the lab frame, $P_{x}=\mp\Omega\cos\theta_{i}$, and the reflected angle observed in the lab frame, $\theta_{r}=\cos^{-1}(\pm p_{x}/\omega)$, is given by $\cos\theta_{r}=\frac{(1+\beta^{2})\cos\theta_{i}\pm 2\beta}{1\pm 2\beta\cos\theta_{i}+\beta^{2}}.$ (12) The incident frequency $\omega^{\prime}$ and angle $\theta^{\prime}$ in the mirror’s rest frame are related to the incident frequency $\Omega$ and angle $\theta_{i}$ in the lab frame by $\displaystyle\omega^{\prime}=\Omega\left(\frac{1\pm\beta\cos\theta_{i}}{\sqrt{1-\beta^{2}}}\right),\quad\cos\theta^{\prime}=\frac{\cos\theta_{i}\pm\beta}{1\pm\beta\cos\theta_{i}}.$ (13) Since $\cos\theta^{\prime}$ should be non-negative, it leads to the constraint: $\cos\theta_{i}>|\beta|$, which is also the condition for frequency red shifting, for a receding mirror. Thus, $\theta_{i}$ can be no greater than $\theta_{m}=\cos^{-1}|\beta|$. Physically, this means that waves with incident angles larger than $\theta_{m}$ are not able to catch up the receding mirror. In addition, when $\theta_{i}=\theta_{m}$, we have $\theta_{r}=-\theta_{m}+\pi$. This shows that the wave can be reflected to the other side of the receding mirror. This is known as the forward reflection. The critical incident angle ($\theta_{i}=\theta_{c}$) beyond which the waves are forwardly reflected can be found by setting $\theta_{r}=\pi/2$ and solve for $\theta_{i}$ in Eq. (12). The result is $\cos\theta_{i}\left.\right|_{\theta_{i}=\theta_{c}}=\cos\theta_{c}=\mp\left(\frac{2\beta}{1+\beta^{2}}\right),$ (14) which corresponds to $\theta^{\prime}=\cos^{-1}|\beta|$ in the rest frame. In summary, for a receding mirror and in the lab frame, incident waves with $\theta_{i}\in[0,\theta_{c})$ remain on the same side of the mirror as the incident ones after the reflection; waves with $\theta_{i}\in(\theta_{c},\theta_{m})$ are reflected forwardly to the other side of the mirror, and waves with $\theta_{i}\in(\theta_{m},\pi/2]$ are not able to hit the mirror. On the other hand, in the mirror’s rest frame, waves with $\theta_{i}\in[0,\theta_{m})$ are on the same side of the mirror after the reflection. For an ultra-relativistically receding mirror, $\beta\sim\mp 1\pm\delta\beta$, $\delta\beta\rightarrow 0^{+}$, the reflected frequency and momentum are $\displaystyle\frac{\omega}{\Omega}$ $\displaystyle\sim\frac{\delta\beta}{2}\ll 1,\quad\frac{p_{x}}{\Omega}\sim\pm\frac{\delta\beta}{2},\quad\text{for }\theta_{i}\ll 1.$ (15) In addition, we have $\theta_{c}\sim\delta\beta\ll 1$ and $\theta_{m}\sim\sqrt{2}\sqrt{\delta\beta}\ll 1$. Thus, although waves with $\theta_{i}\in(\theta_{c},\theta_{m})$ are reflected forwardly to the mirror’s other side, they also experience Doppler red shift. On the other hand, for an ultra-relativistically approaching mirror, $\beta\sim\pm 1\mp\delta\beta$, $\delta\beta\rightarrow 0^{+}$, the reflected frequency and momentum are $\displaystyle\frac{\omega}{\Omega}$ $\displaystyle\sim\frac{1+\cos\theta_{i}}{\delta\beta}\gg 1,\quad\frac{p_{x}}{\Omega}\sim\pm\left(\frac{1+\cos\theta_{i}}{\delta\beta}\right),$ (16) where $\theta_{i}\in[0,\pi/2]$. That is, the frequencies are always Doppler blue shifted, and the reflected waves are mostly longitudinal, i.e., $\omega\sim|p_{x}|\gg|\mathbf{p}_{\perp}|$, and are always on the same side as their incident counterparts. In the case of an accelerated mirror, we expect to have $\displaystyle\phi_{inc}$ $\displaystyle=e^{i\mathbf{P}_{\perp}\cdot\mathbf{x}_{\perp}}\text{exp}\left(-i\Omega t+iP_{x}x\right),$ (17) $\displaystyle\phi_{ref}$ $\displaystyle=e^{i\mathbf{p}_{\perp}\cdot\mathbf{x}_{\perp}}\text{exp}\left(-i\int dt\;\omega(t,\mathbf{x})+i\int dx\;p_{x}(t,\mathbf{x})\right),$ where $\displaystyle\omega(t,\mathbf{x})=\Omega\left(\frac{1\pm 2\beta(t,\mathbf{x})\cos\theta_{i}+\beta^{2}(t,\mathbf{x})}{1-\beta^{2}(t,\mathbf{x})}\right),$ (18) $\displaystyle p_{x}(t,\mathbf{x})=-P_{x}\left(\frac{1\pm 2\beta(t,\mathbf{x})\sec\theta_{i}+\beta^{2}(t,\mathbf{x})}{1-\beta^{2}(t,\mathbf{x})}\right),$ (19) $\displaystyle\cos\theta_{r}(t,\mathbf{x})=\frac{(1+\beta^{2}(t,\mathbf{x}))\cos\theta_{i}\pm 2\beta(t,\mathbf{x})}{1\pm 2\beta(t,\mathbf{x})\cos\theta_{i}+\beta^{2}(t,\mathbf{x})},$ (20) where the additional dependence on $\mathbf{x}_{\perp}$ in $\beta(t,\mathbf{x})$ is due to the fact that $(t,\mathbf{x})$ is related to the retarded time by $t-t_{r}=\pm|\mathbf{x}-\mathbf{x}_{r}(t_{r})|$ for a given reflected light ray. ## IV Quantum particle spectrum In the usual context of quantum field theory in (1+$d$)-dimensional curved spacetime or quantum field theory in (1+$d$)-dimensional flat spacetime in the presence of a suitable external source, the quantum field, say, a massless real scalar field $\phi(t,\mathbf{x})$, can be expanded in terms of the mode $u_{\mathbf{P}}(t,\mathbf{x})$ or the mode $v_{\mathbf{p}}(t,\mathbf{x})$ by $\displaystyle\hat{\phi}(t,\mathbf{x})$ $\displaystyle=\int d^{d}P\left[\hat{a}_{\mathbf{P}}u_{\mathbf{P}}(t,\mathbf{x})+h.c.\right],$ (21) $\displaystyle=\int d^{d}p\left[\hat{b}_{\mathbf{p}}v_{\mathbf{p}}(t,\mathbf{x})+h.c.\right],$ (22) where $\Omega=|\mathbf{P}|$, $\omega=|\mathbf{p}|$, $(\hat{a}_{\mathbf{P}},\hat{b}_{\mathbf{p}})$ are annihilation operators, $h.c.$ denotes Hermitian conjugate, and the two mode bases are related by the Bogoliubov transformations $\displaystyle v_{\mathbf{p}}(t,\mathbf{x})$ $\displaystyle=\int d^{d}P\left[\alpha_{\mathbf{pP}}u_{\mathbf{P}}(t,\mathbf{x})+\beta_{\mathbf{pP}}\bar{u}_{\mathbf{P}}(t,\mathbf{x})\right],$ (23) $\displaystyle u_{\mathbf{P}}(t,\mathbf{x})$ $\displaystyle=\int d^{d}p\left[\alpha_{\mathbf{pP}}v_{\mathbf{p}}(t,\mathbf{x})-\bar{\beta}_{\mathbf{pP}}\bar{v}_{\mathbf{p}}(t,\mathbf{x})\right],$ (24) where $\alpha$ and $\beta$ are the Bogoliubov coefficients and the overbar refers to taking complex conjugation. In the context of quantum field theory with asymptotic in and out regions within a given coordinate system, whether in flat spacetime, e.g., particle creation by a moving mirror, or in curved spacetime, e.g., cosmological particle production, one often encounters $\displaystyle\lim_{t\rightarrow-\infty}u_{\mathbf{P}}(t,\mathbf{x})$ $\displaystyle\sim\frac{1}{(2\pi)^{3/2}\sqrt{2\Omega}}\text{exp}\left(-i\Omega t+i\mathbf{P}\cdot\mathbf{x}\right),$ (25) $\displaystyle\lim_{t\rightarrow+\infty}v_{\mathbf{p}}(t,\mathbf{x})$ $\displaystyle\sim\frac{1}{(2\pi)^{3/2}\sqrt{2\omega}}\text{exp}\left(-i\omega t+i\mathbf{p}\cdot\mathbf{x}\right),$ (26) so by taking the Fourier transformation of the out mode at the infinite past, i.e., $\displaystyle\lim_{t\rightarrow-\infty}v_{\mathbf{p}}(t,\mathbf{x})=\int$ $\displaystyle\frac{d^{3}P}{(2\pi)^{4}\sqrt{\Omega}}\biggl{[}e^{-i\Omega t+i\mathbf{P}\cdot\mathbf{x}}\tilde{v}_{\mathbf{p}}(\Omega,\mathbf{P})$ $\displaystyle+e^{i\Omega t-i\mathbf{P}\cdot\mathbf{x}}\tilde{v}_{\mathbf{p}}(-\Omega,-\mathbf{P})\biggr{]},$ (27) allows us to identify the Bogoliubov coefficients as $\alpha_{\mathbf{pP}}=\frac{\sqrt{2}}{2\pi}\frac{\tilde{v}_{\mathbf{p}}(\Omega,\mathbf{P})}{(2\pi)^{3/2}},\quad\beta_{\mathbf{pP}}=\frac{\sqrt{2}}{2\pi}\frac{\tilde{v}_{\mathbf{p}}(-\Omega,-\mathbf{P})}{(2\pi)^{3/2}}.$ (28) On the other hand, in non-dynamical situations, e.g., Unruh effect in (1+1) dimensions, which involves two different coordinate systems, one encounters $\displaystyle v_{\omega}(t,x)$ $\displaystyle=\frac{1}{(2\pi)^{1/2}\sqrt{2\omega}}\text{exp}\left(-i\omega t+i\omega x\right),$ (29) $\displaystyle u_{\Omega}(\tau,\xi)$ $\displaystyle=\frac{1}{(2\pi)^{1/2}\sqrt{2\Omega}}\text{exp}\left(-i\Omega\tau+i\Omega\xi\right),$ (30) where $(t,x)$ are Minkowski coordinates and $(\tau,\xi)$ are Rindler coordinates, which are related by the transformation: $t=a^{-1}e^{a\xi}\sinh(a\tau)$, $x=a^{-1}e^{a\xi}\cosh(a\tau)$, $a>0$. On the worldline of a Rindler observer at $\xi=0$, the Minkowski mode can be expanded in terms of the Rindler mode by $\displaystyle v_{\omega}(\tau)$ $\displaystyle=\frac{1}{\sqrt{4\pi\omega}}\cdot\text{exp}\left(\frac{i\omega}{a}e^{-a\tau}\right)$ $\displaystyle=\int_{0}^{\infty}\frac{d\Omega}{2\pi}\left[e^{-i\Omega\tau}\tilde{v}_{\omega}(\Omega)+e^{i\Omega\tau}\tilde{v}_{\omega}(-\Omega)\right],$ (31) which allows the identifications $\alpha_{\omega\Omega}=\frac{\sqrt{4\pi\Omega}}{2\pi}\;\tilde{v}_{\omega}(\Omega),\quad\beta_{\omega\Omega}=\frac{\sqrt{4\pi\Omega}}{2\pi}\;\tilde{v}_{\omega}(-\Omega),$ (32) where $\tilde{v}_{\omega}(\Omega)=a^{-1}(-i\omega/a)^{i\Omega/a}\Gamma(-i\Omega/a)/\sqrt{4\pi\omega}$. In this paper, we are interested in particle creation by a relativistic perfectly reflecting mirror, thus only the former dynamical situation will be relevant to our following discussions. ### IV.1 Flying mirror in (1+1)D To mimic Hawking radiation emitted by a black hole formed from gravitational collapse, we shall consider a perfect mirror whose trajectory asymptotes the Davies-Fulling one: $x_{r}(t_{r})\sim-t_{r}-A\text{exp}(-2\kappa t_{r})-B,\left\\{A,B,\kappa\right\\}>0$ at late times [5] (see Fig.1). Since the out mode $v_{\omega}(x_{+}(\mathcal{O}_{1}))\sim\text{exp}(-i\omega x_{-})$ is a plane wave basis (up to a normalization factor) for an (out) observer $\mathcal{O}_{1}$ at $x_{+}=x_{+}(\mathcal{O}_{1})$, the reflected waves, which have experienced Doppler red shift, that reach $x_{+}=x_{+}(\mathcal{O}_{1})$ can be expanded in terms of $v_{\omega}(x_{+}(\mathcal{O}_{1}))$, which further leads to the interpretation of Fourier modes as Bogoliubov coefficients. In this case, we have $\omega(t_{r}(x_{-}))/\Omega\sim A\kappa\cdot\text{exp}(-2\kappa t_{r}(x_{-}))\sim A\kappa\cdot\text{exp}(-\kappa x_{-})$, and the Fourier transformation of in mode at $x_{+}(\mathcal{O}_{1})$: $\displaystyle\tilde{u}_{\Omega}(\omega)=\int_{\mathbb{R}^{1}}dx_{-}e^{i\omega x_{-}}u_{\Omega}(x_{+}(\mathcal{O}_{1}))$ $\displaystyle=-\frac{1}{\sqrt{4\pi\Omega}}\int_{\mathbb{R}^{1}}dx_{-}e^{i\omega x_{-}}\text{exp}\left(-i\int dx_{-}\omega(t_{r}(x_{-}))\right)$ $\displaystyle\sim-\frac{1}{\sqrt{4\pi\Omega}}\int_{\mathbb{R}^{1}}dx_{-}e^{i\omega x_{-}}\text{exp}\left(i\Omega Ae^{-\kappa x_{-}}\right)$ $\displaystyle=-\frac{1}{\sqrt{4\pi\Omega}}\frac{(-i\Omega A)^{i\omega/\kappa}}{\kappa}\Gamma\left(\frac{-i\omega}{\kappa}\right),$ (33) and the Bogoliubov coefficients are $\alpha_{\omega\Omega}=\frac{\sqrt{4\pi\omega}}{2\pi}\tilde{u}_{\Omega}(\omega),\quad\bar{\beta}_{\omega\Omega}=-\frac{\sqrt{4\pi\omega}}{2\pi}\tilde{u}_{\Omega}(-\omega).$ (34) On the other hand, one intuitively expects an observer $\mathcal{O}_{2}$ at $x=x(\mathcal{O}_{2})$ should obtain the same spectrum as that obtained by $\mathcal{O}_{1}$ at $x_{+}=x_{+}(\mathcal{O}_{1})$. Indeed, since $x_{-}=t-x$, the integration over $x_{-}$ in the Fourier transform of $u_{\Omega}(t,x)$ can also be regarded as an integration over the observation time $t$ at a fixed spatial position $x=x(\mathcal{O}_{2})$. In this perspective, we have $\displaystyle\tilde{u}_{\Omega}(\omega)=\int_{\mathbb{R}^{1}}dt\;e^{i\omega x_{-}}u_{\Omega}(x(\mathcal{O}_{2}))$ $\displaystyle=-\frac{1}{\sqrt{4\pi\Omega}}\int_{\mathbb{R}^{1}}dt\;e^{i\omega x_{-}}\text{exp}\left(-i\int dt\;\omega(t_{r}(x_{-}))\right)$ $\displaystyle\sim-\frac{1}{\sqrt{4\pi\Omega}}e^{-i\omega x}\int_{\mathbb{R}^{1}}dt\;e^{i\omega t}\text{exp}\left(i\Omega Ae^{\kappa x}e^{-\kappa t}\right)$ $\displaystyle=-\frac{1}{\sqrt{4\pi\Omega}}\frac{(-i\Omega A)^{i\omega/\kappa}}{\kappa}\Gamma\left(\frac{-i\omega}{\kappa}\right),$ (35) which is identical to the result of $\mathcal{O}_{1}$. Finally, it is obvious that identical result can also be obtained if one integrates along the retarded time $t_{r}$. Figure 1: Reflection of plane waves by a relativistically receding perfect mirror in (1+1)-dimensions. Blue: mirror’s trajectory. Dashed line: observer $\mathcal{O}_{1}$’s worldline $x_{+}=x_{+}(\mathcal{O}_{1})$. Dotted line: observer $\mathcal{O}_{2}$’s worldline $x=x(\mathcal{O}_{2})$. ### IV.2 Flying mirror in (1+3)D In a (1+3)-dimensional spacetime with an infinite-size plane mirror moving along the $x$-axis, the concept of light cone coordinate $x_{\pm}=t\pm x$ is not suitable for discussion. Nevertheless, discussions based on the observation point $(t,\mathbf{x})$ remain valid and more convenient. Thus, we shall consider an observer $\mathcal{O}_{1}$ with the worldline $\mathbf{x}=\mathbf{x}(\mathcal{O}_{1})$, as shown in Fig.2. Since the plane wave mode $v_{\mathbf{p}}(\mathbf{x}(\mathcal{O}_{1}))\sim\text{exp}(-i\omega t+i\mathbf{p}\cdot\mathbf{x})$ can serve as a basis at $\mathbf{x}=\mathbf{x}(\mathcal{O}_{1})$, this validates the action of taking the Fourier transformation of the reflected wave mode and interpreting its Fourier components as Bogoliubov coefficients. Figure 2: Reflection of plane waves by a relativistically receding perfect mirror in higher dimensions. Blue: mirror’s trajectory. Dashed line: observer $\mathcal{O}_{1}$’s worldline $x=x(\mathcal{O}_{1})$. We shall again consider a perfect mirror with trajectory that asymptotes: $x_{r}(t_{r})\sim-t_{r}-A\text{exp}(-2\kappa t_{r})-B,\left\\{A,B,\kappa\right\\}>0$ at late times. For incident in modes catching up the receding mirror, the critical incident angle $\theta_{c}$ and maximum incident angle $\theta_{m}$ are $\displaystyle\theta_{c}(t_{r})$ $\displaystyle\sim 2A\kappa\;e^{-2\kappa t_{r}}+\mathcal{O}(e^{-4\kappa t_{r}})\ll 1,$ (36) $\displaystyle\theta_{m}(t_{r})$ $\displaystyle\sim\sqrt{4A\kappa}\;e^{-\kappa t_{r}}+\mathcal{O}(e^{-3\kappa t_{r}})\ll 1.$ (37) Since $\theta_{m}(t_{r})\ll 1$ at late times, only incident in modes with $\theta_{i}\leq\theta_{m}\ll 1$ can catch up the receding mirror and get reflected. Among the reflected in modes, those with $\theta_{i}<\theta_{c}\ll 1$ are reflected backwardly, while those with $\theta_{c}<\theta_{i}\leq\theta_{m}\ll 1$ are reflected forwardly. For an incident in mode with $\theta_{i}\leq\theta_{m}\ll 1$, its reflected frequency and longitudinal momentum are $\displaystyle\omega(t_{r})$ $\displaystyle\sim\Omega A\kappa\cos^{2}\left(\theta_{i}/2\right)e^{-2\kappa t_{r}}+\mathcal{O}(e^{-4\kappa t_{r}}),$ (38) $\displaystyle p_{x}(t_{r})$ $\displaystyle\sim\Omega A\kappa\cos^{2}\left(\theta_{i}/2\right)e^{-2\kappa t_{r}}+\mathcal{O}(e^{-4\kappa t_{r}}).$ (39) Thus, to an observer at $\mathbf{x}=\mathbf{x}(\mathcal{O}_{1})$, the only non-trivial wave mode he/she will receive is the reflected in mode with negligible transverse momentum which comes from the reflection of an incident in mode with $\theta_{i}<\theta_{c}\ll 1$. Having the above discussions in mind, we decompose the in annihilation operator $\hat{a}_{\mathbf{P}}$ into longitudinal component $\hat{a}_{\mathbf{P}}^{\parallel}\equiv 2\pi\delta^{(2)}(\mathbf{P}_{\perp})\hat{c}_{P_{x}}$ and its complement $\hat{a}_{\mathbf{P}}^{c}$, where the commutation relation for $\hat{c}_{P_{x}}$ is $\left[\hat{c}_{P_{x}},\hat{c}_{P_{x}^{\prime}}^{\dagger}\right]=\frac{\delta(P_{x}-P_{x}^{\prime})}{\mathcal{A}_{\infty}},$ (40) where $\mathcal{A}_{\infty}\rightarrow\infty$ denotes the mirror’s transverse area. Thus, the decomposition of a scalar field now becomes $\displaystyle\hat{\phi}(t,\mathbf{x})$ $\displaystyle=\int_{P_{x}<0}d^{3}P\left[(\hat{a}_{\mathbf{P}}^{\parallel}+\hat{a}_{\mathbf{P}}^{c})u_{\mathbf{P}}(t,\mathbf{x})+h.c.\right]$ $\displaystyle=\int_{\mathbb{R}^{-}}dP_{x}\left[\hat{c}_{P_{x}}\cdot\left.2\pi u_{\mathbf{P}}(t,\mathbf{x})\right|_{\mathbf{P}_{\perp}=0}+h.c.\right]$ $\displaystyle\quad+\int_{P_{x}<0}d^{3}P\left[\hat{a}_{\mathbf{P}}^{c}u_{\mathbf{P}}(t,\mathbf{x})+h.c.\right],$ (41) $\displaystyle=\int_{p_{x}>0}d^{3}p\left[\hat{b}_{\mathbf{p}}v_{\mathbf{p}}(t,\mathbf{x})+h.c.\right],$ (42) to the mirror’s right, where the in mode with $\mathbf{P}_{\perp}=0$ is $\displaystyle\left.u_{\mathbf{P}}(t,\mathbf{x})\right|_{\mathbf{P}_{\perp}=0}=\frac{e^{-i\Omega t+iP_{x}x}-e^{-i\int dx_{-}\omega(t_{r})}}{(2\pi)^{3/2}\sqrt{2\Omega}},$ (43) where $\Omega=|P_{x}|$, and Eq. (38) will later be used as the explicit expression for $\omega(t_{r})$. For the out observer $\mathcal{O}_{1}$ at $\mathbf{x}=\mathbf{x}(\mathcal{O}_{1})$, the (reflected) in mode can be decomposed into Fourier components as $\displaystyle u_{\mathbf{P}}(\mathbf{x}(\mathcal{O}_{1}))=-\frac{e^{-i\int dx_{-}\omega(t_{r}(x_{-}))}}{(2\pi)^{3/2}\sqrt{2\Omega}}$ (44) $\displaystyle=\int_{p_{x}>0}\frac{d^{3}p}{(2\pi)^{4}\sqrt{\omega}}\left[e^{ip\cdot x}\tilde{u}_{\Omega}(\omega,\mathbf{p})+e^{-ip\cdot x}\tilde{u}_{\Omega}(-\omega,-\mathbf{p})\right],$ where $\omega=p^{t}=-p_{t}=|\mathbf{p}|$, and $p\cdot x=-\omega t+\mathbf{p}\cdot\mathbf{x}$, and the Bogoliubov coefficients are given by $\displaystyle\alpha_{\mathbf{pP}}=\frac{\sqrt{2}}{2\pi}\frac{\tilde{u}_{\Omega}(\omega,\mathbf{p})}{(2\pi)^{3/2}},\quad\bar{\beta}_{\mathbf{pP}}=-\frac{\sqrt{2}}{2\pi}\frac{\tilde{u}_{\Omega}(-\omega,-\mathbf{p})}{(2\pi)^{3/2}}.$ (45) Since the first line of Eq. (44) is independent of $\mathbf{x}_{\perp}$, the Fourier components must have the form: $\tilde{u}_{\Omega}(\pm\omega,\pm\mathbf{p})=\delta^{(2)}(\mathbf{p}_{\perp})\underline{u}_{\Omega}(\pm\omega,\pm p_{x})$. Then, the Fourier components are related to the Fourier transformation of the in mode by $\displaystyle\underline{u}_{\Omega}(\omega,p_{x})$ $\displaystyle=-\frac{(2\pi)^{3/2}\sqrt{\Omega}}{\sqrt{2\omega}}e^{-i\Omega x}\int_{\mathbb{R}^{1}}dt\;e^{i\Omega t}e^{-i\int dx_{-}\omega(t_{r})}$ $\displaystyle=-\frac{(2\pi)^{3/2}\sqrt{\Omega}}{\sqrt{2\omega}}\frac{(-i\Omega A)^{i\omega/\kappa}}{\kappa}\Gamma\left(\frac{-i\omega}{\kappa}\right),$ (46) where we have used Eq. (38) and the fact that $\theta_{i}\ll 1$. Thus, the Bogoliubov coefficients of interest are $\displaystyle\alpha_{\mathbf{pP}}$ $\displaystyle=-\frac{\delta^{(2)}(\mathbf{p}_{\perp})}{2\pi\kappa}\frac{\sqrt{\omega}}{\sqrt{\Omega}}(-i\Omega A)^{i\omega/\kappa}\Gamma\left(-\frac{i\omega}{\kappa}\right),$ (47) $\displaystyle\bar{\beta}_{\mathbf{pP}}$ $\displaystyle=\frac{\delta^{(2)}(\mathbf{p}_{\perp})}{2\pi\kappa}\frac{\sqrt{\omega}}{\sqrt{\Omega}}(-i\Omega A)^{-i\omega/\kappa}\Gamma\left(\frac{i\omega}{\kappa}\right).$ (48) The mean occupation number of out particles is thus $\displaystyle\frac{d^{3}N}{d^{3}p}$ $\displaystyle=\left<0;\text{in}\right|\hat{b}_{\mathbf{p}}^{\dagger}\hat{b}_{\mathbf{p}}\left|0;\text{in}\right>$ $\displaystyle=\lim_{\mathbf{p}^{\prime}\rightarrow\mathbf{p}}\frac{(2\pi)^{2}}{\mathcal{A}_{\infty}}\int_{\mathbb{R}^{-}}dP_{x}\beta_{\mathbf{pP}}\bar{\beta}_{\mathbf{p^{\prime}P}}$ $\displaystyle=\frac{(2\pi)^{2}\delta^{(2)}(\mathbf{p}_{\perp})}{\mathcal{A}_{\infty}}\frac{\delta^{(2)}(\mathbf{p}_{\perp})\Delta t}{2\pi}\left(\frac{1}{e^{\omega/T_{H}}-1}\right),$ (49) where $p_{x}=\omega>0$, $\Delta t=2\pi\delta(\omega-\omega^{\prime}=0)\rightarrow\infty$ is the infinite accelerating time period measured by the out observer $\mathcal{O}_{1}$, the first fraction in the last line equals to unity for $\mathbf{p}_{\perp}=\mathbf{0}$, and $T_{H}=\kappa/2\pi$ is the Hawking temperature. This is the analog Hawking radiation spectrum emitted by a perfectly reflecting, infinite transverse area, plane mirror in (1+3)-dimensional flat spacetime at late times and is to be observed by an out observer $\mathcal{O}_{1}$ on the right-hand-side of the leftwardly receding mirror. By using the identity: $\displaystyle\lim_{L\rightarrow\infty}\text{sinc}(pL/2)=2\pi\delta(p)/L,$ the Dirac delta functions can be replaced by sinc functions instead. This alternative expression allows a direct extrapolation to the case of a finite- size, say, square mirror with transverse area: $\mathcal{A}=L\times L$ as $\displaystyle\frac{d^{3}N}{d^{3}p}\approx\frac{\mathcal{A}\Delta t}{(2\pi)^{3}}\;\text{sinc}^{2}\left(\frac{p_{y}L}{2}\right)\text{sinc}^{2}\left(\frac{p_{z}L}{2}\right)\left(\frac{1}{e^{\omega/T_{H}}-1}\right)$ for $\omega=|\mathbf{p}|\sim p_{x}\gg|\mathbf{p}_{\perp}|$. If $|\mathbf{p}_{\perp}|$ is not required to be negligible compared to $p_{x}$, then we should additionally replace $T_{H}$ by $T_{eff}=\kappa/[(1+\cos\theta)\pi]$ since the mirror is neither isotropic nor translational invariant [15, 16]. The appearance of sinc functions is a manifestation of diffraction due to the mirror’s finite transverse geometry. In far field, using $p_{y}\sim\omega y/R$ and $p_{z}\sim\omega z/R$, where $R\gg|\mathbf{x}_{\perp}|$ is the longitudinal distance between the observer $\mathcal{O}_{1}$ and the mirror at the instant of emission, the arguments of the sinc functions become $\omega Ly/(2R)\sim\omega L\theta\cos\phi/2$ and $\omega Lz/(2R)\sim\omega L\theta\sin\phi/2$, which are also justified by $p_{y}=\omega\sin\theta\cos\phi$ and $p_{z}=\omega\sin\theta\sin\phi$ when $\theta\ll 1$. Similarly, for a circular mirror of diameter $D$, the particle spectrum would be $\displaystyle\frac{d^{3}N}{d^{3}p}\approx\frac{\mathcal{A}\Delta t}{(2\pi)^{3}}\left[2\text{jinc}\left(\frac{|\mathbf{p}_{\perp}|D}{2}\right)\right]^{2}\left(\frac{1}{e^{\omega/T_{H}}-1}\right),$ where $\mathcal{A}=D^{2}\pi/4$, $\text{jinc}(x)=J_{1}(x)/x$, $J_{1}(x)$ is the Bessel function of the first kind of order 1, and $|\mathbf{p}_{\perp}|\sim\omega\theta$ with $\theta\ll 1$. This leads to the Airy pattern, and the first minimum of the jinc function occurs at $\theta=\theta_{1}=7.66/(\omega D)=1.22\lambda/D$, where $\lambda=2\pi/\omega$ is the wavelength of the emitted out particle. The total occupation number can be obtained by integrating over the momentum $\mathbf{p}$. For $\mathcal{A}\rightarrow\infty$, only particles with $|\mathbf{p}_{\perp}|=0$ can be emitted, and the above three types of mirrors all lead to the same yield: $\displaystyle N$ $\displaystyle=\frac{\Delta t}{2\pi}\int_{\omega_{1}}^{\omega_{2}}\frac{d\omega}{e^{2\pi\omega/\kappa}-1}$ $\displaystyle=\frac{T_{H}\Delta t}{2\pi}\log\left(\frac{1-e^{-\omega_{2}/T_{H}}}{1-e^{-\omega_{1}/T_{H}}}\right),$ (50) where $(\omega_{1},\omega_{2})$ is the frequency range of interest. On the other hand, for $\mathcal{A}=$ finite, the emitted particles can acquire non-negligible transverse momenta. However, analytic expressions of spectra can only be obtained under certain approximations. For simplicity, in the case of a square mirror, the frequency spectrum in the low frequency regime: $\omega\ll L^{-1}$ is $\displaystyle\frac{dN}{d\omega}\approx\frac{\mathcal{A}\Delta t}{(2\pi)^{2}}\frac{\kappa\omega}{\pi}\log\big{(}1+e^{-\pi\omega/\kappa}\big{)},$ (51) while the angular spectrum with $\theta\ll 1$ is $\displaystyle\frac{dN}{d\Omega}\approx\frac{\mathcal{A}\Delta t}{(2\pi)^{3}}\frac{\kappa^{3}\zeta(3)}{4\pi^{3}}\Big{[}1-\frac{\zeta(5)(\kappa L)^{2}-3\pi^{2}\zeta(3)}{4\pi^{2}\zeta(3)}\theta^{2}\Big{]},$ (52) where $\zeta$’s are Riemann zeta functions. The full spectra are plotted in Figs. 3 $\&$ 4, which are obtained from numerical integrations and they indeed agree well with the above analytic formulae in the corresponding regimes. Figure 3: Frequency spectrum of analog Hawking radiation. Figure 4: Angular spectrum of analog Hawking radiation. The frequency spectrum initially grows linearly in $\omega$ in the low frequency regime. Since small frequency corresponds to long wavelength, the spectrum is simply proportional to the mirror’s area. However, as the frequency gets higher, the wavelength becomes comparable to the area and diffraction emerges, leading to the wiggling in the figure. On the other hand, as indicated in Eq. (52), for $\zeta(5)(\kappa L)^{2}>3\pi^{2}\zeta(3)$, which is the case used in the figure, the angular spectrum has its maximum at $\theta=0$ and decreases quadratically in $\theta$ as $\theta$ gets larger. In the recently proposed AnaBHEL collaboration [17, 18, 19], the flying mirror generated via plasma-laser interaction has a low reflectivity [23] and the event yield per laser shot is estimated as $N\sim 0.3$ [16]. Assuming a perfectly reflecting plane mirror with the same trajectory and parameter values, i.e., $L=254\;\text{eV}^{-1},\;\kappa=0.2\;\text{eV}$, can be generated, which faces both physical and technical challenges, and noting that the Davies-Fulling-like motion takes about $\Delta t\sim 1$ ps $\sim 1520$ eV-1 based on the design in Ref.[17], the yield of analog Hawking particles per laser shot is then $N\sim 0.011\times 1520\sim 16$ events, where $0.011$ eV is the area under the frequency spectrum in Fig. 3. For a petawatt-class laser facility that provides 1 laser shot per minute and 8 hours of operation time per day, a 20-day experiment with an ideal detector efficiency would then give the total yield as $N_{total}\sim 160,000$ events. ## V Conclusion The conventional approach toward radiation induced by a relativistically flying mirror in spacetimes other than (1+1) dimensions is, if possible, difficult. If the mirror is non-relativistic and perfectly reflecting, then it can be solved by perturbing a static mirror solution [12, 13, 14]; if the mirror is relativistic but semitransparent, it can also be solved by perturbing a free field solution [15, 16]. To circumvent the difficulty encountered by a perfectly reflecting mirror in higher dimensional spacetimes, we begin by studying the frequency and momentum of an incident plane wave after reflection from a flying mirror instead. From this, we observe that, in general, the reflected frequency and momentum only depend on the mirror’s velocity, while the phase of the reflected wave will also depend on the acceleration. This viewpoint provides crucial physical insights which allow us to generalize the discussion to higher dimensional spacetimes, which is important for actual laboratory experiments. It should be noted that by striking a flying mirror by a classical plane wave and Fourier transforming the received reflected wave can indeed lead to a frequency spectrum. However, this is only a classical signal in Fourier space. It is for quantum fluctuations that the Fourier components of the reflected wave modes have the interpretation as quantum particle creation from the vacuum (although this quantum radiation spectrum has exactly the same form as that of the previous classical spectrum). Without going through rigorous derivation of the radiation spectrum, one may also heuristically argue the expected expression from dimensional analysis. Since the mean occupation number $d^{3}N/d^{3}p$ has a dimension of length3, and, for a relativistically receding mirror, the reflected wave mode are mostly longitudinal, one expects $d^{3}N/d^{3}p\sim\delta^{(2)}(\mathbf{p}_{\perp})dN/dp_{x}$ up to a dimensionless proportionality constant. By going through rigorous derivation, the proportionality constant can be uniquely determined, which is important to the estimation of event yield for a proposed experiment. The results based on the viewpoint adopted in this paper also allow direct generalization to mirrors with various geometries, e.g., square, circular, etc., and diffraction in the radiation spectrum emerges. Finally, having robust expressions for the analog Hawking radiation spectra, we estimated, in the case of $T_{H}=0.03$ eV, the yield to be $N\sim 16$ events per laser shot for an ideal detector, and $N_{total}\sim 160,000$ in a 20-day experiment, which is roughly 50 times larger than the yield, $N_{total}\sim 3,000$, for a square-root-Lorentzian (SRLD), semitransparent mirror [16]. ###### Acknowledgements. This work is supported by ROC (Taiwan) Ministry of Science and Technology (MOST), National Center for Theoretical Sciences (NCTS), and Leung Center for Cosmology and Particle Astrophysics (LeCosPA) of National Taiwan University. P.C. is in addition supported by U.S. Department of Energy under Contract No. DE-AC03-76SF00515. ## References * Hawking [1975] S. W. Hawking, Particle creation by black holes, Commun. Math. Phys. 43, 199 (1975). * Moore [1970] G. T. Moore, Quantum theory of the electromagnetic field in a variable‐length one‐dimensional cavity, J. Math. Phys. (N.Y.) 11, 2679 (1970). * DeWitt [1975] B. S. DeWitt, Quantum field theory in curved spacetime, Phys. Rep. 19, 295 (1975). * Fulling and Davies [1976] S. A. Fulling and P. C. W. Davies, Radiation from a moving mirror in two dimensional space-time: conformal anomaly, Proc. R. Soc. A 348, 393 (1976). * Davies and Fulling [1977] P. C. W. Davies and S. A. Fulling, Radiation from moving mirrors and from black holes, Proc. R. Soc. A 356, 237 (1977). * Carlitz and Willey [1987] R. D. Carlitz and R. S. Willey, Reflections on moving mirrors, Phys. Rev. D 36, 2327 (1987). * Foo and Good [2021] J. Foo and M. R. R. Good, Hawking radiation particle spectrum of a Kerr-Newman black hole, JCAP 01, 019 (2021), arXiv:2006.09681 [gr-qc] . * Good _et al._ [2020] M. R. R. Good, A. Zhakenuly, and E. V. Linder, Mirror at the edge of the universe: Reflections on an accelerated boundary correspondence with de Sitter cosmology, Phys. Rev. D 102, 045020 (2020), arXiv:2005.03850 [gr-qc] . * Good [2020] M. R. R. Good, Extremal Hawking radiation, Phys. Rev. D 101, 104050 (2020), arXiv:2003.07016 [gr-qc] . * Good and Ong [2020] M. R. R. Good and Y. C. Ong, Particle spectrum of the Reissner–Nordström black hole, Eur. Phys. J. C 80, 1169 (2020), arXiv:2004.03916 [gr-qc] . * Good _et al._ [2017] M. R. R. Good, K. Yelshibekov, and Y. C. Ong, On horizonless temperature with an accelerating mirror, JHEP 03, 013 (2017), arXiv:1611.00809 [gr-qc] . * Ford and Vilenkin [1982] L. H. Ford and A. Vilenkin, Quantum radiation by moving mirrors, Phys. Rev. D 25, 2569 (1982). * Neto and Machado [1996] P. A. M. Neto and L. A. S. Machado, Quantum radiation generated by a moving mirror in free space, Phys. Rev. A 54, 3420 (1996). * Rego _et al._ [2013] A. L. C. Rego, B. W. Mintz, C. Farina, and D. T. Alves, Inhibition of the dynamical Casimir effect with Robin boundary conditions, Phys. Rev. D 87, 045024 (2013), arXiv:1302.4709 [quant-ph] . * Lin _et al._ [2021] K.-N. Lin, C.-E. Chou, and P. Chen, Particle production by a relativistic semitransparent mirror in (1+3)D Minkowski spacetime, Phys. Rev. D 103, 025014 (2021), arXiv:2008.12251 [gr-qc] . * Lin and Chen [2021] K.-N. Lin and P. Chen, Particle production by a relativistic semitransparent mirror of finite transverse size (2021), arXiv:2107.09033 [gr-qc] . * Chen _et al._ [2022] P. Chen _et al._ , [AnaBHEL Collaboration], AnaBHEL (Analog Black Hole Evaporation via Lasers) Experiment: Concept, Design, and Status (2022), arXiv:2205.12195 [gr-qc] . * Chen and Mourou [2017] P. Chen and G. Mourou, Accelerating plasma mirrors to investigate black hole information loss paradox, Phys. Rev. Lett. 118, 045001 (2017), arXiv:1512.04064 [gr-qc] . * Chen and Mourou [2020] P. Chen and G. Mourou, Trajectory of a flying plasma mirror traversing a target with density gradient, Phys. Plasmas 27, 123106 (2020), arXiv:2004.10615 [physics.plasm-ph] . * Einstein [1905] A. Einstein, Zur elektrodynamik bewegter körper, Annalen der Physik 322, 891 (1905). * Van Meter _et al._ [2001] J. R. Van Meter, S. Carlip, and F. V. Hartemann, Reflection of plane waves from a uniformly accelerating mirror, American Journal of Physics 69, 783 (2001). * Gjurchinovski and Skeparovski [2008] A. Gjurchinovski and A. Skeparovski, Einstein’s mirror, The Physics Teacher 46, 416 (2008). * Liu _et al._ [2021] Y.-K. Liu, P. Chen, and Y. Fang, Reflectivity and spectrum of relativistic flying plasma mirrors, Phys. Plasmas 28, 103301 (2021), arXiv:2012.05769 [physics.plasm-ph] .
# Classical-Quantum Combs, their Min-Entropy and their Measurement-Based Applications Isaac D. Smith<EMAIL_ADDRESS>Marius Krumm Lukas J. Fiderer Hendrik Poulsen Nautrup Hans J. Briegel Institute for Theoretical Physics, UIBK, 6020 Innsbruck, Austria ###### Abstract Learning a hidden property of a quantum system typically requires a series of interactions. In this work, we consider a formalisation of such multi-round learning processes that uses a generalisation of classical-quantum states, called classical-quantum combs. Here, “classical” refers to a random variable encoding the hidden property to be learnt, and “quantum” refers to the quantum comb describing the behaviour of the system. By using the quantum combs formalism, the optimal strategy for learning the hidden property can be quantified via the comb min-entropy (Chiribella and Ebler, NJP, $2016$). With such a tool on hand, we focus attention on an array of combs derived from measurement-based quantum computation (MBQC) and related applications. Specifically, we describe a known blind quantum computation (BQC) protocol using the combs formalism and thereby leverage the min-entropy to provide a proof of single-shot security for multiple rounds of the protocol, extending the existing result in the literature. Furthermore, we introduce novel connections between MBQC and quantum causal models and quantum causal inference, which allows for the use of the min-entropy to quantify the optimal strategy for causal discovery. We consider further operationally motivated examples, including one associated to learning a quantum reference frame. ## 1 Introduction With the rapid development of quantum technology, a plethora of increasingly complex quantum devices has become available. Already, a range of noisy- intermediate scale quantum computers (NISQ) [1] are accessible via the internet. Year on year, these devices increase in size and quality, as evidenced by the growing number of addressable qubits, the lengthening coherence times and improving gate fidelity. An important parallel development to that of the devices themselves, is in how they interconnect. The current and future progress of quantum communication networks [2, 3, 4, 5, 6, 7, 8, 9, 10, 11] aims to create a global quantum internet of networked quantum computers [12, 13], with the attendant benefits in communication and cryptography. Current implementations of such networks largely focus on protocols for quantum key distribution, however a variety of other significant applications appear to be feasible in the near-term, including distributed quantum computation [14, 15], quantum position verification [16], and blind quantum computation [17]. With the increasing sophistication of quantum networks, a full description using the usual quantum formalism becomes intractable. Consequently, other formalisms are being developed that allow for the description of the relevant parts of complex networks whilst remaining numerically and analytically manageable. One particularly promising such formalism represents networks as quantum causal models [18, 19, 20], in which the connectivity of the the network is faithfully represented but the nodes are freely described by any quantum device. This allows for investigating the statistics of the network under varying quantum protocols at the nodes. The quantum causal modelling formalism is a quantum generalisation of the sub-field of statistics and machine learning known as (classical) causal modelling, which has been successfully applied to modelling complex systems in various classical domains, such as medicine, social science, and engineering [21, 22]. One key object of study in the quantum causal networks literature [23, 24] is known as a quantum comb. Quantum combs naturally model multi-round quantum communication protocols, and can be understood as a concatenation of quantum channels where some information may be input and output from the comb at each time-step and some may remain within the comb (and inaccessible) between time- steps. These combs are represented by operators acting on the input and output spaces only, which constitutes a distinct advantage of their use: any forwarding of information between time-steps is modelled by an operator of the same size. As such, they can model arbitrary sequential interactions with a quantum environment, and have consequently become a fruitful tool to describe e.g. non-Markovian noise [25]. A crucial discovery of the field of quantum causal networks is that the optimisation of multi-round quantum communication protocols can be framed as a semi-definite program [26], where the optimisation can be with respect to a range of different performance measures. For a given measure, the resultant optimal success probability leads to a notion of min-entropy for quantum combs, which generalises the min-entropy of quantum states [27] in quantum cryptography. In this work, we consider a performance measure based on learning a hidden classical parameters of quantum networks, which are modelled as classical-quantum combs: generalisations of classical-quantum states [28] where a series of quantum combs are indexed by a classical parameter. We analyse properties of combs of this form and investigate a range of examples based on measurement-based quantum computation. Measurement-based quantum computing (MBQC) [29, 30, 31, 32, 33, 34] is a well- known paradigm for quantum computation distinct from the circuit model formalism. It consists of a sequential interaction with a quantum state (via time-ordered single qubit measurements) and forms the basis of a range of cryptographic computation protocols within the sub-field of quantum cryptography known as blind quantum computation (BQC) [35, 36, 37, 38, 17]. Due to existing characterisations of the components of MBQC, such as graph states [39, 40, 41, 42] (see also [43]) and the sequence of measurements [32, 44, 45, 46, 47], as well as to the connection with stabiliser quantum mechanics [48, 49], investigations into certain aspects of MBQC remain tractable despite its multi-partite, high-dimensional nature. As such, MBQC provides an exemplary test-bed for the application of the quantum causal networks framework. The first comb we consider models a specific BQC protocol, that of Mantri et al. [36]. This protocol consists of entirely classical communication between a client and a (quantum) server, allowing the client to obtain the results of a quantum computation while maintaining no quantum capabilities. The classical parameter in this context encodes the secret choice of computation by the client which an untrustworthy server may attempt to discover. Due to the restriction to classical communication, the quantum comb part of the classical-quantum comb, the part that models a round of the protocol, is actually classical (i.e. we consider a classical-classical comb). The security of the protocol is analysed by calculating the min-entropy of this comb, which has the interpretation of quantifying how much information the server has yet to learn about the choice of computation when using the most informative strategy. We provide a security analysis both for a single round and for multiple rounds, thereby extending the analysis of the original paper [36]. We then consider a range of classical-quantum combs where the comb component is truly quantum. In particular, we investigate a variety of related scenarios where an unknown device performing a measurement-based computation is interacted with in order to gain some piece of information regarding its internal functioning. In all these scenarios, the classical-quantum combs are built from a comb based upon the method of measurement adaptation required for MBQC [44, 45]. We demonstrate that this base comb satisfies the conditions of a quantum causal model [20] and moreover, we explicate an interpretation of MBQC in the context of quantum causal inference [50], thereby establishing novel connections between MBQC and quantum causality [20, 19, 18, 50, 51, 52, 53]. One classical-quantum comb we consider explicitly utilises this connection: the classical parameter encodes the different possible causal structures and the comb min-entropy quantifies the optimal causal inferential strategy. Finally, two examples are considered, each concerned with learning an aspect of the MBQC device required for its proper use: in one case, learning the types of measurement that ensure deterministic computation, and in the other, calibrating measurement devices to ensure the correct computation is performed. The latter has an interesting interpretation as learning a quantum reference frame [54]. The remainder of this paper is structured as follows. The next subsection provides a brief and informal presentation of the key results in this work. In Section 2, the required background knowledge regarding the formalisms of MBQC and quantum combs is presented. We introduce and motivate the comb min-entropy in Section 3, including some results regarding the min-entropy for classical- classical combs in Section 3.1. Section 4 outlines the BQC protocol of [36] (Section 4.1), defines the corresponding classical comb (Section 4.2), and conducts an analysis of the security of the protocol (Section 4.3 and Section 4.4). Section 5 contains the definition of the comb describing the corrections required for MBQC (Section 5.1), a discussion of the connection to quantum causal models and inference (Section 5.2), and an analysis of a series of examples (Section 5.3, Section 5.4 and Section 5.5). Section 6 discusses the future applications and limitations of the methods presented in this work and concludes. ### 1.1 Our Contributions The following is a more specific, but informal, summary of the key contributions of this paper. * • 3.2 establishes upper and lower bounds for the min-entropy for classical- quantum combs where the quantum combs are all diagonal in the same basis (i.e. classical-classical combs). These bounds have an interpretation in terms of maximal Bayesian updating, an interpretation which extends from the analogous case for the state min-entropy (i.e. classical-classical states). These bounds play a role in the security proof of the BQC protocol in Section 4. * • 4.2 establishes that the BQC protocol of [36] is (partially) secure in a single-round by proving a general lower bound on the min-entropy of the secret choice of computation given the interactions via the protocol. This lower bound is strictly positive, indicating that the secret computation cannot be known with certainty after only a single round, even if the optimal strategy is used. This result extends the existing result [[, Theorem 2,]]mantri2017flow via use of the min-entropy as opposed to the Shannon entropy. Furthermore, we give a simple example that obtains the lower bound, which gives some indication of its tightness. * • Similarly, 4.3 provides a positive lower bound for the min-entropy under any number of rounds of the protocol, demonstrating that the server will remain uncertain about at least some aspect of the computation even if the protocol is repeated. We use the same example as for the single-round case to demonstrate that the min-entropy can strictly decrease between rounds, indicating that more information is leaked in the multi-round case in general. * • 5.1 establishes the connection between MBQC and quantum causal models, by demonstrating that the conditional adaptation channels defined by gflow [44] satisfy the required commutation conditions of the channels defining a quantum causal model. * • The above connection allows us, via the min-entropy, to frame quantum causal discovery as a semi-definite program, thereby establishing a new tool in the quantum causal inference toolbox. We use this approach to numerically quantify the optimal causal inferential strategy for a given example. * • We demonstrate that in certain circumstances, not even the optimal strategy can learn anything about the causal structure such as when the possible causal structures are related by symmetries. 5.2 establishes this fact for when the causal structures are given by gflows that are related by symmetries of the underlying graph state. ## 2 Preliminaries This section introduces the background information regarding quantum combs and measurement-based quantum computation as required for the introduction of the comb min-entropy in Section 3 and the applications presented in Section 4 and Section 5. Some emphasis is placed on aspects of the discourse deemed important for the sequel, otherwise the aim is for brevity without sacrificing completeness. ### 2.1 Multi-round Quantum Protocols: Quantum Combs The standard framework for quantum information theory typically consists of density matrices, positive operator-valued measures (POVMs) and quantum channels (completely positive trace-preserving maps) to describe the processing of quantum information. Each of these can be defined for or act upon multiple quantum systems, can be composed, and moreover, can be composed on subsystems, such as when two quantum channels are composed over the domain of the latter and a subsystem of the co-domain of the former. Quantum networks, separated nodes with quantum processing capabilities and interconnected via quantum communication, can be modelled via an array of such compositions of quantum channels. In fact, it was realised [24, 23] that it is always possible to represent a quantum network in a canonical form, namely as a sequence of channels with memory, where some of the output of each channel is passed directly as input to the next, and some is available for arbitrary processing as depicted in Figure 1. Such a sequence comes with an explicit ordering on inputs and outputs of the network, from which the interpretation of interacting with a quantum system over multiple time steps can be naturally understood. Throughout this work, we use standard notation: $\mathcal{H}_{A}$ denotes a Hilbert space for system $A$, $\mathcal{L}(\mathcal{H}_{A})$ the space of linear operators on that space, $\rho_{A}$ a state of $\mathcal{H}_{A}$, $\mathcal{E}:\mathcal{L}(\mathcal{H}_{A})\rightarrow\mathcal{L}(\mathcal{H}_{B})$ for a CPTP map, and so on. Instead of explicitly working with a quantum network as a sequence of channels, it is convenient to work with the corresponding operator, a quantum comb, which is a generalisation of the Choi operator [55, 56] to a sequence of quantum channels. A quantum comb is an operator on a Hilbert space consisting of the tensor product of all input spaces $\mathcal{H}_{j}^{\operatorname{in}}$ and output spaces $\mathcal{H}_{j}^{\operatorname{out}}$. Each interaction with the system is associated to an input-output pair of Hilbert spaces, between which a quantum instrument can be inserted specifying the interaction. The distinction between input and output is indicated by superscripts in the notation, with time-steps indicated by subscripts. We take the following as our definition of quantum comb, which is sometimes presented as a theorem establishing the connection to quantum networks instead (see e.g. [[, Theorem 1,]]chiribella2008quantum, [[, Theorem 3,]]chiribella2009theoretical): ###### Definition 2.1. A quantum comb is a positive semi-definite operator $D\in\mathcal{L}\left(\bigotimes_{j=1}^{n}\mathcal{H}_{A_{j}^{\operatorname{in}}}\otimes\mathcal{H}_{A_{j}^{\operatorname{out}}}\right)$ for which there exists a sequence of positive semi-definite operators $D_{k}\in\mathcal{L}\left(\bigotimes_{j=1}^{k}\mathcal{H}_{A_{j}^{\operatorname{in}}}\otimes\mathcal{H}_{A_{j}^{\operatorname{out}}}\right)$, $k=0,...,n$ which satisfy $\displaystyle\operatorname{Tr}_{A_{k}^{\operatorname{out}}}\left[D_{k}\right]=I_{A_{k}^{\operatorname{in}}}\otimes D_{k-1}\quad\quad\forall k\in\\{1,...,n\\},$ (1) with $D_{n}=D$ and $D_{0}=1$. A comb $D$ is called unnormalised if $D_{0}$ instead satisfies the relaxed constraint $D_{0}>0$. The partial trace conditions in the above definition correspond to enforcing trace preservation of the individual component channels and positive semi- definiteness enforces complete positivity (for a more thorough explanation, see Section A.1 or [24, 23]). Since we also consider combs in this work that are classical, which is to say diagonal, it is useful to have an explicit definition for later reference: ###### Definition 2.2. A positive semi-definite operator $C\in\mathcal{L}\left(\bigotimes_{j=1}^{n}\mathcal{H}_{A_{j}^{\operatorname{in}}}\otimes\mathcal{H}_{A_{j}^{\operatorname{out}}}\right)$ is a classical comb if it can be written as $\displaystyle C=\sum_{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}f(\boldsymbol{a}^{\operatorname{out}},\boldsymbol{a}^{\operatorname{in}})\ket{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}\bra$ where $\\{\ket{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}\\}$ is an orthonormal basis for the Hilbert space and where $f$ is a non-negative function for which there exists a sequence of non-negative functions $f^{(k)}$, $k=0,...,n$ that satisfy, for all $j=2,...,n$: $\displaystyle\sum_{a_{j}^{\operatorname{out}}}f^{(j)}$ $\displaystyle(a_{1}^{\operatorname{out}},...,a_{j}^{\operatorname{out}},a_{1}^{\operatorname{in}},...,a_{j}^{\operatorname{in}})$ $\displaystyle=f^{(j-1)}(a_{1}^{\operatorname{out}},...,a_{j-1}^{\operatorname{out}},a_{1}^{\operatorname{in}},...,a_{j-1}^{\operatorname{out}})$ (2) with $f^{(n)}=f$ and $f^{(0)}:=\sum_{a_{1}^{\operatorname{out}}}f^{(1)}(a_{1}^{\operatorname{out}},a_{1}^{\operatorname{in}})>0$. If $f^{(0)}=1$ then the comb is normalised, otherwise it is unnormalised. The conditions described in Equation 2 are a simple rewriting of the trace conditions in Equation 1. In the case where the classical comb is normalised, it is most intuitive to write $f(\boldsymbol{a}^{\operatorname{out}},\boldsymbol{a}^{\operatorname{in}})$ as a conditional probability distribution $P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{a}^{\operatorname{in}})$ and the conditions of Equation 2 then detail independence conditions on this distribution under sequential marginalisation of the $a_{j}^{\operatorname{out}}$. Note that these distributions can be understood as observable quantities by using $C$ simply as a channel, i.e. by inputting all $\boldsymbol{a}^{\operatorname{in}}$ at once and collecting statistics of the outputs. Figure 1: A quantum comb is a representation of a quantum system evolving over multiple time-steps. An example of a comb $D$ (grey outline) is shown, corresponding to the sequence of quantum channels $\mathcal{D}_{1},\mathcal{D}_{2},\mathcal{D}_{3}$, each representing the evolution of the system between adjacent time-steps. Interaction with the system at a given time-step can also be represented by a quantum channel, here denoted $\mathcal{E}$. We use the notation $\operatorname{Comb}(A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}})$ to denote the set of all combs on $\bigotimes_{j=1}^{n}\mathcal{H}_{A_{j}^{\operatorname{in}}}\otimes\mathcal{H}_{A_{j}^{\operatorname{out}}}$, including unnormalised combs. It will be clear from context, whether we are considering classical combs only for a specific example. If a given input or output space is $1$-dimensional, the corresponding $A_{j}^{\operatorname{in}/\operatorname{out}}$ is replaced with $\mathbb{C}$. For example, we will often consider combs where the first input space is $1$-dimensional, that is, elements of $\operatorname{Comb}(\mathbb{C}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}})$. For later sections, particularly Section 5, it useful to use the following notation, which provides a convenient way of writing the contraction of a comb with a compatible operator (for example, for contracting $\mathcal{E}$ with the comb in Figure 1). ###### Definition 2.3 ([23]). For two operators $M\in\mathcal{L}(\mathcal{H}_{A}\otimes\mathcal{H}_{B})$ and $N\in\mathcal{L}(\mathcal{H}_{B}\otimes\mathcal{H}_{C})$, the link product is defined as: $\displaystyle M*N:=\operatorname{Tr}_{B}\left[(M^{T_{B}}\otimes I_{C})\cdot(I_{A}\otimes N)\right],$ (3) where $T_{B}$ denotes the partial transpose over the system $B$. (a) Classical-quantum comb $D$ (b) Multi-round classical-quantum comb $D^{(m)}$ Figure 2: In this work, we consider an extension to classical-quantum states, called classical-quantum combs, where a classical random variable $X$ indexes a set of quantum combs $\sigma_{x}$ (rather than quantum states). (a) The variable $X$ is considered to be unknown and as such is represented as an inaccessible system (upper part). The combs $\sigma_{x}$ are contingent on the value of $X$ and describe the dynamics of a system which can be interacted with (lower part) over a series of time-steps. By interacting with the accessible system, updated knowledge of $X$ can be obtained. (b) Some interactions are naturally modelled as a series of separate rounds, which are independent apart from the dependence on $X$. This is represented by taking each round to be a separate comb $\sigma_{x}^{(j)}$ which have no direct influence on later rounds. An example of such a multi-round interaction is the BQC protocol presented in Section 4. In this work, we consider the comb extension of classical-quantum states, called classical-quantum combs, which consist of a series of combs indexed by a classical random variable. Explicitly, we consider operators of the form $\displaystyle D=\sum_{x\in X}P(x)\ket{x}\\!\\!\bra{x}\otimes\sigma_{x}$ (4) which is an element in $\operatorname{Comb}(A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}},\mathbb{C}\rightarrow X)$ where $X$ is a (finite, discrete) random variable with outcomes defining an orthonormal basis $\\{\ket{x}\\}_{x}$ of the $|X|$-dimensional Hilbert space $\mathcal{H}_{X}$, and where the $\sigma_{x}$ are in combs in $\operatorname{Comb}(A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}})$. Note that, even though the state corresponding to $\mathcal{H}_{X}$ is written on the left in $D$, $\mathcal{H}_{X}$ is always taken to be the last output space of quantum comb, in order to be compatible with the comb min-entropy defined below. The appropriate perspective to take of $D$ for the remainder of this work, which is elaborated upon below, is that we aim to learn something about the classical state on $X$ by interacting only with the $\sigma_{x}$, i.e. we only have access to the spaces $\mathcal{H}_{A_{1}^{\operatorname{in}}},...,\mathcal{H}_{A_{n}^{\operatorname{out}}}$ and not $\mathcal{H}_{X}$. This is portrayed in the pictorial representation of $D$ in Figure 2(a), where the lower part of the comb is accessible and the upper part is not. It is important to note that, due to the special form that $D$ takes, it is both an element of $\operatorname{Comb}(A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}},\mathbb{C}\rightarrow X)$ and an element of $\operatorname{Comb}(\mathbb{C}\rightarrow X,A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}})$, as demonstrated by the proof of the following result (given in Section A.1). ###### Proposition 2.1. The operator $D$ as above is an element of both $\operatorname{Comb}(A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}},\mathbb{C}\rightarrow X)$ and $\operatorname{Comb}(\mathbb{C}\rightarrow X,A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}})$. This result is significant for our purposes since the latter inclusion highlights that there is no back-in-time signalling from the inputs of the $\sigma_{x}$ to the output at $X$ (cf. [[, Section III.C,]]chiribella2009theoretical), a fact that is explicitly used in Section 3.1. At times, we consider interactions with a system that occur in a series of rounds, where the evolution of the system between rounds is independent except for a common dependence on the parameter $X$ (this is represented in Figure 2(b)). For example, repeated instances of a multi-step cryptographic protocol can be considered in this way (see Section 4). Despite being able to represent these interactions using $D$ above, where independence of rounds is merely an extra constraint on the $\sigma_{x}$, we find it more convenient to use $\sigma_{x}$ to denote a single round, and write the multi-round operator as follows (the round number is indicated by the superscript in brackets): $\displaystyle D^{(m)}:=\sum_{x\in X}P(x)\ket{x}\\!\\!\bra{x}\otimes\bigotimes_{j=1}^{m}\sigma_{x}^{(j)}$ (5) which is an operator in $\operatorname{Comb}\left((A_{1}^{\operatorname{in}})^{(1)}\\!\rightarrow\\!(A_{1}^{\operatorname{out}})^{(1)},...,(A_{n}^{\operatorname{in}})^{(m)}\\!\rightarrow\\!(A_{n}^{\operatorname{out}})^{(m)},\mathbb{C}\\!\rightarrow\\!X\right)$. ### 2.2 MBQC and Gflow Measurement-based quantum computation [29, 30, 31, 32, 33] is a model for quantum computation distinct from the circuit picture, where, as the name indicates, the computation is driven by measurement operations rather than unitary ones. One can consider MBQC as consisting of three components: a highly entangled state, namely a graph state, which is the substrate that supports the computation; a sequence of single qubit projective measurements parametrised by an angle in a plane of the Bloch sphere, the positive outcomes of which encode the desired computation; and a correction method which ensures the desired computation is performed even when negative outcomes are obtained. The latter refers to the notion of gflow [44, 45] which plays a key role in the blind quantum computing protocol of [36] as well as in Section 4 and Section 5 below. In this subsection, each of these three components is introduced, with emphasis on the latter due to its importance in the sequel. Graph states take their name from the connection to mathematical graphs. Let $G$ be a (simple, connected) graph on vertex set $V=\\{1,...,n\\}$ and edge set $E$. One can define a graph state, denoted $\ket{G}$, as $\displaystyle\ket{G}:=\prod_{(i,j)\in E}CZ_{ij}\ket{+}^{\otimes n}$ (6) where each vertex of the graph is assigned a qubit in the $\ket{+}$ state and where each edge $(i,j)$ in the graph is associated to a controlled Pauli-$Z$ gate, denoted $CZ_{ij}$, between the corresponding qubits. Equivalently, $\ket{G}$ can be specified as the unique stabiliser state of the set of stabilisers [57] generated by $\displaystyle\left\\{K_{v}:=X_{v}\bigotimes_{v^{\prime}\in N_{v}^{G}}Z_{v^{\prime}}|v\in V\right\\}$ (7) where $N_{v}^{G}$ denotes the set of neighbours of $v$ in $G$. The stabiliser picture of graph states is a fruitful one since the correction of negative measurement outcomes relies upon the associated symmetries (see below). We will use the notation $\rho_{G}$ for $\ket{G}\\!\\!\bra{G}$. The single qubit measurements are restricted to projective measurements defined by states that lie in the intersection of the Bloch sphere with the $XY$-, $XZ$\- and $YZ$-planes. When it is necessary to specify a particular measurement, we denote the corresponding projectors as $\ket{+_{\alpha}}\\!\\!\bra{+_{\alpha}}_{\text{mp}}$ and $\ket{-_{\alpha}}\\!\\!\bra{-_{\alpha}}_{\text{mp}}$ where $\text{mp}\in\\{XY,XZ,YZ\\}$ and $\alpha\in[0,2\pi)$ specifies the angle from one of the axes in the plane. For example, $\ket{+_{\alpha}}_{XY}=\frac{1}{\sqrt{2}}(\ket{0}+e^{-i\alpha}\ket{1})$. If no measurement plane label is given, it should be understood as being an $XY$-plane measurement, as this is the most commonly treated type of measurement in this work. For a fixed graph state $\ket{G}$, the computation is specified by stipulating an angle $\alpha_{v}$ and measurement plane for each qubit/vertex $v$. The key observation for understanding why the restriction to only these planes is made, and also how the correction procedure given by gflow (outlined below) works, is the following set of relations, which hold for all values of $\alpha$: $\displaystyle\begin{split}\ket{-_{\alpha}}\\!\\!\bra{-_{\alpha}}_{XY}&=Z^{\dagger}\ket{+_{\alpha}}\\!\\!\bra{+_{\alpha}}_{XY}Z;\\\ \ket{-_{\alpha}}\\!\\!\bra{-_{\alpha}}_{XZ}&=(XZ)^{\dagger}\ket{+_{\alpha}}\\!\\!\bra{+_{\alpha}}_{XZ}XZ;\\\ \ket{-_{\alpha}}\\!\\!\bra{-_{\alpha}}_{YZ}&=X^{\dagger}\ket{+_{\alpha}}\\!\\!\bra{+_{\alpha}}_{YZ}X.\\\ \end{split}$ (8) Here, one should note that each of the associated unitaries ($X$, $Z$ or their product) appears naturally in the $K_{v}$ stabilisers for $\ket{G}$ (or a product thereof). One can thus interpret the occurrence of a negative outcome as equivalent to first applying a $Z$ (or $X$ or $XZ$) to the graph state and then receiving the correct (positive) measurement outcome. Applying such a gate to the graph state alone changes the computation, however applying the remainder of a stabiliser, of which the $Z$, $X$ or $XZ$ term is constituent, completes a symmetry of the graph state and so the negative measurement outcome is effectively transformed into a positive one. Since the evolution of a state to be measured can equivalently be considered as the evolution of the measurement observables themselves, we can in fact apply the remainder of the stabiliser to the other measurement operators, leaving the graph state untouched. This is significant, since, due to the form of the stabiliser operators and the properties of the measurement operators, the application of the stabiliser can be affected by an appropriate change to the measurement angle. For example, the measurement projectors in the $XY$-plane (the other measurement planes follow similar patterns), we have the following: $\displaystyle\begin{split}X^{\dagger}\ket{\pm_{\alpha}}\\!\\!\bra{\pm_{\alpha}}_{XY}X&\equiv\ket{\pm_{-\alpha\bmod 2\pi}}\\!\\!\bra{\pm_{-\alpha\bmod 2\pi}}_{XY};\\\ Z^{\dagger}\ket{\pm_{\alpha}}\\!\\!\bra{\pm_{\alpha}}_{XY}Z&\equiv\ket{\pm_{\alpha+\pi\bmod 2\pi}}\\!\\!\bra{\pm_{\alpha+\pi\bmod 2\pi}}_{XY}.\end{split}$ (9) So, applying the remainder of a stabiliser to such measurements amounts to applying appropriate reflections or adding $\pi$ phases to the original measurement angles. We thus have two equivalent perspectives for the correction of measurements, both of which play a role in this paper: corrections can either be implemented via applying (conditional) unitaries to the graph state or via adjustments to the classical measurement parameters. The latter perspective is most relevant for Section 4 where as the former is used in Section 5. Due to the importance of this equivalence for this work, a more thorough explanation is provided in Section A.2, which is particularly useful for motivating the discourse in Section 5. To summarise: computations on a graph state are specified by positive outcomes of single qubit measurements; measurements are restricted to measurement planes that exhibit useful symmetries; and negative outcomes are treated via conditional applications of certain symmetries of the graph state. To ensure deterministic computation, the appropriate symmetries are required for every measurement, and the conditionality of their application induces an order in which the measurements must be performed. However, for any given graph state, not every choice of symmetry and measurements order are consistent (i.e. one cannot adapt a measurement that has already been performed). The definition of gflow below addresses this point: a gflow assigns to every measured qubit a choice of corrective stabiliser and assigns a (partial) order of measurements which ensures consistency. ###### Definition 2.4 ([44]). Let $G=(V,E)$ be a graph, $I$ and $O$ be input and output subsets of $V$ respectively, and $\omega:O^{c}\rightarrow\\{XY,XZ,YZ\\}$ be a map assigning measurement planes to qubits (the superscript $c$ denotes set complement). The tuple $(G,I,O,\omega)$ has gflow if there exists a map $g:O^{c}\rightarrow\mathcal{P}(I^{c})$, where $\mathcal{P}$ denotes the powerset, and a partial order over $V$ such that the following hold for all $v\in O^{c}$: 1. 1. if $v^{\prime}\in g(v)$ and $v^{\prime}\neq v$, then $v<v^{\prime}$; 2. 2. if $v^{\prime}\in\operatorname{Odd}(g(v))$ and $v^{\prime}\neq v$, then $v<v^{\prime}$; 3. 3. if $\omega(v)=XY$, then $v\notin g(v)$ and $v\in\operatorname{Odd}(g(v))$; 4. 4. if $\omega(v)=XZ$, then $v\in g(v)$ and $v\in\operatorname{Odd}(g(v))$; 5. 5. if $\omega(v)=YZ$, then $v\notin g(v)$ and $v\notin\operatorname{Odd}(g(v))$; where $\operatorname{Odd}(K):=\\{\tilde{v}\in V:|N_{\tilde{v}}^{G}\cap K|=1\bmod 2\\}$ for any $K\subseteq V$. It is known that the presence of gflow is both necessary and sufficient for deterministic MBQC [[, Theorems 2 and 3,]]browne2007generalized. Furthermore, polynomial time algorithms exist for determining whether a given tuple $(G,I,O,\omega)$ supports gflow [58, 59]. Many distinct gflows for a given $(G,I,O,\omega)$ can exist, and characterising or counting all gflows for a given graph (as the input and output sets vary) in general remains an open problem. We can understand this definition in light of the discussion preceding it as follows. The map $g$ assigns a (product of) stabiliser(s) to each qubit being measured: $g(v)$ is a subset of $V$ and identifies the stabiliser $\displaystyle K_{g(v)}:=\prod_{w\in g(v)}K_{w}.$ Every element in the set $g(v)$ receives an $X$-correction from the above product and every element of $\operatorname{Odd}(g(v))$ receives a $Z$-correction; the vertices in their intersection will receive both. The partial order and the first two conditions enforce that every correction conditioned on the measurement at $v$ happens in the future of that measurement. The remaining three conditions enforce that the component of the product that acts on the qubit $v$ is precisely the required symmetry associated to the measurement plane $\omega(v)$, as given in Equation 8. Derived from the definition of gflow, it is useful to define, for each $v\in V$, the set of vertices whose corrections induce an $X$-operation on $v$ and the set whose correction induce a $Z$-operation: $\displaystyle\begin{split}\mathcal{X}_{v}&:=\\{v^{\prime}\in V:v\in g(v^{\prime})\setminus\\{v\\}\\};\\\ \mathcal{Z}_{v}&:=\\{v^{\prime}\in V:v\in\operatorname{Odd}(g(v^{\prime}))\setminus\\{v\\}\\}.\end{split}$ (10) We allow for $\mathcal{X}_{v}$ or $\mathcal{Z}_{v}$ to be empty (such as when $v\in I$ for example). ## 3 Quantum Comb Min-Entropy In this section, we introduce and motivate the primary tool of analysis used in the remainder of this work: the comb min-entropy [26]. This entropic quantity is an extension of the min-entropy for quantum states (see e.g. [60, 27, 61, 62]), a well-understood information-theoretic quantity with interpretations relating to maximal guessing probability of a random variable or the maximal singlet fraction achievable from a quantum state [27]. Below, we consider the operational meaning of the comb min-entropy, with particular emphasis on the cases relevant for subsequent sections, namely for combs of the forms given in Equations 4 and 5. We commence by stating the definition of the comb min-entropy via three equivalent characterisations, and then contextualise these in the subsequent discussion. ###### Definition 3.1 ([26]). Let $D\in\operatorname{Comb}(A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}})$. The min-entropy relative to $D$ is $\displaystyle H_{\min}$ $\displaystyle(A_{n}|A_{1},...,A_{n-1})_{D}$ $\displaystyle:=-\log\left[\min_{\Gamma}\min\\{\lambda\in\mathbb{R}:I_{A_{n}^{\operatorname{in}}A_{n}^{\operatorname{out}}}\otimes\lambda\Gamma\geq D\\}\right]$ (11) $\displaystyle\equiv-\log\left[\min_{\begin{subarray}{c}\widehat{\Gamma},\\\ I_{A_{n}^{\operatorname{in}}A_{n}^{\operatorname{out}}}\otimes\widehat{\Gamma}\geq D\end{subarray}}\frac{\operatorname{Tr}[\widehat{\Gamma}]}{\prod_{j=1}^{n-1}\dim A_{j}^{\operatorname{in}}}\right]$ (12) $\displaystyle\equiv-\log\left[\max_{E}\operatorname{Tr}\left[DE^{T}\right]\right]$ (13) where $\Gamma\in\operatorname{Comb}(A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n-1}^{\operatorname{in}}\rightarrow A_{n-1}^{\operatorname{out}})$ is a normalised comb, $\widehat{\Gamma}$ is an unnormalised comb over the same space, and $E$ ranges over elements of $\operatorname{Comb}(\mathbb{C}\rightarrow A_{1}^{\operatorname{in}},A_{1}^{\operatorname{out}}\rightarrow A_{2}^{\operatorname{in}},...,A_{n-1}^{\operatorname{out}}\rightarrow A_{n}^{\operatorname{in}}\otimes A_{n}^{\operatorname{out}})$. Each of the three quantities above are equivalent and each can be calculated via semi-definite programming methods. Despite their equivalence, we have chosen to present all three due to the following reasons. The first expression, Equation 11, is consistent with the definition given in [26] (cf. Definition $4$). The third expression as written has a nice interpretation in terms of the generalised Born rule: the term inside the logarithm is the maximum probability given by the Born rule for any comb dual to $D$, that is, a strategy for interacting with $D$. Equivalently, it can be interpreted as maximising over dual combs on a slightly different space that generate correlations with the output space of the comb $D$. This perspective is particularly useful when $D$ is a classical-quantum comb where the output is the variable we aim to learn about, and hence we elaborate upon it below (see also Figure 3(b)). The third expression is equivalent to the first expression via e.g., Theorem $2$ in [26], and reflects the strong duality in the associated semi-definite programs. The second expression is the form of the comb min-entropy that is most used throughout this paper for both analytical and numerical results (such as in Section 3.1, Section 4 and Section 5), due in part to its suitability for implementation in code. This expression is equivalent to the first by combining $\lambda$ and $\Gamma$ into one operator. In the present context, the expression for the min-entropy given in Equation 13 is particularly valuable conceptually . Above, $E$ is defined to be a dual operator on the exact same Hilbert spaces as $D$, since this provides a closed expression for the probability given by $\operatorname{Tr}\left[DE^{T}\right]$. As dual operators, the $E$ can be understood as multi-timestep, adaptive strategies for interacting with the system. However, by considering operators $\widehat{E}\in\operatorname{Comb}(\mathbb{C}\rightarrow A_{1}^{\operatorname{in}},A_{1}^{\operatorname{out}}\rightarrow A_{2}^{\operatorname{in}},...,A_{n-1}^{\operatorname{out}}\rightarrow A_{n}^{\operatorname{in}}\otimes\widehat{A}_{n}^{\operatorname{out}})$ instead, where $\mathcal{H}_{\widehat{A}_{n}^{\operatorname{out}}}\cong\mathcal{H}_{A_{n}^{\operatorname{out}}}$, the min-entropy can equivalently be understood as optimising over strategies $\widehat{E}$ that produce maximal correlations between two different spaces, namely $\mathcal{H}_{\widehat{A}_{n}^{\operatorname{out}}}$ and $\mathcal{H}_{A_{n}^{\operatorname{out}}}$. We find this perspective of maximising correlations between separate spaces especially appropriate for our purposes, since, for the classical-quantum combs considered here, the output space of $D$ (the classical parameter space) is deemed to be inaccessible, and we thus aim to learn the classical parameter via strategies with an output space available to us. To begin to formalise these equivalent perspectives, it can be shown (a combination of [[, Proposition 6,]]chiribella2016optimal and [[, Theorem 2,]]konig2009operational), that $\displaystyle\max_{E}$ $\displaystyle\operatorname{Tr}\left[DE^{T}\right]\equiv\max_{\widehat{E}}\dim A_{n}^{\operatorname{out}}\operatorname{Tr}\left[D\widehat{E}^{T}\ket{\Phi^{+}}\\!\\!\bra{\Phi^{+}}\right]$ (14) where $T$ is to be understood as denoting the transpose over spaces that $D$ and $E$ (respectively $D$ and $\widehat{E}$) share, and where $\displaystyle\ket{\Phi^{+}}:=\frac{1}{\sqrt{\dim A_{n}^{\operatorname{out}}}}\sum_{i}\ket{i}_{A_{n}^{\operatorname{out}}}\ket{i}_{\widehat{A}_{n}^{\operatorname{out}}}$ (15) with $\\{\ket{i}\\}$ an orthonormal basis. The term $\operatorname{Tr}_{A_{1}^{\operatorname{in}},A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}}[D\widehat{E}^{T}]$ is in particular a state on $\mathcal{H}_{A_{n}^{\operatorname{out}}}\otimes\mathcal{H}_{\widehat{A}_{n}^{\operatorname{out}}}$ and thus the full trace term on the right-hand side of Equation 14 is a measure of the overlap with the maximally-entangled state $\ket{\Phi^{+}}\\!\\!\bra{\Phi^{+}}$. Maximising over $\widehat{E}$ amounts to maximising this overlap. In the case where $D$ is a bipartite quantum state, $\widehat{E}$ is the Choi representation of a CPTP map so Equation 14 is directly related to the operational meaning of the state min-entropy (see the proof of [[, Theorem 2,]]konig2009operational). For any comb with classical output, $\ket{\Phi^{+}}\\!\\!\bra{\Phi^{+}}$ can be replaced by its diagonal part $(\dim A_{n}^{\operatorname{out}})^{-1}\sum_{i}\ket{ii}\\!\\!\bra{ii}_{A_{n}^{\operatorname{out}}\widehat{A}_{n}^{\operatorname{out}}}$. Figure 3 depicts this perspective for both the state and comb cases. Throughout this work, when discussing the maximum formulation of the min- entropy, we will often identify $E$ with $\widehat{E}$ for simplicity (they are in one-to-one correspondence) despite the abuse of notation regarding the constituent Hilbert spaces. (a) Min-entropy of a quantum state $\rho_{XA}$ (b) Comb min-entropy of a comb $D$ Figure 3: The min-entropy can be considered as a measure of the maximal amount of correlation achievable from a given state of two systems by interacting with one system alone. (a) The state min-entropy is calculated if the interaction occurs at a single time step. For a quantum state $\rho_{XA}$, the min-entropy is calculated via maximising over all quantum channels $\mathcal{E}$ acting on the system $A$, and comparing the output to a maximally-entangled state $\ket{\Phi}\\!\\!\bra{\Phi}_{XX^{\prime}}$. If $\rho_{XA}$ is a classical-quantum state instead, one simply maximises over POVMs and compares the output to a maximally classically-correlated state (namely $(\dim X)^{-1}\sum_{x}\ket{xx}\\!\\!\bra{xx}_{X^{\prime}X}$). (b) If interactions occur over a series of time steps, the comb min-entropy is calculated instead. For a quantum (respectively, classical-quantum) comb $D$, the min-entropy is calculated via maximising over strategies $\widehat{E}$ (dual combs to $D$) and again comparing the output to a maximally-entangled (maximally classically-correlated) state. The comb min-entropy given for $D$ a classical-quantum as depicted in Figure 2 are the primary concern of this work. Let us now consider more closely the classical-quantum combs as presented in Equation 4 and Equation 5. For classical-quantum states, i.e. for $D=\sum_{x}P(x)\ket{x}\\!\\!\bra{x}\otimes\rho_{x}^{A}$ where $\rho_{x}^{A}\in\mathcal{L}(\mathcal{H}_{A})$, the term $\max_{E}\operatorname{Tr}[DE^{T}]$ is called the guessing probability [27] since it quantifies the best probability of guessing $X$ after interacting with the state on $A$. We use the same terminology for the combs case: $P_{\text{guess}}(X|A_{1},...,A_{n})_{D}$ is the probability of guessing $X$ after interacting with $A_{1}^{\operatorname{in}},A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}},A_{n}^{\operatorname{out}}$ as given by $\displaystyle P_{\text{guess}}(X|A_{1},...,A_{n})_{D}:=\max_{E}\sum_{x}P(x)\operatorname{Tr}[\left(\ket{x}\\!\\!\bra{x}\otimes\sigma_{x}\right)E^{T}]$ where $E$ ranges over elements of $\operatorname{Comb}(\mathbb{C}\rightarrow A_{1}^{\operatorname{in}},A_{1}^{\operatorname{out}}\rightarrow A_{2}^{\operatorname{in}},...,A_{n}^{\operatorname{out}}\rightarrow X)$. Since $P_{\text{guess}}(X|A_{1},...,A_{n})_{D}=2^{-H_{\min}(X|A_{1},...,A_{n})_{D}}$, we will often use the terms ‘guessing probability’ and ‘min-entropy’ interchangeably. Later sections will compare the comb min-entropy across different numbers of rounds of interactions, i.e. comparing the min-entropy for $D^{(m)}$ for different values of $m$. Intuitively, given the specific structure of $D^{(m)}$, one would expect that as the number of rounds increases, the min- entropy should not increase. This is indeed the case, as stated in the following lemma whose proof is given in Appendix B. ###### Lemma 3.1. For $D^{(m)}$ and $D^{(l)}$ of the form given in Equation 5, where $m\geq l$, $\displaystyle H_{\min}(X|A_{1}^{(1)},...,A_{n}^{(l)})_{D^{(l)}}\geq H_{\min}(X|A_{1}^{(1)},...,A_{n}^{(m)})_{D^{(m)}}.$ For emphasis, we reiterate comments made above: throughout this work, we take the perspective that the min-entropy relates to the maximal guessing probability under the optimal strategy of interacting with a comb, which we find more intuitive, however calculations will largely be done via the expressions involving the minimum (typically Equation 12), which is easier to work with since the optimisation is over fewer Hilbert spaces. ### 3.1 Min-Entropy for Classical-Classical Combs In the previous subsection, we considered the relation between the comb min- entropy and guessing probability for classical-quantum combs $D$. Due to their relevance for the analysis that will take place in Section 4, we present here a general result regarding the guessing probability for a further restricted class of combs: classical-classical combs. The result is 3.2 and in particular provides upper and lower bounds for the guessing probability. Classical-classical are combs of the form $\displaystyle D=\sum_{x}P(x)\ket{x}\\!\\!\bra{x}\otimes\sigma_{x}$ (16) where $D\in\operatorname{Comb}(A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}},\mathbb{C}\rightarrow X)$ and where all the $\sigma_{x}\in\operatorname{Comb}(A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}})$ are normalised and diagonal in the same basis (cf. 2.2). Due to normalisation, we can adopt the probability distribution notation and write $\displaystyle\sigma_{x}:=\sum_{x,\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}P(\boldsymbol{a}^{\operatorname{out}}|x,\boldsymbol{a}^{\operatorname{in}})\ket{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}\\!\\!\bra{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}$ (17) where $P(\boldsymbol{a}^{\operatorname{out}}|x,\boldsymbol{a}^{\operatorname{in}})$ satisfies the required marginalisation conditions (Equation 2) for each $x$. This allows $D$ to be rewritten as $\displaystyle D$ $\displaystyle=\sum_{x}P(x)\ket{x}\\!\\!\bra{x}\otimes\sum_{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}P(\boldsymbol{a}^{\operatorname{out}}|x,\boldsymbol{a}^{\operatorname{in}})\ket{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}\\!\\!\bra{..}$ (18) $\displaystyle=\sum_{x,\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}P(x)P(\boldsymbol{a}^{\operatorname{out}}|x,\boldsymbol{a}^{\operatorname{in}})\ket{x,\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}\\!\\!\bra{..}.$ (19) In the case where all input spaces are trivial, that is, when $D$ is a classical-classical state (all the $\sigma_{x}$ are states that are diagonal in the same basis), the guessing probability is known to have a nice interpretation in terms of a maximal Bayesian update (this interpretation is explicated in Appendix B; see also [[, Section 6.1.4,]]tomamichel2015quantum). In an attempt to arrive at similar interpretation for the case where the inputs are non-trivial, let us consider the conditional version of Bayes’ rule, namely: $\displaystyle P(X|\boldsymbol{A}^{\operatorname{in}},\boldsymbol{A}^{\operatorname{out}})P(\boldsymbol{A}^{\operatorname{out}}|\boldsymbol{A}^{\operatorname{in}})=P(\boldsymbol{A}^{\operatorname{out}}|X,\boldsymbol{A}^{\operatorname{in}})P(X|\boldsymbol{A}^{\operatorname{in}})$ (20) where the capitalised letters denote random variables (as opposed to their values denoted by the lower case letters as above). One notes that the right- hand side is almost what appears in the form of $D$ above, except instead of $P(X)$ we have $P(X|\boldsymbol{A}^{\operatorname{in}})$. A consequence of 2.1 is that $X$ is independent of $\boldsymbol{A}^{\operatorname{in}}$, so we do in fact have $\displaystyle P(X|\boldsymbol{A}^{\operatorname{in}})=P(X).$ (21) This allows $D$ to be written as $\displaystyle D=\sum_{\begin{subarray}{c}x,\boldsymbol{a}^{\operatorname{in}},\\\ \boldsymbol{a}^{\operatorname{out}}\end{subarray}}P(x|\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}})P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{a}^{\operatorname{in}})\ket{x,\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}\\!\\!\bra{..}.$ (22) Using this form for $D$, we can establish the following: ###### Proposition 3.2. Let $D$ be as above. Then $\displaystyle\sum_{\begin{subarray}{c}\boldsymbol{a}^{\operatorname{in}},\\\ \boldsymbol{a}^{\operatorname{out}}\end{subarray}}\max_{x}P(x|\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}})$ $\displaystyle P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{a}^{\operatorname{in}})\leq\min_{\begin{subarray}{c}\Gamma,\\\ I_{X}\otimes\Gamma\geq D\end{subarray}}\operatorname{Tr}\left[\Gamma\right]$ (23) $\displaystyle\leq\sum_{\begin{subarray}{c}\boldsymbol{a}^{\operatorname{in}},\\\ \boldsymbol{a}^{\operatorname{out}}\end{subarray}}\max_{x,\boldsymbol{\tilde{a}}}P(x|\boldsymbol{\tilde{a}},\boldsymbol{a}^{\operatorname{out}})P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{\tilde{a}})$ (24) where the minimisation is over unnormalised combs $\Gamma$. By normalising each term by the dimension of the input spaces, and in light of the definition of the comb min-entropy(specifically Equation 12), the above provides bounds for the guessing probability (this is discussed in more depth below). The proof is given in Appendix B and demonstrates the left-hand inequality by showing that every positive semi-definite operator $\Gamma$, which clearly includes all unnormalised combs that satisfies $I_{X}\otimes\Gamma\geq D$, must have trace greater than the left-hand quantity above. The right-hand inequality is obtained by showing that $\displaystyle\Gamma:=I_{\boldsymbol{A}^{\operatorname{in}}}\otimes\sum_{\boldsymbol{a}^{\operatorname{out}}}\max_{x,\boldsymbol{\tilde{a}}}P(x|\boldsymbol{\tilde{a}},\boldsymbol{a}^{\operatorname{out}})P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{\tilde{a}})\ket{\boldsymbol{a}^{\operatorname{out}}}\bra{\boldsymbol{a}^{\operatorname{out}}}$ (25) is indeed an unnormalised classical comb. It also is worth noting that, for $D$ classical, the minimum over unnormalised combs can always be achieved by a diagonal $\Gamma$ (see B.1 and B.2). Explicitly, the consequences for the guessing probability are as follows: $\displaystyle P_{\text{guess}}$ $\displaystyle(X|\boldsymbol{A}^{\operatorname{in}},\boldsymbol{A}^{\operatorname{out}})\geq\sum_{\begin{subarray}{c}\boldsymbol{a}^{\operatorname{in}},\\\ \boldsymbol{a}^{\operatorname{out}}\end{subarray}}\max_{x}P(x|\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}})\frac{P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{a}^{\operatorname{in}})}{\dim\boldsymbol{A}^{\operatorname{in}}}$ (26) $\displaystyle\equiv\sum_{\begin{subarray}{c}\boldsymbol{a}^{\operatorname{in}},\\\ \boldsymbol{a}^{\operatorname{out}}\end{subarray}}\max_{x}P(x|\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}})P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{a}^{\operatorname{in}})P_{\text{unif}}(\boldsymbol{a}^{\operatorname{in}})$ (27) where $P_{\text{unif}}(\cdot)$ denotes the uniform probability distribution, and $\displaystyle P_{\text{guess}}(X|\boldsymbol{A}^{\operatorname{in}},\boldsymbol{A}^{\operatorname{out}})$ $\displaystyle\leq\sum_{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}\max_{x,\boldsymbol{\tilde{a}}}P(x|\boldsymbol{\tilde{a}},\boldsymbol{a}^{\operatorname{out}})\frac{P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{\tilde{a}})}{\dim\boldsymbol{A}^{\operatorname{in}}}$ (28) $\displaystyle=\sum_{\boldsymbol{a}^{\operatorname{out}}}\max_{x,\boldsymbol{\tilde{a}}}P(x|\boldsymbol{\tilde{a}},\boldsymbol{a}^{\operatorname{out}})P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{\tilde{a}}).$ (29) If, for each $\boldsymbol{a}^{\operatorname{out}}$, the factor $\max_{x}P(x|\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}})P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{a}^{\operatorname{in}})$ is the same for every $\boldsymbol{a}^{\operatorname{in}}$, then the upper and lower bounds coincide, thus specifying the guessing probability exactly. As discussed in Appendix B, this occurs trivially for the state case since the input space is one-dimensional, but we will see in Section 4.3 a simple comb example with non-trivial input spaces where this also holds. The above proposition highlights that, even in the entirely classical case, the generalisation of the maximum Bayesian updating interpretation of the guessing probability for states to the classical comb case is more subtle. This aligns with what one would expect: the fact that the comb case allows for inputs to the system plays a non-trivial role in the interpretation of the guessing probability. For the case of quantum combs, further subtleties likely exist, not in the least since certain formulations of conditional quantum Bayesian updating are known to be potentially problematic (see e.g. [[, Section VII,]]leifer2013towards). A final comment regarding the multi-round case. Since a classical multi-round comb (recall Equation 5) can be viewed as a special case of Equation 19 above with extra independence conditions on the distribution $P(\boldsymbol{a}^{\operatorname{out}}|x,\boldsymbol{a}^{\operatorname{in}})$, namely that $\displaystyle P(\boldsymbol{a}^{\operatorname{out}}|x,\boldsymbol{a}^{\operatorname{in}})=\prod_{j=1}^{m}P(\boldsymbol{a}^{\operatorname{out},(j)}|x,\boldsymbol{a}^{\operatorname{in},(j)})$ (30) where $\boldsymbol{a}^{\operatorname{in},(j)}$ denotes the inputs for round $j$ and similarly for $\boldsymbol{a}^{\operatorname{out},(j)}$, applying the results of 3.2 to the multi-round case gives $\displaystyle P_{\text{guess}}(X|\boldsymbol{A}^{\operatorname{in},(1)},\boldsymbol{A}^{\operatorname{out},(1)},...,\boldsymbol{A}^{\operatorname{in},(m)},\boldsymbol{A}^{\operatorname{out},(m)})$ $\displaystyle\geq\sum_{\begin{subarray}{c}\boldsymbol{a}^{\operatorname{in},(1:m)},\\\ \boldsymbol{a}^{\operatorname{out},(1:m)}\end{subarray}}\max_{x}P(x|\boldsymbol{a}^{\operatorname{in},(1:m)},\boldsymbol{a}^{\operatorname{out},(1:m)})\prod_{j=1}^{m}\frac{P(\boldsymbol{a}^{\operatorname{out},(j)}|\boldsymbol{a}^{\operatorname{in},(j)})}{\dim\boldsymbol{A}^{\operatorname{in},(j)}}$ (31) and $\displaystyle P_{\text{guess}}(X|\boldsymbol{A}^{\operatorname{in},(1)},\boldsymbol{A}^{\operatorname{out},(1)},...,\boldsymbol{A}^{\operatorname{in},(m)},\boldsymbol{A}^{\operatorname{out},(m)})$ $\displaystyle\leq\sum_{\boldsymbol{a}^{\operatorname{out},(1:m)}}\max_{x,\boldsymbol{\tilde{a}}^{(1:m)}}P(x|\boldsymbol{\tilde{a}}^{(1:m)},\boldsymbol{a}^{\operatorname{out},(1:m)})\prod_{j=1}^{m}P(\boldsymbol{a}^{\operatorname{out},(j)}|\boldsymbol{\tilde{a}}^{(j)})$ (32) where the shorthand $\boldsymbol{a}^{\operatorname{in},(1:m)}$ has been used for the inputs for all rounds and similarly for $\boldsymbol{a}^{\operatorname{out},(j)}$. ## 4 Blind Quantum Computing Protocol Analysis Blind quantum computing is a term that encompasses an array of cryptographic quantum computational protocols (see e.g. [35, 36, 37, 38] and the review [17]), many of which build upon, or are inspired by, aspects derived from MBQC. In very broad terms, a BQC protocol consists of a client, who has limited computational power, interacting with a server (multi-party variants also exist), who has quantum computational capabilities, in order to carry out a desired computation in such a way that the latter is “blind” to the details. A range of theoretical tools for modelling quantum cryptographic protocols and analysing their security exist (see e.g. [64, 65] and the review [66]). In this work, we consider one specific BQC protocol, that of Mantri et al. [36] (discussed in the next subsection), and our main method of security analysis consists of calculating the comb min-entropy applied to the corresponding classical comb, $D_{\operatorname{client}}$, which we define below (Section 4.2). In Section 4.3, by establishing a strictly positive lower bound for the min-entropy of $D_{\operatorname{client}}$, which means the server must still learn more information in order to know the client’s choice of computation with certainty, we prove the partial security of the protocol for a single round. We further provide a simple example of the protocol run on a three-qubit graph state that obtains the lower bound. In Section 4.4, we extend the security analysis to multiple rounds by considering a comb $D_{\operatorname{client}}^{(m)}$, where a positive lower bound is again shown for all rounds $m$, and comment on the consequences for the blindness of the protocol in this case. ### 4.1 A Classically Driven BQC Protocol The protocol of [36] is distinct from many other BQC proposals in that it considers purely classical communication between client and server. At its core, the protocol leverages various properties of gflow and MBQC outlined in Section 2.2, such as the non-uniqueness of gflows existing for a given graph state and the symmetries of measurement projections as in Equation 9, along with cryptographic primitives such as the use of random bits as one-time pads. The protocol parameters consist of a choice of graph $G$ along with a total order on vertices, and a discrete set of angles $\mathcal{A}$ which satisfies the following property: $\displaystyle\mathcal{A}=\\{(-1)^{x}\alpha+z\pi\bmod 2\pi:\alpha\in\mathcal{A};x,z\in\mathbb{Z}_{2}\\}.$ (33) All measurements are made in the $XY$-plane (which is no restriction on the universality of the resulting computations - see [67]) and so the above property enforces that $\mathcal{A}$ is closed under the angle transformations given in Equation 9. The specific graph, total order and angle set are agreed upon collectively by both the client and server prior to the commencement of a specific computation. In secret, the client also chooses their desired computation, that is, a list of measurement angles $\boldsymbol{\alpha}\in\mathcal{A}^{n}$ (where $n$ is the number of vertices in $G$) and designated input and output sets $I$ and $O$, a bit-string one- time pad $\boldsymbol{r}\in\mathbb{Z}_{2}^{n}$ uniformly at random, and a gflow compatible with $(G,I,O)$ and the total order. One round of computation proceeds as follows: 1. 1. The server initialises the graph state $\rho_{G}$. 2. 2. For $i=1,...,n$ according to the total order, the following sequence is repeated: 1. (a) The user reports a measurement angle $\alpha^{\prime}_{i}$ to the server, where $\displaystyle\alpha^{\prime}_{i}:=(-1)^{\bigoplus_{j\in\mathcal{X}_{i}}c_{j}}\alpha_{i}+\left(r_{i}\oplus\bigoplus_{j\in\mathcal{Z}_{i}}c_{j}\right)\pi\bmod 2\pi$ (34) where $c_{j}$ denotes the measurement outcome for qubit $j<i$ recorded by the user based on the outcome reported by the server and where $\mathcal{X}_{i}$ and $\mathcal{Z}_{i}$ are the corrections sets defined by the gflow (as in Equation 10). 2. (b) The server measures $\mathcal{M}_{\alpha^{\prime}_{i}}$ and reports $c^{\prime}_{i}=0$ for a positive outcome and $c^{\prime}_{i}=1$ for a negative outcome. 3. (c) The user records $c_{i}=c^{\prime}_{i}\oplus r_{i}$. 3. 3. The outcomes pertaining to the output qubits (which are known only to the user) are processed to obtain the results of the computation. There are two things worth noting about the one-time pads. Firstly, their utility for obscuring angles is a direct consequence of the $Z$-relation given in Equation 9 (i.e. a positive measurement outcome for $\alpha$ and negative measurement outcome for $\alpha+\pi$ are equivalent). Secondly, the presence of $r_{i}$ in both the equation for $\alpha^{\prime}_{i}$ as well as for any $\alpha^{\prime}_{k}$ for which $i\in\mathcal{X}_{k}$ or $i\in\mathcal{Z}_{k}$ places constraints on the set of possible reported angles $\boldsymbol{\alpha^{\prime}}$ for a given choice of true angles $\boldsymbol{\alpha}$. The protocol is known to be correct if both the client and server behave accordingly [[, Theorem 1,]]mantri2017flow, however it also known that this protocol is not verifiable: the client has no way of knowing whether the server actually prepares a graph state, measures according to the reported angles, and communicates the actual measurement outcomes. The server could in fact do none of these and report random outcomes $c^{\prime}$ instead, with the client only being aware of this fact if they could compute the measurement statistics of the final state themselves anyway. We review the proof of blindness for a single round of the protocol [[, Theorem 2,]]mantri2017flow and provide our own analysis in Section 4.3 after introducing the classical comb that models the protocol. Some final remarks regarding minor differences between the analysis in [36] and the one below. Mantri et al. place a further condition on the definition of gflow, largely for the purpose of simplifying a counting argument (see [[, Theorem 3,]]mantri2017flow), which effectively singles out one gflow for every pair $(I,O)$ given $G$ and they thus identify a choice of computation with a choice of angles and choice of gflow. Here, we work with the unconstrained definition of gflow and so identify a choice of computation as a choice of $\boldsymbol{\alpha}$ and $(I,O)$ (again for fixed $G$), for which any of the compatible gflows can be chosen. Moreover, since the protocol is entirely classical, the client can only prepare an input state on the qubits in $I$ via measurement-based state preparation, which is thus indistinguishable from the part of the measurement sequence implementing the unitary. As such, computations are specified here purely by angles $\boldsymbol{\alpha}$ and a choice of output set $O$ for which $(G,I,O)$ supports gflow for some $I$. ### 4.2 The Classical BQC Protocol as a Classical Comb The protocol parameters outlined in the previous section also form the starting point for defining the corresponding classical comb: we fix a choice of graph $G$, say on $n$ vertices, a total order on the vertices $1<2<...<n$, and a set of allowed angles $\mathcal{A}$ satisfying Equation 33. Since all communication in the protocol is classical, we represent the reported angles and measurement outcomes as basis states in corresponding Hilbert spaces: $\ket{\alpha^{\prime}_{i}}\bra{\alpha^{\prime}_{i}}\in\mathcal{H}_{A^{\prime}_{i}}$ for the reported angle at step $i$ and $\ket{c^{\prime}_{i}}\bra{c^{\prime}_{i}}\in\mathcal{H}_{C^{\prime}_{i}}$ for the reported measurement outcome, where $\dim\mathcal{H}_{A^{\prime}_{i}}=|\mathcal{A}|$ and $\dim\mathcal{H}_{C^{\prime}_{i}}=2$. Prior to the commencement of the protocol, the client selects the output space $O$, the “true” angles for the computation $\boldsymbol{\alpha}\in\mathcal{A}^{n}$, the one-time pads $\boldsymbol{r}\in\mathbb{Z}_{2}^{n}$, and a gflow $g$ compatible with the graph and total order and chosen output set. We denote compatibility with an output set $O$ as $g\sim O$. We define the classical comb corresponding to the protocol for the specific choice of $\boldsymbol{\alpha},\boldsymbol{r}$ and $g$ as $\displaystyle\sigma_{\operatorname{BQC}}^{\boldsymbol{\alpha},\boldsymbol{r},g}:=\sum_{\begin{subarray}{c}\boldsymbol{\alpha^{\prime}}\in\mathcal{A}^{n},\\\ \boldsymbol{c^{\prime}}\in\mathbb{Z}_{2}^{n}\end{subarray}}P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)\ket{\boldsymbol{\alpha^{\prime}}\boldsymbol{c^{\prime}}}\bra{\boldsymbol{\alpha^{\prime}}\boldsymbol{c^{\prime}}}$ (35) where $P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)$ is simply a deterministic distribution that encodes the angle adaptations: $\displaystyle P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)=\begin{cases}1,\text{ if \lx@cref{creftypecap~refnum}{eq:adapt_meas_angles} holds for all $i$}\\\ 0,\text{ otherwise }\end{cases}.$ (36) Note that Equation 34 contains the notation $c_{i}$, but since $c_{i}:=c^{\prime}_{i}\oplus r_{i}$, we use only the $\boldsymbol{c^{\prime}}$ and $\boldsymbol{r}$ notation instead. For a given computation, that is, for a given $\boldsymbol{\alpha}$ and $O$, any choice of $g\sim O$ and $\boldsymbol{r}$ produce the desired result. Since these choices are made in secret, the protocol for a given computation is modelled as $\displaystyle\sum_{g\sim O,\boldsymbol{r}}P(g|O)P(\boldsymbol{r})\sigma_{\operatorname{BQC}}^{\boldsymbol{\alpha},\boldsymbol{r},g}$ (37) where $P(g|O)$ is the probability that the gflow $g$ is chosen and $P(\boldsymbol{r})$ is the probability of the one-time pad $\boldsymbol{r}$ is chosen. According to the above protocol, the latter distribution is taken to be uniform, $P(\boldsymbol{r})=\frac{1}{2^{n}}$, and independent of all other variables (see [[, Lemma 4,]]mantri2017flow). Since the specific choice of gflow is irrelevant for the computation once $O$ is fixed, we take $P(g|O)$ to have uniform weight for all $g\sim O$ and $0$ otherwise, and moreover we take $P(g|O)=P(g|\boldsymbol{\alpha},O)$. We thus have the following: $\displaystyle P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)$ $\displaystyle=P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g,O)\quad\forall g\sim O$ (38) $\displaystyle P(\boldsymbol{r},g|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},O)$ $\displaystyle=P(\boldsymbol{r})P(g|O)$ (39) where Equation 39 has used the independence of $g$ and $\boldsymbol{r}$ from $\boldsymbol{c^{\prime}}$ which follows from similar reasoning as for Equation 21 in the previous section. We can thus expand Equation 37 as follows: $\displaystyle\sum_{g\sim O,\boldsymbol{r}}$ $\displaystyle P(g|O)P(\boldsymbol{r})\sigma_{\operatorname{BQC}}^{\boldsymbol{\alpha},\boldsymbol{r},g}$ $\displaystyle=\sum_{\begin{subarray}{c}\boldsymbol{\alpha^{\prime}},\boldsymbol{c^{\prime}},\\\ g\sim O,\boldsymbol{r}\end{subarray}}P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g,O)P(\boldsymbol{r},g|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},O)\ket{\boldsymbol{\alpha^{\prime}}\boldsymbol{c^{\prime}}}\\!\\!\bra{..}$ (40) $\displaystyle=\sum_{\boldsymbol{\alpha^{\prime}},\boldsymbol{c^{\prime}}}P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},O)\ket{\boldsymbol{\alpha^{\prime}}\boldsymbol{c^{\prime}}}\\!\\!\bra{\boldsymbol{\alpha^{\prime}}\boldsymbol{c^{\prime}}}$ (41) $\displaystyle=:\sigma_{\boldsymbol{\alpha},O}$ (42) where we have used $\bra{..}$ when its contents are the same as the accompanying ket. Having summed over $g\sim O$ and $\boldsymbol{r}$, the probability distribution $P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},O)$ is no longer a simple deterministic distribution, but in general a more complicated one. In particular, it depends on the set of gflows compatible with $O$ and the set of angles they can mutually report. The final step to arrive at a comb of the form of Equation 4 is to include the random variable. In this case, the random variable encodes the choice of computation made by the client. Denoting the set of all possible output sets for the fixed graph by $\mathcal{O}$, the values that the computation random variable can take are given by $\ket{\boldsymbol{\alpha},O}\\!\\!\bra{\boldsymbol{\alpha},O}\in\mathcal{H}_{\mathcal{A}^{n}}\otimes\mathcal{H}_{\mathcal{O}}$. We can now write down the classical comb that fully represents the BQC protocol for a fixed $G$ and $\mathcal{A}$ as: $\displaystyle D_{\operatorname{client}}=\sum_{\boldsymbol{\alpha}\in\mathcal{A}^{n},O\in\mathcal{O}}$ $\displaystyle P(\boldsymbol{\alpha},O)\ket{\boldsymbol{\alpha},O}\\!\\!\bra{\boldsymbol{\alpha},O}\otimes\sigma_{\boldsymbol{\alpha},O}.$ (43) This is a classical comb in $\operatorname{Comb}(\mathbb{C}\rightarrow A^{\prime}_{1},C^{\prime}_{1}\rightarrow A^{\prime}_{2},...,C^{\prime}_{n}\rightarrow\mathbb{C},\mathbb{C}\rightarrow\boldsymbol{A}\times\boldsymbol{O})$ and as such the results of Section 3.1 regarding the min-entropy apply, a fact that will be used in the blindness proof in the next section. ### 4.3 Blindness in a Single Round We begin this subsection with a discussion of the security proof given in [36] for the protocol. The proof proceeds by showing that, for a single round, the mutual information between the information received by the server and the client’s choice of computation is bounded away from the required amount to learn the computation with certainty. We state their blindness theorem below, using notation as close to the original as possible, and then consider a related form more suitable for comparison with our analysis using the min- entropy. Recalling the comments at the end of Section 4.1, since in [36] a restricted version of gflow was used, a computation is stipulated by a choice of angles and a choice of restricted gflow (instead of the choice of $O$ as above). We let $\boldsymbol{A}$ denote the random variable for the angles, which takes values in $\mathcal{A}^{n}$, and let $\boldsymbol{F}$ denote the random variable which takes values in the set of restricted gflows. The random variables for the reported angles and measurement outcomes are denoted $\boldsymbol{A^{\prime}}$ and $\boldsymbol{C^{\prime}}$ respectively. With this notation, the blindness theorem is: ###### Theorem 4.1 (Theorem 2, [36]). In a single instance of the protocol, the mutual information between the client’s secret input $\\{\boldsymbol{\alpha},\boldsymbol{f}\\}$ and the information received by the server is bounded by $\displaystyle I(\boldsymbol{C^{\prime}},\boldsymbol{A^{\prime}};\boldsymbol{A},\boldsymbol{F})\leq H(\boldsymbol{A^{\prime}})$ (44) where $I(\cdot;\cdot)$ denotes the mutual information and $H(\cdot)$ denotes the Shannon entropy. Using part of the proof of the above theorem which in particular shows that $H(\boldsymbol{A^{\prime}},\boldsymbol{C^{\prime}}|\boldsymbol{A},\boldsymbol{F})\geq n$ (see [[, Lemma 4,]]mantri2017flow) along with the properties of the mutual information and Shannon entropy, it is possible to arrive at a related entropic inequality which indicates the at least partial security of the protocol (‘partial’ because some information is leaked about the choice of angles in a given round): $\displaystyle H(\boldsymbol{A},\boldsymbol{F}|\boldsymbol{C^{\prime}},\boldsymbol{A^{\prime}})\geq\log_{2}|\boldsymbol{F}|>0.$ (45) One important aspect that we want to emphasise is that the Shannon entropy is not a single-shot entropy and therefore only provides limited insight about the security of the protocol in a single round. The min-entropy however, is a single-shot quantity, and since it deals with optimal learning strategies (optimal from the inquisitive server’s perspective), it provides a more appropriate measure of security in this case: if it can be shown that the min- entropy equivalent of the quantity $H(\boldsymbol{A},\boldsymbol{F}|\boldsymbol{C^{\prime}},\boldsymbol{A^{\prime}})$ is bounded above zero, then the server is guaranteed to retain some uncertainty about the true computation. This is the content of the following theorem. It should first be noted that, in deriving Equation 45, the assumption is made that the computation is chosen uniformly at random (i.e. $H(\boldsymbol{A},\boldsymbol{F})=\log_{2}|\boldsymbol{A}|+\log_{2}|\boldsymbol{F}|$) and we make the equivalent assumption here (however for the corresponding change in variable from $\boldsymbol{F}$ to $\boldsymbol{O}$; explicitly, we assume $P(\boldsymbol{\alpha},O)=\frac{1}{|\mathcal{A}|^{n}|\mathcal{O}|}$). ###### Theorem 4.2. For $D_{\operatorname{client}}$ as above, and under the given assumptions, the following holds for any choice of graph $G$ and angle set $\mathcal{A}$ that satisfy the required conditions: $\displaystyle H_{\min}(\boldsymbol{A},\boldsymbol{O}|\boldsymbol{A^{\prime}},\boldsymbol{C^{\prime}})_{D_{\operatorname{client}}}\geq n+\log_{2}(|\mathcal{O}|).$ (46) The proof of this theorem is given in Section C.1. The key step of the proof consists of providing an upper bound for the guessing probability (which provides a corresponding lower bound for the min-entropy) via the upper bound of 3.2 applied to the comb $D_{\operatorname{client}}$. Due to the specific form of $D_{\operatorname{client}}$, this bound has an interpretation in terms of the number of different ways each $\boldsymbol{\alpha^{\prime}}$ can be reported for a given classical message (that is, for different gflows $g$ and one-time pads $\boldsymbol{r}$) with a larger number corresponding to a lower min-entropy bound (this is explained further after the proof of the theorem in Section C.1). One can rightfully ask how good this bound is in practice. To conclude this subsection, we consider a simple example for which the gflows are characterised and the min-entropy is in fact given exactly by the right-hand side of Equation 46. Let $G$ be the three-vertex graph as shown in Figure 4 with total order given by the natural order on vertex labels. As demonstrated in Section C.2, there is exactly one non-trivial output set ($O=\\{2,3\\}$) that supports gflow for this example, and exactly two corresponding gflows exist. It can further be shown that the upper and lower bounds of 3.2 for $D_{\operatorname{client}}$ coincide, which establishes $n+\log_{2}(|\mathcal{O}|)$ as the true min-entropy value. For this example, $n=3$ and $|\mathcal{O}|=1$, so $H_{\min}(\boldsymbol{A},\boldsymbol{O}|\boldsymbol{A^{\prime}},\boldsymbol{C^{\prime}})_{D_{\operatorname{client}}}=3$, which is corroborated by the numerical results (for two choices of $\mathcal{A}$, one such that $|\mathcal{A}|=4$ and the other such that $|\mathcal{A}|=8$). Figure 4: An example graph state for which the security for a single round of the BQC protocol considered here is maximally bad (that is, the min-entropy for the associated quantum comb obtains the lower bound given in 4.2). All potential gflows for the protocol are required to be compatible with the order $1<2<3$ on vertices. The only output set that supports such gflows is $O=\\{2,3\\}$. ### 4.4 Regarding Multi-Round Blindness In this subsection, we extend the blindness analysis to the situation where the client and server engage in multiple rounds of the protocol, for example, in order to collect measurement statistics of the output. It is not clear that the protocol does indeed stay secure in this case (multi-round security was not treated in the original work [36]). As an extension of 4.2 above, we demonstrate that the min-entropy is bounded away from zero for all rounds. We use similar assumptions to 4.2, namely that $P(g|O)$ and $P(\boldsymbol{r})$ are uniform, however we make a slightly weaker assumption on $P(\boldsymbol{\alpha},O)$: we take $P(\boldsymbol{\alpha},O)=\frac{P(O)}{|\mathcal{A}|^{n}}$ for $P(O)$ an arbitrary distribution over the output sets. Furthermore, we make the additional assumptions that the gflow and one-time pads are chosen independently for each round, allowing the multi-round protocol to be represented as $\displaystyle D_{\operatorname{client}}^{(m)}=\sum_{\boldsymbol{\alpha},O}\frac{P(O)}{|\mathcal{A}|^{n}}\ket{\boldsymbol{\alpha},O}\\!\\!\bra{\boldsymbol{\alpha},O}\bigotimes_{j=1}^{m}\sigma_{\boldsymbol{\alpha},O}^{(j)}.$ (47) We have the following theorem: ###### Theorem 4.3. For any $m$, the following holds: $\displaystyle H_{\min}(\boldsymbol{A},\boldsymbol{O}|\boldsymbol{A^{\prime}}^{(1)},\boldsymbol{C^{\prime}}^{(1)},...,$ $\displaystyle\boldsymbol{A^{\prime}}^{(m)},\boldsymbol{C^{\prime}}^{(m)})_{D_{\operatorname{client}}^{(m)}}$ $\displaystyle\geq-\log\left(\sum_{O\in\mathcal{O}}\frac{P(O)}{2^{|O|}}\right).$ (48) A key point of the proof, given in Section C.3, consists of the observation that, for any choice of output set $O$, the one-time pads assigned to the qubits in the output set cannot be learnt by the server since they do not appear in the adaptation of any other angle (in contrast to the one-time pads assigned to non-output qubits). The consequence is that the server can, at best, only learn the output angles of $\boldsymbol{\alpha}$ up to a $\pi$ phase. However, the server may be able to learn all other measurement angles with certainty, which correspond to the computation proper, in which case the only remaining uncertainty consists of not knowing how to interpret the measurement outcomes on the output state. This level of uncertainty is likely insufficient to deem the protocol as secure in multiple rounds. The above theorem does not say anything about whether the min-entropy does in fact change from round to round, let alone whether the above bound can ever be reached. To provide an answer to the former, we conclude by returning to the minimal example presented above and consider the associated min-entropy for two rounds. As outlined in Section C.2, for a specific choice of angle set $\mathcal{A}$, the two round entropy is upper bounded by $-\log_{2}(0.140625)\approx 2.830(0)$ and thus the min-entropy value does indeed decrease from the single round to the two round case. ## 5 Grey Box MBQC In this section, we consider a set of combs that are once again defined via gflow, however there is now a mixture of quantum and classical interactions. The inputs to the combs defined here are the same as for the BQC case (measurement outcomes), but the outputs are now single qubit states of the graph state, potentially acted upon by correction operators. We demonstrate that these combs can be understood as quantum causal models in the style of [20]. Moreover, the measurement channels inserted into the comb to perform a measurement-based computation bear close resemblance to the notion of observation put forward in [50] in the context of quantum causal inference. As such, a key goal of this section is to explicate a novel connection between measurement-based quantum computation and quantum causal models and causal discovery. In keeping with the rest of this work, we again apply the comb min-entropy to quantify optimal learning of an unknown property, now of a “grey box MBQC device” - a device which both prepares the graph state and conducts measurement corrections (so that we simply need to insert measurement channels) and for which we have only partial knowledge. Specifically, we consider an example where the graph state and designated input and output vertex sets are known but the particular gflow defining the corrections is not. In light of the above connection, learning the gflow corresponds to learning the causal structure of the device, i.e. a type of causal discovery. The min-entropy for this example provides a benchmark for optimal causal learning to which e.g. a learning strategy involving only observations (defined below) can be compared. In the final two subsections, we introduce two further pragmatic examples that continue to demonstrate the utility of the quantum comb and min-entropic approach to the type of partial-information analysis considered here. The first considers the same specific graph as the causal discovery example, however instead of learning the gflow itself, we aim now only to learn the required measurement planes for which the gflow is compatible. For the given level of knowledge about the device, i.e. knowing the graph state being prepared and the input and output sets, this represents the minimal information required to use the MBQC device for deterministic computation - knowing the full details of the specific gflow is not required. However, to ensure that any particular computation can be correctly performed, we also need to calibrate the frame of reference of our measurement channels (outside the device) with the frame of reference of the graph state preparation (inside the device). Said another way, we want to ensure that measuring in, say, the $XY$-plane at an angle $\alpha$ to the positive $X$-axis in our measuring device does indeed correspond to an angle $\alpha$ with respect to the positive $X$-axis implicit in the preparation of $\ket{G}$. The final example is concerned with this scenario by aiming to learn the angle discrepancy between the positive $X$-axes in the internal and external frame of references under the assumption that the $Z$-axis is known. This can be considered as an example of correlating reference frames in the quantum reference frames literature [54]. ### 5.1 Gflow Quantum Combs In Section 2.2, and reinforced by the discourse in Section A.2, much emphasis was placed on the two equivalent ways of correcting for a negative measurement outcome at a given qubit in an MBQC: by classically adapting measurement angles for measurement on other qubits or by applying a partial stabiliser on the graph state. The latter perspective is the relevant one for our present purposes. An MBQC on a graph state $\rho_{G}$ that prepares an output (quantum) state on the output set $O$ can be written as $\displaystyle\operatorname{Tr}\left[\left(\bigotimes_{v\in V\setminus O}\ket{\pm_{\alpha_{v}}}\\!\\!\bra{\pm_{\alpha_{v}}}_{v}\right)\rho_{G}\right]$ (49) where the trace is over all qubits in $V\setminus O$ and the measurement planes can be considered as all $XY$ for simplicity. For a given vertex $v_{0}\in V\setminus O$, there are two possibilities: either a positive measurement outcome obtains in which case no corrections are required, or a negative outcome obtains, for which corrections can be performed by applying a partial stabiliser to $\rho_{G}$: $\displaystyle\operatorname{Tr}$ $\displaystyle\left[\left(\ket{-_{\alpha_{v_{0}}}}\\!\\!\bra{-_{\alpha_{v_{0}}}}_{v_{0}}\bigotimes_{v\in V\setminus(O\cup v_{0})}\ket{\pm_{\alpha_{v}}}\\!\\!\bra{\pm_{\alpha_{v}}}_{v}\right)\rho_{G}\right]$ $\displaystyle\equiv\operatorname{Tr}\Bigg{[}\left(\ket{+_{\alpha_{v_{0}}}}\\!\\!\bra{+_{\alpha_{v_{0}}}}_{v_{0}}\bigotimes_{v\in V\setminus(O\cup v_{0})}\ket{\pm_{\alpha_{v}}}\\!\\!\bra{\pm_{\alpha_{v}}}_{v}\right)$ $\displaystyle\quad\quad\quad\quad\quad\quad\quad\quad\cdot K_{g(v_{0})}|_{\setminus v_{0}}\rho_{G}K_{g(v_{0})}|_{\setminus v_{0}}^{\dagger}\Bigg{]}$ (50) where $K_{g(v_{0})}$ is the (product of) stabiliser(s) assigned by a gflow $g$ to correct for measurements at $v_{0}$ and the notation $K_{g(v_{0})}|_{\setminus v_{0}}$ denotes the operator derived from $K_{g(v_{0})}$ with the $v_{0}$ tensor factor replaced with the identity. So, for each measurement and conditional on the measurement outcome, we are applying an operator on the graph state, either the identity (no correction implemented) or the partial stabiliser (correction implemented). The ability to correct for a negative outcome at every qubit via stabilisers only works if each measurement is performed in the assigned plane for which the graph and measurement plane assignment has gflow. If gflow exists and the restriction on the measurement planes is followed, one can take the perspective that the correction method defined from gflow is a mapping from measurement channels to unitary channels, i.e. a higher order quantum map (a quantum comb). To make this perspective precise, let us consider a tuple $(G,I,O,\omega)$ for which a gflow $g$ exists. Our aim is then to write down an operator $\sigma_{\operatorname{MBQC}}^{g}$ corresponding to the correction method induced by $g$ such that the following quantity is the same for every set of measurement outcomes (which in particular includes the all positive outcome instance, and thus is equivalent to the desired computation): $\displaystyle\left(\bigotimes_{v\in V\setminus O}\mathcal{M}_{v,\alpha_{v},\omega(v)}\right)\ast\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}$ (51) where the projections $\ket{\pm_{\alpha_{v}}}\bra{\pm_{\alpha_{v}}}$ in measurement plane $\omega(v)$ are now described by appropriately defined measurement channels $\mathcal{M}_{v,\alpha_{v},\omega(v)})$ (see below) and where $\ast$ denotes the link product (2.3) which in particular includes the trace over all spaces not related to the output space. The next few paragraphs are devoted to defining the operator $\sigma_{\operatorname{MBQC}}^{g}$. Let $g$ be a gflow for $(G,I,O,\omega)$, where $G$ is a graph on $n$ vertices and as such we identify $V$ with $\\{1,...,n\\}$. The operator $\sigma_{\operatorname{MBQC}}^{g}$ is defined on the Hilbert spaces $\bigotimes_{i=1}^{n}\mathcal{H}_{A_{i}}\otimes\mathcal{H}_{C_{i}}\otimes\mathcal{H}_{A^{\prime}_{i}}$, where $\mathcal{H}_{A_{i}}\cong\mathcal{H}_{C_{i}}\cong\mathcal{H}_{A^{\prime}_{i}}\cong\mathbb{C}^{2}$ for each $i$. Recalling from Equation 10 that $g$ defines the correction sets $\mathcal{X}_{i}$ and $\mathcal{Z}_{i}$ for each $i\in V$, we define the correction operator for $i$ as $\displaystyle U_{\text{corr}(\boldsymbol{c}),i}:=X_{A_{i}}^{\bigoplus_{j\in\mathcal{X}_{i}}c_{j}}Z_{A_{i}}^{\bigoplus_{j\in\mathcal{Z}_{i}}c_{j}}$ (52) and for notational simplicity in the following, we take $\displaystyle U_{\text{corr}(\boldsymbol{c})}:=\bigotimes_{i=1}^{n}U_{\text{corr}(\boldsymbol{c}),i}.$ (53) We can then define $\sigma_{\operatorname{MBQC}}^{g}$ as $\displaystyle\sigma_{\operatorname{MBQC}}^{g}:=\sum_{\boldsymbol{a},\boldsymbol{b},\boldsymbol{c}}U_{\text{corr}(\boldsymbol{c})}\ket{\boldsymbol{a}}\\!\\!\bra{\boldsymbol{b}}_{A}U_{\text{corr}(\boldsymbol{c})}^{\dagger}\otimes\ket{\boldsymbol{c},\boldsymbol{a}}\\!\\!\bra{\boldsymbol{c},\boldsymbol{b}}_{CA^{\prime}}$ (54) where the subscripts $A$, $C$ and $A^{\prime}$ denote that the corresponding state is in $\mathcal{H}_{A}:=\bigotimes_{i=1}^{n}\mathcal{H}_{A_{i}}$, $\mathcal{H}_{C}:=\bigotimes_{i=1}^{n}\mathcal{H}_{C_{i}}$ or $\mathcal{H}_{A^{\prime}}:=\bigotimes_{i=1}^{n}\mathcal{H}_{A^{\prime}_{i}}$ respectively, and where the sum is over the computational basis of these spaces. The space $\mathcal{H}_{A^{\prime}}$ receives the graph state $\rho_{G}$ via the link product as in Equation 51. To connect back to the discussion above, $\sigma_{\operatorname{MBQC}}^{g}$ is the Choi operator of the channel that conditionally applies the appropriate corrections via partial stabiliser to the graph state. See Figure 5(a) for a depiction of $\sigma_{\operatorname{MBQC}}^{g}$ for one of the gflows considered in the example in Section 5.3. It remains then only to define the measurement channels $\mathcal{M}_{\alpha_{i},i,\omega(i)}$. Previously, we denoted measurements simply via projection operators $\ket{\pm_{\alpha}}\\!\\!\bra{\pm_{\alpha}}_{\text{mp}}$ for $\text{mp}\in\\{XY,XZ,YZ\\}$, however for our present purposes we would like to consider channels from $\mathcal{H}_{A_{i}}$ to $\mathcal{H}_{C_{i}}$ which moreover prepare classical measurement outcomes ($\ket{0}\\!\\!\bra{0}$ or $\ket{1}\\!\\!\bra{1}$) at $\mathcal{H}_{C_{i}}$. We define $\mathcal{M}_{\alpha_{i},i,\omega(i)}:\mathcal{H}_{A_{i}}\rightarrow\mathcal{H}_{C_{i}}$ via its Choi representation: $\displaystyle\mathcal{M}_{\alpha_{i},i,\omega(i)}:=$ $\displaystyle\ket{0}\\!\\!\bra{0}_{C_{i}}\otimes\ket{+_{\alpha_{i}}}\\!\\!\bra{+_{\alpha_{i}}}_{\omega(i),A_{i}}^{T}$ $\displaystyle+\ket{1}\\!\\!\bra{1}_{C_{i}}\otimes\ket{-_{\alpha_{i}}}\\!\\!\bra{-_{\alpha_{i}}}_{\omega(i),A_{i}}^{T}.$ (55) (a) $\sigma_{\operatorname{MBQC}}^{g_{1}}$ (b) $D_{\text{gflow}}$ Figure 5: The classical-quantum combs defined here for the consideration of an MBQC device have quantum comb component based on the flow of corrections required for measurement-based computation (known as gflow). (a) An example of such a quantum comb is $\sigma_{\operatorname{MBQC}}^{g_{1}}$ corresponding to a gflow $g_{1}$ for a four-qubit graph state (see the example in Section 5.3 for more details). The graph state is input on the left and passed through the conditional correction channels to each of the output spaces on the right. Measurement channels performing single qubit measurements can be inserted into the gaps (such as between $A_{1}$ AND $C_{1}$) (b) An example of the full MBQC device (the grey outline) is modelled as a classical-quantum comb built from the classical parameter that takes values $g$ denoting the gflow (equivalently causal structure) and quantum comb $\sigma_{\operatorname{MBQC}}^{g}$ along with the preparation of the graph state $\rho_{G}$. As with all classical- quantum combs in this work, only the lower systems (those extending out of $\sigma_{\operatorname{MBQC}}^{g}$) can be interacted with, and the upper system is to be learnt about. It should be noted that, since quantum combs are typically considered with a total order, we endow $\sigma_{\operatorname{MBQC}}^{g}$ with a choice of total order for which the partial order given by the gflow is compatible. Any such total order is sufficient and we leave it implicit in the labelling of the Hilbert spaces. Explicitly, we consider $\sigma_{\operatorname{MBQC}}^{g}\in\operatorname{Comb}(\mathcal{H}_{A^{\prime}}\rightarrow\mathcal{H}_{A_{1}},H_{C_{1}}\rightarrow\mathcal{H}_{A_{2}},...,\mathcal{H}_{C_{|V\setminus O|}}\rightarrow\bigotimes_{i\in O}\mathcal{H}_{A_{i}})$. A proof that $\sigma_{\operatorname{MBQC}}^{g}$ is indeed correct in the sense of ensuring that Equation 51 is equivalent for all measurement outcomes is given in Section D.1. This proof is related to the existing theorem establishing sufficiency of gflow for deterministic computation [[, Theorem 2,]]browne2007generalized and also provides justification for why the ordering of operators in Equation 52 is valid. ### 5.2 Quantum Causal Models and Inference Having defined $\sigma_{\operatorname{MBQC}}^{g}$ for an arbitrary $g$, we can now make two connections to the literature of quantum causal models [20, 19, 68] and quantum causal inference [50, 52, 51]. Quantum causal models (QCMs) are an extension to the classical causal modelling literature (see e.g. [21, 22, 69]) due, in part, to the realisation that the classical methodology is insufficient for treating quantum correlations [70], and are typically represented in the process matrix formalism [71]. In the first part of this subsection, we show that each $\sigma_{\operatorname{MBQC}}^{g}$ satisfies the conditions required for being a quantum causal model as per [20]. A key aspect of the classical causality literature consists of regarding the differences between the probability distribution obtained under observations and that obtained under an intervention on one or more of the causal mechanisms generating the distribution. A full introduction and treatment of these notions is beyond the scope of this work, however suffice it to say that their counterparts for quantum causal inference are far more nuanced [50, 52, 51]. For our present purposes, we focus on the notion of observation as given in [50] (discussed below) and showcase in the second part of this subsection the close resemblance between this perspective of observation and the measurement channels $\mathcal{M}_{\alpha_{i},i,\omega(i)}$ defined in the previous subsection for MBQC. As a first step towards the upcoming definition of quantum causal models, it is worthwhile to briefly review the main components of their classical counterparts. Classical causal models are specified by a directed acyclic graph (DAG), with each node assigned a random variable, and functional equations relating the value of a given node to the values of its parents. In the quantum case, random variables are replaced by quantum nodes $B_{i}$, which consist of two Hilbert spaces $\mathcal{H}_{B_{i}}$ and $\mathcal{H}_{B_{i}}^{*}$ specifying incoming and outgoing state spaces of the node (just as for combs), and with functional equations replaced by quantum channels. ###### Definition 5.1 (Definition 3.3, [20]). A quantum causal model is given by a directed acyclic graph over quantum nodes $B_{1},...,B_{n}$ and for each node $B_{i}$, a quantum channel $\rho_{B_{i}|\operatorname{Pa}(B_{i})}\in\mathcal{L}(\mathcal{H}_{B_{i}}\otimes\mathcal{H}_{\operatorname{Pa}(B_{i})}^{*})$ (where $\operatorname{Pa}(B_{i})$ denotes the set of parent nodes to $B_{i}$ with respect to the DAG) such that all channels mutually commute. This defines a process operator $\displaystyle\sigma_{B_{1},...,B_{n}}:=\prod_{i=1}^{n}\rho_{B_{i}|\operatorname{Pa}(B_{i})}.$ A note regarding notation: the asterisk on the $\mathcal{H}_{B_{i}}^{*}$ is not (necessarily) to indicate that this is the dual space of $\mathcal{H}_{B_{i}}$. It is merely to differentiate the outgoing space of the node from the incoming one. We are avoiding using superscripts involving “in” and “out” since the convention here would clash with that used in the combs notation (ingoing to the node is outgoing from the comb and vice versa). To show that $\sigma_{\operatorname{MBQC}}^{g}$ is a valid QCM, we first realise that each gflow $g$ (including the partial order) induces a DAG on the vertices of the corresponding graph $G$ (see D.2). In particular, for each $i\in V$, the parents of $i$ are given by the union of the correction sets: $\displaystyle\operatorname{Pa}(i):=\mathcal{X}_{i}\cup\mathcal{Z}_{i}.$ (56) Using this notation, one can see how to write $\sigma_{\operatorname{MBQC}}^{g}$ as given in Equation 54 as a product of channels as in the definition above and what those channels are: $\displaystyle\sigma_{\operatorname{MBQC}}^{g}=\prod_{i=1}^{n}\rho_{A_{i}|C_{j:j\in\operatorname{Pa}(i)},A^{\prime}_{i}}$ (57) where $\displaystyle\rho_{A_{i}|C_{j:j\in\operatorname{Pa}(i)},A^{\prime}_{i}}:=\sum_{\begin{subarray}{c}a_{i},b_{i}\\\ \boldsymbol{c}_{\operatorname{Pa}(i)}\end{subarray}}$ $\displaystyle U_{\text{corr}(\boldsymbol{c}_{\operatorname{Pa}(i)}),i}\ket{a_{i}}\\!\\!\bra{b_{i}}U_{\text{corr}(\boldsymbol{c}_{\operatorname{Pa}(i)}),i}^{\dagger}$ $\displaystyle\quad\quad\otimes\ket{\boldsymbol{c}_{\operatorname{Pa}(i)}a_{i}}\\!\\!\bra{\boldsymbol{c}_{\operatorname{Pa}(i)}b_{i}}$ (58) with $\boldsymbol{c}_{\operatorname{Pa}(i)}$ denoting the classical message that are input into $\sigma_{\operatorname{MBQC}}^{g}$ for the vertices in $\operatorname{Pa}(i)$. The channel $\rho_{A_{i}|C_{j:j\in\operatorname{Pa}(i)},A^{\prime}_{i}}$ clearly has a very similar structure to Equation 54, but contains only the relevant spaces and corrections pertaining to $A_{i}$. It is immediate to see that the remaining condition, that the $\rho_{A_{i}|C_{j:j\in\operatorname{Pa}(i)},A^{\prime}_{i}}$ commute, is satisfied: the only overlap between any two such channels occurs on the classical message spaces, upon which the operators are diagonal. We have thus shown: ###### Proposition 5.1. For each gflow $g$, $\sigma_{\operatorname{MBQC}}^{g}$ is a quantum causal model. Having demonstrated the first connection between MBQC and quantum causality, we return to the common theme running through this work: learning about a random variable via interactions with a comb. Let the random variable $X$ take values in the set of all gflows for a given graph $G$ (or perhaps a given tuple $(G,I,O)$ or $(G,I,O,\omega)$) and consider $\displaystyle D_{\text{gflow}}:=\sum_{g\in X}P(g)\ket{g}\\!\\!\bra{g}\otimes\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}.$ (59) In this case, we can understand learning about $X$ as a type of quantum causal discovery: $X$ encodes information about $g$ and $g$ corresponds to a unique QCM, so learning about $X$, via interactions with the $\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}$, is equivalent to learning the causal structure. In the classical literature, a distinction is made between learning the causal structure via observations and via interventions. For the quantum case, Ried et al. [50] (see also [53]) define an observation as a standard projective measurement channel (i.e. one the performs a projective measurement and then prepares the post-measurement state according to the standard update rule) which takes a maximally mixed state as input. The requirement of a maximally mixed input is necessary and sufficient to enforce an “informational symmetry” between the updated belief of the input state and the knowledge of the output state of the channel (see Section III.A in the supplementary information of [50]). The measurement channels $\mathcal{M}_{\alpha_{i},i,\omega_{i}}$ defined in Section 5.1 are not exactly the measurement channels as stipulated above, but the difference is only cosmetic: the standard projective measurement channel can be recovered simply by post-composing $\mathcal{M}_{\alpha_{i},i,\omega_{i}}$ with a unitary channel (that depends on $\alpha_{i}$ and $\omega_{i}$). In particular, the guiding principle of informational symmetry between input and output is preserved. Regarding the requirement of a maximally mixed input state, this is indeed ensured provided that a graph state is input into $\sigma_{\operatorname{MBQC}}^{g}$ (each single qubit reduced state of a graph state is maximally mixed). Thus, using the present terminology, one could consider a measurement-based computation, such as that described by Equation 51, as a specific set of observations of a particular quantum causal model. For the reader familiar with the classical causality literature, in particular the work of Pearl (eg., [21]), the $A^{\prime}_{i}$ are analogous to the exogenous variables for a classical causal model, and thus $\rho_{G}$ can be considered as the equivalent of a distribution over the exogenous variables and moreover one that does not factorise into a product distribution as is often assumed. The last sentence of the previous paragraph can then be more thoroughly reformulated as MBQC being a specific set of observations of a particular quantum causal model for a certain state of the exogenous nodes. See Section D.2 for more details of this correspondence. ### 5.3 Causal Discovery Example Let us now consider a concrete example, which acts both as an explanatory aid to the preceding subsection and as a vehicle to discuss the comparison between the optimal strategy for causal discovery and an observational strategy. We fix a graph and input and output sets for which all corresponding gflows are characterised, and thereafter numerically calculate the min-entropy for $D_{\text{gflow}}$ (Equation 59) as well as for a related comb with a restriction on the included gflows. With this benchmark for the optimal causal learning, we optimise over a restricted set of strategies constrained to consist of observations only (as per the above notion of observation) thereby quantifying the difference between the best such observational strategy and the best overall strategy. We further demonstrate that, for cases where the causal structures (gflows) differ only by symmetries of the input state (graph state), even the optimal strategy is completely uninformative for causal discovery. Figure 6(a) depicts the four-vertex graph and choice of input and output sets that forms the basis for our example. For this $(G,I,O)$ there are $15$ different possible gflows, which are catalogued in Appendix D by stipulating the map $g$ and corresponding DAG for each (an example is depicted in Figure 6(b) for $g_{1}$; the labelling is given in the appendices). (a) $(G,I,O)$ (b) DAG for $g_{1}$ Figure 6: To give a concrete example, we consider a specific graph state and the set of all corresponding gflows. (a) The four qubit graph $G$ and choice of input set $I=\\{1\\}$ and output set $O=\\{3,4\\}$. There are $15$ gflows compatible with this choice of $(G,I,O)$ which are catalogued in Section D.4 including the corresponding directed acyclic graph and correction operators. (b) The DAG for gflow $g_{1}$, which measures qubits $1$ and $2$ in the $XY$-plane following the order $1<2$. The labels on the directed arrows depict the conditional correction operators, with the head and tail of the arrow denoting the target and control of the operation respectively. We write the $D_{\text{gflow}}$ for this example as $\displaystyle D_{\text{gflow}}=\sum_{j=1}^{15}P(g_{j})\ket{j}\\!\\!\bra{j}\otimes\sigma_{\operatorname{MBQC}}^{g_{j}}\ast\rho_{G}$ (60) which is a comb in $\operatorname{Comb}(\mathbb{C}\rightarrow A_{1},C_{1}\rightarrow A_{2},C_{2}\rightarrow A_{3,4},\mathbb{C}\rightarrow X)$ where $A_{1}$ and $A_{2}$ label the Hilbert spaces for graph state qubits $1$ and $2$ respectively, $A_{3,4}$ labels the two-qubit Hilbert space for the remaining graph state qubits, and $X$ labels a $15$-dimensional Hilbert space for which the elements of a choice of orthonormal basis $\\{\ket{j}\\}_{j=1}^{15}$ correspond to the different gflows. Note that no measurement outcomes on the output qubits are input into $D_{\text{gflow}}$. In this example, we once again assume a uniform prior: $P(g_{j})=\frac{1}{15}$. We will also consider a related comb to $D_{\text{gflow}}$, but instead of including all $15$ gflows that exist for the choice of graph and input and output sets, we consider only those for which the assigned measurement planes for qubits $1$ and $2$ are both in the $XY$-plane and moreover only the gflows which have partial order such that $1<2$. Using the labelling convention of Appendix D, the gflows $g_{1},g_{2},g_{4}$ and $g_{5}$ satisfy these criteria. We define $\displaystyle D_{XY,1<2}:=\sum_{j\in\\{1,2,4,5\\}}P(g_{j})\ket{j}\\!\\!\bra{j}\otimes\sigma_{\operatorname{MBQC}}^{g_{j}}\ast\rho_{G}$ (61) with $P(g_{j})=\frac{1}{4}$. This is a comb in $\operatorname{Comb}(\mathbb{C}\rightarrow A_{1},C_{1}\rightarrow A_{2},C_{2}\rightarrow A_{3,4},\mathbb{C}\rightarrow X)$ where $\mathcal{H}_{X}$ is $4$-dimensional in this case (we overload $X$ for notational simplicity). The guessing probability for each of these combs is given below, and, as for all numerical calculations of guessing probabilities in this work, it is calculated via the minimisation over unnormalised combs (recall Equation 12). However, to conduct the comparison with an optimal observational strategy, it is more convenient to use the maximisation form instead (c.f. Equation 13): $\displaystyle P_{\text{guess}}=\max_{E}\operatorname{Tr}[DE^{T}]$ (62) where $D$ is replaced by $D_{\text{gflow}}$ or $D_{XY,1<2}$ for the relevant calculation, and $E$ is an element of $\operatorname{Comb}(A_{1}\rightarrow C_{1},A_{2}\rightarrow C_{2},A_{3,4}\rightarrow X)$ with $X$ the appropriate space for each case. The guessing probability under observational strategies only is given by an analogous quantity to Equation 62 except that $E$ is now restricted to having the form $\displaystyle E:=\sum_{c_{1},c_{2}}\mathcal{M}_{C_{1}|A_{1}}^{c_{1}}(\ket{\psi})\otimes\mathcal{M}_{C_{2}|A_{2}}^{c_{2}}(\phi)\otimes\mathcal{E}_{X|A_{3,4}}^{c_{1}c_{2}}$ (63) where $\displaystyle\mathcal{M}_{C_{i}|A_{i}}^{0}(\ket{\psi})$ $\displaystyle:=\ket{0}\\!\\!\bra{0}_{C_{i}}\otimes\ket{\psi}\\!\\!\bra{\psi}_{A_{i}}$ $\displaystyle\mathcal{M}_{C_{i}|A_{i}}^{1}(\ket{\psi})$ $\displaystyle:=\ket{1}\\!\\!\bra{1}_{C_{i}}\otimes(I-\ket{\psi}\\!\\!\bra{\psi})_{A_{i}}$ (64) denote the observations, i.e. measurement channels $\mathcal{M}_{C_{i}|A_{i}}(\psi))=\mathcal{M}_{C_{i}|A_{i}}^{0}(\ket{\psi})+\mathcal{M}_{C_{i}|A_{i}}^{1}(\ket{\psi})$, and $\mathcal{E}_{X|A_{3,4}}^{c_{1}c_{2}}$ denotes the Choi state for a CPTP map $\mathcal{E}:\mathcal{H}_{A_{3,4}}\rightarrow\mathcal{H}_{X}$ for the relevant $\mathcal{H}_{X}$ for the corresponding $D$. We allow for the choice of CPTP map to depend on the measurement outcomes, hence the $c_{1},c_{2}$ in the superscript. Numerically, the optimal guessing probability for $D_{\operatorname{gflow}}$, which is to say, for selecting the correct causal structure after a single round of interaction with the device, stands at $0.373(2)$ (see Appendix E for details of the numerical implementation). This is a clear improvement over the uniform distribution prior to interacting with the device ($\frac{1}{15}\approx 0.067(1)$). To calculate the guessing probability with the restriction to observational strategies, a different approach is required since the extra restriction on the set of operators that is maximised over imposes non-linear constraints on the problem, so SDP methods no longer apply. Namely, a brute force search was conducted over a mesh lattice of single qubit projectors. Using this method, an approximate value for the guessing probability under the optimal observational strategy is $0.199(1)$ which indicates that there exist strategies for causal discovery that are more informative than any observational strategy for this case. For $D_{XY,1<2}$, the situation is different and rather interesting: the guessing probability is $0.250(0)$. That is, in a single round even the optimal learning strategy can not distinguish between the four different causal structures. A trivial consequence of this fact is that, in this case, any observational strategy is optimal (but still completely uninformative). In fact, one can understand why the lack of distinguishability is to be expected by considering the following proposition. ###### Proposition 5.2. For any $c_{j}\in\\{0,1\\}$ for $j=1,...,|V\setminus O|$, $\displaystyle\left(\bigotimes_{j=1}^{|V\setminus O|}\ket{c_{j}}\\!\\!\bra{c_{j}}_{C_{j}}\right)\ast\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}$ (65) is the same state for each gflow $g$ compatible with $(G,I,O,\omega)$ and have mutually compatible partial orders. The proof is given in Section D.3, but the key point is that for any given measurement outcomes, the correction operators induced by each $g\sim(G,I,O,\omega)$ are related to each other by some stabiliser and hence are in fact equivalent. There are a number of conclusions one can draw from the above analysis. Firstly, as expected, an observational strategy for quantum causal discovery is typically not optimal in general (at least in a single round). Secondly, in certain cases, it is not guaranteed that anything about the causal structure can be learnt in a single round by any strategy. This arises due to the fine- tuned situation where the possible causal structures coincide precisely with symmetries of the particular prepared state. This serves to highlight the important role that noise likely plays in the ability to learn a causal structure (a topic for future work). Finally, we must emphasise that the numerical results here are for a single round of interaction only, largely due to implementation constraints (see Appendix E). ### 5.4 Learning Measurement Planes The example of $D_{XY,1<2}$ above demonstrates that barriers to learning the exact flow of corrections exist, but fortunately, these details are not required if our objective is to use the device to perform computations. In this subsection, we consider the same example as given above, however we are no longer concerned with learning the causal structure of the device but rather learning enough information about the gflow being implemented in order to ensure deterministic computation. Specifically, we consider the case where the assignment of measurement planes is unknown: that is, we know $(G,I,O)$ and need to learn $\omega$. We continue with the same graph $(G,I,O)$ as in Figure 6 which has gflows catalogued in Appendix D. Due to the properties of gflow, the input qubits are always assigned the $XY$-plane by $\omega$, so for our example $\omega(1)\equiv XY$. Since we allow for any measurements on the output qubits, it remains only to learn the value of $\omega(2)$. It turns out that, for the current choice of $(G,I,O)$, all three measurement planes on qubit $2$ are possible, with an equal number of gflows ($5$) for each. Under the labelling as given in Appendix D, gflows $g_{1},...,g_{5}$ correspond to $\omega(2)=XY$, $g_{6},...,g_{10}$ to $\omega(2)=XZ$ and $g_{11},...,g_{15}$ to $\omega(2)=YZ$. For this example, we let $X$ be the random variable that encodes the measurement plane choice for the second qubit, i.e. takes values in $\\{XY,XZ,YZ\\}$, and we denote an arbitrary measurement plane by mp. The notation $g\sim\text{mp}$ indicates the gflow $g$ is defined for $(G,I,O,\omega)$ where $\omega(2)=\text{mp}$. We consider the comb $\displaystyle D_{\text{mp}}:=\sum_{\text{mp}\in X}P(\text{mp})\ket{\text{mp}}\\!\\!\bra{\text{mp}}\otimes\sum_{g\sim\text{mp}}P(g|\text{mp})\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}$ (66) in $\operatorname{Comb}(\mathbb{C}\rightarrow A_{1},C_{1}\rightarrow A_{2},C_{2}\rightarrow A_{3,4},\mathbb{C}\rightarrow X)$ where $\dim\mathcal{H}_{X}=3$. For simplicity, we once again make uniformity assumptions on the distributions: $P(\text{mp})=\frac{1}{3}$ and $P(g|\text{mp})=\frac{1}{5}$. Under these assumptions, the guessing probability for $D_{\text{mp}}$ reaches $1.000(0)$ indicating that the measurement planes can in fact be determined with certainty after just a single use of the device. This is perhaps unsurprising since the correction operators of gflows that differ in the measurement plane assignment for the second qubit (e.g. for $g_{1}$ and $g_{6}$) are not related by a stabiliser and so the type of indistinguishability problem that arose for $D_{XY,1<2}$ does not apply here. ### 5.5 Calibrating Measurements In the previous example, the aim was to learn the correct measurement planes for the proper functioning of the device, which in particular allows the user to insert measurement channels $\mathcal{M}_{\alpha_{v},v,\text{mp}_{v}}$ for the correct $\text{mp}_{v}$ for each $v$ (i.e. $\text{mp}_{v}=\omega(v)$), thereby performing deterministic computation. However, this is not yet enough to ensure the correctness of a specific computation: we need to ensure that the $\alpha_{v}$ are also correct for each $v$. To explain this further, consider a computation where all measurements are in the $XY$-plane, meaning that the projection operators are given by $\displaystyle\ket{\pm_{\alpha_{v}}}\\!\\!\bra{\pm_{\alpha_{v}}}=R_{Z}(\alpha_{v})\ket{\pm}\\!\\!\bra{\pm}R_{Z}(\alpha)^{\dagger}$ (67) for each $v$. The $\ket{\pm}$ denote the eigenvectors of the $X$ operator and moreover, there is an implicit assumption that the $\ket{+}$ state is the same as that used in the graph state preparation (recall Equation 6). This assumptions is typically a very reasonable one, however, since in this case the state preparation occurs inside the device and the measurement channels occur outside of it, there is no guarantee that this assumption is valid. If a measurement channel $\mathcal{M}_{\alpha_{v},v,XY}$ , thought to consist of projectors as in Equation 67, instead consists of projectors $\displaystyle R_{Z}(\alpha_{v})\ket{\pm_{\text{meas}}}\\!\\!\bra{\pm_{\text{meas}}}R_{Z}(\alpha_{v})^{\dagger}$ (68) where $\displaystyle\ket{\pm_{\text{meas}}}\\!\\!\bra{\pm_{\text{meas}}}=R_{Z}(\theta)\ket{\pm}\\!\\!\bra{\pm}R_{Z}^{\dagger}(\theta)$ (69) for some $\theta$, then the effective measurement on the graph state qubit is for an angle $\alpha_{v}+\theta$ rather than $\alpha_{v}$, meaning an incorrect computation is performed. Even if the measurement planes for the device are known, the specific measurement channels required to calibrate the device in order to give the computations as expected (i.e. specifying the computation using $\alpha_{v}-\theta$ recovers the correct computation). It is interesting to note that this calibration can be considered as correlating two quantum reference frames [54], the one inside the device with the one outside of it. Figure 7: The guessing probability for $D_{\text{calibr}}$ in a single round as the size of possible angles set $\mathcal{A}$ varies (shown in blue). The angles of $\mathcal{A}$ are taken to be evenly spaced. The probability of guessing the correct value for the angle off-set $\theta$ prior to interacting with the device is shown in black. In this subsection, we consider a simple example along the lines of the explanation above: we consider a scenario where the positive $Z$-axis of the Bloch sphere is known, but the positive $X$-axis is not (we assume a right- handed frame meaning that the positive $Y$-axis is determined if the positive $X$\- and $Z$-axes are). In other words, we aim to learn $\theta$ in Equation 69 above. We assume $\theta$ can take values in a discrete set of angles which are evenly spaced between $0$ and $2\pi$. We consider here a simple three- vertex linear graph (i.e. with vertices $V=\\{1,2,3\\}$ and edges $\\{(1,2),(2,3)\\}$), along with a single choice of gflow $g$ defined as $1\mapsto\\{2\\}$ and $2\mapsto\\{3\\}$, which gives the correction sets (only the non-empty such sets are shown): $\displaystyle\mathcal{X}_{2}$ $\displaystyle=\\{1\\};$ (70) $\displaystyle\mathcal{X}_{3}$ $\displaystyle=\\{2\\};$ (71) $\displaystyle\mathcal{Z}_{3}$ $\displaystyle=\\{1\\}.$ (72) For the purposes of writing down the comb $D_{\text{calibr}}$ for this example, we take the basis $\ket{\pm_{\text{meas}}}\\!\\!\bra{\pm_{\text{meas}}}$ as reference and write the graph state $\rho_{G}$ and the correction operators $U_{\text{corr}(\boldsymbol{c}),i}$ (recall Equation 52) with respect to this basis: $\displaystyle\rho_{G}^{\theta}$ $\displaystyle:=R_{Z}^{\otimes n}(-\theta)\rho_{G}\left(R_{Z}^{\otimes n}(-\theta)\right)^{\dagger};$ (73) $\displaystyle U_{\text{corr}(\boldsymbol{c}),i}^{\theta}$ $\displaystyle:=\left(R_{Z}(-\theta)X_{A_{i}}^{\bigoplus_{j\in\mathcal{X}_{i}}c_{j}}R_{Z}(-\theta)^{\dagger}\right)Z_{A_{i}}^{\bigoplus_{j\in\mathcal{Z}_{i}}c_{j}}.$ (74) Denoting the comb defined from the $U_{\text{corr}(\boldsymbol{c}),i}^{\theta}$ in analogy to Equation 54 as $\sigma_{\operatorname{MBQC}}^{\theta}$ and the (discrete) set of possible (regularly spaced) values for $\theta$ by $\mathcal{A}$, the comb of interest for this example is $\displaystyle D_{\text{calibr}}:=\sum_{\theta\in\mathcal{A}}P(\theta)\ket{\theta}\\!\\!\bra{\theta}\otimes\sigma_{\operatorname{MBQC}}^{\theta}\ast\rho_{G}^{\theta}.$ (75) As per usual, we take the prior distribution over $\theta$ to be uniform: $P(\theta)=\frac{1}{|\mathcal{A}|}$. The guessing probability values for $D_{\text{calibr}}$ for different sizes of $\mathcal{A}$, ranging from $|\mathcal{A}|=2$ to $|\mathcal{A}|=32$, are shown in Figure 7. For $|\mathcal{A}|=2$, the correct direction can be known with certainty in a single round, and almost so for the case where $|\mathcal{A}|=3$ ($P_{\text{guess}}\approx 0.992(5)$), but otherwise there is a steep decrease in the guessing probability as the size of $\mathcal{A}$ increases. ## 6 Discussion By interacting with a system, possibly via some complicated sequence of actions, one can learn a certain property of the system that influences the details of the interaction. By modelling such a situation within the quantum combs formalism, it is possible to leverage the comb min-entropy [26], to quantify how much one can learn about the unknown property in question. Due to the general, and also natural, modelling of both quantum and classical interactions as combs of the form considered above, the methodology showcased here has broad applicability. In this work, we restricted our attention to a novel set of combs defined by the paradigm of measurement-based quantum computation. In this first instance, we defined a classical comb which models a specific BQC protocol [36] and by so doing, give a proof of partial security based on the comb min-entropy for both a single round of the protocol as well as the multi-round case, extending the security analysis in the existing literature. We further defined a series of combs that model interactions with an MBQC device under varying levels of knowledge regarding the inner working of the device. This establishes a previously unmade connection between MBQC and quantum causality. By observing that the gflow component from which the MBQC combs are built satisfies the criteria of a quantum causal model, we were able to utilise the comb min-entropy to quantify the optimal strategy for causal discovery in specific cases, including cases where no information about the causal structure could be determined by any strategy. We investigated further examples related to learning the minimal extra information required in order to ensure the correct functioning of the device as a computer. ### 6.1 Limitations of the Min-Entropy Approach Despite the broad applicability and operational meaning of the methodology used in this work, there are certain limitations of which one should be made aware. Primarily, when using a numerical SDP solver, size issues quickly start to play a role. For example, in the three-qubit graph state example for the BQC protocol in Section 4.3, as a square matrix, the operator $D_{\operatorname{client}}^{(m)}$ has dimension $|\mathcal{A}|^{3+3m}2^{3m}$ where $m$ is the number of rounds. For the minimal possible choice of angle set, $|\mathcal{A}|=4$ and for a single round, this already equals $32768$. For larger angle sets or greater number of rounds, the size problems became prohibitive, disallowing any numerical calculations of the guessing probability on the available hardware (see Appendix E). ### 6.2 Future Work As mentioned above, the generality of the combs framework ensures that the methodology considered in this work can be applied in a wide variety of contexts (keeping in mind the limitations outlined above). We conclude by outlining a couple of avenues for future work. One feature of many SDP solvers that was not utilised in this work, regards the possibility to return an optimal solution instance (and its dual) along with the optimal value (e.g. the min-entropy) for the problem at hand. Depending on the exact formulation used for the min-entropy (recall Equations 11, 12 and 13) either the primal solution or its dual will be a matrix representing a strategy for interacting with the system in question. It would be interesting to analyse these matrices in order to infer what the corresponding strategy would be. For example, in Section 5.3, the min-entropy quantified the best strategy for inferring the causal structure in a single round of interaction. The solution that gives the optimal value could in principle be deconstructed into components acting at each node of the quantum causal model. This may help identify a type of quantum instrument or sequence of instruments that are optimal for causal inference in general (rather than just for the specific example). A second interesting type of analysis that could be performed relates to the effect of noise on the min-entropy. The examples considered in Section 5 all included the preparation of a graph state $\rho_{G}$ and the operators $\sigma_{\operatorname{MBQC}}$ which were specifically tailored to the graph $G$. In realistic situations, these preparations and the implementations of the corrections $U_{\text{corr}(\boldsymbol{c})}$ could be afflicted by noise. Using the methodology here, one could, for example, investigate the effect that different noise models could have on the guessing probability for determining e.g. the correct measurement planes for the MBQC device. Since adding noise to $\rho_{G}$ would not change the dimension of the comb to which it is constituent, there is no added computational overhead and so the noise analysis could be conducted for any case in which the noiseless case can be calculated. ## 7 Acknowledgements We would like to thank Atul Mantri for useful discussions about the security of the BQC protocol of [36], Simon Milz for insightful comments regarding quantum combs, Joshua Morris for invaluable suggestions regarding semi- definite programming resources, and Sofiene Jerbi for useful conversations regarding gflow and MBQC. We acknowledge support from the Austrian Science Fund (FWF) through DK-ALM: W$1259$-N$27$ and SFB BeyondC F$7102$, and from the University of Innsbruck. This work was also co-funded by the European Union (ERC, QuantAI, Project No. $101055129$). Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union or the European Research Council. Neither the European Union nor the granting authority can be held responsible for them. ## References * [1] John Preskill “Quantum Computing in the NISQ era and beyond” In _Quantum_ 2 Verein zur Förderung des Open Access Publizierens in den Quantenwissenschaften, 2018, pp. 79 DOI: 10.22331/q-2018-08-06-79 * [2] H.-J. Briegel, W. Dür, J. I. Cirac and P. Zoller “Quantum Repeaters: The Role of Imperfect Local Operations in Quantum Communication” In _Phys. Rev. Lett._ 81 American Physical Society, 1998, pp. 5932–5935 DOI: 10.1103/PhysRevLett.81.5932 * [3] Rodney Van Meter “Quantum Networking” ISTE Ltd/John WileySons Inc, Hoboken, NJ, 2014 DOI: 10.1002/9781118648919 * [4] Davide Castelvecchi “The quantum internet has arrived (and it hasn’t).” In _Nature_ 554.7690 Nature Publishing Group, 2018, pp. 289–293 DOI: 10.1038/d41586-018-01835-3 * [5] Sheng-Kai Liao et al. “Satellite-relayed intercontinental quantum network” In _Physical review letters_ 120.3 APS, 2018, pp. 030501 DOI: 10.1103/PhysRevLett.120.030501 * [6] CT Nguyen et al. “Quantum network nodes based on diamond qubits with an efficient nanophotonic interface” In _Physical review letters_ 123.18 APS, 2019, pp. 183602 DOI: 10.1103/PhysRevLett.123.183602 * [7] Peter C Humphreys et al. “Deterministic delivery of remote entanglement on a quantum network” In _Nature_ 558.7709 Nature Publishing Group, 2018, pp. 268–273 DOI: 10.1038/s41586-018-0200-5 * [8] Boris Korzh et al. “Provably secure and practical quantum key distribution over 307 km of optical fibre” In _Nature Photonics_ 9.3 Nature Publishing Group, 2015, pp. 163–168 DOI: 10.1038/nphoton.2014.327 * [9] Rachel Courtland “China’s 2,000-km quantum link is almost complete” In _IEEE Spectrum_ 53.11 IEEE, 2016, pp. 11–12 DOI: 10.1109/MSPEC.2016.7607012 * [10] Mohamed Elboukhari, Mostafa Azizi and Abdelmalek Azizi “Quantum Key Distribution Protocols: A Survey.” In _International Journal of Universal Computer Science_ 1.2, 2010 DOI: 10.1109/ICWT.2018.8527822 * [11] Angela Sara Cacciapuoti et al. “Quantum internet: networking challenges in distributed quantum computing” In _IEEE Network_ 34.1 IEEE, 2019, pp. 137–143 DOI: 10.1109/MNET.001.1900092 * [12] Quantum Internet Alliance, https://quantum-internet.team, 2022 * [13] Antonio Acín et al. “The quantum technologies roadmap: a European community view” In _New Journal of Physics_ 20.8 IOP Publishing, 2018, pp. 080201 DOI: 10.1088/1367-2630/aad1ea * [14] Robert Beals et al. “Efficient distributed quantum computing” In _Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences_ 469.2153 The Royal Society Publishing, 2013, pp. 20120686 DOI: 10.1098/rspa.2012.0686 * [15] Vasil S Denchev and Gopal Pandurangan “Distributed quantum computing: A new frontier in distributed systems or science fiction?” In _ACM SIGACT News_ 39.3 ACM New York, NY, USA, 2008, pp. 77–95 * [16] Rene Allerstorfer, Harry Buhrman, Florian Speelman and Philip Verduyn Lunel “On the Role of Quantum Communication and Loss in Attacks on Quantum Position Verification” arXiv, arXiv:2208.04341, 2022 DOI: 10.48550/ARXIV.2208.04341 * [17] Joseph F Fitzsimons “Private quantum computation: an introduction to blind quantum computing and related protocols” In _npj Quantum Information_ 3.1 Nature Publishing Group, 2017, pp. 1–11 DOI: 10.1038/s41534-017-0025-3 * [18] John-Mark A Allen et al. “Quantum common causes and quantum causal models” In _Physical Review X_ 7.3 APS, 2017, pp. 031021 DOI: 10.1103/PhysRevX.7.031021 * [19] Fabio Costa and Sally Shrapnel “Quantum causal modelling” In _New Journal of Physics_ 18.6 IOP Publishing, 2016, pp. 063032 DOI: 10.1088/1367-2630/18/6/063032 * [20] Jonathan Barrett, Robin Lorenz and Ognyan Oreshkov “Quantum causal models” In _arXiv preprint arXiv:1906.10726_ , 2019 DOI: 10.48550/arXiv.1906.10726 * [21] Judea Pearl “Causality” Cambridge University Press, 2009 DOI: 10.1017/CBO9780511803161 * [22] Peter Spirtes, Clark N Glymour, Richard Scheines and David Heckerman “Causation, prediction, and search” MIT press, 2000 DOI: 10.7551/mitpress/1754.001.0001 * [23] Giulio Chiribella, Giacomo Mauro D’Ariano and Paolo Perinotti “Theoretical framework for quantum networks” In _Physical Review A_ 80.2 APS, 2009, pp. 022339 DOI: 10.1103/PhysRevA.80.022339 * [24] Giulio Chiribella, G Mauro D’Ariano and Paolo Perinotti “Quantum circuit architecture” In _Physical review letters_ 101.6 APS, 2008, pp. 060401 DOI: 10.1103/PhysRevLett.101.060401 * [25] Felix A Pollock et al. “Non-Markovian quantum processes: Complete framework and efficient characterization” In _Physical Review A_ 97.1 APS, 2018, pp. 012127 DOI: 10.1103/PhysRevA.97.012127 * [26] Giulio Chiribella and Daniel Ebler “Optimal quantum networks and one-shot entropies” In _New Journal of Physics_ 18.9 IOP Publishing, 2016, pp. 093053 DOI: 10.1088/1367-2630/18/9/093053 * [27] Robert Konig, Renato Renner and Christian Schaffner “The operational meaning of min-and max-entropy” In _IEEE Transactions on Information theory_ 55.9 IEEE, 2009, pp. 4337–4347 DOI: 10.1109/TIT.2009.2025545 * [28] Mark M Wilde “From classical to quantum Shannon theory” In _arXiv preprint arXiv:1106.1445_ , 2011 DOI: 10.48550/arXiv.1106.1445 * [29] Robert Raussendorf and Hans J Briegel “A one-way quantum computer” In _Physical review letters_ 86.22 APS, 2001, pp. 5188 DOI: 10.1103/PhysRevLett.86.5188 * [30] Hans J Briegel et al. “Measurement-based quantum computation” In _Nature Physics_ 5.1 Nature Publishing Group, 2009, pp. 19–26 DOI: 10.1038/nphys1157 * [31] Robert Raussendorf, Daniel E Browne and Hans J Briegel “Measurement-based quantum computation on cluster states” In _Physical review A_ 68.2 APS, 2003, pp. 022312 DOI: 10.1103/PhysRevA.68.022312 * [32] Robert Raussendorf and Hans Briegel “Computational model underlying the one-way quantum computer” In _arXiv preprint quant-ph/0108067_ , 2001 DOI: 10.48550/arXiv.quant-ph/0108067 * [33] Richard Jozsa “An introduction to measurement based quantum computation” In _NATO Science Series, III: Computer and Systems Sciences. Quantum Information Processing-From Theory to Experiment_ 199, 2006, pp. 137–158 * [34] D Gross and J Eisert “Novel schemes for measurement-based quantum computation” In _Physical review letters_ 98.22 APS, 2007, pp. 220503 DOI: 10.1103/PhysRevLett.98.220503 * [35] Anne Broadbent, Joseph Fitzsimons and Elham Kashefi “Universal blind quantum computation” In _2009 50th Annual IEEE Symposium on Foundations of Computer Science_ , 2009, pp. 517–526 IEEE DOI: 10.1109/FOCS.2009.36 * [36] Atul Mantri, Tommaso F Demarie, Nicolas C Menicucci and Joseph F Fitzsimons “Flow ambiguity: A path towards classically driven blind quantum computation” In _Physical Review X_ 7.3 APS, 2017, pp. 031004 DOI: 10.1103/PhysRevX.7.031004 * [37] Tomoyuki Morimae and Keisuke Fujii “Blind quantum computation protocol in which Alice only makes measurements” In _Physical Review A_ 87.5 APS, 2013, pp. 050301 DOI: 10.1103/PhysRevA.87.050301 * [38] Tomoyuki Morimae “Verification for measurement-only blind quantum computing” In _Physical Review A_ 89.6 APS, 2014, pp. 060302 DOI: 10.1103/PhysRevA.89.060302 * [39] Hans J Briegel and Robert Raussendorf “Persistent entanglement in arrays of interacting particles” In _Physical Review Letters_ 86.5 APS, 2001, pp. 910 DOI: 10.1103/PhysRevLett.86.910 * [40] Marc Hein et al. “Entanglement in graph states and its applications” In _Volume 162: Quantum Computers, Algorithms and Chaos_ , Proceedings of the International School of Physics ”Enrico Fermi” IOS Press Ebooks, 2006, pp. 115–218 DOI: 10.3254/978-1-61499-018-5-115 * [41] Marc Hein, Jens Eisert and Hans J Briegel “Multiparty entanglement in graph states” In _Physical Review A_ 69.6 APS, 2004, pp. 062311 DOI: 10.1103/PhysRevA.69.062311 * [42] Mehdi Mhalla et al. “Which graph states are useful for quantum information processing?” In _Conference on Quantum Computation, Communication, and Cryptography_ , 2011, pp. 174–187 Springer DOI: 10.1007/978-3-642-54429-3“–˙˝12 * [43] Marc Hein et al. “Entanglement in graph states and its applications” In _arXiv preprint quant-ph/0602096_ , 2006 DOI: 10.48550/arXiv.quant-ph/0602096 * [44] Daniel E Browne, Elham Kashefi, Mehdi Mhalla and Simon Perdrix “Generalized flow and determinism in measurement-based quantum computation” In _New Journal of Physics_ 9.8 IOP Publishing, 2007, pp. 250 DOI: 10.1088/1367-2630/9/8/250 * [45] Vincent Danos and Elham Kashefi “Determinism in the one-way model” In _Physical Review A_ 74.5 APS, 2006, pp. 052310 DOI: 10.1103/PhysRevA.74.052310 * [46] Damian Markham and Elham Kashefi “Entanglement, flow and classical simulatability in measurement based quantum computation” In _Horizons of the Mind. A Tribute to Prakash Panangaden_ Springer, 2014, pp. 427–453 DOI: 10.1007/978-3-319-06880-0“–˙˝22 * [47] Vincent Danos, Elham Kashefi and Prakash Panangaden “The measurement calculus” In _Journal of the ACM (JACM)_ 54.2 ACM New York, NY, USA, 2007, pp. 8–es DOI: 10.1145/1219092.1219096 * [48] Maarten Nest, Jeroen Dehaene and Bart De Moor “Graphical description of the action of local Clifford transformations on graph states” In _Physical Review A_ 69.2 APS, 2004, pp. 022316 DOI: 10.1103/PhysRevA.69.022316 * [49] Maarten Nest, Jeroen Dehaene and Bart De Moor “Local unitary versus local Clifford equivalence of stabilizer states” In _Physical Review A_ 71.6 APS, 2005, pp. 062323 DOI: 10.1103/PhysRevA.75.032325 * [50] Katja Ried et al. “A quantum advantage for inferring causal structure” In _Nature Physics_ 11.5 Nature Publishing Group, 2015, pp. 414–420 DOI: 10.1038/nphys3266 * [51] Joseph F Fitzsimons, Jonathan A Jones and Vlatko Vedral “Quantum correlations which imply causation” In _Scientific reports_ 5.1 Nature Publishing Group, 2015, pp. 1–7 DOI: 10.1038/srep18281 * [52] Giulio Chiribella and Daniel Ebler “Quantum speedup in the identification of cause–effect relations” In _Nature communications_ 10.1 Nature Publishing Group, 2019, pp. 1–8 DOI: 10.1038/s41467-019-09383-8 * [53] Jonas M Kübler and Daniel Braun “Two-qubit causal structures and the geometry of positive qubit-maps” In _New Journal of Physics_ 20.8 IOP Publishing, 2018, pp. 083015 DOI: 10.1088/1367-2630/aad612 * [54] Stephen D Bartlett, Terry Rudolph and Robert W Spekkens “Reference frames, superselection rules, and quantum information” In _Reviews of Modern Physics_ 79.2 APS, 2007, pp. 555 DOI: 10.1103/RevModPhys.79.555 * [55] Man-Duen Choi “Completely positive linear maps on complex matrices” In _Linear algebra and its applications_ 10.3 Elsevier, 1975, pp. 285–290 DOI: 10.1016/0024-3795(75)90075-0 * [56] Andrzej Jamiołkowski “Linear transformations which preserve trace and positive semidefiniteness of operators” In _Reports on Mathematical Physics_ 3.4 Elsevier, 1972, pp. 275–278 DOI: 10.1016/0034-4877(72)90011-0 * [57] Daniel Gottesman “Stabilizer codes and quantum error correction” California Institute of Technology, 1997 * [58] Mehdi Mhalla and Simon Perdrix “Finding optimal flows efficiently” In _International Colloquium on Automata, Languages, and Programming_ , 2008, pp. 857–868 Springer DOI: 10.1007/978-3-540-70575-8“–˙˝70 * [59] Niel De Beaudrap “Finding flows in the one-way measurement model” In _Physical Review A_ 77.2 APS, 2008, pp. 022328 DOI: 10.1103/PhysRevA.77.022328 * [60] Renato Renner “Security of quantum key distribution” In _International Journal of Quantum Information_ 6.01 World Scientific, 2008, pp. 1–127 DOI: 10.1142/S0219749908003256 * [61] Marco Tomamichel, Roger Colbeck and Renato Renner “A fully quantum asymptotic equipartition property” In _IEEE Transactions on information theory_ 55.12 IEEE, 2009, pp. 5840–5847 DOI: 10.1109/TIT.2009.2032797 * [62] Marco Tomamichel “Quantum information processing with finite resources: mathematical foundations” Springer, 2015 DOI: 10.1007/978-3-319-21891-5 * [63] Matthew S Leifer and Robert W Spekkens “Towards a formulation of quantum theory as a causally neutral theory of Bayesian inference” In _Physical Review A_ 88.5 APS, 2013, pp. 052130 DOI: 10.1103/PhysRevA.88.052130 * [64] Christopher Portmann et al. “Causal boxes: quantum information-processing systems closed under composition” In _IEEE Transactions on Information Theory_ 63.5 IEEE, 2017, pp. 3277–3305 DOI: 10.1109/TIT.2017.2676805 * [65] Ueli Maurer “Abstract models of computation in cryptography” In _IMA International Conference on Cryptography and Coding_ , 2005, pp. 1–12 Springer DOI: 10.1007/11586821“–˙˝1 * [66] Christopher Portmann and Renato Renner “Security in quantum cryptography” In _Reviews of Modern Physics_ 94.2 APS, 2022, pp. 025008 DOI: 10.1103/RevModPhys.94.025008 * [67] Atul Mantri, Tommaso F Demarie and Joseph F Fitzsimons “Universality of quantum computation with cluster states and (X, Y)-plane measurements” In _Scientific reports_ 7.1 Nature Publishing Group, 2017, pp. 1–7 DOI: 10.1038/srep42861 * [68] Giacomo Mauro D’Ariano “Causality re-established” In _Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences_ 376.2123 The Royal Society Publishing, 2018, pp. 20170313 DOI: 10.1098/rsta.2017.0313 * [69] Bernhard Schölkopf et al. “On causal and anticausal learning” In _arXiv preprint arXiv:1206.6471_ , 2012 DOI: 10.48550/arXiv.1206.6471 * [70] Christopher J Wood and Robert W Spekkens “The lesson of causal discovery algorithms for quantum correlations: Causal explanations of Bell-inequality violations require fine-tuning” In _New Journal of Physics_ 17.3 IOP Publishing, 2015, pp. 033002 DOI: 10.1088/1367-2630/17/3/033002 * [71] Ognyan Oreshkov, Fabio Costa and Časlav Brukner “Quantum correlations with no causal order” In _Nature communications_ 3.1 Nature Publishing Group, 2012, pp. 1–8 DOI: 10.1038/ncomms2076 * [72] Isaac D. Smith and Marius Krumm “Min-Entropy and MBQC”, https://github.com/IsaacDSmith/Min-Entropy_and_MBQC * [73] Steven Diamond and Stephen Boyd “CVXPY: A Python-embedded modeling language for convex optimization” In _Journal of Machine Learning Research_ 17.83, 2016, pp. 1–5 * [74] Akshay Agrawal, Robin Verschueren, Steven Diamond and Stephen Boyd “A rewriting system for convex optimization problems” In _Journal of Control and Decision_ 5.1, 2018, pp. 42–60 * [75] Brendan O’Donoghue, Eric Chu, Neal Parikh and Stephen Boyd “SCS: Splitting Conic Solver, version 3.2.2”, https://github.com/cvxgrp/scs, 2021 * [76] Brendan O’Donoghue, Eric Chu, Neal Parikh and Stephen Boyd “Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding” In _Journal of Optimization Theory and Applications_ 169.3, 2016, pp. 1042–1068 URL: http://stanford.edu/~boyd/papers/scs.html ## Appendix A Preliminaries Section Supplementary This appendix houses some supporting discussion for Section 2. ### A.1 Motivating Quantum Combs In this subsection, we briefly motivate the definition of quantum comb given in the main text (2.1), including why this is an appropriate representation of quantum networks and what the conditions of the definition correspond to. We provide these explanations here for completeness; they are drawn from references such as [24, 23] which should be referred to for further details. The definition of quantum comb given earlier is as an operator on a tensor product of Hilbert spaces, which represents a series of connected quantum channels. This correspondence between operators on a tensor product of spaces and maps between these spaces, that is, between elements of $\mathcal{L}(\mathcal{H}_{A}\otimes\mathcal{H}_{B})$ and $\mathcal{L}(\mathcal{L}(\mathcal{H}_{A}),\mathcal{L}(\mathcal{H}_{B}))$ respectively, makes use of the Choi-Jamiolkowski isomorphism [55, 56]. Since we are most interested in the operator form, we make the following definition. ###### Definition A.1. For a linear map $\mathcal{E}:\mathcal{L}(\mathcal{H}_{A})\rightarrow\mathcal{L}(\mathcal{H}_{B})$, its Choi operator is defined as: $E:=(\mathcal{E}\otimes I_{\mathcal{H}_{A^{\prime}}})(\ket{\Phi^{+}}\\!\\!\bra{\Phi^{+}}).$ (76) Here, $\mathcal{H}_{A^{\prime}}\simeq\mathcal{H}_{A}$ is a copy of the input space, and $\ket{\Phi^{+}}:=\sum_{i=0}^{d_{A}-1}\ket{ii}_{A^{\prime}A}$ is an unnormalised maximally entangled state on $\mathcal{H}_{A^{\prime}}\otimes\mathcal{H}_{A}$. The defining properties of CPTP maps correspond to properties of the Choi operator: ###### Lemma A.1. The Choi operator $E$ of a linear map $\mathcal{E}$ satisfies the following properties: 1. 1. $E$ is positive semi-definite, denoted $E\geq 0$, iff $\mathcal{E}$ is completely positive; 2. 2. $E$ is Hermitian iff $\mathcal{E}$ is Hermitian-preserving; 3. 3. $\operatorname{Tr}_{B}\left[E\right]\leq I_{A}$ iff $\mathcal{E}$ is trace non-increasing; 4. 4. $\operatorname{Tr}_{B}\left[E\right]=I_{A}$ iff $\mathcal{E}$ is trace- preserving. One notices that the first and last items of the above lemma feature in 2.1, however iteratively for the latter item. To arrive at the iterative constraints, consider a simple example of two CPTP maps $\mathcal{E}_{1}:\mathcal{L}(\mathcal{H}_{A_{1}})\rightarrow\mathcal{L}(\mathcal{H}_{B_{1}}\otimes\mathcal{H}_{C})$ and $\mathcal{E}_{2}:\mathcal{L}(\mathcal{H}_{A_{2}}\otimes\mathcal{H}_{C})\rightarrow\mathcal{H}_{B_{2}}$ with Choi operators $E_{1}$ and $E_{2}$ respectively. The composition of the two maps over the space $H_{C}$, denoted $\mathcal{E}_{2}\circ_{C}\mathcal{E}_{1}$ is a linear operator from $\mathcal{L}(\mathcal{H}_{A_{1}}\otimes\mathcal{H}_{A_{2}})$ to $\mathcal{L}(\mathcal{H}_{B_{1}}\otimes\mathcal{H}_{B_{2}})$, and the corresponding Choi operator is given by $\displaystyle E=E_{1}\ast_{C}E_{2}$ (77) where $\ast_{C}$ denotes the link product over the space $C$, which is the analogue of composition in the Choi operator picture. By tracing over $B_{2}$, we get $\displaystyle\operatorname{Tr}_{B_{2}}\left[E\right]$ $\displaystyle=\operatorname{Tr}_{B_{2}}\left[E_{1}\ast_{C}E_{2}\right]$ (78) $\displaystyle=E_{1}\ast_{C}\operatorname{Tr}_{B_{2}}\left[E_{2}\right]$ (79) $\displaystyle=E_{1}\ast_{C}(I_{C}\otimes I_{A_{2}})$ (80) $\displaystyle=I_{A_{2}}\otimes\operatorname{Tr}_{C}\left[E_{1}\right]$ (81) where we have used the trace-preservation criterion of A.1. Noting that $\operatorname{Tr}_{C}\left[E_{1}\right]$ is a positive semi-definite operator, we see that the last line above is indeed of the form of the constraints in 2.1. By tracing also over $B_{1}$, we can use the trace- preservation again to obtain the terminal constraint of normalisation to $1$. This reasoning extends to any number of composed maps, which helps demonstrate the conciseness of the comb notation. The remainder of this appendix subsection provides the proof of 2.1 given in Section 2.1. ###### Proof of 2.1. The proof consists of showing that the sequence of partial trace constraints (2.1) are satisfied for both the sequence where $X$ is traced over first then $A_{n}^{\operatorname{out}},...,A_{1}^{\operatorname{out}}$ and the sequence $A_{n}^{\operatorname{out}},...,A_{1}^{\operatorname{out}},X$. Starting with the former, tracing over $X$ gives $\displaystyle\operatorname{Tr}_{X}[D]=\sum_{x}P(x)\sigma_{x}\equiv I_{\mathbb{C}}\otimes\sum_{x}P(x)\sigma_{x}.$ (82) Since the $\sigma_{x}$ are normalised combs, and since the set of normalised combs is convex, $\sum_{x}P(x)\sigma_{x}$ is also a normalised comb and so the remaining trace conditions are satisfied, thus showing that $D\in\operatorname{Comb}(A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}},\mathbb{C}\rightarrow X)$. For the other sequence, define a sequence of operators $D_{k}$, $k=0,...,n+1$ via $\displaystyle D_{k}$ $\displaystyle:=\sum_{x}P(x)\ket{x}\\!\\!\bra{x}\otimes C_{x,k-1}\quad\quad\forall k=1,...,n+1$ (83) $\displaystyle D_{0}$ $\displaystyle:=\sum_{x}P(x)$ (84) where $C_{x,0},...,C_{x,n}$ denote the positive semi-definite operators that satisfy the comb conditions for $\sigma_{x}$, that is: $\displaystyle\sigma_{x}$ $\displaystyle=C_{x,n};$ (85) $\displaystyle\operatorname{Tr}_{A_{j}^{\operatorname{out}}}[C_{x,j}]$ $\displaystyle=I_{A_{j}^{\operatorname{in}}}\otimes C_{x,j-1}\quad\quad\forall j=1,...,n;$ (86) $\displaystyle C_{x,0}$ $\displaystyle=1.$ (87) It follows that $D_{n+1}=D$, that $\displaystyle\operatorname{Tr}_{A_{k-1}^{\operatorname{out}}}\left[D_{k}\right]$ $\displaystyle=\sum_{x}P(x)\ket{x}\\!\\!\bra{x}\otimes\operatorname{Tr}_{A_{k-1}^{\operatorname{out}}}[C_{x,k-1}]$ (88) $\displaystyle=\sum_{x}P(x)\ket{x}\\!\\!\bra{x}\otimes I_{A_{k-1}^{\operatorname{in}}}\otimes C_{x,k-2}$ (89) $\displaystyle=I_{A_{k-1}^{\operatorname{in}}}\otimes D_{k-1}$ (90) for all $k=2,...,n+1$, that $\displaystyle\operatorname{Tr}_{X}[D_{1}]$ $\displaystyle=\operatorname{Tr}_{X}\left[\sum_{x}P(x)\ket{x}\\!\\!\bra{x}\otimes 1\right]$ (91) $\displaystyle=\sum_{x}P(x)$ (92) $\displaystyle=I_{\mathbb{C}}\otimes D_{0}$ (93) and that $D_{0}=1$. Thus, $D$ is also in $\operatorname{Comb}(\mathbb{C}\rightarrow X,A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}})$. ∎ ### A.2 Equivalent Correction Methods for Measurement Outcomes Here, we elaborate on the discussion in Section 2.2 regarding the correction of measurement outcomes in MBQC via products of stabilisers $K_{v}$ of graph state $\ket{G}$. To recap what is written in the main text, the correction for negative measurement outcomes starts by noticing that, due to the restriction of the allowed measurement to the planes of the Bloch sphere, the negative projection operators are related to the positive ones via conjugation by $X$, $Z$ or their product, which each exist as a factor of a (product of) stabiliser(s). By applying the remainder of an appropriate stabiliser in the case when a negative outcome obtains, a symmetry of the graph state is completed and in effect, a positive projection has been enacted instead. We demonstrate this with the following example where a vertex $v\in V$ is measured in the $XY$-plane (the same reasoning applies for the other measurement planes). We let $v^{\prime}\in V$ be a distinguished neighbour of $v$ and denote all pending measurements on $\ket{G}$ (also in the $XY$-plane for simplicity) as $\ket{\pm_{\alpha_{\tilde{v}}}}\bra{\pm_{\alpha_{\tilde{v}}}}$ for $\tilde{v}\in V\setminus\\{v\\}$. The notation $K_{v^{\prime}}|_{\setminus v}$ denotes the stabiliser $K_{v^{\prime}}$ with the tensor factor corresponding to $v$ replaced by $I_{v}$. $\displaystyle\operatorname{Tr}$ $\displaystyle\left[\left(\ket{-_{\alpha_{v}}}\\!\\!\bra{-_{\alpha_{v}}}_{v}\bigotimes_{\tilde{v}\in V\setminus\\{v\\}}\ket{\pm_{\alpha_{\tilde{v}}}}\\!\\!\bra{\pm_{\alpha_{\tilde{v}}}}_{\tilde{v}}\right)\rho_{G}\right]=\operatorname{Tr}\left[\left(Z_{v}^{\dagger}\ket{+_{\alpha_{v}}}\\!\\!\bra{+_{\alpha_{v}}}_{v}Z_{v}\bigotimes_{\tilde{v}\in V\setminus\\{v\\}}\ket{\pm_{\alpha_{\tilde{v}}}}\\!\\!\bra{\pm_{\alpha_{\tilde{v}}}}_{\tilde{v}}\right)\rho_{G}\right]$ (94) $\displaystyle=\operatorname{Tr}\left[\left(\ket{+_{\alpha_{v}}}\\!\\!\bra{+_{\alpha_{v}}}_{v}\bigotimes_{\tilde{v}\in V\setminus\\{v\\}}\ket{\pm_{\alpha_{\tilde{v}}}}\\!\\!\bra{\pm_{\alpha_{\tilde{v}}}}_{\tilde{v}}\right)Z_{v}\rho_{G}Z_{v}^{\dagger}\right]$ (95) $\displaystyle=\operatorname{Tr}\left[\left(\ket{+_{\alpha_{v}}}\\!\\!\bra{+_{\alpha_{v}}}_{v}\bigotimes_{\tilde{v}\in V\setminus\\{v\\}}\ket{\pm_{\alpha_{\tilde{v}}}}\\!\\!\bra{\pm_{\alpha_{\tilde{v}}}}_{\tilde{v}}\right)K_{v^{\prime}}|_{\setminus v}\rho_{G}K_{v^{\prime}}|_{\setminus v}^{\dagger}\right]$ (96) $\displaystyle=\operatorname{Tr}\left[\left(\ket{+_{\alpha_{v}}}\\!\\!\bra{+_{\alpha_{v}}}_{v}\otimes X_{v^{\prime}}^{\dagger}\ket{\pm_{\alpha_{v^{\prime}}}}\\!\\!\bra{\pm_{\alpha_{v^{\prime}}}}X_{v^{\prime}}^{\dagger}\bigotimes_{\tilde{v}\in N_{v^{\prime}}^{G}\setminus\\{v\\}}Z_{\tilde{v}}^{\dagger}\ket{\pm_{\alpha_{\tilde{v}}}}\\!\\!\bra{\pm_{\alpha_{\tilde{v}}}}_{\tilde{v}}Z_{\tilde{v}}\bigotimes_{\hat{v}\in V\setminus(N_{v^{\prime}}^{G}\cup\\{v^{\prime}\\})}\ket{\pm_{\alpha_{\hat{v}}}}\\!\\!\bra{\pm_{\alpha_{\hat{v}}}}_{\hat{v}}\right)\rho_{G}\right]$ (97) $\displaystyle=\operatorname{Tr}\left[\left(\ket{+_{\alpha_{v}}}\\!\\!\bra{+_{\alpha_{v}}}_{v}\otimes X_{v^{\prime}}^{\dagger}\ket{\pm_{\alpha_{v^{\prime}}}}\\!\\!\bra{\pm_{\alpha_{v^{\prime}}}}X_{v^{\prime}}^{\dagger}\bigotimes_{\tilde{v}\in N_{v^{\prime}}^{G}\setminus\\{v\\}}Z_{\tilde{v}}^{\dagger}\ket{\pm_{\alpha_{\tilde{v}}}}\\!\\!\bra{\pm_{\alpha_{\tilde{v}}}}_{\tilde{v}}Z_{\tilde{v}}\bigotimes_{\hat{v}\in V\setminus(N_{v^{\prime}}^{G}\cup\\{v^{\prime}\\})}\ket{\pm_{\alpha_{\hat{v}}}}\\!\\!\bra{\pm_{\alpha_{\hat{v}}}}_{\hat{v}}\right)\rho_{G}\right]$ (98) $\displaystyle=\operatorname{Tr}\left[\left(\ket{+_{\alpha_{v}}}\\!\\!\bra{+_{\alpha_{v}}}_{v}\otimes\ket{\pm_{-\alpha_{v^{\prime}}}}\\!\\!\bra{\pm_{-\alpha_{v^{\prime}}}}\bigotimes_{\tilde{v}\in N_{v^{\prime}}^{G}\setminus\\{v\\}}\ket{\pm_{\alpha_{\tilde{v}}+\pi}}\\!\\!\bra{\pm_{\alpha_{\tilde{v}}+\pi}}_{\tilde{v}}\bigotimes_{\hat{v}\in V\setminus(N_{v^{\prime}}^{G}\cup\\{v^{\prime}\\})}\ket{\pm_{\alpha_{\hat{v}}}}\\!\\!\bra{\pm_{\alpha_{\hat{v}}}}_{\hat{v}}\right)\rho_{G}\right]$ (99) where the last line uses the useful correspondence between applying $X$ or $Z$ and the update of measurement angles presented in Equation 9 (with the modulo $2\pi$ notation suppressed). Thus, a positive outcome at $v$ is recovered at the cost of applying the remainder of the stabiliser $K_{v^{\prime}}$ to either the graph state $\rho_{G}$ (Equation 96) or the measurement operators on $v^{\prime}\cup N_{v^{\prime}}^{G}\setminus\\{v\\}$ via a classical change of measurement parameter (Equation 99). The former perspective is used in Section 5 and the latter is key to the BQC protocol in Section 4. Furthermore, we believe the above discussion can provide a more intuitive understanding of corrections work in MBQC and motivate the definition of gflow (2.4). ## Appendix B Min-Entropy Section Supplementary In Section 3.1, we considered combs of the form $\displaystyle D=\sum_{x}P(x)\ket{x}\\!\\!\bra{x}\otimes\sigma_{x}\in\operatorname{Comb}(A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}},\mathbb{C}\rightarrow X)$ (100) where all the $\sigma_{x}\in\operatorname{Comb}(A_{1}^{\operatorname{in}}\rightarrow A_{1}^{\operatorname{out}},...,A_{n}^{\operatorname{in}}\rightarrow A_{n}^{\operatorname{out}})$ are classical. The ultimate aim was to provide an interpretation of the guessing probability, $P_{\text{guess}}(X|\boldsymbol{A}^{\operatorname{in}},\boldsymbol{A}^{\operatorname{out}})$, in terms of Bayesian updating since such an interpretation exists for the state min-entropy (see eg., [[, Section 6.1.4,]]tomamichel2015quantum). For completeness, we present this interpretation here and discuss the similarities and differences with the combs case. Let $D$ be a classical-quantum state on $\mathcal{H}_{X}\otimes\mathcal{H}_{Y}$ (i.e. $D\in\operatorname{Comb}(\mathbb{C}\rightarrow Y,\mathbb{C}\rightarrow X)$) such that each $\sigma_{x}\in\mathcal{L}(\mathcal{H}_{Y})$ is classical, that is, they can be written as $\displaystyle\sigma_{x}=\sum_{y}P(y|x)\ket{y}\\!\\!\bra{y}$ (101) for $\\{\ket{y}\\}_{y}$ a common choice of orthonormal basis for $\mathcal{H}_{Y}$ and for $P(y|x)$ a conditional probability distribution. We can thus write $\displaystyle D=\sum_{x,y}P(x)P(y|x)\ket{xy}\\!\\!\bra{xy}.$ (102) Applying Bayes’ rule, it follows that $\displaystyle D=\sum_{x,y}P(x|y)P(y)\ket{xy}\\!\\!\bra{xy}.$ (103) By maximising over $x$ for each $y$, we obtain the inequality: $\displaystyle D$ $\displaystyle\leq\sum_{x,y}\left[\max_{\tilde{x}}P(\tilde{x}|y)P(y)\right]\ket{xy}\\!\\!\bra{xy}$ (104) $\displaystyle=I_{X}\otimes\sum_{y}\max_{\tilde{x}}P(\tilde{x}|y)P(y)\ket{y}\\!\\!\bra{y}.$ (105) The second tensor factor above is an (in general) unnormalised state on $Y$, and moreover it can be shown that this unnormalised state is a minimal such state $\rho_{Y}$ for which $D\leq I_{X}\otimes\rho_{Y}$ holds (this follows from the proof of the left-hand inequality of 3.2). Thus, using the unnormalised version of the min-entropy (recall Equation 12), we arrive at $\displaystyle P_{\text{guess}}(X|Y)_{D}=\sum_{y}\max_{\tilde{x}}P(\tilde{x}|y)P(y)$ (106) and hence also at the desired interpretation of the guessing probability in terms of Bayesian updating: the guessing probability is the maximal Bayesian update for each interaction (denoted by $y$) averaged over all possible interactions. Clearly, for $P_{\text{guess}}$ to take value $1$, we must have perfect updates for every interaction (in the support of $P(y)$). It is worthwhile emphasising here that, in the state case, the states $\sigma_{x}$ can be considered as combs with trivial input spaces, ie. $\sigma_{x}\in\operatorname{Comb}(\mathbb{C}\rightarrow Y)$, and consequentially, we are guaranteed that $\sum_{y}\max_{x}P(x|y)P(y)\ket{y}\\!\\!\bra{y}$ is an unnormalised state. For the general case where the $\sigma_{x}$ are classical combs with non-trivial input spaces, we have no analogous guarantee as we will now discuss. In the main text, we used the conditional Bayes’ rule and an independence condition to write the classical-classical comb $D$ as $\displaystyle D=\sum_{x,\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}P(x|\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}})P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{a}^{\operatorname{in}})\ket{x,\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}\bra{x,\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}$ (107) Similarly to above, we can obtain the inequality: $\displaystyle D\leq I_{X}\otimes\sum_{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}\max_{x}P(x|\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}})P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{a}^{\operatorname{in}})\ket{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}\\!\\!\bra{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}$ (108) However, unlike above, we are not guaranteed that the second tensor factor is an unnormalised classical comb: $\sum_{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}\max_{x}P(x|\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}})P(\boldsymbol{a}^{\operatorname{out}}|\boldsymbol{a}^{\operatorname{in}})\ket{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}\\!\\!\bra{\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}}}$ may fail the required marginalisation conditions due to the dependence on the inputs $\boldsymbol{a}^{\operatorname{in}}$ that persists in the distribution $P(x|\boldsymbol{a}^{\operatorname{in}},\boldsymbol{a}^{\operatorname{out}})$ in the maximum. The trace of this operator (appropriately normalised by the dimension of the input spaces) still provides a lower bound for the guessing probability, just as for the state case above, but the assurance that this bound can be reached is lacking. To obtain an upper bound, we can construct an operator that removes the dependence on the inputs by also maximising over the $\boldsymbol{a}^{\operatorname{in}}$, which then trivially satisfies the required marginalisation conditions. The proof of these bounds are given more formally in the following proof of 3.2. Note that, for notational simplicity, we have replaced $\boldsymbol{a}^{\operatorname{in}}$ by $\boldsymbol{a}$ and $\boldsymbol{a}^{\operatorname{out}}$ by $\boldsymbol{b}$ in the following. Also note that the proof makes use of lemmas that are stated and proved after the current proof. ###### Proof of 3.2. The lower bound is established by showing that any positive semi-definite operator $\Gamma$ on $\bigotimes_{i=1}^{n}\mathcal{H}_{A_{i}}\otimes\mathcal{H}_{B_{i}}$ that satisfies $I_{X}\otimes\Gamma\geq D$ must have trace greater than or equal to $\sum_{\boldsymbol{a},\boldsymbol{b}}\max_{x}P(x|\boldsymbol{a},\boldsymbol{b})P(\boldsymbol{b}|\boldsymbol{a})$. Let $\Gamma$ be any positive semi-definite operator on $\bigotimes_{i=1}^{n}\mathcal{H}_{A_{i}}\otimes\mathcal{H}_{B_{i}}$ such that $I_{X}\otimes\Gamma\geq D$. By B.1, it follows that $\operatorname{diag}(I_{X}\otimes\Gamma)=I_{X}\otimes\operatorname{diag}(\Gamma)\geq D$. We write $\displaystyle I_{X}\otimes\operatorname{diag}(\Gamma)=\sum_{x,\boldsymbol{a},\boldsymbol{b}}\alpha_{\boldsymbol{a}\boldsymbol{b}}\ket{x\boldsymbol{a}\boldsymbol{b}}\\!\\!\bra{x\boldsymbol{a}\boldsymbol{b}}$ where the $\alpha_{\boldsymbol{a}\boldsymbol{b}}$ are all real and non- negative by positive semi-definiteness of $\Gamma$. The condition $I_{X}\otimes\operatorname{diag}(\Gamma)\geq D$ induces a further condition on the $\alpha$ terms: we must have for all $\boldsymbol{a},\boldsymbol{b}$ that, for all $x$, $\displaystyle\alpha_{\boldsymbol{a}\boldsymbol{b}}\geq P(x|\boldsymbol{a},\boldsymbol{b})P(\boldsymbol{b}|\boldsymbol{a})$ which in particular enforces that $\displaystyle\alpha_{\boldsymbol{a}\boldsymbol{b}}\geq\max_{x}P(x|\boldsymbol{a},\boldsymbol{b})P(\boldsymbol{b}|\boldsymbol{a})$ and it thus follows that $\displaystyle\operatorname{Tr}[\Gamma]\geq\sum_{\boldsymbol{a},\boldsymbol{b}}\max_{x}P(x|\boldsymbol{a},\boldsymbol{b})P(\boldsymbol{b}|\boldsymbol{a}).$ The upper bound is established by showing that $\displaystyle\Upsilon:=\sum_{\boldsymbol{a},\boldsymbol{b}}\max_{x,\boldsymbol{\tilde{a}}}P(x|\boldsymbol{\tilde{a}},\boldsymbol{b})P(\boldsymbol{b}|\boldsymbol{\tilde{a}})\ket{\boldsymbol{a}\boldsymbol{b}}\\!\\!\bra{\boldsymbol{a}\boldsymbol{b}}$ (109) is a valid unnormalised comb, since $I_{X}\otimes\Upsilon\geq D$ clearly holds. Defining $f$ via $f(\boldsymbol{b},\boldsymbol{a})=\max_{x,\boldsymbol{\tilde{a}}}P(x|\boldsymbol{\tilde{a}},\boldsymbol{b})P(\boldsymbol{b}|\boldsymbol{\tilde{a}})$, non-negativity is immediate. Moreover, due to the maximum over $\boldsymbol{\tilde{a}}$, $f$ has no dependence on $\boldsymbol{a}$: $f(\boldsymbol{b},\boldsymbol{a})=f(\boldsymbol{b},\boldsymbol{a^{\prime}})$ for all $\boldsymbol{a},\boldsymbol{a^{\prime}}$. By defining the functions $f^{(k)}$ via $\displaystyle f^{(k)}(b_{1},...,b_{k},a_{1},...,a_{k}):=\sum_{b_{n},...,b_{k+1}}f(\boldsymbol{b},\boldsymbol{a})$ (110) for all $k=1,...,n$, the required conditions (non-negativity and the marginalisation conditions - recall Equation 2) are trivially satisfied due to the non-negativity and independence from $\boldsymbol{a}$ exhibited by $f$. The sum over $b_{1}$ of $f^{(1)}$ so defined is also clearly positive since $\max_{x,\boldsymbol{a}}P(x|\boldsymbol{a},\boldsymbol{b})P(\boldsymbol{b}|\boldsymbol{a})$ must be non-zero for some $\boldsymbol{b}$, and so $\Upsilon$ is indeed an unnormalised classical comb. ∎ The following two lemmas establish that, for any classical comb $D$, we need only (un)normalised classical combs $\Gamma$ in the minimum formulations of the min-entropy since if some non-classical comb achieves the minimum, then there exists a related classical comb that does also. ###### Lemma B.1. Let $A$ and $B$ be operators on $\mathcal{H}$, where $A$ is diagonal in a specific basis and $B$ is an arbitrary positive semi-definite operator. If $B-A\geq 0$ then $\operatorname{diag}(B)-A\geq 0$. ###### Proof. Let $\mathcal{E}:\mathcal{L}(\mathcal{H})\rightarrow\mathcal{L}(\mathcal{H})$ be a decohering channel in the basis for which $A$ is diagonal. Since this is in particular a positive map and $B-A$ is positive semi-definite by assumption, it follows that $\mathcal{E}(B-A)$ is also positive semi-definite Linearity of $\mathcal{E}$ ensures that $\mathcal{E}(B-A)=\mathcal{E}(B)-\mathcal{E}(A)=\operatorname{diag}(B)-A$ giving the result. ∎ ###### Lemma B.2. If $\Gamma$ is a normalised (resp. unnormalised) quantum comb, then $\operatorname{diag}(\Gamma)$ is a normalised (resp. unnormalised) classical comb. ###### Proof. The proof is essentially immediate by definition but the details are spelt out nonetheless. Let $\Gamma$ be a normalised (resp. unnormalised) quantum comb in $\operatorname{Comb}(A_{1}\rightarrow B_{1},...,A_{n}\rightarrow B_{n})$ and let $C^{(k)}$, $k=0,...,n$ be positive semi-definite operators that satisfy $C^{(n)}=\Gamma$, $\displaystyle\operatorname{Tr}_{B_{k}}\left[C^{(k)}\right]=I_{A_{k}}\otimes C^{(k-1)}$ (111) for $k=1,...,n$ and $C^{(0)}=1$ (resp. $C^{(0)}>0$). Denoting the $\boldsymbol{a}\boldsymbol{b}^{\text{th}}$ diagonal element of $\Gamma$ by $\alpha_{\boldsymbol{a}\boldsymbol{b}}$, we define $f$ via $f(\boldsymbol{b},\boldsymbol{a})=\alpha_{\boldsymbol{a}\boldsymbol{b}}$. Thus, we can write $\displaystyle\operatorname{diag}(\Gamma)=\sum_{\boldsymbol{a},\boldsymbol{b}}f(\boldsymbol{b},\boldsymbol{a})\ket{\boldsymbol{a}\boldsymbol{b}}\\!\\!\bra{\boldsymbol{a}\boldsymbol{b}}.$ (112) Since $\Gamma$ is positive semi-definite every $\alpha_{\boldsymbol{a}\boldsymbol{b}}$, and hence every $f(\boldsymbol{b},\boldsymbol{a})$, is non-negative. By denoting the diagonal elements of $C^{(k)}$ similarly as $\alpha_{\boldsymbol{a}_{1:k}\boldsymbol{b}_{1:k}}$, where $\boldsymbol{a}_{1:k}:=a_{1}a_{2}...a_{k}$ and similarly for $\boldsymbol{b}_{1:k}$, we define $f^{(k)}(\boldsymbol{b}_{1:k},\boldsymbol{a}_{1:k}):=\alpha_{\boldsymbol{a}_{1:k}\boldsymbol{b}_{1:k}}$ for all $k=1,...,n$ and take $f^{(0)}:=C^{(0)}$. Since taking the partial trace and applying $\operatorname{diag}(\cdot)$ commute, that is, $\displaystyle\operatorname{diag}\left(\operatorname{Tr}_{B_{k}}\left[C^{(k)}\right]\right)=\operatorname{Tr}_{B_{k}}\left[\operatorname{diag}(C^{(k)})\right]$ (113) it follows that $\displaystyle\operatorname{Tr}_{B_{k}}\left[\operatorname{diag}(C^{(k)})\right]=\operatorname{diag}(I_{A_{k}}\otimes C^{(k-1)})=I_{A_{k}}\otimes\operatorname{diag}(C^{(k-1)}).$ (114) The left-hand side can be written as $\displaystyle\sum_{\boldsymbol{a}_{1:k},\boldsymbol{b}_{1:k-1}}\left(\sum_{b_{k}}\alpha_{\boldsymbol{a}_{1:k}\boldsymbol{b}_{1:k}}\right)\ket{\boldsymbol{a}_{1:k}\boldsymbol{b}_{1:k-1}}\\!\\!\bra{\boldsymbol{a}_{1:k}\boldsymbol{b}_{1:k-1}}=\sum_{\boldsymbol{a}_{1:k},\boldsymbol{b}_{1:k-1}}\left(\sum_{b_{k}}f^{(k)}(\boldsymbol{b}_{1:k},\boldsymbol{a}_{1:k})\right)\ket{\boldsymbol{a}_{1:k}\boldsymbol{b}_{1:k-1}}\\!\\!\bra{\boldsymbol{a}_{1:k}\boldsymbol{b}_{1:k-1}}$ (115) and the right-hand side as $\displaystyle\sum_{\boldsymbol{a}_{1:k},\boldsymbol{b}_{1:k-1}}\alpha_{\boldsymbol{a}_{1:k-1}\boldsymbol{b}_{1:k-1}}\ket{\boldsymbol{a}_{1:k}\boldsymbol{b}_{1:k-1}}\\!\\!\bra{\boldsymbol{a}_{1:k}\boldsymbol{b}_{1:k-1}}=\sum_{\boldsymbol{a}_{1:k},\boldsymbol{b}_{1:k-1}}f^{(k-1)}(\boldsymbol{b}_{1:k-1},\boldsymbol{a}_{1:k-1})\ket{\boldsymbol{a}_{1:k}\boldsymbol{b}_{1:k-1}}\\!\\!\bra{\boldsymbol{a}_{1:k}\boldsymbol{b}_{1:k-1}}$ (116) which establishes the required marginalisation condition on the $f^{(k)}$: $\sum_{b_{k}}f^{(k)}(\boldsymbol{b}_{1:k},\boldsymbol{a}_{1:k})=f^{(k-1)}(\boldsymbol{b}_{1:k-1},\boldsymbol{a}_{1:k-1})$ for $k=2,...,n$ as well as the edge case of $\sum_{b_{1}}f^{(1)}(b_{1},a_{1})=f^{(0)}$. ∎ The following is the proof of 3.1 which states that the min-entropy for multi- round combs, i.e. of the form $D^{(m)}$, is non-increasing as the number of rounds $m$ increases. ###### Proof of 3.1. Let $\lambda\in\mathbb{R}$ and $\Gamma\in\operatorname{Comb}\left((A_{1}^{\operatorname{in}})^{(1)}\rightarrow(A_{1}^{\operatorname{out}})^{(1)},...,(A_{n}^{\operatorname{in}})^{(m)}\rightarrow(A_{n}^{\operatorname{out}})^{(m)}\right)$ be such that $I_{X}\otimes\lambda\Gamma\geq D^{(m)}$ and that $-\log(\lambda)$ gives the min-entropy for $D^{(m)}$. Let $\Gamma^{\prime}\in\operatorname{Comb}\left((A_{1}^{\operatorname{in}})^{(1)}\rightarrow(A_{1}^{\operatorname{out}})^{(1)},...,(A_{n}^{\operatorname{in}})^{(l)}\rightarrow(A_{n}^{\operatorname{out}})^{(l)}\right)$ be given by $\displaystyle\Gamma^{\prime}:=\frac{1}{\prod_{t=l+1}^{m}\dim(\boldsymbol{A}^{\operatorname{in}})^{(t)}}\operatorname{Tr}_{\boldsymbol{A}^{(l+1)}...\boldsymbol{A}^{(m)}}\left[\Gamma\right]$ (117) where $\dim(\boldsymbol{A}^{\operatorname{in}})^{(t)}:=\prod_{k=1}^{n}\dim(A_{k}^{\operatorname{in}})^{(t)}$ and the subscripts $\boldsymbol{A}^{(j)}$ in the trace indicate that the trace is over all subspace related to $\sigma_{x}^{(j)}$, i.e. $(\mathcal{H}_{A_{1}^{\operatorname{in}}})^{(j)},(\mathcal{H}_{A_{1}^{\operatorname{out}}})^{(j)},...,(\mathcal{H}_{A_{n}^{\operatorname{in}}})^{(j)},(\mathcal{H}_{A_{n}^{\operatorname{out}}})^{(j)}$. It remains only to show that $I_{X}\otimes\lambda\Gamma^{\prime}\geq D^{(l)}$. It suffices to show that $\displaystyle\frac{1}{\prod_{t=l+1}^{m}\dim(\boldsymbol{A}^{\operatorname{in}})^{(t)}}\operatorname{Tr}_{\boldsymbol{A}^{(l+1)}...\boldsymbol{A}^{(m)}}\left[D^{(m)}\right]=D^{(l)}.$ (118) Starting from the left-hand side: $\displaystyle\frac{1}{\prod_{t=l+1}^{m}\dim(\boldsymbol{A}^{\operatorname{in}})^{(t)}}\operatorname{Tr}_{\boldsymbol{A}^{(l+1)}...\boldsymbol{A}^{(m)}}\left[D^{(m)}\right]$ $\displaystyle=\frac{1}{\prod_{t=l+1}^{m}\dim(\boldsymbol{A}^{\operatorname{in}})^{(t)}}\sum_{x}P(x)\ket{x}\\!\\!\bra{x}\otimes\operatorname{Tr}_{\boldsymbol{A}^{(l+1)}...\boldsymbol{A}^{(m)}}\left[\bigotimes_{j=1}^{m}\sigma_{x}^{(j)}\right]$ (119) $\displaystyle=\frac{1}{\prod_{t=l+1}^{m}\dim(\boldsymbol{A}^{\operatorname{in}})^{(t)}}\sum_{x}P(x)\ket{x}\\!\\!\bra{x}\otimes\left(\prod_{k=l+1}^{m}\dim(\boldsymbol{A}^{\operatorname{in}})^{(k)}\right)\bigotimes_{j=1}^{l}\sigma_{x}^{(j)}$ (120) $\displaystyle=D^{(l)}$ (121) where we have used the fact that each $\sigma_{x}^{(j)}$ is a comb, so $\operatorname{Tr}[\sigma_{x}^{(j)}]$ is equal to the product of input space dimensions by definition. So, it holds that $I_{X}\otimes\lambda\Gamma^{\prime}\geq D^{(l)}$, which entails that $\displaystyle H_{\min}(X|\boldsymbol{A}^{(1)},...,\boldsymbol{A}^{(l)})_{D^{(l)}}\geq-\log(\lambda)=H_{\min}(X|\boldsymbol{A}^{(1)},...,\boldsymbol{A}^{(m)})$ (122) thus proving that the min-entropy is non-increasing with increasing round number. ∎ ## Appendix C Blind Quantum Computing Protocol Supplementary This appendix contains supporting results for the blindness theorems of Section 4 and for the accompanying examples. ### C.1 Single Round Theorem Supporting Results The proof of 4.2 makes use of the following lemma which demonstrates that for $\boldsymbol{\alpha^{\prime}},\boldsymbol{\alpha},\boldsymbol{c^{\prime}}$ and $g$ fixed, there is at most one $\boldsymbol{r}$ such that Equation 34 is satisfied for each $i$. ###### Lemma C.1. If there exists an $\boldsymbol{r}$ such that $P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)=1$ for all other variables fixed, then it is unique, otherwise $P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)=0$. ###### Proof. Since $P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)$ is a deterministic distribution, it takes values either $0$ or $1$. Let $\boldsymbol{r}$ be such that $P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)=1$. Recalling Equation 34 and the fact that a total order is imposed on communication, we have that for each $i$ $\displaystyle\alpha^{\prime}_{i}:=(-1)^{\bigoplus_{j\in\mathcal{X}_{i}}c^{\prime}_{j}\oplus r_{j}}\alpha_{i}+\left(r_{i}\oplus\bigoplus_{j\in\mathcal{Z}_{i}}c^{\prime}_{j}\oplus r_{j}\right)\pi\bmod 2\pi$ (123) where $\mathcal{X}_{i}$ and $\mathcal{Z}_{i}$ are necessarily subsets of $\\{1,...,i-1\\}$. In particular, this means that $\displaystyle\alpha^{\prime}_{1}=\alpha_{1}+r_{1}\pi\bmod 2\pi$ (124) and hence there is a unique value for $r_{1}$ given $\boldsymbol{\alpha}$ and $\boldsymbol{\alpha^{\prime}}$ are fixed. The result then follows by induction: since $\boldsymbol{c^{\prime}}$ and $g$ are fixed, for fixed values $\boldsymbol{r}_{1:i-1}$ the quantities $\bigoplus_{j\in\mathcal{X}_{i}}c^{\prime}_{j}\oplus r_{j}$ and $\bigoplus_{j\in\mathcal{Z}_{i}}c^{\prime}_{j}\oplus r_{j}$ are determined for each $i$, and thus there is a unique $r_{i}$ for which Equation 123 holds. ∎ We can now give the proof of the theorem. ###### Proof of 4.2. From 3.2, we know that $\displaystyle H_{\min}(\boldsymbol{A},\boldsymbol{O}|\boldsymbol{A^{\prime}},\boldsymbol{C^{\prime}})_{D_{\operatorname{client}}}\geq-\log\left[\sum_{\boldsymbol{\alpha^{\prime}}}\max_{\boldsymbol{\alpha},O,\boldsymbol{c^{\prime}}}P(\boldsymbol{\alpha},O|\boldsymbol{\alpha^{\prime}},\boldsymbol{c^{\prime}})P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}})\right].$ (125) Instead of dealing directly with $P(\boldsymbol{\alpha},O|\boldsymbol{\alpha^{\prime}},\boldsymbol{c^{\prime}})P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}})$, it is easier to revert back to the form before the Bayes’ rule was applied. That is, we aim to find $\displaystyle\sum_{\boldsymbol{\alpha^{\prime}}}\max_{\boldsymbol{\alpha},O,\boldsymbol{c^{\prime}}}P(\boldsymbol{\alpha},O)P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},O).$ (126) Using the uniformity assumptions for choosing the computation, gflow and one- time pads, as well as the definition of $P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},O)$ in the previous subsection, we have $\displaystyle P(\boldsymbol{\alpha},O)P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},O)=\frac{1}{|\mathcal{A}|^{n}|\mathcal{O}|}\sum_{g\sim O,\boldsymbol{r}}\frac{P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)}{|g\sim O|2^{n}}$ (127) where $g\sim O$ indicates that the gflow is defined for the output set $O$ and $|g\sim O|$ denotes the number of all such gflows. Recalling that $P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)$ is a deterministic distribution, it can be shown (C.1) that for fixed $g$, $\boldsymbol{\alpha}$ and $\boldsymbol{c^{\prime}}$ there is at most one $\boldsymbol{r}$ for which $P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)=1$. Thus, $\displaystyle\max_{\boldsymbol{\alpha},O,\boldsymbol{c^{\prime}}}\sum_{g\sim O,\boldsymbol{r}}\frac{P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)}{|g\sim O|2^{n}}$ (128) can be interpreted as selecting the $\boldsymbol{\alpha}$, $O$ and $\boldsymbol{c^{\prime}}$ for which $\boldsymbol{\alpha^{\prime}}$ is reportable from $\boldsymbol{\alpha}$ for the greatest number of pairs $(g,\boldsymbol{r})$ for gflows $g\sim O$. This quantity is clearly upper- bounded by a situation where $\boldsymbol{\alpha^{\prime}}$ is reportable under all gflows, hence $\displaystyle\max_{\boldsymbol{\alpha},O,\boldsymbol{c^{\prime}}}\sum_{g\sim O,\boldsymbol{r}}\frac{P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)}{|g\sim O|2^{n}}$ $\displaystyle\leq\frac{1}{2^{n}}$ (129) which holds for all $\boldsymbol{\alpha^{\prime}}$. Thus, returning to Equation 126: $\displaystyle\sum_{\boldsymbol{\alpha^{\prime}}}\max_{\boldsymbol{\alpha},O,\boldsymbol{c^{\prime}}}P(\boldsymbol{\alpha},O)P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},O)$ $\displaystyle\leq\sum_{\boldsymbol{\alpha^{\prime}}}\frac{1}{|\mathcal{A}|^{n}|\mathcal{O}|2^{n}}$ (130) $\displaystyle=\frac{1}{|\mathcal{O}|2^{n}}$ (131) which proves the theorem. ∎ It is worthwhile making some further comments regarding the interpretation of Equation 128 and the related inequality Equation 129 in the proof above. Firstly, it is possible to find a simple example, namely that given in the main text (see also below), for which Equation 129 is equality for every $\boldsymbol{\alpha^{\prime}}$. That is, for every $\boldsymbol{\alpha^{\prime}}$, there exists an $\boldsymbol{\alpha}$, $O$ and $\boldsymbol{c^{\prime}}$ that $\boldsymbol{\alpha^{\prime}}$ is reportable from $\boldsymbol{\alpha}$ for every gflow $g\sim O$. Said another way, the pre-images of $\boldsymbol{\alpha^{\prime}}$ under the gflows $g\sim O$ (for some fixed $\boldsymbol{c^{\prime}}$ and as $\boldsymbol{r}$ varies) have non- empty mutual intersection. Since the pre-image of each gflow has a fixed size (this follows from C.1), a larger mutual intersection corresponds to a smaller total set of angles $\boldsymbol{\alpha}$ from which $\boldsymbol{\alpha^{\prime}}$ can be reported. Since $\boldsymbol{\alpha}$ is one part of the secret information, it is intuitive that a smaller set of possible $\boldsymbol{\alpha}$ given the evidence $\boldsymbol{\alpha^{\prime}},\boldsymbol{c^{\prime}}$ corresponds to a lower min-entropy. This also suggests that, to improve security, considering a graph where the corresponding gflows have smaller mutual intersection is beneficial. Obtaining equality in Equation 129 for all $\boldsymbol{\alpha^{\prime}},\boldsymbol{c^{\prime}}$ accordingly means that equality is also achieved in Equation 46, however, it is unlikely that this bound is obtained in general. The above discussion highlights how much is dependent on the specific properties of gflows for the graph chosen for the protocol. Since (to the best knowledge of the authors’) there is no characterisation of gflows in terms of graph-theoretic properties, and since the given example achieves the lower bound, it is unlikely that a better lower bound that that given in the theorem can be given in general. ### C.2 BQC Minimum Example This appendix provides further details regarding the minimal example given in the main text (recall Figure 4) which obtains the lower bound in 4.2. We begin by demonstrating that $O=\\{2,3\\}$ is indeed the only choice of output set for which gflows compatible with the total order exist. ###### Lemma C.2. For $G$ as in Figure 4 with partial order $1<2<3$ and the $XY$-plane the only allowed measurement plane, $O=\\{2,3\\}$ is the only non-trivial output set for which there exist $I\subset\\{1,2,3\\}$ such that a gflow compatible with the total order exists. Moreover, there are just two possible gflows, $g_{1}$ and $g_{2}$, defined by $\displaystyle g_{1}:1$ $\displaystyle\mapsto\\{2\\},$ (132) $\displaystyle g_{2}:1$ $\displaystyle\mapsto\\{3\\},$ (133) where $g_{1}$ is compatible with $(I,O)$ for $I$ equal to $\\{1\\}$, $\\{3\\}$ or $\\{1,3\\}$ and $g_{2}$ is compatible with $(I,O)$ for $I=\\{1\\}$, $\\{2\\}$ and $\\{1,2\\}$, for $O$ as above. ###### Proof. We begin by demonstrating $O=\\{2,3\\}$ is the only valid non-trivial output set (note that we do not consider the trivial output set $O=\\{1,2,3\\}$ since this does not allow for any computation). Due to the total order and measurement plane restriction, $3$ must be in the output set, since if this was not the case, then $3\in O^{c}$ meaning that any gflow would map $3$ to $\\{1\\}$, $\\{2\\}$ or $\\{1,2\\}$. In any of these cases, we would require that $3<1$ or $3<2$ which contradicts the compatibility with total order. Thus $\\{1\\},\\{2\\}$ and $\\{1,2\\}$ cannot be output sets. Suppose either $\\{3\\}$ or $\\{1,3\\}$ was a valid output set. Any gflow compatible with either output must then map $2$ to some subset of $\\{1,2,3\\}$, however no such set exists for which some contradiction does not arise, as follows. If $2$ maps to $\\{1\\}$, $2<1$ which is a contradiction to the total order. If $2$ maps to $\\{2\\}$, $\\{1,2\\}$ or $\\{2,3\\}$, the gflow requirement $2\notin g(2)$ is contradicted. If $2$ maps to $\\{3\\}$, $1\in\operatorname{Odd}(g(2))$ implying $2<1$, contradicting the total order. If $2$ maps to $\\{1,3\\}$, then $2\notin\operatorname{Odd}(g(2))$ which contradicts a gflow requirement. This leaves $\\{2,3\\}$ as the only remaining possible output set. We now show it does indeed support gflow and characterise them. For output set $O=\\{2,3\\}$, the domain of any gflow map is $O^{c}=\\{1\\}$, so we can begin to characterise valid gflows by where they map $1$. Due to the $XY$-plane restriction, we can’t have $1\in g(1)$, so a valid gflow cannot map $1$ to $\\{1\\},\\{1,2\\}$ or $\\{1,3\\}$. Since $\operatorname{Odd}(\\{2,3\\})=\\{2,3\\}$ and since we require $1\in\operatorname{Odd}(g(1))$, no valid gflow maps $1$ to $\\{2,3\\}$. This leaves just the options $1\mapsto\\{2\\}$ and $1\mapsto\\{3\\}$. Both of these are valid gflows since in both cases, $1\notin g(1)$ is satisfied and the corresponding implications for the partial order, $1<2$ and $1<3$ respectively, are compatible with the total order. Similarly, for both maps $1\in\operatorname{Odd}(g(1))$ is satisfied ($\operatorname{Odd}(\\{2\\})=\\{1,3\\}$ and $\operatorname{Odd}(\\{3\\})=\\{1,2\\}$) and the again the implications for the partial order are compatible with the total order. Denote by $g_{1}$ the gflow that maps $1$ to $\\{2\\}$ and by $g_{2}$ the gflow that maps $1$ to $\\{3\\}$. Since a gflow is a map from $O^{c}$ to $\mathcal{P}(I^{c})$, $g_{1}$ is compatible with all sets $I$ for which $\\{2\\}\in\mathcal{P}(I^{c})$, i.e. $I=\\{1\\},\\{3\\}$ and $\\{1,3\\}$, and similarly, $g_{2}$ is compatible with $I=\\{1\\},\\{2\\}$ and $\\{1,2\\}$. ∎ Returning to the discussion of the BQC protocol for the example, it is useful to explicitly write out the correction sets given by $g_{1}$ and $g_{2}$ (only those that are non-empty are shown): $\displaystyle\mathcal{X}_{2}^{g_{1}}$ $\displaystyle=\mathcal{Z}_{3}^{g_{1}}=\\{1\\};$ $\displaystyle\mathcal{Z}_{2}^{g_{2}}$ $\displaystyle=\mathcal{X}_{3}^{g_{2}}=\\{1\\}.$ Let $\mathcal{A}$ be any agreed upon set of angles that satisfies Equation 33. For a single round of the protocol, if $g_{1}$ is chosen by the client, then the reported angles are given by $\displaystyle\alpha^{\prime}_{1}$ $\displaystyle=\alpha_{1}+r_{1}\pi\bmod 2\pi,$ (134) $\displaystyle\alpha^{\prime}_{2}$ $\displaystyle=(-1)^{r_{1}\oplus c^{\prime}_{1}}\alpha_{2}+r_{2}\pi\bmod 2\pi,$ (135) $\displaystyle\alpha^{\prime}_{3}$ $\displaystyle=\alpha_{3}+(r_{3}\oplus r_{1}\oplus c^{\prime}_{1})\pi\bmod 2\pi,$ (136) where $\alpha_{1},\alpha_{2},\alpha_{3}\in\mathcal{A}$ are the chosen (and secret) angles for the computation, $c^{\prime}_{1}$ is the classical message reported by the server after the first measurement and $\boldsymbol{r}=r_{3}r_{2}r_{1}\in\\{0,1\\}^{3}$ is the one-time pad. If $g_{2}$ is used instead, the angles are reported as $\displaystyle\alpha^{\prime}_{1}$ $\displaystyle=\alpha_{1}+r_{1}\pi\bmod 2\pi,$ (137) $\displaystyle\alpha^{\prime}_{2}$ $\displaystyle=\alpha_{2}+(r_{2}\oplus r_{1}\oplus c^{\prime}_{1})\pi\bmod 2\pi,$ (138) $\displaystyle\alpha^{\prime}_{3}$ $\displaystyle=(-1)^{r_{1}\oplus c^{\prime}_{1}}\alpha_{3}+r_{3}\pi\bmod 2\pi.$ (139) Note that the only classical message relevant to this example is $c^{\prime}_{1}$. Let $\boldsymbol{\alpha^{\prime}}$ and $\boldsymbol{c^{\prime}}$ be reported in a single round of the protocol. We can write the pre-image of $\boldsymbol{\alpha^{\prime}}$ under $g_{1}$ given $\boldsymbol{c^{\prime}}$ by inverting Equations 134, 135 and 136 (we drop the $\bmod 2\pi$ notation and leave it implicit): $\displaystyle\alpha_{1}$ $\displaystyle=\alpha^{\prime}_{1}+r_{1}\pi;$ (140) $\displaystyle\alpha_{2}$ $\displaystyle=(-1)^{r_{1}\oplus c^{\prime}_{1}}\alpha^{\prime}_{2}+r_{2}\pi;$ (141) $\displaystyle\alpha_{3}$ $\displaystyle=\alpha^{\prime}_{3}+(r_{3}\oplus r_{1}\oplus c^{\prime}_{1})\pi.$ (142) Restricting our focus to a subset of the pre-image defined by $r_{1}=c^{\prime}_{1}$, that is, the angles $\displaystyle\alpha_{1}$ $\displaystyle=\alpha^{\prime}_{1}+c^{\prime}_{1}\pi,$ (143) $\displaystyle\alpha_{2}$ $\displaystyle=\alpha^{\prime}_{2}+r_{2}\pi,$ (144) $\displaystyle\alpha_{3}$ $\displaystyle=\alpha^{\prime}_{3}+r_{3}\pi,$ (145) one observes that $\boldsymbol{\alpha^{\prime}}$ can be reported from any one of these angles under $g_{2}$ given $\boldsymbol{c^{\prime}}$, namely by the one-time pads $\boldsymbol{\hat{r}}=\hat{r}_{3}\hat{r}_{2}\hat{r}_{1}=r_{3}r_{2}c^{\prime}_{1}$ (as can be shown via simple substitution into Equations 137, 138 and 139). Since this holds for any $\boldsymbol{\alpha^{\prime}}$ and $\boldsymbol{c^{\prime}}$, we have thus shown both that the bound Equation 129 in the proof of 4.2 above is in fact equality for every $\boldsymbol{\alpha^{\prime}}$ and moreover that the maximum is obtained for every $\boldsymbol{c^{\prime}}$. Recalling the discussion after 3.2, this means that we have found the guessing probability (equivalently min-entropy) exactly for this example and that any strategy used by the server is equally informative. These results are corroborated numerically (see Appendix E for details of the implementation using an SDP solver): the min-entropy is given as $-\log_{2}(0.125)=3$, which is consistent with $|\mathcal{O}|=1$ and $n=3$ for this example (this result was obtained for different choices of $\mathcal{A}$: for $|\mathcal{A}|=4$ and for $|\mathcal{A}|=8$). We now consider two rounds of the protocol in order to calculate the guessing probability for $D_{\operatorname{client}}^{(2)}$, which, for this example, can be written as $\displaystyle D_{\operatorname{client}}^{(2)}$ $\displaystyle=\sum_{\boldsymbol{\alpha},O}\frac{1}{|\mathcal{A}|^{n}|\mathcal{O}|}\ket{\boldsymbol{\alpha},O}\\!\\!\bra{\boldsymbol{\alpha},O}\otimes\sigma_{\boldsymbol{\alpha},O}^{(1)}\otimes\sigma_{\boldsymbol{\alpha},O}^{(2)}$ (146) $\displaystyle=\sum_{\boldsymbol{\alpha}}\frac{1}{|\mathcal{A}|^{3}}\ket{\boldsymbol{\alpha}}\\!\\!\bra{\boldsymbol{\alpha}}\otimes\sum_{\begin{subarray}{c}\boldsymbol{\alpha^{\prime}}^{(1)},\boldsymbol{\alpha^{\prime}}^{(2)},\\\ \boldsymbol{c^{\prime}}^{(1)},\boldsymbol{c^{\prime}}^{(2)}\end{subarray}}\sum_{\begin{subarray}{c}g^{(1)},g^{(2)},\\\ \boldsymbol{r}^{(1)},\boldsymbol{r}^{(2)}\end{subarray}}\frac{1}{2^{8}}P(\boldsymbol{\alpha^{\prime}}^{(1)}|\boldsymbol{c^{\prime}}^{(1)},\boldsymbol{\alpha},\boldsymbol{r}^{(1)},g^{(1)})P(\boldsymbol{\alpha^{\prime}}^{(2)}|\boldsymbol{c^{\prime}}^{(2)},\boldsymbol{\alpha},\boldsymbol{r}^{(2)},g^{(2)})\ket{\boldsymbol{\alpha^{\prime}}^{(1)}\boldsymbol{c^{\prime}}^{(1)}\boldsymbol{\alpha^{\prime}}^{(2)}\boldsymbol{c^{\prime}}^{(2)}}\\!\\!\bra{..}$ (147) where $g^{(j)}$ take values in $\\{g_{1},g_{2}\\}$ where $g_{1}$ and $g_{2}$ are as above, and where we have used $P(g^{(j)})=\frac{1}{2}$ and $P(\boldsymbol{r}^{(j)})=\frac{1}{2^{3}}$. For a choice of angle set $\mathcal{A}=\\{\frac{\pi}{5},\frac{\pi}{3},\frac{-\pi}{3}+\pi,\frac{-\pi}{5}+\pi,\frac{\pi}{5}+\pi,\frac{\pi}{3}+\pi,\frac{-\pi}{3},\frac{-\pi}{5}\\}$, the guessing probability for $D_{\operatorname{client}}^{(2)}$ takes value between $0.140625$ and $0.28125$ as computed via numeric methods (see [72]) and as explained in the following. Note that various other choices of angle set give the same or similar results. These bounds are derived directly from 3.2 for the specific comb under consideration. The lower bound (which includes the normalisation over the input spaces) can be written as $\displaystyle\sum_{\begin{subarray}{c}\boldsymbol{\alpha^{\prime}}^{(1)},\boldsymbol{\alpha^{\prime}}^{(2)},\\\ \boldsymbol{c^{\prime}}^{(1)},\boldsymbol{c^{\prime}}^{(2)}\end{subarray}}\frac{1}{|\mathcal{A}|^{3}2^{6}}\max_{\boldsymbol{\alpha}}\sum_{\begin{subarray}{c}g^{(1)},g^{(2)},\\\ \boldsymbol{r}^{(1)},\boldsymbol{r}^{(2)}\end{subarray}}\frac{1}{2^{8}}P(\boldsymbol{\alpha^{\prime}}^{(1)}|\boldsymbol{c^{\prime}}^{(1)},\boldsymbol{\alpha},\boldsymbol{r}^{(1)},g^{(1)})P(\boldsymbol{\alpha^{\prime}}^{(2)}|\boldsymbol{c^{\prime}}^{(2)},\boldsymbol{\alpha},\boldsymbol{r}^{(2)},g^{(2)}).$ (148) Due to the redundancy of $(c^{\prime}_{2})^{(j)}$ and $(c^{\prime}_{3})^{(j)}$ (they don’t appear in Equations 134, 135, 136, 137, 138 and 139), this reduces to $\displaystyle\sum_{\begin{subarray}{c}\boldsymbol{\alpha^{\prime}}^{(1)},\boldsymbol{\alpha^{\prime}}^{(2)},\\\ (c^{\prime}_{1})^{(1)},(c^{\prime}_{1})^{(2)}\end{subarray}}\frac{1}{|\mathcal{A}|^{3}2^{2}}$ $\displaystyle\max_{\boldsymbol{\alpha}}\sum_{\begin{subarray}{c}g^{(1)},g^{(2)},\\\ \boldsymbol{r}^{(1)},\boldsymbol{r}^{(2)}\end{subarray}}\frac{1}{2^{8}}P(\boldsymbol{\alpha^{\prime}}^{(1)}|(c^{\prime}_{1})^{(1)},\boldsymbol{\alpha},\boldsymbol{r}^{(1)},g^{(1)})P(\boldsymbol{\alpha^{\prime}}^{(2)}|(c^{\prime}_{1})^{(2)},\boldsymbol{\alpha},\boldsymbol{r}^{(2)},g^{(2)}).$ (149) The upper bound is given by $\displaystyle\sum_{\boldsymbol{\alpha^{\prime}}^{(1)},\boldsymbol{\alpha^{\prime}}^{(2)}}\frac{1}{|\mathcal{A}|^{3}}$ $\displaystyle\max_{\boldsymbol{\alpha},(c^{\prime}_{1})^{(1)},(c^{\prime}_{1})^{(2)}}\sum_{\begin{subarray}{c}g^{(1)},g^{(2)},\\\ \boldsymbol{r}^{(1)},\boldsymbol{r}^{(2)}\end{subarray}}\frac{1}{2^{8}}P(\boldsymbol{\alpha^{\prime}}^{(1)}|(c^{\prime}_{1})^{(1)},\boldsymbol{\alpha},\boldsymbol{r}^{(1)},g^{(1)})P(\boldsymbol{\alpha^{\prime}}^{(2)}|(c^{\prime}_{1})^{(1)},\boldsymbol{\alpha},\boldsymbol{r}^{(2)},g^{(2)}).$ (150) Computing these quantities for $\mathcal{A}$ above via the code in the repository [72] gives the results as listed above. ### C.3 Multi-round Blindness Supplementary For this section, it is useful to have the following notation. Let a graph $G$ on vertices $V$ be given and $O\subset V$ a choice of output set. For a given $\boldsymbol{\alpha}\in\mathcal{A}^{n}$, define the set $\mathcal{A}_{\boldsymbol{\alpha},O}:=\\{\boldsymbol{\alpha}+(0,...,0,b_{o_{1}},...,b_{o_{|O|}})\pi\bmod 2\pi:b_{o_{i}}\in\\{0,1\\}\\}$ where the zero entries of $(0,...,0,b_{o_{1}},...,b_{o_{|O|}})$ correspond to $V\setminus O$ and the labels $o_{i}$ correspond to the elements of $O$. We define an equivalence relation $\sim_{O}$ by $\boldsymbol{\alpha}\sim_{O}\boldsymbol{\widetilde{\alpha}}$ for $\widetilde{\alpha}\in\mathcal{A}_{\boldsymbol{\alpha},O}$. The set obtained by quotienting $\mathcal{A}^{n}$ by the equivalence relation, i.e. $\mathcal{A}^{n}/\sim_{O}$ has $\frac{|\mathcal{A}|^{n}}{2^{|O|}}$ elements. ###### Lemma C.3. Consider $\sum_{g\sim O,\boldsymbol{r}}P(g|O)p(\boldsymbol{r})\sigma_{\operatorname{BQC}}^{\boldsymbol{\alpha},\boldsymbol{r},g}$ as in Equation 37 with $P(g|O)$ and $P(\boldsymbol{r})$ both uniform. Then, for all $\boldsymbol{\widetilde{\alpha}}\sim_{O}\boldsymbol{\alpha}$: $\displaystyle\sum_{g\sim O,\boldsymbol{r}}P(g|O)P(\boldsymbol{r})\sigma_{\operatorname{BQC}}^{\boldsymbol{\alpha},\boldsymbol{r},g}=\sum_{g\sim O,\boldsymbol{r}}P(g|O)P(\boldsymbol{r})\sigma_{\operatorname{BQC}}^{\boldsymbol{\widetilde{\alpha}},\boldsymbol{r},g}$ (151) ###### Proof. We begin by noting that, by definition of gflow, for any $g\sim O$, no element of $O$ is in the domain of $g$. This means that, for all $i\in O$, $i\notin\mathcal{X}_{k}$ and $i\notin\mathcal{Z}_{k}$ for all $k\in V$. The consequence of this being that for any $r_{i}$ for $i\in O$, the only equation for $\alpha^{\prime}_{k}$ (recall Equation 34) that contains $r_{i}$ is when $k=i$. This results in the following symmetry: $\displaystyle\alpha^{\prime}_{i}=(-1)^{\bigoplus_{j\in\mathcal{X}_{i}}c^{\prime}_{j}\oplus r_{j}}\alpha_{i}+(r_{i}\oplus\bigoplus_{j\in\mathcal{Z}_{i}}c^{\prime}_{j}\oplus r_{j})\pi\bmod 2\pi=(-1)^{\bigoplus_{j\in\mathcal{X}_{i}}c^{\prime}_{j}\oplus r_{j}}(\alpha_{i}+\pi)+((r_{i}\oplus 1)\oplus\bigoplus_{j\in\mathcal{Z}_{i}}c^{\prime}_{j}\oplus r_{j})\pi\bmod 2\pi$ (152) Thus, for fixed gflow $g$ and classical messages $\boldsymbol{c^{\prime}}$, if $\boldsymbol{\alpha^{\prime}}$ is reportable from $\boldsymbol{\alpha}$ for one-time pads $\boldsymbol{r}$, i.e. $P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)=1$, then $\boldsymbol{\alpha^{\prime}}$ is reportable from $\boldsymbol{\alpha}+(0,...,0,b_{o_{1}},...,b_{o_{|O|}})\pi$ for one-time pads $\boldsymbol{r}+(0,...,0,b_{o_{1}},...,b_{o_{|O|}})\pi$, i.e. $P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha}+(0,...,0,b_{o_{1}},...,b_{o_{|O|}})\pi,\boldsymbol{r}+(0,...,0,b_{o_{1}},...,b_{o_{|O|}})\pi,g)=1$. In light of C.1, and with the assumption that $P(\boldsymbol{r})$ is uniform, this in particular means that $\displaystyle\sum_{\boldsymbol{r}}P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)P(\boldsymbol{r})=\sum_{\boldsymbol{r}}P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}}.\boldsymbol{\alpha}+(0,...,0,b_{o_{1}},...,b_{o_{|O|}})\pi,\boldsymbol{r},g)P(\boldsymbol{r})$ (153) and so $\displaystyle\sum_{g\sim O,\boldsymbol{r}}P(g|O)P(\boldsymbol{r})\sigma_{\operatorname{BQC}}^{\boldsymbol{\alpha},\boldsymbol{r},g}$ $\displaystyle=\sum_{\begin{subarray}{c}g\sim O,\boldsymbol{r},\\\ \boldsymbol{\alpha^{\prime}},\boldsymbol{c^{\prime}}\end{subarray}}P(g|O)P(\boldsymbol{r})P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha},\boldsymbol{r},g)\ket{\boldsymbol{\alpha^{\prime}}\boldsymbol{c^{\prime}}}\\!\\!\bra{\boldsymbol{\alpha^{\prime}}\boldsymbol{c^{\prime}}}$ (154) $\displaystyle=\sum_{\begin{subarray}{c}g\sim O,\boldsymbol{r},\\\ \boldsymbol{\alpha^{\prime}},\boldsymbol{c^{\prime}}\end{subarray}}P(g|O)P(\boldsymbol{r})P(\boldsymbol{\alpha^{\prime}}|\boldsymbol{c^{\prime}},\boldsymbol{\alpha}+(0,...,0,b_{o_{1}},...,b_{o_{|O|}})\pi,\boldsymbol{r},g)\ket{\boldsymbol{\alpha^{\prime}}\boldsymbol{c^{\prime}}}\\!\\!\bra{\boldsymbol{\alpha^{\prime}}\boldsymbol{c^{\prime}}}$ (155) $\displaystyle=\sum_{g\sim O,\boldsymbol{r}}P(g|O)P(\boldsymbol{r})\sigma_{\operatorname{BQC}}^{\boldsymbol{\alpha}+(0,...,0,b_{o_{1}},...,b_{o_{|O|}})\pi,\boldsymbol{r},g}$ (156) ∎ In the notation of Equation 42, the above lemma states that $\sigma_{\boldsymbol{\alpha},O}=\sigma_{\boldsymbol{\widetilde{\alpha}},O}$ for all $\widetilde{\alpha}\sim_{O}\boldsymbol{\alpha}$ which is the notation used in the following proof. ###### Proof of 4.3. For this proof, we work with the maximisation formulation for the guessing probability, and moreover in the form given by maximising over pairs of dual combs and CPTP maps $(\widehat{E},\mathcal{F})$ as given in Equation 14. That is, we consider $\displaystyle\max_{\widehat{E},\mathcal{F}}\operatorname{Tr}_{\boldsymbol{A},\boldsymbol{O},\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}\left[(I_{\boldsymbol{A},\boldsymbol{O}}\otimes\mathcal{F})\left(D_{\operatorname{client}}^{(m)}\widehat{E}^{T}\right)\left(\sum_{\boldsymbol{\alpha},O}\ket{\boldsymbol{\alpha},O,\boldsymbol{\alpha},O}\\!\\!\bra{\boldsymbol{\alpha},O,\boldsymbol{\alpha},O}_{\boldsymbol{A},\boldsymbol{O},\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}\right)\right]$ (157) where $\widehat{E}\in\operatorname{Comb}(\mathbb{C}\rightarrow\mathbb{C},(A^{\prime}_{1})^{(1)}\rightarrow(C^{\prime}_{1})^{(1)},...,(A^{\prime}_{n})^{(m)}\rightarrow(C^{\prime}_{n})^{(m)},\mathbb{C}\rightarrow B)$ is a normalised dual comb and $\mathcal{F}:\mathcal{H}_{B}\rightarrow\mathcal{H}_{\boldsymbol{A^{\prime}}}\otimes\mathcal{H}_{\boldsymbol{O^{\prime}}}\cong\mathcal{H}_{\boldsymbol{A}}\otimes\mathcal{H}_{\boldsymbol{O}}$ is a CPTP map. The transpose $T$ is over all spaces $\boldsymbol{A^{\prime}}_{1},\boldsymbol{C^{\prime}}_{1},...,\boldsymbol{A^{\prime}}_{n},\boldsymbol{C^{\prime}}_{n}$. Let $(\widehat{E},\mathcal{F})$ be any such pair. Writing $D_{\operatorname{client}}^{(m)}$ as $\sum_{\boldsymbol{\alpha},O}P(\boldsymbol{\alpha},O)\ket{\boldsymbol{\alpha},O}\\!\\!\bra{\boldsymbol{\alpha},O}\bigotimes_{j=1}^{m}\sigma_{\boldsymbol{\alpha},O}^{(j)}$, we get $\displaystyle\operatorname{Tr}_{\boldsymbol{A},\boldsymbol{O},\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}$ $\displaystyle\left[(I_{\boldsymbol{A},\boldsymbol{O}}\otimes\mathcal{F})\left(D_{\operatorname{client}}^{(m)}\widehat{E}^{T}\right)\left(\sum_{\boldsymbol{\alpha},O}\ket{\boldsymbol{\alpha},O,\boldsymbol{\alpha},O}\\!\\!\bra{\boldsymbol{\alpha},O,\boldsymbol{\alpha},O}_{\boldsymbol{A},\boldsymbol{O},\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}\right)\right]$ $\displaystyle=\sum_{\boldsymbol{\alpha},O}\bra{\boldsymbol{\alpha},O}_{\boldsymbol{A},\boldsymbol{O}}\bra{\boldsymbol{\alpha},O}_{\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}\left[\sum_{\boldsymbol{\alpha},O}P(\boldsymbol{\alpha},O)\ket{\boldsymbol{\alpha},O}\\!\\!\bra{\boldsymbol{\alpha},O}\otimes\mathcal{F}\left(\left(\bigotimes_{j=1}^{m}\sigma_{\boldsymbol{\alpha},O}^{(j)}\right)\widehat{E}^{T}\right)\right]\ket{\boldsymbol{\alpha},O}_{\boldsymbol{A},\boldsymbol{O}}\ket{\boldsymbol{\alpha},O}_{\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}$ (158) By the fact that $\widehat{E}$ is a (normalised) dual comb, $\left(\bigotimes_{j=1}^{m}\sigma_{\boldsymbol{\alpha},O}^{(j)}\right)\widehat{E}^{T}$ is a (normalised) state in $\mathcal{H}_{B}$. Moreover, since $\bigotimes_{j=1}^{m}\sigma_{\boldsymbol{\alpha},O}^{(j)}=\bigotimes_{j=1}^{m}\sigma_{\boldsymbol{\widetilde{\alpha}},O}^{(j)}$ for all $\boldsymbol{\widetilde{\alpha}}\sim_{O}\boldsymbol{\alpha}$ by C.3, $\left(\bigotimes_{j=1}^{m}\sigma_{\boldsymbol{\alpha},O}^{(j)}\right)\widehat{E}^{T}$ is the same state for each $\widetilde{\alpha}\in\mathcal{A}_{\boldsymbol{\alpha},O}$, and hence so is $\mathcal{F}\left(\left(\bigotimes_{j=1}^{m}\sigma_{\boldsymbol{\alpha},O}^{(j)}\right)\widehat{E}^{T}\right)$. For simplicity, let us replace the notation $\mathcal{F}\left(\left(\bigotimes_{j=1}^{m}\sigma_{\boldsymbol{\alpha},O}^{(j)}\right)\widehat{E}^{T}\right)$ by $\rho_{\sim_{O}\boldsymbol{\alpha}}$. With this notation and invoking the assumption on $P(\boldsymbol{\alpha},O)$, Equation 158 becomes $\displaystyle\sum_{\boldsymbol{\alpha},O}\bra{\boldsymbol{\alpha},O}_{\boldsymbol{A},\boldsymbol{O}}\bra{\boldsymbol{\alpha},O}_{\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}\left(\sum_{\boldsymbol{\alpha},O}\frac{P(O)}{|\mathcal{A}|^{n}}\ket{\boldsymbol{\alpha},O}\\!\\!\bra{\boldsymbol{\alpha},O}\otimes\rho_{\sim_{O}\boldsymbol{\alpha}}\right)$ $\displaystyle\ket{\boldsymbol{\alpha},O}_{\boldsymbol{A},\boldsymbol{O}}\ket{\boldsymbol{\alpha},O}_{\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}$ $\displaystyle=\frac{1}{|\mathcal{A}|^{n}}\sum_{\boldsymbol{\alpha},O}P(O)\bra{\boldsymbol{\alpha},O}_{\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}\rho_{\sim_{O}\boldsymbol{\alpha}}\ket{\boldsymbol{\alpha},O}$ (159) $\displaystyle=\frac{1}{|\mathcal{A}|^{n}}\sum_{O}P(O)\sum_{\boldsymbol{\alpha}\in\mathcal{A}^{n}/\sim_{O}}\sum_{\boldsymbol{\widetilde{\alpha}}\sim_{O}\boldsymbol{\alpha}}\bra{\boldsymbol{\widetilde{\alpha}},O}_{\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}\rho_{\sim_{O}\boldsymbol{\alpha}}\ket{\boldsymbol{\widetilde{\alpha}},O}$ (160) where we are abusing notation slightly by denoting by $\boldsymbol{\alpha}$ the equivalence class $[\boldsymbol{\alpha}]\in\mathcal{A}^{n}/\sim_{O}$. Since $\rho_{\sim_{O}\boldsymbol{\alpha}}$ is a normalised state for every $\boldsymbol{\alpha},O$ and the term $\sum_{\boldsymbol{\widetilde{\alpha}}\sim_{O}\boldsymbol{\alpha}}\bra{\boldsymbol{\widetilde{\alpha}},O}_{\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}\rho_{\sim_{O}\boldsymbol{\alpha},O}\ket{\boldsymbol{\widetilde{\alpha}},O}$ can be interpreted as part of the trace over $\rho_{\sim_{O}\boldsymbol{\alpha}}$, we thus have $\displaystyle\sum_{\boldsymbol{\widetilde{\alpha}}\sim_{O}\boldsymbol{\alpha}}\bra{\boldsymbol{\widetilde{\alpha}},O}_{\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}\rho_{\sim_{O}\boldsymbol{\alpha},O}\ket{\boldsymbol{\widetilde{\alpha}},O}\leq 1$ (161) Thus, $\displaystyle\frac{1}{|\mathcal{A}|^{n}}\sum_{O}P(O)\sum_{\boldsymbol{\alpha}\in\mathcal{A}^{n}/\sim_{O}}\sum_{\boldsymbol{\widetilde{\alpha}}\sim_{O}\boldsymbol{\alpha}}\bra{\boldsymbol{\widetilde{\alpha}},O}_{\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}\rho_{\sim_{O}\boldsymbol{\alpha}}\ket{\boldsymbol{\widetilde{\alpha}},O}$ $\displaystyle\leq\frac{1}{|\mathcal{A}|^{n}}\sum_{O}P(O)\sum_{\boldsymbol{\alpha}\in\mathcal{A}^{n}/\sim_{O}}1$ (162) $\displaystyle=\frac{1}{|\mathcal{A}|^{n}}\sum_{O}\frac{P(O)|\mathcal{A}|^{n}}{2^{|O|}}$ (163) $\displaystyle=\frac{1}{|\mathcal{O}|}\sum_{O}\frac{P(O)}{2^{|O|}}$ (164) Since this holds for any pair $(\widehat{E},\mathcal{F})$, it in particular also holds for the one that maximises the trace, so $\displaystyle\max_{E}\operatorname{Tr}_{\boldsymbol{A},\boldsymbol{O},\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}\left[D_{\operatorname{client}}^{(m)}E^{T}\left(\sum_{\boldsymbol{\alpha},O}\ket{\boldsymbol{\alpha},O,\boldsymbol{\alpha},O}\\!\\!\bra{\boldsymbol{\alpha},O,\boldsymbol{\alpha},O}_{\boldsymbol{A},\boldsymbol{O},\boldsymbol{A^{\prime}},\boldsymbol{O^{\prime}}}\right)\right]\leq\sum_{O}\frac{P(O)}{2^{|O|}}$ (165) proving the theorem. ∎ ## Appendix D Grey Box MBQC Supplementary ### D.1 Correctness of $\sigma_{\operatorname{MBQC}}^{g}$ ###### Proposition D.1. Let $g$ be a gflow for $(G,I,O,\omega)$ where $G$ is a graph on vertices $V$, let $\sigma_{\operatorname{MBQC}}^{g}$ be defined as in Equation 54, and let $\mathcal{M}_{\alpha_{i},i,\omega(i)}^{c_{i}}$ denote the $c_{i}$ measurement outcome of the measurement channels defined in Equation 55 ($c_{i}=0$ for positive outcome and $c_{i}=1$ for the negative outcome). Then $\displaystyle\left(\bigotimes_{i\in V\setminus O}\mathcal{M}_{\alpha_{i},i,\omega(i)}^{c_{i}}\right)\ast\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}$ (166) is the same state for all $\boldsymbol{c}=c_{1}...c_{|V\setminus O|}$. ###### Proof. The proposition is a consequence of the fact that $\sigma_{\operatorname{MBQC}}^{g}$ is defined directly from gflow, and essentially follows the same reasoning as that presented in Section A.2 but written in the combs formalism and for general measurement planes. We write out the details in part to highlight that no issues arise with the choice of ordering of $X$ and $Z$ operators in Equation 52. We proceed by showing that the state produced by any series of measurement outcomes is the same as that produced by all positive measurement outcomes (which encodes the computation for the MBQC): $\displaystyle\left(\bigotimes_{i\in V\setminus O}\mathcal{M}_{\alpha_{i},i,\omega(i)}^{c_{i}}\right)\ast\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}=\left(\bigotimes_{i\in V\setminus O}\mathcal{M}_{\alpha_{i},i,\omega(i)}^{0}\right)\ast\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}$ (167) Consider first only a single measurement channel for qubit $i\in V\setminus O$, which obtains the negative outcome: $\displaystyle\mathcal{M}_{\alpha_{i},i,\omega(i)}^{1}\ast\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}$ $\displaystyle=\left(\ket{1}\\!\\!\bra{1}\otimes\ket{-_{\alpha_{i}}}\\!\\!\bra{-_{\alpha_{i}}}_{\omega(i)}\right)\ast\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}$ (168) By the definition of gflow, we are guaranteed that $\displaystyle\ket{-_{\alpha_{i}}}\\!\\!\bra{-_{\alpha_{i}}}_{\omega(i)}\equiv K_{g(i)}|_{i}^{\dagger}\ket{+_{\alpha_{i}}}\\!\\!\bra{+_{\alpha_{i}}}_{\omega(i)}K_{g(i)}|_{i}$ (169) where $K_{g(i)}|_{i}$ denotes the operator which appears as the tensor factor of $i$ in the stabiliser $K_{g(i)}$. Thus, using the properties of the link product and also that $K_{g(i)}|_{i}^{\dagger}\equiv K_{g(i)}|_{i}^{T}$ (in fact $K_{v}^{\dagger}=K_{v}^{T}$ for all stabilisers $K_{v}$), we have $\displaystyle\left(\ket{1}\\!\\!\bra{1}\otimes\ket{-_{\alpha_{i}}}\\!\\!\bra{-_{\alpha_{i}}}_{\omega(i)}\right)\ast\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}$ $\displaystyle=\left(\ket{1}\\!\\!\bra{1}\otimes K_{g(i)}|_{i}^{\dagger}\ket{+_{\alpha_{i}}}\\!\\!\bra{+_{\alpha_{i}}}_{\omega(i)}K_{g(i)}|_{i}\right)\ast\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}$ (170) $\displaystyle=\left(\ket{1}\\!\\!\bra{1}\otimes\ket{+_{\alpha_{i}}}\\!\\!\bra{+_{\alpha_{i}}}_{\omega(i)}\right)\ast K_{g(i)}|_{i}\sigma_{\operatorname{MBQC}}^{g}K_{g(i)}|_{i}^{\dagger}\ast\rho_{G}$ (171) By contracting the link product over $C_{i}$ and writing $\sigma_{\operatorname{MBQC}}^{g}$ in the form of Equation 54, we get $\displaystyle\ket{+_{\alpha_{i}}}\\!\\!\bra{+_{\alpha_{i}}}_{\omega(i)}\ast\left(\sum_{\boldsymbol{a},\boldsymbol{b},\boldsymbol{c}|_{\setminus i}}K_{g(i)}|_{i}U_{\text{corr}(1\boldsymbol{c}|_{\setminus i})}\ket{\boldsymbol{a}}\\!\\!\bra{\boldsymbol{b}}U_{\text{corr}(1\boldsymbol{c}|_{\setminus i})}^{\dagger}K_{g(i)}|_{i}^{\dagger}\otimes\ket{\boldsymbol{c}|_{\setminus i}\boldsymbol{a}}\\!\\!\bra{\boldsymbol{c}|_{\setminus i}\boldsymbol{b}}\right)\ast\rho_{G}$ (172) where the sum over $\boldsymbol{c}|_{\setminus i}$ indicates the sum over basis element of all the $\mathcal{H}_{C_{j}}$ for $j\neq i$ and $1\boldsymbol{c}|_{\setminus i}$ denotes the binary string $\boldsymbol{c}$ with the entry corresponding to $C_{i}$ a $1$ and the other entries given by $\boldsymbol{c}|_{\setminus i}$. By the definition of the correction sets $\mathcal{X}_{j}$ and $\mathcal{Z}_{j}$, as well as the definition of $U_{\text{corr}(\boldsymbol{c})}$, we can write $\displaystyle U_{\text{corr}(1\boldsymbol{c}|_{\setminus i})}=(-1)^{f(\boldsymbol{c}|_{\setminus i})}U_{\text{corr}(0\boldsymbol{c}|_{\setminus i})}K_{g(i)}|_{\setminus i}$ (173) where the exponent $f(\boldsymbol{c}|_{\setminus i})$ encodes the coefficient that arises from commuting the factors of $K_{g(i)}|_{\setminus i}$ through the other factors of $U_{\text{corr}(1\boldsymbol{c}|_{\setminus i})}$ and where $U_{\text{corr}(0\boldsymbol{c}|_{\setminus i})}$ denotes the correction operator with $c_{i}=0$. In particular, this means that $\displaystyle K_{g(i)}|_{i}U_{\text{corr}(1\boldsymbol{c}|_{\setminus i})}$ $\displaystyle=(-1)^{f(\boldsymbol{c}|_{\setminus i})+t_{i}}U_{\text{corr}(0\boldsymbol{c}|_{\setminus i})}K_{g(i)}|_{i}K_{g(i)}|_{\setminus i}$ (174) $\displaystyle=(-1)^{f(\boldsymbol{c}|_{\setminus i})+t_{i}}U_{\text{corr}(0\boldsymbol{c}|_{\setminus i})}K_{g(i)}$ (175) where $t_{i}$ encodes the coefficient arises from commuting $K_{g(i)}|_{i}$ through $U_{\text{corr}(0\boldsymbol{c}|_{\setminus i})}$. Since the same coefficient arises from the conjugate term (i.e. $U_{\text{corr}(1\boldsymbol{c}|_{\setminus i})}^{\dagger}K_{g(i)}|_{i}^{\dagger}$), the net result is that no $-1$ factor can appear. That is, we have $\displaystyle K_{g(i)}|_{i}U_{\text{corr}(1\boldsymbol{c}|_{\setminus i})}(\cdot)U_{\text{corr}(1\boldsymbol{c}|_{\setminus i})}^{\dagger}K_{g(i)}|_{i}^{\dagger}=U_{\text{corr}(0\boldsymbol{c}|_{\setminus i})}K_{g(i)}(\cdot)K_{g(i)}^{\dagger}U_{\text{corr}(0\boldsymbol{c}|_{\setminus i})}^{\dagger}$ (176) The cancelling of any phase factor arising from commuting $X$ and $Z$ factors in $U_{\text{corr}(\boldsymbol{c})}$ is the reason why the ordering in Equation 52 is justified. Continuing from Equation 172, we have $\displaystyle\ket{+_{\alpha_{i}}}\\!\\!\bra{+_{\alpha_{i}}}_{\omega(i)}\ast$ $\displaystyle\left(\sum_{\boldsymbol{a},\boldsymbol{b},\boldsymbol{c}|_{\setminus i}}U_{\text{corr}(0\boldsymbol{c}|_{\setminus i})}K_{g(i)}\ket{\boldsymbol{a}}\\!\\!\bra{\boldsymbol{b}}K_{g(i)}^{\dagger}U_{\text{corr}(0\boldsymbol{c}|_{\setminus i})}^{\dagger}\otimes\ket{\boldsymbol{c}|_{\setminus i}\boldsymbol{a}}\\!\\!\bra{\boldsymbol{c}|_{\setminus i}\boldsymbol{b}}\right)\ast\rho_{G}$ (177) $\displaystyle=\ket{+_{\alpha_{i}}}\\!\\!\bra{+_{\alpha_{i}}}_{\omega(i)}\ast\left(\sum_{\boldsymbol{a},\boldsymbol{b},\boldsymbol{c}|_{\setminus i}}U_{\text{corr}(0\boldsymbol{c}|_{\setminus i})}\ket{\boldsymbol{a}}\\!\\!\bra{\boldsymbol{b}}U_{\text{corr}(0\boldsymbol{c}|_{\setminus i})}^{\dagger}\otimes\ket{\boldsymbol{c}|_{\setminus i}}\\!\\!\bra{\boldsymbol{c}|_{\setminus i}}\otimes K_{g(i)}\ket{\boldsymbol{a}}\\!\\!\bra{\boldsymbol{b}}K_{g(i)}^{\dagger}\right)\ast\rho_{G}$ (178) $\displaystyle\equiv\left(\ket{0}\\!\\!\bra{0}\otimes\ket{+_{\alpha_{i}}}\\!\\!\bra{+_{\alpha_{i}}}_{\omega(i)}\right)\ast K_{g(i),A^{\prime}}\sigma_{\operatorname{MBQC}}^{g}K_{g(i),A^{\prime}}^{\dagger}\ast\rho_{G}$ (179) $\displaystyle=\mathcal{M}_{\alpha_{i},i,\omega(i)}^{0}\ast\sigma_{\operatorname{MBQC}}^{g}\ast K_{g(i)}\rho_{G}K_{g(i)}^{\dagger}$ (180) $\displaystyle=\mathcal{M}_{\alpha_{i},i,\omega(i)}^{0}\ast\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}$ (181) where the extra subscript $A^{\prime}$ on the operators in Equation 179 indicate that they act on the appropriate spaces $A^{\prime}_{j}$ rather than $A_{j}$ as before, and can thus be transferred to the $\rho_{G}$ via the properties of the link product (which also uses the fact that $K_{g(i)}^{\dagger}=K_{g(i)}^{T}$). We have also made use of properties of the Choi operator in Equation 178 to transfer the $K_{g(i)}$ from the $A$-spaces to the $A^{\prime}$-spaces in the first place. Since the left-hand side of Equation 167 contains a tensor product of terms $\mathcal{M}_{\alpha_{i},i,\omega(i)}^{c_{i}}$, we can apply the same reasoning as above to each factor for which $c_{i}=1$ which thus establishes that Equation 167 does indeed hold for any measurement outcomes $\boldsymbol{c}$. This means that once all the link products are evaluated, which in particular includes the trace over all spaces except for $\bigotimes_{i\in O}\mathcal{H}_{A_{i}}$, the same state is produced on the output space. ∎ ### D.2 Gflow-Induced Quantum Causal Models This subsection provides some further details regarding the the quantum causal model (QCM) induced by gflow for MBQC and thus supports Section 5.2 in the main text. Firstly, we confirm here that gflow does indeed define a DAG, which is a requirement for showing 5.1. Thereafter we provide a table that elucidates the comparisons between the components of the QCM defined for MBQC and the components of classical causal models as presented eg., by Pearl in [21]. ###### Proposition D.2. Let $(G,I,O,\omega)$ be such that a gflow exists and let $(g,<)$ be a choice of such a gflow. The directed graph $\overline{G}$ on vertex set $V$ and edge set defined from the gflow via $\displaystyle E$ $\displaystyle:=\left\\{(i,j)\in V\times V|j\in V,i\in\mathcal{X}_{j}\cup\mathcal{Z}_{j}\right\\}$ is acyclic. ###### Proof. Suppose for a contradiction that $\overline{G}$ is not acyclic, that is, there exists a directed path $v_{0}\rightarrow v_{1}\rightarrow...\rightarrow v_{k}=v_{0}$ for some sequence of vertices $v_{0},...,v_{k}$. By the definition of the edge set of $\overline{G}$, it follows that either $v_{i+1}\in g(v_{i})\setminus\\{v_{i}\\}$ or $v_{i}\neq v_{i+1}\in\operatorname{Odd}(g(v_{i}))$. In either case $v_{i}<v_{i+1}$ in the ordering of the given gflow. Thus, $v_{0}<v_{k}=v_{0}$ by transitivity of the order, which gives the desired contradiction as the order is strict by definition. ∎ Classical causal models are typically presented as consisting of a set of observed variables $V=\\{V_{1},...,V_{n}\\}$ (those within the model), a set of unobserved variables $U=\\{U_{1},...,U_{n}\\}$ (those determined by factors outside of the model), a DAG that specifies which variables (both unobserved and observed) that have a causal influence on a given observed variable, and set of functions $F=\\{f_{i}:\operatorname{Pa}(V_{i})\cup U_{i}\rightarrow V_{i}\\}_{i=1}^{n}$ that map from the parents of a variable to the variable itself. These functions are often called structural equations and uniquely specify the values for the observed variables given values of the unobserved variables. A probabilistic causal model further includes a distribution $P(U)$ over the unobserved variables. Using this terminology, the following table outlines the correspondence between classical causal models and the QCM $\sigma_{\operatorname{MBQC}}^{g}$ defined in Section 5.2. | Comparing the QCM $\sigma_{\operatorname{MBQC}}^{g}$ to Classical Causal Models ---|--- Classical Causal Model | Unobserved Variable $U_{i}$ | Observed Variable $V_{i}$ | $f_{i}:\operatorname{Pa}(V_{i})\cup U_{i}\rightarrow V_{i}$ | DAG MBQC QCM $\sigma_{\operatorname{MBQC}}^{g}$ | Input Space $\mathcal{H}_{A^{\prime}_{i}}$ | Quantum Node $\mathcal{H}_{A_{i}}\otimes\mathcal{H}_{C_{i}}$ | $\rho_{A_{i}|C_{j:j\in\operatorname{Pa}(i)},A^{\prime}_{i}}$ | $\overline{G}$ ### D.3 Causal Equivalence of Gflows ###### Proof of 5.2. Consider $\displaystyle\left(\bigotimes_{j=1}^{|V\setminus O|}\ket{c_{j}}\\!\\!\bra{c_{j}}_{C_{j}}\right)\ast\sigma_{\operatorname{MBQC}}^{g}\ast\rho_{G}$ (182) for some $g\sim(G,I,O,\omega)$. Evaluating the link product over the $C_{j}$ gives $\displaystyle\left(\sum_{\boldsymbol{a},\boldsymbol{b}}U_{\text{corr}(\boldsymbol{c})}^{g}\ket{\boldsymbol{a}}\\!\\!\bra{\boldsymbol{b}}(U_{\text{corr}(\boldsymbol{c})}^{g})^{\dagger}\otimes\ket{\boldsymbol{a}}\\!\\!\bra{\boldsymbol{b}}\right)\ast\rho_{G}=\left(\sum_{\boldsymbol{a},\boldsymbol{b}}\ket{\boldsymbol{a}}\\!\\!\bra{\boldsymbol{b}}\otimes\ket{\boldsymbol{a}}\\!\\!\bra{\boldsymbol{b}}\right)\ast U_{\text{corr}(\boldsymbol{c})}^{g}\rho_{G}(U_{\text{corr}(\boldsymbol{c})}^{g})^{\dagger}$ (183) where $\boldsymbol{c}=c_{1}...c_{|V\setminus O|}$. To prove the proposition, it suffices to show that all the $U_{\text{corr}(\boldsymbol{c})}^{g}$ as $g$ varies are mutually related by stabilisers of $G$ for each $\boldsymbol{c}$, up to a phase (which is cancelled by the corresponding conjugate phase from the adjoint $(U_{\text{corr}(\boldsymbol{c})}^{g})^{\dagger}$). Let $g,g^{\prime}\sim(G,I,O,\omega)$ be arbitrary. By definition of $U_{\text{corr}(\boldsymbol{c})}$, we have $\displaystyle U_{\text{corr}(\boldsymbol{c})}^{g}$ $\displaystyle\propto\prod_{j=1}^{|V\setminus O|}K_{g(j)}^{c_{j}}|_{\setminus j}$ (184) $\displaystyle U_{\text{corr}(\boldsymbol{c})}^{g^{\prime}}$ $\displaystyle\propto\prod_{j=1}^{|V\setminus O|}K_{g^{\prime}(j)}^{c_{j}}|_{\setminus j}$ (185) where $\propto$ denotes that a $-1$ phase may arise from commuting $X$ and $Z$ operators to arrive at the above form of the operator in terms of $K_{g(j)}$ from the canonical form of $U_{\text{corr}(\boldsymbol{c})}^{g}$ (respectively $U_{\text{corr}(\boldsymbol{c})}^{g^{\prime}}$) as in Equation 53 and Equation 52. Despite that $K_{g(j)}$ and $K_{g^{\prime}(j)}$ may be different stabilisers, by the fact that both $g$ and $g^{\prime}$ observe the same measurement planes, the $j$th tensor factor of each is the same. Thus, $\displaystyle U_{\text{corr}(\boldsymbol{c})}^{g}$ $\displaystyle\propto\left(\bigotimes_{l=1}^{|V\setminus O|}K_{g(l)}^{c_{l}}|_{l}\right)\prod_{j=1}^{|V\setminus O|}K_{g(j)}$ (186) $\displaystyle=\left(\bigotimes_{l=1}^{|V\setminus O|}K_{g(l)}^{c_{l}}|_{l}\right)\prod_{j=1}^{|V\setminus O|}K_{g^{\prime}(j)}K_{g^{\prime}(j)}K_{g(j)}$ (187) $\displaystyle=\left(\bigotimes_{l=1}^{|V\setminus O|}K_{g^{\prime}(l)}^{c_{l}}|_{l}\right)\left(\prod_{j=1}^{|V\setminus O|}K_{g^{\prime}(j)}\right)\left(\prod_{i=1}^{|V\setminus O|}K_{g^{\prime}(i)}K_{g(i)}\right)$ (188) $\displaystyle\propto U_{\text{corr}(\boldsymbol{c})}^{g^{\prime}}\left(\prod_{i=1}^{|V\setminus O|}K_{g(i)}K_{g^{\prime}(i)}\right)$ (189) The restriction to only those gflows $g\sim(G,I,O,\omega)$ that have mutually compatible partial orders that is made in the statement of the proposition is required when considering $\sigma_{\operatorname{MBQC}}^{g}$ with a total order on input and output spaces. ∎ ### D.4 Gflow Catalogue Figure 8, Figure 9, and Figure 10 depict the DAGs corresponding to the $15$ different gflows for the four-vertex graph considered in Section 5 and depicted in Figure 6(a), grouped by the assigned measurement plane for the second qubit (the first is always measured in the $XY$-plane). The details of the gflow and corresponding $U_{\text{corr}(\boldsymbol{c})}$ for each DAG are gives in the caption. (a) $g_{1}:1\mapsto\\{2\\},2\mapsto\\{3,4\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{1}}=X_{2}^{c_{1}}\otimes X_{3}^{c_{2}}Z_{3}^{c_{2}}\otimes X_{4}^{c_{2}}Z_{4}^{c_{1}c_{2}}$ (b) $g_{2}:1\mapsto\\{3\\},2\mapsto\\{3,4\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{2}}=X_{3}^{c_{1}c_{2}}Z_{3}^{c_{2}}\otimes X_{4}^{c_{2}}Z_{4}^{c_{1}c_{2}}$ (c) $g_{3}:1\mapsto\\{3\\},2\mapsto\\{4\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{2}}=Z_{1}^{c_{2}}\otimes X_{3}^{c_{1}}Z_{3}^{c_{2}}\otimes X_{4}^{c_{2}}Z_{4}^{c_{1}}$ (d) $g_{4}:1\mapsto\\{4\\},2\mapsto\\{3,4\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{4}}=Z_{2}^{c_{1}}\otimes X_{3}^{c_{2}}Z_{3}^{c_{1}c_{2}}\otimes X_{4}^{c_{1}c_{2}}Z_{4}^{c_{2}}$ (e) $g_{5}:1\mapsto\\{2,3,4\\},2\mapsto\\{3,4\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{5}}=X_{2}^{c_{1}}Z_{2}^{c_{1}}\otimes X_{3}^{c_{1}c_{2}}Z_{3}^{c_{1}c_{2}}\otimes X_{4}^{c_{1}c_{2}}Z_{4}^{c_{2}}$ Figure 8: (a) - (e) display the directed acyclic graphs and correction operators for the gflows $g_{1},...,g_{5}$ respectively which are compatible with $(G,I,O,\omega)$ where $\omega(1)=\omega(2)=XY$. The corresponding captions detail the gflows themselves and the associated $U_{\text{corr}(\boldsymbol{c})}$. The partial order for $g_{1},g_{2},g_{4}$ and $g_{5}$ is given by $1<2$ , and that for $g_{3}$ is $2<1$. (a) $g_{6}:1\mapsto\\{2\\},2\mapsto\\{2,4\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{6}}=X_{2}^{c_{1}}\otimes Z_{3}^{c_{2}}\otimes X_{4}^{c_{2}}Z_{4}^{c_{1}c_{2}}$ (b) $g_{7}:1\mapsto\\{3\\},2\mapsto\\{2,4\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{7}}=X_{3}^{c_{1}}Z_{3}^{c_{2}}\otimes X_{4}^{c_{2}}Z_{4}^{c_{1}c_{2}}$ (c) $g_{8}:1\mapsto\\{3\\},2\mapsto\\{2,3,4\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{8}}=Z_{1}^{c_{2}}\otimes X_{3}^{c_{1}c_{2}}Z_{3}^{c_{2}}\otimes X_{4}^{c_{2}}Z_{4}^{c_{1}}$ (d) $g_{9}:1\mapsto\\{4\\},2\mapsto\\{2,4\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{9}}=Z_{2}^{c_{1}}\otimes Z_{3}^{c_{1}c_{2}}\otimes X_{4}^{c_{1}c_{2}}Z_{4}^{c_{2}}$ (e) $g_{10}:1\mapsto\\{2,3,4\\},2\mapsto\\{2,4\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{10}}=X_{2}^{c_{1}}Z_{2}^{c_{1}}\otimes X_{3}^{c_{1}}Z_{3}^{c_{1}c_{2}}\otimes X_{4}^{c_{1}c_{2}}Z_{4}^{c_{2}}$ Figure 9: (a) - (e) display the directed acyclic graphs and correction operators for the gflows $g_{6},...,g_{10}$ respectively which are compatible with $(G,I,O,\omega)$ where $\omega(1)=XY$ and $\omega(2)=XZ$. The corresponding captions detail the gflows themselves and the associated $U_{\text{corr}(\boldsymbol{c})}$. The partial order for $g_{6},g_{7},g_{9}$ and $g_{10}$ is given by $1<2$ , and that for $g_{8}$ is $2<1$. (a) $g_{11}:1\mapsto\\{2\\},2\mapsto\\{2,3\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{11}}=X_{2}^{c_{1}}\otimes X_{3}^{c_{2}}\otimes Z_{4}^{c_{1}}$ (b) $g_{12}:1\mapsto\\{3\\},2\mapsto\\{2\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{12}}=Z_{1}^{c_{2}}\otimes X_{3}^{c_{1}}\otimes Z_{4}^{c_{1}c_{2}}$ (c) $g_{13}:1\mapsto\\{3\\},2\mapsto\\{2,3\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{13}}=X_{3}^{c_{1}c_{2}}\otimes Z_{4}^{c_{1}}$ (d) $g_{14}:1\mapsto\\{4\\},2\mapsto\\{2,3\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{14}}=Z_{2}^{c_{1}}\otimes X_{3}^{c_{2}}Z_{3}^{c_{1}}\otimes X_{4}^{c_{1}}$ (e) $g_{15}:1\mapsto\\{2,3,4\\},2\mapsto\\{2,3\\}$, $U_{\text{corr}(\boldsymbol{c})}^{g_{15}}=X_{2}^{c_{1}}Z_{2}^{c_{1}}\otimes X_{3}^{c_{1}c_{2}}Z_{3}^{c_{1}}\otimes X_{4}^{c_{1}}$ Figure 10: (a) - (e) display the directed acyclic graphs and correction operators for the gflows $g_{11},...,g_{15}$ respectively which are compatible with $(G,I,O,\omega)$ where $\omega(1)=XY$ and $\omega(2)=YZ$. The corresponding captions detail the gflows themselves and the associated $U_{\text{corr}(\boldsymbol{c})}$. The partial order for $g_{11},g_{13},g_{14}$ and $g_{15}$ is given by $1<2$ , and that for $g_{12}$ is $2<1$. ## Appendix E Details of SDP Implementation The numerical calculations of the guessing probabilities throughout this work made use of the convex optimisation library CVXPY [73, 74]. We primarily used the Splitting Conic Solver (SCS) [75, 76]. Our code is provided at [72]. Calculations were run on an HP $Z4$ $G4$ $9980XE$ workstation. To provide some indication of the limitations of the numerical approach in its current form, to complement those detailed in Section 6.1, we document here some of hurdles we faced when calculating guessing probabilities. For the BQC examples, we generated the classical combs as $1$-dimensional objects representing the corresponding diagonals. Despite this, size issues played a role when storing the combs even before calling the SDP solver: $D_{\operatorname{client}}$ for a single round and for any $\mathcal{A}$ of size greater than $12$ and $D_{\operatorname{client}}^{(2)}$ for $\mathcal{A}$ of size greater than $4$ caused problems. The guessing probability for $D_{\operatorname{client}}^{(1)}$ for $|\mathcal{A}|=4$ was calculated within a day, whereas for $|\mathcal{A}|=8$, it took on the order of a week and required approximately $110GB$ of RAM. Attempting to calculate the guessing probability for $D_{\operatorname{client}}^{(2)}$ with $|\mathcal{A}|=4$ exceeded the available RAM of the machine. For the Grey Box MBQC examples, the guessing probabilities for all single round combs could be calculated quickly (on the order of minutes) however all multi-round combs again caused size problems. Typically, the primary bottleneck occurred when enforcing the comb constraints (i.e. sequential partial trace constraints) on the variable in the SDP solver, which tended to dominate the runtime of the algorithm. Otherwise, some size error would occur during the Cone Matrix Stuffing reduction phase of the solver (see the CVXPY documentation for details). ## Appendix F Whose Prior? Throughout this work, we have considered combs of the form $\sum_{x}P(x)\ket{x}\\!\\!\bra{x}\otimes\sigma_{x}$. As made particularly apparent in Section 3.1 (in conjunction with Appendix B), the distribution $P(x)$ can be considered as a prior distribution and the min-entropy for such a comb can be interpreted in terms of a maximum over posterior distributions. One might then ask, whose prior is it? The natural answer seems to be that it is the prior of the agent interacting with the system in each case: the server in the BQC protocol or the user of the MBQC device. As such, the operators $D_{\operatorname{client}}$, $D_{\text{gflow}}$, $D_{\text{mp}}$ or $D_{\text{calibr}}$ can be considered as a model of the client internal to the server in the former case, or as models of the MBQC device internal to the user in the latter three cases. With this interpretation, it is interesting to reconsider $D_{\text{gflow}}$ and $D_{\text{mp}}$: they essentially consist of the same components but via a slight difference in their composition, which is to say, a slight difference in the internal representation of the system by the agent, the agent learns about a different property of the system. From a practical point of view, the presence of the prior distribution, as in any Bayesian updating, also allows for the inputting of specific knowledge of the problem into the calculation of the min-entropy. Throughout, we have typically taken prior distributions to be uniform in order to simplify the numerics, however this need not be the case. For example, in the BQC example, the assumption that the client chooses a computation uniformly at random is not necessarily a reasonable one - most known quantum algorithms have specific structure which would correspond to the requirement that certain entries of $\boldsymbol{\alpha}$ to specific values. This does, however, seem to pose a slight dilemma - since changing the prior changes the value for the min- entropy, is it really acceptable to have a proof of security for a protocol that depends on the (subjective) prior of the server?
11institutetext: Dept. of Computer & Systems Sciences, Stockholm University, Stockholm, Sweden 11email<EMAIL_ADDRESS><EMAIL_ADDRESS><EMAIL_ADDRESS> # Early prediction of the risk of ICU mortality with Deep Federated Learning Korbinian Randl 11 0000-0002-7938-2747 Núria Lladós Armengol 11 0000-0002-8584-5058 Lena Mondrejevski 11 0000-0002-1790-3842 Ioanna Miliou 11 0000-0002-1357-1967 ###### Abstract Intensive Care Units usually carry patients with a serious risk of mortality. Recent research has shown the ability of Machine Learning to indicate the patients’ mortality risk and point physicians toward individuals with a heightened need for care. Nevertheless, healthcare data is often subject to privacy regulations and can therefore not be easily shared in order to build Centralized Machine Learning models that use the combined data of multiple hospitals. Federated Learning is a Machine Learning framework designed for data privacy that can be used to circumvent this problem. In this study, we evaluate the ability of deep Federated Learning to predict the risk of Intensive Care Unit mortality at an early stage. We compare the predictive performance of Federated, Centralized, and Local Machine Learning in terms of AUPRC, F1-score, and AUROC. Our results show that Federated Learning performs equally well as the centralized approach and is substantially better than the local approach, thus providing a viable solution for early Intensive Care Unit mortality prediction. In addition, we show that the prediction performance is higher when the patient history window is closer to discharge or death. Finally, we show that using the F1-score as an early stopping metric can stabilize and increase the performance of our approach for the task at hand. ###### Keywords: Federated Learning Early Mortality Prediction Recurrent Neural Networks Multivariate Time Series Intensive Care Unit ## 1 Introduction Intensive Care Units (ICUs) usually treat patients with a heightened mortality risk. A study conducted on patients admitted to 167 ICUs from 17 European countries during four weeks in 2011, and 2012 registered that, out of $5,834$ patients, $1,113\,(19\,\%)$ died in the ICU alone and a total of $1,397$ patients $(24\%)$ died in the whole hospital [3]. Mortality increases even further when considering patients admitted to an ICU during 2020 and 2021. Among $1,686$ patients admitted to the ICU with COVID-19, the mortality rate was $30\%$ [1]. The ICU is also one of the units where medical errors are most likely to occur, given the complexity of care. ICU patients are severely ill and usually subject to multiple complex interventions and treatments, thus leading to a high risk of an adverse outcome. In order to enable clinicians to take action and prevent such an outcome, the risk of patients’ mortality has to be predicted not only as accurately as possible but also as early as possible; we refer to this concept as early ICU mortality prediction. ICUs are therefore equipped with advanced diagnostic and therapeutic resources in order to enable quick response to changes in patients’ health. Traditionally, hospitals use the collected Electronic Health Records (EHRs) to assess individual mortality risks in the ICU with the help of scores like APACHE [13] and SAPS [5]. Recent research in the field of Machine Learning (ML) has shown improvements in these scores in terms of predictive performance [2, 11]. The results of Awad et al. [2] clearly showcase the performance improvement of classical ML methods over clinical severity scores for early ICU mortality prediction. Similarly, Johnson and Mark [11] show that classical ML models trained using data from the first $24\,\mathrm{h}$ of the ICU stay can outperform clinical scores. These ML models rely on big amounts of data in order to learn correlations between current patient data and their risk of mortality within a specific time window in the future. Conventionally, ML approaches use all the available data to train the model in a centralized manner. However, as patient data are usually subject to privacy regulations, like the European GDPR, the data cannot simply be shared between hospitals to train Centralized ML (CML) models. Alternatively, locally available data could be used by each hospital to set up its own independent Local ML (LML) early warning system. However, this paradigm could suffer from a lack of sufficient training data and would not consider the heterogeneity of patients across multiple medical centers. A promising alternative to LML or CML is presented by Federated Learning (FL). Instead of exchanging data, this ML framework relies on training many local models of identical structure at the location of the data, which are then combined into a global model. In our case, the data are stored at different hospitals, which we refer to as clients. Apart from ensuring privacy by design, FL enables parallel training using the clients as computational units [14, 17]. FL can be perfectly integrated into hospital facilities that store patients’ information in the form of EHRs. An advantage of EHRs is that data are collected in the form of Multivariate Time-Series (MTS), which means that each feature consists of a stream of values changing over time. Deep Learning (DL) architectures like Recurrent Neural Networks (RNNs) represent powerful tools for dealing with this kind of data, as they also take into account the history encoded by the time series data. One major challenge when training a DL model is to find a trade-off between training the neural network enough to learn the features of the training set and stopping before it overfits the training data. One possible technique to find the optimal number of training epochs is Early Stopping (ES). In case of overfitting, the validation performance of the model begins to degrade, and the training process stops. Different metrics, such as loss or F1-score, can be used to trigger ES. Their suitability depends on the task. Some approaches in literature use RNNs for binary classification on MTS from the ICU; however, they do not tackle the problem in combination with FL. Pattalung et al. [16] compare different RNN architectures to create benchmark values for ICU mortality prediction tasks on different publicly accessible databases. The authors take MTS data over a $48\,\mathrm{h}$ period in order to predict whether a patient dies at the end of this period. Ge et al. [6] combine Logistic Regression and LSTM for early ICU mortality prediction after $48\,\mathrm{h}$ of ICU admission. Their results show that DL achieves higher accuracy than Logistic Regression in identifying patients at high risk of death. Despite the benefits of using an FL setup, few papers address the problem of mortality prediction with FL. In this study, we built upon the research of Mondrejevski et al. [15] that propose FLICU, an FL workflow for mortality prediction tasks in the ICU. However, their paper does not focus on predicting ICU mortality at an early stage but it is rather a retrospective study. This paper tackles the current limitations by combining FL with early prediction of ICU mortality. The main contributions are summarized below: 1. 1. We propose a workflow for predicting early ICU mortality using deep FL. 2. 2. We analyze the predictive performance of our proposed solution on different time windows and data cohorts. 3. 3. We compare the predictive performance of the FL setup (with $2$, $4$, and $8\,$clients) against CML and LML. 4. 4. We compare two early-stopping criteria during training: loss and F1-score. ## 2 Method In this paper, we propose a workflow for early ICU mortality prediction that comprises three main phases: (i) Data Preparation, (ii) Window Selection, and (iii) Modeling. A schematic view of our workflow is shown in Fig. 0a. (a) (b) Patient data is available from the admission to the ICU $t_{adm}$ to death or discharge from the ICU $t_{end}$. Training data is sampled during $\Delta t_{data}$ (from $t_{adm}$ to $t_{eval}$). This leaves a prediction window $\Delta t_{pred}$, during which data is unknown to the model. (c) Data are split into $F$ equally sized folds. Over $F$ iterations, each of the folds is used $F-1$ times for training and validation and once for testing. For FL and LML, we further split the combined training and validation data into the number of clients. These partitions are, in turn, divided into training and validation data. Figure 1: (a) Schematic Workflow, (b) Time-window representation (c), Data splits. ### 2.1 Problem Formulation Our basic research problem can be formulated as a binary classification problem for early ICU mortality prediction: given a patient cohort $D$ that consists of $n$ patients, we aim to estimate the real class label $y_{i}\in[0,1]$ for each patient $i\in[0,1,\dotsc,n]$ by predicting a label $\hat{y}_{i}$. The label $y_{i}$ denotes whether patient $i$ died or survived the ICU stay and $\hat{y}_{i}$ indicates the patient’s predicted mortality risk. The label $\hat{y}_{i}$ is estimated based on an MTS feature stream $X_{i}(t)=[x_{i0}(t),x_{i1}(t),\dotsc,x_{im}(t)]$, that consists of $m$ features (e.g. vital signs or laboratory values). Each $x_{ij}(t),~{}j\in[0,1,\dotsc,m]$ represents the value of a univariate time series at time $t$. Since the main focus of this study is early prediction, the observations $X_{i}(t)$ that are used are collected only during the first hours of the patient’s ICU stay. Furthermore, we assume that the patients are randomly distributed over $K$ ICUs/hospitals. $D_{k}$ is the local cohort of patients (i.e., the set of locally available $X_{i}(t)$) in each hospital $k$. Thus, $D=D_{0}~{}\cup~{}D_{1}~{}\cup~{}\dotsc~{}\cup~{}D_{K}$ is the global dataset of $n$ patients with $D_{a}\cap D_{b}=\emptyset,\forall(a,b)\in[1,K]\times[1,K]$. In this paper, we compare the efficiency of CML, LML, and FL on this problem. ### 2.2 Data Preparation In the first phase of our workflow, we are preparing the data for the predictive modeling, following five steps (based on the FLICU-workflow [15]): (i) Patient selection: Initially, we select patients with an ICU stay and filter the cohort according to the criteria below: 1. 1. We dismiss all but the first ICU stay of each patient. 2. 2. We dismiss patients with data recorded for less than $\Delta t_{min}$. 3. 3. We dismiss patients staying longer than $\Delta t_{max}$. The windows $\Delta t_{min}$ and $\Delta t_{max}$ allow us to define the patient cohorts. $\Delta t_{min}$ denotes the minimum length of MTS data per patient. Additionally, it guarantees that all patients will have at least the required history length. The upper bound $\Delta t_{max}$ limits the prediction window in order to dismiss patients that stay in the ICU for an extensive period of time. (ii) Feature selection: Consecutively, we extract vital signs and lab values in the form of MTS. We extract vitals explicitly connected to ICU stays and consider the time of the first vital of each patient to be the time of admission to the ICU. If no vitals are recorded for an ICU stay, we use the admission time logged by the hospital staff. Lab values are sampled during the patient’s whole ICU stay, from ICU admission to ICU discharge or death. (iii) Re-sampling: Vital signs and lab values are re-sampled to fixed length intervals. The length of these intervals depends on the data collection frequency in the hospitals and may differ for vitals and labs. If more than one measurement falls in the same interval, the median is used for data aggregation. (iv) Imputation: Missing values are treated with forward, and then backward imputation, starting from the beginning of the ICU stay. Non-observed features for a patient’s ICU stay are replaced with $-1$. (v) Labeling: Finally, if a patient dies in the ICU, we assign the class label $y_{i}=1$ (death). Otherwise, we assume that the patient survived the ICU and assign the label $y_{i}=0$ (discharge). ### 2.3 Window Selection In the second phase of our workflow, we are interested in selecting the patient history window that will be used as input to the predictive models, defined as $\Delta t_{data}\leq\Delta t_{min}$. $\Delta t_{data}$ starts at the time of ICU admission $t_{adm}$ and ends at the time of evaluation $t_{eval}$ (see Fig. 0b). Since the main focus of this study is early ICU mortality prediction, we aim to predict the label $\hat{y}_{i}$ ahead of the patient’s time of ICU discharge or death, thus $\Delta t_{data}$ is considered to end before the end of the ICU stay $t_{end}$. More precisely, there is a prediction window $\Delta t_{pred}\geq 0$ that starts at $t_{eval}$ and ends at $t_{end}$. ### 2.4 Modeling To avoid bias induced by data partitioning, we use $F$-fold cross-validation as the model evaluation technique. We split the data into $F$ partitions, where each split is used as a testing dataset once, while the remaining folds are split into training and validation sets for each client, respectively (see Fig. 0c). To allocate the data points throughout the clients (in LML and FL), we assume $K$ horizontal, stratified splits. This means that the data in each hospital $k\in[0,1,\dotsc,K]$ has the same number of patients $|D_{k}|=\frac{n}{K}$ with the same class distribution and that each patient’s records are kept in only one hospital at the same time. Before being passed into the model, all data streams are normalized according to the global minima and maxima found in the available training and validation data. To deal with class imbalance during training, class weights are added to the training data. This method down-weighs the impact of classes with more examples (in our case: the discharged patients) and increases the importance of classes with fewer data examples during error back-propagation. Thus, applying class weights enables us to achieve optimization-level class balance. The weights $w_{c}$ for each class $c$ are calculated according to the following formula: $w_{c}=\frac{\text{\\# of samples in dataset}}{\text{\\# of samples in class }c}$. This setup is aggregated to the three configurations of our predictive model, FL, CML, and LML. As both main gated RNN architectures, LSTM [8], and GRU [4] have shown to perform equally well on tasks comparable to ours [15, 16], we focus on the less resource-intensive GRU in this study. Our basic DL model architecture (similar to [15] and [16]) consists of two parallel input layers, one for the vital signs and one for the laboratory variables, each followed by three recurrent layers of 16 GRUs. Consecutively, we perform batch normalization and combine the resulting outputs using two fully connected layers. Finally, we add a sigmoid-layer for estimating the patient risk of ICU mortality with a value between $0$ and $1$. Matching to this binary output of the system, we use binary cross-entropy as the loss function $L(\cdot)$. In CML, the model $f^{CML}(\cdot,\theta)$ is trained on the whole training data $D$, where $\theta$ is a set of DL weights defining the function of the model. The goal is to find the optimal set of weights $\theta^{CML}$ that minimizes the error between $y_{i}$ and $\hat{y}_{i}=f^{CML}(X_{i}(t),\theta)$ for a given $X_{i}(t)$. More formally: $\theta^{CML}~{}=~{}\underset{\theta^{*}\in\mathbb{R}}{\mathrm{argmin}}~{}L(D,\theta^{*})\vspace{-2mm}$ (1) After each epoch of training, we evaluate the predictive performance of $f^{CML}(\cdot,\theta)$ on the validation data using a previously selected metric $M(\bf{\hat{y}},\bf{y})$, comparing a vector of predictions $\bf{\hat{y}}$ with the vector of true class labels $\bf{y}$. In order to avoid overfitting the training data, we stop the training and reset $\theta$ to the time of the best score $s^{*}$ if $s=M(\cdot)$ does not improve for a predefined number of epochs $P$. We refer to this as early stopping (ES) with patience $P$. In LML, each hospital $k$ trains a local model $f^{LML}_{k}(\cdot,\theta)$, using only the data in the local dataset $D_{k}$, where $\hat{y}_{i}=f^{LML}_{k}(X_{i}(t),\theta_{k})$. As before, the goal is to minimize the prediction error by producing an optimal set of weights $\theta^{LML}_{k}$ for each client $k$: $\theta^{LML}_{k}~{}=~{}\underset{\theta^{*}\in\mathbb{R}}{\mathrm{argmin}}~{}L(D_{k},\theta^{*})$ (2) Similarly to CML, we calculate local validation scores $s_{k}=M(\bf{\hat{y}}_{k},\bf{y}_{k})$ on each client $k$ in every epoch and monitor them for ES. For the FL approach, we use a slightly modified version of Federated Averaging (FedAvg) [14]: (i) First, all local models of the $K$ hospitals are initialized with the same set of starting weights $\theta^{0}$. (ii) By performing local training for $E$ local epochs, each of the participating hospitals derives an updated set of weight parameters $\theta_{k}$. The fraction of participating hospitals per round is $C$. The local objective is similar to Equation 2. (iii) After local training, a global model $f^{FL}(\cdot,\theta)$ is created by averaging all $\theta_{k}$ to one set of parameters $\theta$. (iv) These averaged weights are then sent back to the hospitals, which overwrite their local weights with the new ones. (v) Afterwards, an ES score $s$ is calculated by evaluating the local models on the respective validation set of each client $k$ and then averaging the results: $s=\sum^{K}_{k=1}\frac{n_{k}}{n}M(\bf{\hat{y}}_{k},\bf{y}_{k})$. The above is repeated from step (ii) until the validation scores suggest an optimal set of weights $\theta^{FL}$ and ES is activated. Here, the objective is to optimize the global model $f^{FL}(\cdot,\theta)$ by calculating: $\theta^{FL}~{}=~{}\underset{\theta^{*}\in\mathbb{R}}{\mathrm{argmin}}~{}\sum\limits^{K}_{k=1}\frac{n_{k}}{n}F_{k}(\theta^{*})\vspace{-2mm}$ (3) where $F_{k}(\theta^{*})=\frac{1}{n_{k}}L(D_{k},\theta^{*})$, $n_{k}=|D_{k}|$. ## 3 Empirical Evaluation ### 3.1 Data Description In this paper, we use the MIMIC-III (version 1.4) clinical dataset provided by PhysioNet [7]. This dataset provides de-identified data collected from ICU patients in Beth Israel Deaconess Medical Center in Boston, Massachusetts. The data was collected from $46,476$ patients during the years 2001 to 2012 [9, 10]. The database includes patient information such as demographics, vital sign measurements, and laboratory test results. Initially, we select the patients based on the criteria described in Section 0, and in addition, we dismiss patients from the Neonatal Intensive Care Unit (NICU) and Pediatric Intensive Care Unit (PICU). For this study, we use two different cohorts of patients: (i) Cohort 1 is identified by $\Delta t_{min}=24\,\mathrm{h}$ and $\Delta t_{max}=72\,\mathrm{h}$, and we compare different $\Delta t_{data}\in[8\,\mathrm{h},16\,\mathrm{h},24\,\mathrm{h}]$; and (ii) Cohort 2 is identified by $\Delta t_{min}=48\,\mathrm{h}$ and $\Delta t_{max}=96\,\mathrm{h}$, and we compare $\Delta t_{data}\in[8\,\mathrm{h},16\,\mathrm{h},24\,\mathrm{h},32\,\mathrm{h},40\,\mathrm{h},48\,\mathrm{h}$]. Both of these cohorts use $\Delta t_{max}=\Delta t_{min}+48\,\mathrm{h}$. For simplicity, we therefore, refer to the cohorts by their $\Delta t_{min}$ only. For the pre-processing and feature selection, we follow the approach from [15, 16]. Initially, we extract demographic information, such as gender and age, that is used for describing the cohorts. We also extract $7$ vital signs and $16$ lab values, shown in Tab. 0, in the form of MTS. In this paper, vital signs are re-sampled in $1\,\mathrm{h}$ intervals, while we use a sampling interval of $8\,\mathrm{h}$ for lab values. Table 1: Vital and Laboratory Values Vital Signs: (1 per hour) | heart-rate, systolic blood-pressure, diastolic blood-pressure, mean blood-pressure, respiratory rate, core temperature, blood oxigen saturation (spo2) ---|--- Lab Values: (1 per 8 hours) | albumin, blood urea nitrogen (bun), bilirubin, lactate, bicarbonate, band neutrophils (bands), chloride, creatinine, glucose, hemoglobin, hematocrit, platelets, potassium, partial thromboplastin time (ptt), sodium, white blood-cells Finally, we extract the label indicating whether a patient died or not based on the column deathtime in table admissions of MIMIC-III. If a deathtime is recorded for patient $i$ during the ICU stay, we assign the label $y_{i}=1$. Otherwise, we assume that the patient survived the ICU and assign the label $y_{i}=0$. As shown in Tab. 0, there is a heavily imbalanced class distribution as there are more patients in the discharge class than in the death class. Table 2: Cohort Sizes and Class Distribution | Cohort 1 [$\Delta t_{min}=24\,\mathrm{h}$] | Cohort 2 [$\Delta t_{min}=48\,\mathrm{h}$] ---|---|--- Deaths | Discharges | Total | Deaths | Discharges | Total absolute | percent | absolute | percent | absolute | percent | absolute | percent Patients | $804$ | $4.4\%$ | $17,477$ | $95.6\%$ | $18,281$ | $547$ | $5.4\%$ | $9,496$ | $94.6\%$ | $10,043$ Male | $420$ | $2.3\%$ | $10,075$ | $55.1\%$ | $10,495$ | $287$ | $2.9\%$ | $5,255$ | $52.3\%$ | $5,542$ Female | $384$ | $2.1\%$ | $7,402$ | $40.5\%$ | $7,786$ | $260$ | $2.6\%$ | $4,241$ | $42.2\%$ | $4,501$ Age 0 to 29 | $18$ | $0.1\%$ | $892$ | $4.9\%$ | $910$ | $19$ | $0.2\%$ | $421$ | $4.2\%$ | $440$ Age 30 to 59 | $177$ | $1.0\%$ | $5,777$ | $31.6\%$ | $5,954$ | $133$ | $1.3\%$ | $2,863$ | $28.5\%$ | $2,996$ Age 60 to 89 | $525$ | $2.9\%$ | $9,920$ | $54.3\%$ | $10,445$ | $352$ | $3.5\%$ | $5,692$ | $56.7\%$ | $6,044$ Age 90+ | $84$ | $0.5\%$ | $888$ | $4.9\%$ | $972$ | $43$ | $0.4\%$ | $520$ | $5.2\%$ | $563$ ### 3.2 Model Training and Evaluation In this study, we use $F=5$ folds for stratified cross-validation, where we evaluate CML, LML, and FL on the same testing data splits within each cross- validation round. In each round, the data of the remaining four folds are again partitioned in LML and FL models, as those are trained with different numbers of clients $K\in[2,4,8]$, whilst all remaining data are used in CML. The available data in each scenario (either all remaining data in CML or each client’s data in LML and FL) are split into $80\%$ training and $20\%$ validation (see Fig. 0c). We use Adaptive Moment Estimation (ADAM) [12] as an optimizer for updating the network weights in CML, FL, and LML. Additionally, we apply an initial learning rate $\eta$ of $0.01$, which is reduced by $50\%$ every five epochs. Lastly, we use ES via monitoring the loss or F1-score on the validation set with patience $P=30$ and the maximum number of epochs set to $100$. In case the F1-score is undefined, i.e., recall and precision are zero, we set it to $-1$. The local minibatch size $B$ depends on the number of clients $K$ participating in each configuration: we use $B=512/K$ for performance reasons. For FL, we use $E=1$ number of local epochs. The fraction of clients computing in each FL round is $C=1$, meaning that all clients participate in each iteration. To compare the performance of the different settings, we use the evaluation metrics AUPRC, AUROC, Precision, Recall, and F1-score. AUROC is chosen as it is commonly used to assess the performance of ICU mortality prediction. However, AUPRC and F1-score are more suitable for highly imbalanced classes, which is the case for our problem. The entire code produced in this paper is publicly available on GitHub 111https://github.com/randlbem/Early˙ICU˙mortality˙prediction˙with˙deep˙FL.git. ### 3.3 Results & Discussion As previously described, we evaluate the ability of our proposed FL workflow to predict the risk of ICU mortality at an early stage. We compare it with the CML and LML approaches on two cohorts of patients ($\Delta t_{min}=24\,\mathrm{h}$ and $48\,\mathrm{h}$), using two ES metrics (loss and F1-score) and different time windows $\Delta t_{data}$. The results for cohort $\Delta t_{min}=24\,\mathrm{h}$, using the minimum loss for ES, are shown in Tab. 0. Overall, the results show that our model performs well in the task of early ICU mortality prediction. For example, for FL with $\Delta t_{data}=24\,\mathrm{h}$, we obtain an average AUROC of $0.90\pm 0.01$ and an AUPRC of $0.47\pm 0.04$. It’s important to highlight that due to the class imbalance of $4.4\,\%$ towards the positive class, the baseline value for the AUPRC is $0.044$. In addition, we yield three conclusions from Tab. 0: (i) While an increasing number of clients results in a decrease in predictive performance over all the metrics in LML, the performance of FL remains close to that of CML, regardless of the number of clients. (ii) With growing $\Delta t_{data}$, and respectively shrinking $\Delta t_{pred}$, the model performance increases. (iii) The relation between precision and recall is very unstable, which is shown in the fluctuation of the averages and the high standard deviations. While (i) clearly shows that FL has the potential to improve on LML for early ICU mortality prediction, (ii) and (iii) are further explored in the following experiments. Table 3: ES with min. loss ($\Delta t_{min}=24\,\mathrm{h}$). score | CML | FL | LML ---|---|---|--- | | $2~{}clients$ | $4~{}clients$ | $8~{}clients$ | $2~{}clients$ | $4~{}clients$ | $8~{}clients$ $\bf\Delta t_{data}=8\,\mathrm{h}$; avg. $\Delta t_{pred}=35.2\,\mathrm{h}$ AUROC | $0.87\pm 0.02$ | $0.86\pm 0.01$ | $0.86\pm 0.01$ | $0.87\pm 0.01$ | $0.85\pm 0.01$ | $0.82\pm 0.01$ | $0.81\pm 0.01$ AUPRC | $0.36\pm 0.03$ | $0.37\pm 0.02$ | $0.37\pm 0.02$ | $0.37\pm 0.03$ | $0.34\pm 0.04$ | $0.31\pm 0.03$ | $0.29\pm 0.02$ F1 | $0.29\pm 0.15$ | $0.38\pm 0.01$ | $0.36\pm 0.05$ | $0.38\pm 0.04$ | $0.34\pm 0.07$ | $0.34\pm 0.02$ | $0.25\pm 0.04$ precision | $0.62\pm 0.21$ | $0.38\pm 0.07$ | $0.39\pm 0.11$ | $0.41\pm 0.08$ | $0.53\pm 0.14$ | $0.41\pm 0.06$ | $0.48\pm 0.08$ recall | $0.24\pm 0.14$ | $0.41\pm 0.07$ | $0.42\pm 0.14$ | $0.39\pm 0.09$ | $0.29\pm 0.10$ | $0.30\pm 0.07$ | $0.17\pm 0.04$ $\bf\Delta t_{data}=16\,\mathrm{h}$; avg. $\Delta t_{pred}=27.2\,\mathrm{h}$ AUROC | $0.89\pm 0.01$ | $0.88\pm 0.01$ | $0.88\pm 0.01$ | $0.88\pm 0.01$ | $0.87\pm 0.01$ | $0.85\pm 0.01$ | $0.82\pm 0.01$ AUPRC | $0.44\pm 0.04$ | $0.41\pm 0.04$ | $0.41\pm 0.04$ | $0.42\pm 0.03$ | $0.40\pm 0.04$ | $0.35\pm 0.05$ | $0.32\pm 0.02$ F1 | $0.42\pm 0.01$ | $0.35\pm 0.07$ | $0.41\pm 0.02$ | $0.38\pm 0.08$ | $0.29\pm 0.05$ | $0.29\pm 0.12$ | $0.22\pm 0.02$ precision | $0.51\pm 0.10$ | $0.58\pm 0.13$ | $0.47\pm 0.05$ | $0.46\pm 0.07$ | $0.55\pm 0.18$ | $0.50\pm 0.14$ | $0.55\pm 0.09$ recall | $0.38\pm 0.07$ | $0.27\pm 0.09$ | $0.37\pm 0.04$ | $0.34\pm 0.10$ | $0.21\pm 0.03$ | $0.22\pm 0.10$ | $0.14\pm 0.02$ $\bf\Delta t_{data}=24\,\mathrm{h}$; avg. $\Delta t_{pred}=19.2\,\mathrm{h}$ AUROC | $0.89\pm 0.01$ | $0.90\pm 0.01$ | $0.90\pm 0.01$ | $0.89\pm 0.01$ | $0.89\pm 0.01$ | $0.87\pm 0.01$ | $0.83\pm 0.01$ AUPRC | $0.48\pm 0.03$ | $0.48\pm 0.03$ | $0.47\pm 0.04$ | $0.45\pm 0.04$ | $0.46\pm 0.03$ | $0.41\pm 0.05$ | $0.37\pm 0.03$ F1 | $0.10\pm 0.11$ | $0.33\pm 0.17$ | $0.34\pm 0.07$ | $0.30\pm 0.17$ | $0.22\pm 0.12$ | $0.26\pm 0.07$ | $0.32\pm 0.05$ precision | $0.66\pm 0.38$ | $0.70\pm 0.18$ | $0.75\pm 0.06$ | $0.72\pm 0.16$ | $0.51\pm 0.18$ | $0.47\pm 0.10$ | $0.59\pm 0.15$ recall | $0.05\pm 0.07$ | $0.27\pm 0.16$ | $0.22\pm 0.05$ | $0.23\pm 0.14$ | $0.15\pm 0.08$ | $0.19\pm 0.06$ | $0.23\pm 0.05$ #### 3.3.1 Studying the influence of the size of $\Delta t_{data}$ and $\Delta t_{pred}$. To assess whether the performance improves with an increasing $\Delta t_{data}$, decreases with an increasing $\Delta t_{pred}$, or both, we compare the test scores of the two cohorts ($\Delta t_{min}=24\,\mathrm{h}$ and $\Delta t_{min}=48\,\mathrm{h}$). The comparison is shown in Fig. 0. The figure shows that cohort $\Delta t_{min}=24\,\mathrm{h}$ achieves higher performance than cohort $\Delta t_{min}=48\,\mathrm{h}$, at the same $\Delta t_{data}$ (Fig. 0a). Nevertheless, both cohorts’ performance increases alongside $\Delta t_{data}$. When comparing the performance over the length of $\Delta t_{pred}$, we see that with a rising $\Delta t_{pred}$, all the curves are decreasing in a similar manner (Fig. 0b). This behavior can be seen over different models and metrics and is a strong indicator that the size of $\Delta t_{pred}$ is more important than the size of $\Delta t_{data}$, since $\Delta t_{pred}$ is bigger in cohort $\Delta t_{min}=48\,\mathrm{h}$ than cohort $\Delta t_{min}=24\,\mathrm{h}$ for the same $\Delta t_{data}$. This means that even for small $\Delta t_{data}$, $8\,\mathrm{h}$ or just $6\,\mathrm{h}$ as Awad et al. [2] demonstrate, prediction should be possible, if $\Delta t_{pred}$ is small enough. Table 4: ES with max. F1 ($\Delta t_{min}=24\,\mathrm{h}$). score | CML | FL | LML ---|---|---|--- | | $2~{}clients$ | $4~{}clients$ | $8~{}clients$ | $2~{}clients$ | $4~{}clients$ | $8~{}clients$ $\bf\Delta t_{data}=8\,\mathrm{h}$; avg. $\Delta t_{pred}=35.2\,\mathrm{h}$ AUROC | $0.87\pm 0.01$ | $0.87\pm 0.01$ | $0.87\pm 0.01$ | $0.87\pm 0.01$ | $0.86\pm 0.01$ | $0.84\pm 0.01$ | $0.80\pm 0.01$ AUPRC | $0.39\pm 0.04$ | $0.38\pm 0.04$ | $0.37\pm 0.04$ | $0.37\pm 0.03$ | $0.36\pm 0.05$ | $0.32\pm 0.04$ | $0.28\pm 0.02$ F1 | $0.40\pm 0.03$ | $0.38\pm 0.02$ | $0.39\pm 0.04$ | $0.39\pm 0.04$ | $0.39\pm 0.03$ | $0.36\pm 0.03$ | $0.35\pm 0.02$ precision | $0.42\pm 0.07$ | $0.38\pm 0.05$ | $0.35\pm 0.06$ | $0.35\pm 0.08$ | $0.37\pm 0.05$ | $0.37\pm 0.05$ | $0.35\pm 0.04$ recall | $0.41\pm 0.08$ | $0.41\pm 0.11$ | $0.46\pm 0.08$ | $0.47\pm 0.09$ | $0.43\pm 0.04$ | $0.36\pm 0.03$ | $0.34\pm 0.03$ $\bf\Delta t_{data}=16\,\mathrm{h}$; avg. $\Delta t_{pred}=27.2\,\mathrm{h}$ AUROC | $0.89\pm 0.01$ | $0.88\pm 0.01$ | $0.88\pm 0.01$ | $0.88\pm 0.01$ | $0.87\pm 0.01$ | $0.85\pm 0.01$ | $0.82\pm 0.01$ AUPRC | $0.44\pm 0.05$ | $0.41\pm 0.04$ | $0.41\pm 0.04$ | $0.42\pm 0.03$ | $0.40\pm 0.05$ | $0.37\pm 0.04$ | $0.33\pm 0.03$ F1 | $0.44\pm 0.02$ | $0.41\pm 0.01$ | $0.40\pm 0.05$ | $0.43\pm 0.03$ | $0.41\pm 0.02$ | $0.39\pm 0.03$ | $0.37\pm 0.02$ precision | $0.46\pm 0.05$ | $0.41\pm 0.04$ | $0.43\pm 0.06$ | $0.43\pm 0.04$ | $0.38\pm 0.08$ | $0.38\pm 0.05$ | $0.41\pm 0.03$ recall | $0.42\pm 0.05$ | $0.42\pm 0.07$ | $0.40\pm 0.11$ | $0.43\pm 0.06$ | $0.46\pm 0.08$ | $0.42\pm 0.06$ | $0.33\pm 0.04$ $\bf\Delta t_{data}=24\,\mathrm{h}$; avg. $\Delta t_{pred}=19.2\,\mathrm{h}$ AUROC | $0.90\pm 0.01$ | $0.90\pm 0.01$ | $0.90\pm 0.00$ | $0.89\pm 0.01$ | $0.89\pm 0.01$ | $0.87\pm 0.01$ | $0.83\pm 0.01$ AUPRC | $0.50\pm 0.04$ | $0.49\pm 0.04$ | $0.47\pm 0.04$ | $0.47\pm 0.03$ | $0.47\pm 0.04$ | $0.43\pm 0.04$ | $0.37\pm 0.03$ F1 | $0.49\pm 0.04$ | $0.48\pm 0.03$ | $0.46\pm 0.03$ | $0.44\pm 0.03$ | $0.45\pm 0.02$ | $0.42\pm 0.02$ | $0.40\pm 0.02$ precision | $0.55\pm 0.05$ | $0.52\pm 0.05$ | $0.49\pm 0.07$ | $0.45\pm 0.07$ | $0.48\pm 0.08$ | $0.39\pm 0.04$ | $0.46\pm 0.02$ recall | $0.46\pm 0.08$ | $0.46\pm 0.05$ | $0.44\pm 0.05$ | $0.46\pm 0.11$ | $0.46\pm 0.10$ | $0.48\pm 0.09$ | $0.35\pm 0.03$ (a)(b) The dots represent the mean values over the 5-fold cross-validation iterations. The vertical bars show the standard deviation. The scores were calculated on the test sets, while the ES metric is the loss. Figure 2: Mean AUPRC of the two scenarios ($\Delta t_{min}=24\,\mathrm{h}$ and $\Delta t_{min}=48\,\mathrm{h}$). #### 3.3.2 Stabilizing precision & recall. The curves represent the mean values of the validation scores over the 5-fold cross-validation iterations using loss as the ES metric. The borders of the shaded areas mark the standard deviation. The best loss and F1 values recognized by ES fall within the red and purple regions. Figure 3: Learning progress of CML ($\Delta t_{min}=24\,\mathrm{h}$). To better understand the interplay between precision and recall, we examine their fluctuations from the CML model, as it represents the most basic case. Its learning curve in Fig. 0 shows that there is a trade-off between precision and recall: precision increases during the first 12 epochs and then shows a decline, while recall increases steadily. We can also see that the loss shows a very flat minimum while precision is still stabilizing, and recall has just begun to increase. This means that minimal changes in the loss’s progression can greatly impact the models’ precision and recall. The F1-score, however, shows a more defined maximum. In order to stabilize precision and recall, we re-train the models whose performance is shown in Tab. 0, using the F1-score as the ES metric. As the F1-Score is the harmonic mean between precision and recall, using it for stopping the training should create a model with an optimal balance between precision and recall. The results are shown in Tab. 0. A comparison of tables Tab. 0 and Tab. 0 shows that ES with the highest F1-score produces better and more stable results for all models. For example, for FL with $\Delta t_{data}=24\,\mathrm{h}$, we now obtain an average F1-score of $0.46\pm 0.03$, instead of $0.32\pm 0.15$ (as shown in Tab. 0) in addition to marginally higher AUROC and AUPRC values. ## 4 Conclusion We present an FL workflow that allows for early ICU mortality prediction. Our results show that FL performs equally well as the CML approach and substantially better than the LML, especially as the number of clients increases. While the performance remains stable in FL with 2, 4, and 8 clients, in LML, the performance decreases considerably with an increasing number of clients. These findings are based on the AUROC score - widely used in the literature but ill-suited for the heavily imbalanced data in this problem - but also on the more meaningful AUPRC and F1-score. Furthermore, our results indicate, in agreement with literature [2, 16], that the size of the prediction window is much more important for the performance in early prediction tasks than the length of patient history during which data is collected. Thus, the results show better predictive performance when the patient history window is closer to the end of the ICU stay. Lastly, we show that using the F1-score as an ES metric can stabilize and increase the predictive performance in tasks like ours. Nevertheless, this study also creates the basis for future work. Since we limit ourselves to horizontal and stratified client splits for comparability reasons, it is necessary to re-evaluate our findings in more realistic settings. In addition, our FL workflow performs considerably well in predicting ICU mortality at an early stage using the MIMIC-III dataset. However, the generalizability of the approach needs to be tested beyond this specific dataset. Furthermore, it would be interesting to explore how far ahead of death or discharge early ICU mortality can be reasonably predicted and hereby expand on our findings. ## References * [1] Auld, S.C., Harrington, K.R.V., Adelman, M.W., Robichaux, C.J., Overton, E.C., Caridi-Scheible, M., Coopersmith, C.M., Murphy, D.J.: Research collaborative trends in icu mortality from coronavirus disease 2019: A tale of three surges. the Emory COVID-19 Quality and Clinical 50(2), 245–255 (2022) * [2] Awad, A., Bader-El-Den, M., Briggs, J., McNicholas, J., El-Sonbaty, Y.: Predicting hospital mortality for intensive care unit patients: Time-series analysis. Health Informatics Journal 26(2), 1043–1059 (2020) * [3] Capuzzo, M., Volta, C., Tassinati, T., Moreno, R., Valentin, A., Guidet, B., Iapichino, G., Martin, C., Perneger, T., Combescure, C., Poncet, A., Rhodes, A., Oeyen, S., Matejovic, M., Toft, P., Wrigge, H., et al.: Hospital mortality of adults admitted to intensive care units in hospitals with and without intermediate care units: A multicentre european cohort study. Critical Care 18(5) (2014) * [4] Cho, K., van Merrienboer, B., Bahdanau, D., Bengio, Y.: On the properties of neural machine translation: Encoder-decoder approaches (2014) * [5] Gall, J.R., Lemeshow, S., Saulnier, F.: A new simplified acute physiology score (saps ii) based on a european/north american multicenter study. JAMA: The Journal of the American Medical Association 270(24), 2957–2963 (1993) * [6] Ge, W., Huh, J.W., Rang, Y.P., Lee, J., Kim, Y.H., Turchin, A.: An interpretable icu mortality prediction model based on logistic regression and recurrent neural networks with lstm units. In: AMIA annual symposium. vol. 1, pp. 460 – 469. American Medical Informatics Association (2018) * [7] Goldberger, A.L., Amaral, L.A., Glass, L., Hausdorff, J.M., Ivanov, P.C., Mark, R.G., Mietus, J.E., Moody, G.B., Peng, C.K., Stanley, H.E.: Physiobank, physiotoolkit, and physionet: components of a new research resource for complex physiologic signals. Circulation 101(23), E215–220 (2000) * [8] Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Computation 9(8), 1735 (1997) * [9] Johnson, A.E.W., Pollard, T.J., Lehman, L.W.H., Feng, M., Ghassemi, M., Moody, B., Celi, L.A., Mark, R.G., Shen, L., Szolovits, P.: Mimic-iii, a freely accessible critical care database. Scientific Data 3 (2016) * [10] Johnson, A.E.W., Pollard, T.J., Mark, R.G.: Mimic-iii clinical database (version 1.4). PhysioNet (2016) * [11] Johnson, A.E., Mark, R.G.: Real-time mortality prediction in the intensive care unit. In: AMIA Annual Symposium Proceedings. vol. 2017, p. 994. American Medical Informatics Association (2017) * [12] Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014) * [13] Knaus, W., Wagner, D., Draper, E., Zimmerman, J., Bergner, M., Bastos, P., Sirio, C., Murphy, D., Lotring, T., Damiano, A., Harrell Jr., F.: The apache iii prognostic system: Risk prediction of hospital mortality for critically iii hospitalized adults. Chest 100(6), 1619–1636 (1991) * [14] McMahan, H.B., Moore, E., Ramage, D., Hampson, S., Arcas, B.A.y.: Communication-efficient learning of deep networks from decentralized data (2016) * [15] Mondrejevski, L., Miliou, I., Montanino, A., Pitts, D., Hollmen, J., Papapetrou, P.: Flicu: A federated learning workflow for intensive care unit mortality prediction. In: 2022 IEEE 35th International Symposium on Computer-Based Medical Systems (CBMS). pp. 32–37. IEEE Computer Society, Los Alamitos, CA, USA (2022) * [16] Na Pattalung, T., Ingviya, T., Chaichulee, S.: Feature explanations in recurrent neural networks for predicting risk of mortality in intensive care patients. Journal of Personalized Medicine 11(9) (2021) * [17] Shokri, R., Shmatikov, V.: Privacy-preserving deep learning. In: Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security. p. 1310–1321. CCS ’15, Association for Computing Machinery, New York, NY, USA (2015)
# Quantum targeted energy transfer through machine learning tools I. Andronis<EMAIL_ADDRESS>Department of Physics, University of Crete, Heraklion 70013, Greece G. Arapantonis<EMAIL_ADDRESS>Department of Physics, University of Crete, Heraklion 70013, Greece William H. Miller III Department of Physics & Astronomy, Johns Hopkins University, Baltimore, Maryland 21218, USA G. D. Barmparis Department of Physics, University of Crete, Heraklion 70013, Greece G. P. Tsironis Department of Physics, University of Crete, Heraklion 70013, Greece John A. Paulson School of Engineering and Applied Sciences, Harvard University, Cambridge, Massachusetts 02138, USA ###### Abstract In quantum targeted energy transfer, bosons are transferred from a certain crystal site to an alternative one, utilizing a nonlinear resonance configuration similar to the classical targeted energy transfer. We use a novel computational method based on machine learning algorithms in order to investigate selectivity as well as efficiency of the quantum transfer in the context of a dimer and a trimer system. We find that our method identifies resonant quantum transfer paths that allow boson transfer in unison. The method is readily extensible to larger lattice systems involving nonlinear resonances. ## I Introduction Nonlinear dynamical systems are notoriously difficult to study analytically and in most cases, one needs to resort to numerical methods for their analysis [1]. In the classical realm, they are typically described mathematically through coupled nonlinear differential equations that very scarcely admit exact solutions. In this case, one resorts to direct numerical integration of the equations of motion. In the quantum domain, on the other hand, even though the equations are linear, one needs to engage a very large part of the Hilbert space in order to find good, yet approximate, solutions [2]. Direct numerical methods, either in the classical or quantum domain, are relatively straightforward, yet might fail in large or strongly coupled systems. The recent widespread of Machine Learning (ML) techniques and their implementation in the domain of dynamical systems aim to both facilitate and also improve the discovery process in these systems [3]. The aim of the present work is to use techniques motivated by ML and obtain results that would be otherwise very complex to derive. The specific model we work with is that of Targeted Energy Transfer (TET) that was inspired by energy transfer processes in chlorophyll [4]. We have two targets here; the first one is to show how ML-motivated techniques may be superior to standard numerical methods when applied in quantum complex systems and the second is to find explicit results that would be otherwise much more difficult to obtain. In the semiclassical TET model, we focus on resonant exciton transfer between non-identical molecules [4]. In each molecule a single energy state participates in the process, thus we have non-identical energy states coupled together via a non-zero transfer matrix element. In the simplest case of two molecules, we deal with a non-degenerate linear dimer system. Due to the energy mismatch, the exciton transfer from the first site to the second is non-resonant and thus occurs only partially. In order to make the transfer resonant we need to add local interaction with additional degrees of freedom, such as phonons. In the anti-adiabatic approximation, this procedure introduces effective qubic nonlinearities [5]. In TET, complete resonant transfer is restored for specific nonlinearity parameter configurations linking the local interaction with anti-adiabatic phonons and the actual energy mismatch. The analysis is done within the context of the discrete nonlinear Schrödinger (DNLS) equation, a ubiquitous model for a large class of nonlinear phenomena [6],[7]. A number of analytical results known for DNLS dimers are particularly useful for ML implementations in nonlinear systems [8, 9, 10]. While in the case of the semiclassical TET dimer system the resonant transfer regime can be found analytically, similar analysis in larger systems is an arduous task. In order to bypass this difficulty, ML-based approaches have been introduced that enable the discovery of nonlinear resonances in a straightforward way [11, 12]. The method was tested both in the TET dimer and also in other analytically known DNLS equation results. Also, it was applied to the case of a TET trimer model. This ML approach found readily the trimer resonances that were very difficult to obtain differently. The implementation of ML in this semiclassical regime shows that its application can be very beneficial. The next challenge is that of addressing the fully quantum TET regime; this is the aim of the present article. When the TET dimer is quantized with bosonic degrees of freedom a more general resonant condition, which involves the number of quanta, arises [13]. When at resonance, these bosons may transfer collectively from the first site to the second in a way similar to the semiclassical TET, although with rates depending on the boson number and the energy difference. As in the semiclassical case, in the fully quantum TET, the resonant condition can be found analytically in the dimer model [13], but any extension to larger systems is prohibitive analytically and involves a high computational cost. We show in the present work that the implementation of ML methods can help in overcoming these difficulties and be able to obtain readily the required resonant transfer properties. The structure of the present article is thus the following. In the next section II, we introduce a general form of the DNLS equation that upon quantization leads to fully quantum TET models. Once we define clearly the problem in the quantum case, we discuss in the following section III the ML technique we use. More specifically, we discuss the choice of the Loss Function (LF) that enables the analysis of the resonant transfer both in the dimer and also more generally to arbitrary chains. In section IV we detail the optimization method, and how the quantum resonant paths are found for different boson numbers. Section V is the central section of the article; we not only recover the exact dimer results but also apply the method to a trimer configuration. This gives not only the new result of the resonant paths but also shows that the method is fully able to investigate in detail the specifics of the resonant transfer. Finally, in section VI we conclude and summarize the findings of the work and comment on possible extensions. ## II From the semiclassical to the fully quantum TET It is known that the semiclassical DNLS equation can be derived from a classical Hamiltonian through the use of Hamilton’s equations [5]. This classical DNLS Hamiltonian is ${}H=\sum_{k=1}^{f}\omega_{k}|\psi_{k}|^{2}+\frac{1}{2}\chi_{k}|\psi_{k}|^{4}-\lambda\sum_{k=1}^{f-1}(\psi^{*}_{k}\psi_{k+1}+\psi^{*}_{k+1}\psi_{k}),$ (1) where $\omega_{k}$ and $\chi_{k}$ denote the frequency and nonlinearity parameter of the oscillator at site $k$ respectively. Also, ($\psi^{*}_{k},i\psi_{k})$ is a pair of conjugate variables, and the parameter $\lambda$ is the coupling among neighboring sites. For simplicity, we assume that it is the same between every pair of adjacent oscillators. In order to move to the quantum mechanical case we need to focus on the Bose-Hubbard operator [14]. The DNLS model for the quantum domain may be seen to arise also from the Bose-Hubbard model by using the time-dependent variational principle [15]. A simple way to quantize the Hamiltonian of Eq. (1) is by substituting $\psi^{*}_{k},\psi_{k}$ with the creation and annihilation operators $a^{\dagger}_{k},a_{k}$ respectively, as expressed in the second quantization formalism [16]. These operators obey to the commutation relations $[a_{k},a^{\dagger}_{m}]=\delta_{km}$, and $[a_{k},a_{m}]=0$, where $\delta_{km}$ is the Kronecker delta. Therefore, the Hamiltonian operator in the quantum case becomes ${}\hat{H}=\sum_{k=1}^{f}\omega_{k}\hat{N}_{k}+\frac{1}{2}\chi_{k}\hat{N}_{k}^{2}-\lambda\sum_{k=1}^{f-1}(\hat{a}^{\dagger}_{k}\hat{a}_{k+1}+\hat{a}^{\dagger}_{k+1}\hat{a}_{k}),$ (2) where $\hat{N}_{k}=\hat{a}^{\dagger}_{k}\hat{a}_{k}$ is the boson number operator for the site $k$. The dimension of the Hilbert space $\mathcal{H}_{N}$ for this problem is finite. Each state corresponds to an allowed configuration of $N$ indistinguishable bosons occupying $f$ distinguishable sites-nonlinear oscillators, with repetitions. Thus the dimension of the Hilbert space is ${}\mathcal{D}=\frac{(N+f-1)!}{N!(f-1)!}.$ A basis associated with that problem is the one composed of the so-called Fock states $\ket{n}\equiv\ket{n_{1},n_{2},\dots,n_{f}}$, where $n_{1},n_{2},\dots,n_{f}$ is the number of bosons at each respective site $1,2,\dots,f$ at the state indexed as n. We discuss the procedure of labeling the Fock states in section III. The occupation numbers $\\{n_{i}\\}$ are restricted to $\sum_{i}n_{i}=N$, while $n_{i}=0,1,\dots,N$. Additionally, the Fock states are orthonormal, meaning $\bra{n}\ket{m}=\delta_{n_{1}m_{1}}\dots\delta_{n_{f}m_{f}}.$ Continuing, the actions of the operators $\hat{a}_{k},\hat{a}^{\dagger}_{k},\hat{N}_{k}$ on each component $\ket{n}$ of the basis are described by $\hat{a}_{k}\ket{\dots,n_{k},\dots}=\sqrt{n_{k}}\ket{\dots,n_{k}-1,\dots},$ (3a) $\hat{a}^{\dagger}_{k}\ket{\dots,n_{k},\dots}=\sqrt{n_{k}+1}\ket{\dots,n_{k}+1,\dots},$ (3b) $\hat{N}_{k}\ket{\dots,n_{k},\dots}=n_{k}\ket{\dots,n_{k},\dots}.$ (3c) ## III Determination of an appropriate Loss Function We now focus on the choice of an appropriate loss function. We assume that the donor site has the lowest energy with nonlinearity parameter $\chi_{D}$ while the highest energy site is the acceptor site with nonlinearity parameter $\chi_{A}$. We further assume that all the bosons are placed initially (at time $t=0$) to the donor, and investigate the TET configurations that allow the complete transfer of these bosons to the acceptor site. To achieve this, we employ an algorithm relying on the same principle as the ones presented in [17, 11], where an optimization algorithm is used to minimize a quantity that is defined as the LF. The LF is usually associated with some physical parameters and thus, the problem becomes one where the algorithm, has to tune the parameters. The first step towards defining the LF is to construct a numerical scheme for calculating the matrix elements of the Hamiltonian in Eq. (2). While this is usually a simple process, there are special intricacies, with the main one being the labeling of the Fock states. We manage to overcome this obstacle by implementing a similar technique to [18], where the authors rank the states in lexicographic order and assign indices $1,2,\dots,\mathcal{D}$ to each configuration of bosons among the sites. For instance, assuming $N=2$ and $f=3$, state 1 corresponds to $\ket{1}=\ket{2,0,0}$, state 2 corresponds to $\ket{2}=\ket{1,1,0}$ and so on, assigning every state-configuration to a distinct index. Under this indexing policy, it is now straightforward to compute the elements $\hat{H}_{ij}$ similarly to [13], considering that ${}\hat{H}_{ij}=\bra{i}\hat{H}\ket{j}\overset{(\ref{bosehubbard_hamiltonian})}{=}T_{1}+T_{2},$ (4a) $T_{1}\equiv\bra{i}\sum_{k=1}^{f}\left[\omega_{k}\hat{N_{k}}+\frac{1}{2}\chi_{k}(\hat{N_{k}})^{2}\right]\ket{j},$ (4b) $T_{2}\equiv-\lambda\bra{i}\sum_{k=1}^{f-1}\left[\hat{a}^{\dagger}_{k}\hat{a}_{k+1}+\hat{a}^{\dagger}_{k+1}\hat{a}_{k}\right]\ket{j}.$ (4c) Calculating $T_{1}$ is simple since it involves the simple action of the boson number operator on state $\ket{j}$, as shown in Eq. (3c). That said, Eq. (4b) is equivalent to $\displaystyle T_{1}=\sum_{k=1}^{f}\left[\omega_{k}j_{k}+\frac{1}{2}\chi_{k}(j_{k})^{2}\right]\delta_{ij},$ where $j_{k}$ stands for the number of bosons on site $k$ for the state $\ket{j}$. Evaluating the second term is nontrivial because it involves the action of the operators $\hat{a}^{\dagger}_{k}\hat{a}_{k+1}$, $\hat{a}^{\dagger}_{k+1}\hat{a}_{k}$. The consecutive action of these operators can be explored by referring to Eq. (3a) and Eq. (3b). The operators $\hat{a}^{\dagger}_{k}$, $\hat{a}_{k}$ create and annihilate a boson at a given site $k$ respectively. However, when a pair of operators like $\hat{a}^{\dagger}_{k}\hat{a}_{k+1}$ (or $\hat{a}^{\dagger}_{k+1}\hat{a}_{k}$) acts on a Fock state $\ket{j}$, it creates a boson at the site $k$ (or $k+1$) but also destroys a boson at the site $k+1$ (or $k$). Thus, their action on a Fock state conserves the total number of bosons and the resulting state is going to be, up to a constant, another state in $\mathcal{H}_{N}$. Specifically, $\bra{i}\hat{a}^{\dagger}_{k+1}\hat{a}_{k}\ket{j}=\sqrt{j_{k}(j_{k+1}+1)}\delta_{ip}\equiv C_{k}^{(p)}\delta_{ip},$ (5) $\bra{i}a_{k}^{\dagger}a_{k+1}\ket{j}=\sqrt{j_{k+1}(j_{k}+1)}\delta_{im}\equiv D_{k}^{(m)}\delta_{im},$ where the two new states $\ket{p},\ket{m}$ are ${}\ket{p}=\ket{j_{1},\dots,j_{k}-1,j_{k+1}+1\dots,j_{f}},$ (6) ${}\ket{m}=\ket{j_{1},\dots,j_{k}+1,j_{k+1}-1\dots,j_{f}}.$ Combining the above yields $\displaystyle{}T_{2}=-\lambda\sum_{k=1}^{f-1}\ \left[C_{k}^{(p)}\delta_{ip}+D_{k}^{(m)}\delta_{im}\ \right].$ Figure 1: Time evolution of the expectation value of the number of bosons for the two sites of the dimer under the parameters ($\chi_{A}$, $\chi_{D}$, $\omega_{A}$, $\omega_{D}$, $\lambda$, $N$, maxt) = (-2,2, 3, -3, 0.1, 3, 25). The blue line denotes the donor’s expectation value, while the red one the acceptor’s. The matrix representation of the Hamiltonian can be produced and subsequently, the eigenstates and eigenvalues can be calculated. Everything is developed in Python, using the Tensorflow [19] library. The initial distribution of bosons $\ket{\Psi(0)}$ can be expanded to the basis of the eigenstates $\ket{\psi_{i}}$ $\ket{\Psi(0)}=\sum_{i=1}^{\mathcal{D}}C_{i}\ket{\psi_{i}},\ \ C_{i}=\bra{\psi_{i}}\ket{\Psi(0)}.$ Also, these eigenstates can be expanded to the basis of the Fock states $\ket{\psi_{i}}=\sum_{j=1}^{\mathcal{D}}b_{j,i}\ket{j},\ \ b_{j,i}=\bra{j}\ket{\psi_{i}}.$ We can now express the time evolution of the initial distribution $\ket{\Psi(0)}$ by applying the time evolution operator $\hat{U}(t)=e^{-\textbf{i}\hat{H}t}:$ $\ket{\Psi(t)}=e^{-\textbf{i}\hat{H}t}\ket{\Psi(0)}=\sum_{i,j}^{\mathcal{D}}C_{i}b_{j,i}e^{-\textbf{i}E_{i}t}\ket{j},$ (7) where $E_{i}$ is the $i$-th eigenvalue and i is the imaginary unit. Similarly, the time evolution of the average number of bosons at the site k is given by ${}\langle{}\hat{N}_{k}(t)\rangle{}\ \ =\ \ \bra{\Psi(t)}\hat{N_{k}}\ket{\Psi(t)}.$ (8) Combining Eq. (7) and Eq. (8), $\langle\hat{N}_{k}\rangle$ can be assessed in the following way $\displaystyle\langle\hat{N}_{k}(t)\rangle=\sum_{i,j,n}^{\mathcal{D}}j_{k}C_{n}^{*}C_{i}b^{*}_{j,n}b_{j,i}e^{\textbf{i}(E_{n}-E_{i})t}.$ (9) We use Eq. (9) in order to compute the LF for the quantum TET problem. Specifically, we time-evolve Eq. (9) for the acceptor energy level until a pre-defined time `maxt`. During this period, the oscillator of the $f^{th}$ site (acceptor) has completed a few oscillations. The next step in computing the LF is to extract the maximum value from that time evolution. Concluding, the LF is defined as ${}LF=N-max\\{\langle\hat{N}_{f}(t)\rangle\\}=N-max\\{\langle\hat{N}_{A}(t)\rangle\\}.$ (10) In Fig. 1 one can observe the characteristic oscillatory behavior, in the dimer system, when the complete transfer occurs. In the present work, we fix the frequency of the oscillators and optimize for the nonlinearity parameters of the oscillators. While the LF might appear not to have any explicit connection to the nonlinearity parameters we want to optimize for, we can use Tensorflow’s $GradientTape$ to compute the derivatives with respect to these parameters. In that way, we keep track of gradient information in terms of the trainable variables throughout the whole process described in this section; from creating the Hamiltonian to calculating $\langle\hat{N}_{k}(t)\rangle$. Using this information, an optimizer like $ADAM$ [20] can now update the parameters accordingly, so that the LF is minimized, signifying complete transfer. It is important to note that, while many other optimization algorithms (simulated annealing [21], particle swarm [22], differential evolution [23, 24]) were tried on this problem, we were not able to produce adequate results with none of them. The parameter `maxt` is of major importance, because of the oscillation of the bosons between the sites. If it is not set large enough, TET can be missed since the system would not have time to complete an oscillation, while it also has to be small enough, so that precious computational time is saved. Thus, it has to be large enough to obtain at least one complete oscillation, producing this way essential information. It is also observed that the period of each oscillator is proportional to $\lambda^{-1}$, so as one would expect, changes to these parameters should be made concurrently. ## IV Optimization of the Loss Function Once we define the LF we may now proceed with its minimization through the procedure outlined in Fig. 2. We emphasize the first step of this process graph, i.e. a proper update method, which is essential for the present work. We noticed by making several computational runs that optimizers did not work very well when initialized with a random set of parameters; this is due to the extreme selectivity of the TET resonant condition. Specifically, the parameter space that the LF maps to, exhibits slowly varying gradients everywhere, except for areas close to the optimal parameters for transfer, where relatively large gradients are present. Figure 2: Schematic representation of the parameter optimization procedure. Our approach for bypassing this problem relies on exploring a wide range of the parameter space at the same time. To be more specific, the optimization procedure is comprised of two distinct phases. First, many test optimizers are assigned some initial guesses and left to run simultaneously, as different processes, until a stopping condition is reached. Then we gather the LF and the parameters that each test optimizer yielded at the last iteration. For the second part of the optimization procedure, a main optimizer is employed. Its initial guesses are the parameters of the test optimizer with the best performance (i.e smallest LF). The purpose of this additional step is to further minimize the LF, if possible. Two methods were developed for defining the initial guesses of the test optimizers, given that we choose the range of the parameter space they will investigate. In the first approach, which is represented by Fig. 3a, we define a grid of points in the parameter space and use the lattice points as initial guesses for the test optimizers. In the second method, we split the parameter space into a number of regions and chose random initial combinations of parameters from each region. On the one hand, the former method has a high computational cost, since it requires a large number of optimizers running at the same time, but has a greater potential to derive the optimal parameters. On the other hand, the latter is fast but less robust. It is useful in cases of stronger coupling, as the gradients of the parameter space smooth out. If the first run of any method does not produce favorable results we have the option of redefining the limits of the parameter space around the best parameters provided by the test optimizers. Nevertheless, choosing one of the above methodologies relies on the problem at hand. In our case, both of them produce accurate results. Moreover, we need to disambiguate that the main optimizer is an optional step, that aims to corroborate TET, by producing a LF lower than the one deduced from the test optimizers. Figure 3: Defining the initial guesses of the test optimizers. (a) An example using the grid method, where we define 400 different initial guesses for the same number of test optimizers. (b) Representative example of the second method, where splitting of the parameter space into regions is employed. The black lines represent the boundaries among the four different regions from which 16 different guesses are sampled. In both graphs, the test optimizers that were able to derive a $\text{LF}<1$ are displayed with normal opacity and their trajectories, while the rest are faded. Both figures refer to a dimer system with ($N$, $\omega_{A}$, $\omega_{D}$, maxt) = (3, 3, -3, 25). For the purposes of explaining the difference between the two methods, the system presented in the left figure has a coupling parameter of $\lambda=0.1$, while for the system on the right figure $\lambda=1$. Figure 4: Optimization Results. (a) The trajectories and initial guesses of our algorithm, using the grid method, for $N=3$ bosons. The optimal parameters found -by the main optimizer- in this example are ($\chi_{A}$, $\chi_{D}$) = (-1.99, 2). Only the initial guesses of the test optimizers that resulted in a LF smaller than 0.5 are displayed. (b) Parameters for TET from the quantum limit to the semiclassical limit, as deduced from the grid method and the main optimizer, in the dimer system. (c) Comparison of the predicted nonlinearity parameters $\chi_{D}=-\chi_{A}$, illustrated in the figure as red dots, with the reference value derived from Eq. (11) represented in the figure as a dashed line. The system’s constant parameters are ($\omega_{A}$, $\omega_{D}$, $\lambda$, maxt) = (3, -3, 0.1, 25). Regardless of whether we use the grid or the splitting into regions method, the trainable parameters are updated in the same way for both the test optimizers and the main optimizer. Given the initial guesses of $\\{\chi_{k}\\}$, $k=1,2,\dots,f$, the nonlinearity parameters after the mth iteration are updated as following: $\chi_{j}^{(m+1)}=\chi_{j}^{(m)}-\alpha\vec{\nabla}_{j}(LF),$ where the gradient is computed with Tensorflow’s $GradientTape$, as discussed in section III. The learning rate $\alpha$ is a positive real number that defines the rate of change at each iteration while moving towards the minimum of the LF. With the new set of nonlinearity parameters, the optimization procedure moves to the next iteration and it will be interrupted either due to slow convergence or because of reaching maximum iterations. The threshold parameter of Fig. 2 has a dual role. On one hand, it checks whether the nonlinearity parameters of the current iteration yield a LF close to zero, signifying TET. On the other hand, it determines the slow convergence of the algorithm, and its role is to pause the optimization procedure when there is no significant improvement in minimizing the LF. Thus, its value should be small enough (close to zero) to manifest TET, but still nonzero because otherwise, the stochastic optimization procedure we introduce will lead to an infinite loop. The latter problem is also resolved by terminating the optimization procedure after reaching a predefined maximum number of iterations. The $Python$ code implementing this procedure is located in our GitHub repository [25]. ## V Results ### V.1 TET Quantum Dimer In the preceding section, we describe the optimization technique, including the possible alternatives. We may now apply this scheme to the dimer realm and test our method from the semiclassical limit to the fully quantum one, using the grid method described earlier. As we discussed in section I, the optimal parameters for this case are already known for both the semiclassical [4] and the quantum regime [13]: ${}\chi_{D}=-\chi_{A}=\frac{\omega_{A}-\omega_{D}}{N}.$ (11) Our method is successful in obtaining the TET configurations. We fix the frequencies of the donor and acceptor to $\omega_{D}=-3$ and $\omega_{A}=3$ respectively, the coupling parameter to $\lambda=0.1$, while the time evolution of Eq. (9) for the acceptor is performed until `maxt` = 25. The Fig. 4a displays the outcome of the optimization procedure for $N=3$ bosons, with the optimal nonlinearity parameters being: $(\chi_{D},\chi_{A})=(2,-1.99).$ The illustrated test optimizers produce a LF lower than the threshold mentioned in Fig. 2, which is set arbitrarily (in this case 0.5). The results validate the sensitivity of each optimizer to the initial guesses. We keep the same values for $\omega_{A},\omega_{D},\lambda$, and apply the grid method for a variety total of bosons in the dimer system. The outcome of the optimization scheme is shown in Fig. 4b, while we compare our results with Eq. (11) in Fig. 4c. In every case, the proposed method succeeds in identifying the TET paths. It is worth mentioning that we can deduce the same results with the method of splitting the parameter space. Moreover, our analysis proves that the fully quantum case of $N=1$ boson is of special interest since TET occurs for a whole set of nonlinearity parameters $\chi_{A},\chi_{D}$ instead of a single, very limited configuration. Specifically, we identify this set as ${}\chi_{D}=\chi_{A}+2(\omega_{A}-\omega_{D}).$ (12) This result is in agreement with previous analytical calculations. To be more specific, Maniadis et al. in [13] prove that the condition for having TET is for the detuning function to vanish. The latter is defined as the variation of the energy of the oscillators during a transfer: $\epsilon=[H_{D}(N)+H_{A}(0)]-[H_{D}(i)+H_{A}(N-i)].$ (13) In Eq. (13) $i=0,1,\dots,N$ while $H_{D}$, $H_{A}$ are the donor and acceptor parts of the Hamiltonian in Eq. (2). In this case, any nonzero $\lambda$ can raise the degeneracy of the system and complete TET occurs in the limiting of zero coupling. We can easily prove that the detuning function of Eq. (13) for one boson vanishes under the parameters of Eq. (12). For all the cases where the detuning function vanishes, the Hamiltonian becomes quadratic of the bosons operators: ${}\hat{H}=\hat{H}_{D}(N)-\lambda(a^{\dagger}_{D}a_{A}+a^{\dagger}_{A}a_{D}).$ (14) Figure 5: Time evolution of the average number of bosons for the three sites of the trimer system. (a) The time evolution of the expectation values of the boson number operators for nonlinearity parameters that produce near complete TET where ($\chi_{A}$, $\chi_{M}$, $\chi_{D}$, $\omega_{D}$, $\omega_{M}$, $\omega_{A}$, $\lambda$, $N$, maxt) = (-1.5, -38.39, 1.5, -3, -3, 3, 1, 4, 40). (b) The time evolution of the expectation values of the boson number operators for a system with parameters that don’t produce complete TET, where ($\chi_{A}$, $\chi_{M}$, $\chi_{D}$, $\omega_{D}$, $\omega_{M}$, $\omega_{A}$, $\lambda$, $N$, maxt) = (-1.5, 1.5, 1.5, -3, -3, 3, 1, 4, 40). For both figures, the blue, yellow, and red lines refer to the expectation value of the boson number operator for the donor, middle site, and acceptor respectively. ### V.2 TET Quantum Trimer The trimer system has been investigated in [26, 11] in the context of a single electron or boson. We aim to expand this investigation to the arbitrary case of $N$ bosons. While the method is able to optimize the nonlinearity parameters of all three sites, it proved computationally consuming. To circumvent this, we set the acceptor and donor sites to their dimer values, and optimize the parameter of the middle layer (labeled as “M”). Similarly to the dimer case, we can set an arbitrary threshold for the LF (LF$<$ 0.2) and begin the optimization process. We observe that for TET to occur in the quantum realm, the middle layer had to exhibit extremely high nonlinearity. One of the trimer systems examined in this section has the following properties: It is strongly coupled, $\lambda=1$, with frequencies $\omega_{D}=3,\;\omega_{M}=-3,\;\omega_{A}=-3$ and a maximum number of bosons $N=4$ initially at the donor site. The nonlinearity parameter required to have TET, in this case, is $\chi_{M}\approx 38.39$, much higher than that of Eq. (11), compared to the parameters of the donor and acceptor sites $\chi_{D}=-\chi_{A}=1.5$. It is important to mention that the opposite value $\chi_{M}\approx-38.39$ will produce a similar LF value, but still small enough to exceed the threshold and stop the iterative process. We carry out the same procedure for a variety of system parameters, as seen in Fig. 7. Figure 6: Lowest value of the LF in a trimer system with the parameters displayed on the figure. The other system parameters are ($\omega_{A}$, $\omega_{D}$, $\lambda$, maxt) = ($\omega_{M}$-1, -$\omega_{M}$+1, 1, 40). Figure 7: Trimer Results. Absolute values of the nonlinearity parameter of the middle layer with respect to (a) different values of the maximum number of bosons in the system, (b) the frequency of the middle layer $\omega_{M}$. In both cases we use the grid method while the system’s constant parameters for both figures are ($\omega_{A}$, $\omega_{D}$, $\lambda$, maxt) = ($\omega_{M}$-1, -$\omega_{M}$+1, 1, 40). . We display the absolute value of $\chi_{M}$ as both positive and negative values produce the desired result. We observe that the relation between $\chi_{M}$ and the frequencies or the number of bosons of the system seems to differ from the dimer case, as there appears to be a greater correlation with the former rather than the latter. In the graph some points seem to differ a lot from others, a fact that is attributed to an optimization process that was not able to minimize the loss function enough before being terminated by the rules we defined earlier in section IV. We also have plotted the minimum value of the LF that we observed in Fig. 6. The time evolution of the boson number operators for this system is shown in Fig. 5. As we can see in subplot (a) the expectation value of the boson number operator for the middle layer, in the optimal case, appears to be zero for all time steps which indicates that bosons do not stay on this site for any significant amount of time, or at all. However, this is not observed in cases where the nonlinearity parameters are set to nonoptimal values, as seen in subplot (b), where the expectation value is nonzero. It is apparent that the oscillation frequency of the number operator is larger in the non-resonant system and that complete TET cannot be achieved. ## VI Discussion In this work, we introduced a ML method in the context of a quantum many-body system and showed that its efficient implementation can produce results that are very hard to obtain with more conventional methods. We focused on the quantized version of the DNLS equation with arbitrary local energies and nonlinearities and addressed the question of optimal transfer in between different sites for the dimer and trimer cases. Since the fully quantum transfer dimer case is known analytically we compared our method with these results and showed perfect agreement. This successful comparison between analytics and ML methods shows that the latter can be used confidently in more complex cases where results are not known. Subsequently, we applied the method to the trimer case that cannot be solved analytically. Our method enabled a detailed search showing the specifics of the resonant transfer, the different parameter regimes as well as the transfer efficiencies. In terms of physics, we found in the trimer system that in the nonresonant transfer regime from donor to acceptor sites the intermediate state retains some of the probability. In the resonant case, on the other hand, the intermediate site is essentially not populated. This shows that this site acts as some form of a barrier between the donor and acceptor sites that can be completely bypassed in the fully resonant regime. It is noteworthy to point out that the bosons move in unison over to the acceptor site showing a very interesting collective behavior in the transfer. The collective boson transfer can be investigated also in more general chains with a larger number of sites. The computational challenge is now larger since the dimensionality of the system becomes large and the calculation of the Hamiltonian and the evolution of $\langle\hat{N}_{A}(t)\rangle$ slows down. In this regime, one needs to explore other loss functions that could improve scalability and/or implementation of meta-learning methods described in [27]. In this work, we tested also alternative optimizers such as ones with momentum. We found that they were more efficient in finding the resonant transfer parameter regime but were highly dependent on hyper-parameters that needed to be also optimized. Finally, it is possible that variants of the gradient descent algorithm might help in reducing the computational cost[28]. The phenomenon of the collective transfer of bosons in the trimer case opens up very interesting new questions on the interplay of nonlinearity and disorder in the fully quantum regime for more extended systems. The ML method provided in the present work can be readily generalized to this case and be utilized to investigate this very exciting problem with applications in condensed matter physics as well as quantum optics. ## ACKNOWLEDGMENTS We acknowledge the cofinancing of this research by the European Union and Greek national funds through the Operational Program Crete 2020-2024, under the call ”Partnerships of Companies with Institutions for Research and Transfer of Knowledge in the Thematic Priorities of RIS3Crete”, with project title “Analyzing urban dynamics through monitoring the city magnetic environment” (project KPHP1 - 0029067). ## AUTHOR CONTRIBUTIONS I. Andronis and G. Arapantonis contributed equally to this work. ## References * Ott [2002] E. Ott, _Chaos in Dynamical Systems_, 2nd ed. (Cambridge University Press, 2002). * Akulin [2014] V. M. Akulin, _Dynamics of Complex Quantum Systems_, 2nd ed. (Springer Dordrecht, 2014). * Neofotistos _et al._ [2019] G. Neofotistos, M. Mattheakis, G. D. Barmparis, J. Hizanidis, G. P. Tsironis, and E. Kaxiras, Frontiers in Physics 7, 10.3389/fphy.2019.00024 (2019). * Kopidakis _et al._ [2001] G. Kopidakis, S. Aubry, and G. P. Tsironis, Phys. Rev. Lett. 87, 165501 (2001). * Hennig and Tsironis [1999] D. Hennig and G. Tsironis, Physics Reports 307, 333 (1999). * Eilbeck _et al._ [1985] J. Eilbeck, P. Lomdahl, and A. Scott, Physica D: Nonlinear Phenomena 16, 318 (1985). * Kenkre [2022] V. M. N. Kenkre, _Interplay of Quantum Mechanics and Nonlinearity_, 1st ed. (Springer Cham, 2022). * Kenkre and Campbell [1986] V. M. Kenkre and D. K. Campbell, Phys. Rev. B 34, 4959 (1986). * Tsironis and Kenkre [1988] G. Tsironis and V. Kenkre, Physics Letters A 127, 209 (1988). * Tsironis [1993] G. Tsironis, Physics Letters A 173, 381 (1993). * Barmparis and Tsironis [2021] G. D. Barmparis and G. P. Tsironis, JOSA B, Vol. 38, Issue 9, pp. C120-C126 38, C120 (2021). * Tsironis _et al._ [2022] G. P. Tsironis, G. D. Barmparis, and D. K. Campbell, International Journal of Modern Physics B 36, 2240001 (2022), https://doi.org/10.1142/S021797922240001X . * Maniadis _et al._ [2004] P. Maniadis, G. Kopidakis, and S. Aubry, Physica D: Nonlinear Phenomena 188, 153 (2004). * Gersch and Knollman [1963] H. A. Gersch and G. C. Knollman, Physical Review 129, 959 (1963). * Amico and Penna [1998] L. Amico and V. Penna, Phys. Rev. Lett. 80, 2189 (1998). * Pearsall [2020] T. P. Pearsall, _Quantum Photonics_, 2nd ed. (Springer International Publishing, 2020). * Karniadakis _et al._ [2021] G. E. Karniadakis, I. G. Kevrekidis, L. Lu, P. Perdikaris, S. Wang, and L. Yang, Nature Reviews Physics 3, 422 (2021). * Zhang and Dong [2010] J. M. Zhang and R. X. Dong, European Journal of Physics 31, 591 (2010). * Abadi _et al._ [2015] M. Abadi _et al._ , TensorFlow: Large-scale machine learning on heterogeneous systems (2015), software available from tensorflow.org. * Kingma and Ba [2014] D. P. Kingma and J. L. Ba, 3rd International Conference on Learning Representations, ICLR 2015 - Conference Track Proceedings 10.48550/arxiv.1412.6980 (2014). * Bertsimas and Tsitsiklis [1993] D. Bertsimas and J. Tsitsiklis, Statistical Science 8, 10 (1993). * Kennedy and Eberhart [1995] J. Kennedy and R. Eberhart, in _Proceedings of ICNN’95 - International Conference on Neural Networks_, Vol. 4 (1995) pp. 1942–1948 vol.4. * Brest _et al._ [2006] J. Brest, S. Greiner, B. Boskovic, M. Mernik, and V. Zumer, IEEE transactions on evolutionary computation 10, 646 (2006). * Elsayed _et al._ [2011] S. M. Elsayed, R. A. Sarker, and D. L. Essam, in _2011 IEEE Congress of Evolutionary Computation (CEC)_ (IEEE, 2011) pp. 1041–1048. * Andronis and Arapantonis [2022] J. Andronis and G. Arapantonis, Quantum-targeted-energy-transfer, https://github.com/JAndronis/Quantum-Targeted-Energy-Transfer (2022). * S.Aubry and G.Kopidakis [2003] S.Aubry and G.Kopidakis, A non adiabatic theory for ultrafast and catalytic transfer of electrons at low temperature, in _Localization and Energy Transfer in Nonlinear Systems_ (WORLD SCIENTIFIC, 2003) pp. 1–27. * Li and Malik [2016] K. Li and J. Malik, Learning to optimize (2016), arXiv:1606.01885v1 . * Ruder [2016] S. Ruder, An overview of gradient descent optimization algorithms (2016), arXiv:1609.04747 .
# Deep Kernel Learning for Mortality Prediction in the Face of Temporal Shift ††thanks: Citation: Rios, M., Abu-Hanna, A. (2021). Deep Kernel Learning for Mortality Prediction in the Face of Temporal Shift. In: Artificial Intelligence in Medicine. AIME 2021. Lecture Notes in Computer Science, vol 12721. Springer, Cham. https://doi.org/10.1007/978-3-030-77211-6_22 Miguel Rios Centre for Translation Studies University of Vienna <EMAIL_ADDRESS> &Ameen Abu-Hanna Department of Medical Informatics, Amsterdam UMC University of Amsterdam <EMAIL_ADDRESS> ###### Abstract Neural models, with their ability to provide novel representations, have shown promising results in prediction tasks in healthcare. However, patient demographics, medical technology, and quality of care change over time. This often leads to drop in the performance of neural models for prospective patients, especially in terms of their calibration. The deep kernel learning (DKL) framework may be robust to such changes as it combines neural models with Gaussian processes, which are aware of prediction uncertainty. Our hypothesis is that out-of-distribution test points will result in probabilities closer to the global mean and hence prevent overconfident predictions. This in turn, we hypothesise, will result in better calibration on prospective data. This paper investigates DKL’s behaviour when facing a temporal shift, which was naturally introduced when an information system that feeds a cohort database was changed. We compare DKL’s performance to that of a neural baseline based on recurrent neural networks. We show that DKL indeed produced superior calibrated predictions. We also confirm that the DKL’s predictions were indeed less sharp. In addition, DKL’s discrimination ability was even improved: its AUC was 0.746 $(\pm$0.014 std), compared to 0.739 ($\pm$ 0.028 std) for the baseline. The paper demonstrated the importance of including uncertainty in neural computing, especially for their prospective use. _K_ eywords Deep Kernel Learning, temporal shift, time series, calibration, Gaussian process, mortality prediction. ## 1 Introduction In the ICU, the prediction of in-hospital mortality is the task of providing probabilities for Intensive Care patients to die in the hospital, either in the ICU or after discharge to another ward. The (early) detection of such patients is relevant for clinical decision making. Mortality prediction models (MPMs) are often trained with large collections of electronic health records (EHR) that contain structured patient information such as demographics and physiological variables. MPMs based on deep learning are becoming prevalent in medical applications [18]. One reason for this is that NNs automatically derive representations for time series data, which may provide predictive ability superior to that of standard regression models [21, 9]. Specifically, neural models learn features from the input data by the incremental composition of simpler layers, resulting in complex representations for non- linear prediction models [1]. However, patient characteristics, medical technology, and clinical guidelines change over time, thus forming a challenge for the validity of MPMs for prospective patients, as these models were learned on historical data [14]. In particular, due to their flexibility, NNs have the ability to leverage on slight patterns in the data, but such patterns may not be stable over time and hence NN models may be sensitive to such temporal shifts causing a change (usually a drop) in performance [16]. For prediction models of a binary outcome, not only the discriminatory capability of the model may suffer, but especially its (mis)calibration. Calibration refers to the correspondence between the predicted probabilities and the true probabilities. The true probabilities are estimated on the test set by some measure of averaging the number of events for a set of patients. Performance drift has consequences for the task at hand, and the detrimental effects on benchmarking ICUs have been demonstrated [14]. One way to tackle this problem is to augment NNs with the notion of uncertainty: whenever the data distribution changes due to shift, the predictions should be more uncertain [13]. In contrast to NNs, The Gaussian process (GP) is a probabilistic framework for time series modelling that is able to increase model capacity with the amount of available data, and to produce uncertainty estimates. A GP characterises a distribution over possible functions that fit the input data. It is defined by a Gaussian function with a certain mean and, more importantly, a kernel function that captures the correlations between any two observations. The kernel encompasses the notion of uncertainty by performing a pairwise computation among all input data using some notion of similarity between the observations. The kernel can be viewed as providing a probability distribution over all possible models fitting the data. The prediction models based on GPs successfully model time series data, incorporate confidence regions to predictions, and offer interpretability of the variables with the kernel function [20]. Moreover, the GP framework has been used to develop clinical prediction models [5, 2]. In particular, Ghassemi et al. [8] use a multitask GP to model time series with physiological variables and clinical notes for mortality prediction. Directly relevant to our paper is the proposition in [23] to combine both NNs and GPs on a common framework of deep kernel learning (DKL). DKL leverages inductive biases from the NNs and from the non-parametric GPs. In this paper, we investigate the behaviour of mortality prediction models based on DKL. In particular, we are interested in inspecting the robustness of the DKL model to a temporal shift. We also compare it to a strong NN-based baseline. Our hypothesis is that incorporation of uncertainty improves predictions. More specifically, we expect the DKL, when faced with uncertainty in the test set, to provide less extreme predictions that are closer to the global mean rather than providing overconfident predictions. In turn, the resultant prediction set would be less sharp than for the baseline model. Sharpness, which is also referred to refinement in weather forecast [15] measures the tendency of predictions to be close to 0 and 1. We therefore also compare the sharpness of both models but check that this does not come at the cost of discrimination. Finally, we also performed internal validation of the DKL model with all the population (i.e. no temporal shift) to understand whether the DKL’s behaviour is specific to temporal validation. Our main contribution in this paper is the introduction of a DKL model for in- hospital mortality prediction based on the first hours of an ICU stay in the context of temporal validation. The GP component in the DKL is shown to be robust to the shift in population and produces better calibrated predictions, without sacrificing discrimination. Our feature extraction is based on an open source benchmark [9] using the publicly available MIMIC-III [11] database. This facilitates the reproducibility of our results111Code is available at: https://github.com/mriosb08/dkl-temporal-shift.git. ## 2 Deep Kernel Learning The Gaussian Process [19] is a Bayesian non-parametric framework based on kernels for regression and classification. The set of functions that describes a given input data is possibly infinite and the GP assigns a probability to each one. For a dataset $\mathcal{X}=\left\\{\left(\mathbf{x}_{1},y_{1}\right),\left(\mathbf{x}_{2},y_{2}\right),\ldots,\left(\mathbf{x}_{n},y_{n}\right)\right\\}$ where $\mathbf{x}$ is an input vector and $y$ a corresponding output, we want to learn a function $f$ that is inferred from a GP prior: $\displaystyle f(\mathbf{x})\thicksim\operatorname{GP}(m(\mathbf{x}),k(\mathbf{x},\mathbf{x}^{\prime}))$ (1a) where $m(\mathbf{x})$ defines a mean (often set to 0) and $k(\mathbf{x},\mathbf{x}^{\prime})$ defines the covariance in the form of a kernel function. The kernel function models the covariance between all possible pairs $(\mathbf{x},\mathbf{x}^{\prime})$ and provides a measure of uncertainty. The choice of kernel determines properties of the function that we want to learn, usually this choice is based on background knowledge of the problem. Wilson et al. [22] propose kernels based on deep learning architectures for GP regression. The DKL employs a GP with a base kernel as the last hidden layer of a NN. In other words, the DKL is a pipeline for learning complex NN features, and a distribution over functions that fit our input data. The base kernel $k\left(\mathbf{x},\mathbf{x}^{\prime}\mid\theta\right)$ with hyperparameters $\theta$ is parameterized by a non-linear function. $\displaystyle k\left(\mathbf{x},\mathbf{x}^{\prime}\mid\theta\right)\rightarrow k\left(g\left(\mathbf{x},\omega\right),g\left(\mathbf{x}^{\prime},\omega\right)\mid\theta,\omega\right),$ (2a) where $g(\mathbf{x},\omega)$ is a NN architecture with weights $\omega$. In addition, the DKL jointly learns the NN weights and kernel hyperparameters under the GP probabilistic framework. Learning a GP involves computing the kernel function, and finding the best kernel hyperparameters. The DKL optimises both the kernel hyperparameters and the NN weights, by maximising the marginal likelihood. In Figure 1, we define the architecture for extracting features $g(\mathbf{x},\omega)$, $\mathbf{x}_{i}$ denotes the input vector in the ith element of $\mathcal{X}$. Figure 1: NN architecture $g(\mathbf{x},\omega)$ for extracting features $\mathbf{f}_{i}$ for the GP prediction layer. The input features are first projected with an affine layer ($\operatorname{linear}(.)$), then fed to a bidirectional LSTM ($\operatorname{birnn}(.)$) [10] for encoding time series. Next the result goes through an affine layer with a non-linearity ($\operatorname{ReLU}(.)$) that combines the hidden states of the bidirectional LSTM. Next the features $\mathbf{f}_{i}$ are summarised by averaging ($\operatorname{avg}(.)$) and then fed to the GP layer. ## 3 Experiments The Medical Information Mart for Intensive Care (MIMIC-III) database includes over 60,000 ICU stays across 40,000 critical care patients [11]. Harutyunyan et al. [9] propose a public benchmark and baselines based on MIMIC-III for modelling mortality, length of stay, physiologic decline, and phenotype classification. We use the benchmark for predicting in-hospital mortality based on the first 48 hours of an ICU stay. The cohort excludes all ICU stays with unknown length-of-stay, patients under 18, multiple ICU stays, stays less than 48 hours, and no observations during the first 48 hours. The in-hospital mortality class is defined by comparing the date of death against hospital admissions and discharge times with a resulting mortality rate of 13.23%. We use the benchmark to extract $17$ input physiological variables (i.e. features), that are a subset of the Physionet challenge 222https://physionet.org/content/challenge-2012/1.0.0/. The benchmark [9] code processes the time series data with imputation of missing values with the previous hour, and normalisation from MIMIC-III. The normalisation of the features is performed by subtracting the mean and dividing by the standard deviation. The features also provide a binary mask for each variable indicating which time-step is imputed. All categorical variables are encoded using one-hot vectors (e.g. Glasgow coma scales). The final feature vector is formed by the concatenation of the clinical variables and the one-hot vectors with a total of $76$ features. The clinical variables are shown in Table1. Variable --- Capillary refill rate Diastolic blood pressure Fraction inspired oxygen Glascow coma scale eye opening Glascow coma scale motor response Glascow coma scale total Glascow coma scale verbal response Glucose Heart Rate Height Mean blood pressure Oxygen saturation Respiratory rate Systolic blood pressure Temperature Weight pH Table 1: Clinical variables used in our experiments from MIMIC-III. We use the architecture $g(.)$ as the baseline defined as: BiLSTM, which is based on a bidirectional LSTM for feature representation, and a linear prediction layer. We implement the DKL model with GPyTorch [7], with the following components: the RBF kernel as the base kernel, feature extractor $g(.)$, and grid size $100$ which is the number of inducing points used to approximate the GP for faster computations. The computation of the posterior distribution in the GP is expensive and several methods have been proposed to accelerate it by approximating it with a function over a set of inducing points [17, 24]. In addition, we perform a simple ablation on the architecture by replacing the bidirectional LSTM with a LSTM for both models, baseline and DKL defined as: LSTM, and DKL-LSTM. We use the following hyperparameters: optimiser Adam [12], learning rate $1\mathrm{e}{-3}$, epochs 30, encoder size 16, hidden size 16, batch size 100, dropout 0.3 applied after the $\operatorname{linear}$ layer. We perform model selection with the validation dataset based on AUC-ROC. ### 3.1 Temporal shift: strategy and results The MIMIC-III dataset includes data using the CareVue electronic patient record (EPR) system from 2001 to 2008. From 2008 to 2012 the MetaVision system was used instead. In the first experiment for inspecting temporal shift, we split the datataset into the CareVue period for training with $9,646$ instances and $1,763$ for validation (for tuning the hyper-parameters), and the data in the MetaVision period with $7,689$ as the test set. We excluded patients present in both registries. This constitutes a temporal validation strategy in which the model is tested on data collected in the future relative to the data on which it has learned. This means that the model faces possible temporal shift due to changes that occur in time, and indeed possibly also due to the change of the EPR system that collects the data that could have affected the workflow and/or the way of registration. Performance was measured in terms of: Discrimination, by the AUC-ROC; the balance between the positive predicted value and sensitivity, by the AUC-PR; the accuracy of predictions by the Brier score; and calibration by calibration graphs and the Cox recalibration approach [3] in which the observed outcome in the test set is regressed using logistic regression on the log odds of the predictions. If the predictions were perfectly calibrated then the linear predictor of this model would have an intercept of 0 and a slope of 1. We test deviations from these ideal value of 0 and 1, respectively. To test our hypothesis whether the DKL approach provides more conservative predictions due to uncertainty for areas in the test set, we measure the (un)sharpness of the predictions. We use the following measure of unsharpness: $\frac{\sum_{1}^{N}{p_{i}(1-p_{i})}}{N}$ where $p_{i}$ is the $i$th prediction and N is number of observations. | Validation | Test ---|---|--- Model | AUC-ROC | AUC-PR | AUC-ROC | AUC-PR LSTM | $0.838\pm 0.003$ | $0.532\pm 0.006$ | $0.693\pm 0.027$ | $0.317\pm 0.037$ BiLSTM | $0.857\pm 0.002$ | $0.572\pm 0.007$ | $0.739\pm 0.028$ | $\textbf{0.386}\pm 0.018$ DKL-LSTM | $0.854\pm 0.002$ | $0.562\pm 0.010$ | $0.701\pm 0.033$ | $0.327\pm 0.026$ DKL | $0.856\pm 0.002$ | $0.569\pm 0.004$ | $\textbf{0.746}\pm 0.014$ | $0.373\pm 0.018$ Table 2: In-hospital mortality results with a temporal population shift over 10 runs $\pm$ one standard deviation. The training and validation datasets are on CareVue (2001-2008), and the test on MetaVision (2008-2012). (a) (b) Figure 2: Receiver operating characteristic curve (a) and calibration curve (b) for in-hospital mortality with temporal shift in population. Table 2 shows the AUC-ROC and AUC-PR results for in-hospital mortality with a temporal shift in population. The baseline outperforms the DKL model on the validation (tuning) dataset for both metrics. On the test dataset, however, the DKL shows competitive performance on the AUC-ROC. We use the best run from the validation based on the AUC-ROC for reporting the ROC and calibration curves. In addition, we select the best performing models from Table 2 based on AUC-ROC, namely BiLSMT and DKL, for comparing the calibration and ROC curves. The LSTM models consistently underperform compared to the bidirectional ones. Figure 2 shows the ROC and calibration curves for in- hospital mortality with a temporal shift. The Brier score for the DKL is 0.101 which is better that the 0.109 of the BiLSTM. The DKL outperforms the baseline and it shows better calibration. In the Cox re-calibration on both models the BiLSTM had a calibration intercept of 1.965 (1.88, 2.049), and slope of 0.538 (0.5, 0.577) compared to the DKL’s of 0.6615 (0.586, 0.734), 0.712 (0.652, 0.772). Although both models deviated significantly from the ideal values (of 0 and 1), the DKL showed significantly much better calibration. The DKL’s predictions were also much less sharp: unsharpness of 0.061 for DKL versus 0.025 for BiLSTM. ### 3.2 Experiment 2: Internal validation We report the results with all the sources (2001-2012) for in-hospital mortality, with no shift in population. The training, validation and test datasets consisted of respectively $14,681$, $3,222$, and $3,236$ instances. | Validation | Test ---|---|--- Model | AUC-ROC | AUC-PR | AUC-ROC | AUC-PR LSTM | $0.843\pm 0.003$ | $0.513\pm 0.006$ | $0.840\pm 0.005$ | $0.434\pm 0.008$ BiLSTM | $0.858\pm 0.004$ | $0.549\pm 0.010$ | $\textbf{0.851}\pm 0.004$ | $\textbf{0.478}\pm 0.016$ DKL-LSTM | $0.838\pm 0.002$ | $0.485\pm 0.014$ | $0.841\pm 0.003$ | $0.425\pm 0.013$ DKL | $0.854\pm 0.004$ | $0.536\pm 0.010$ | $0.847\pm 0.005$ | $0.454\pm 0.018$ Table 3: In-hospital mortality results over 10 runs $\pm$ one standard deviation. Validation and test dataset from all sources (2001-2012). (a) (b) Figure 3: Receiver operating characteristic curve (a) and calibration curve (b) for in-hospital mortality with all sources. Table 3 shows the AUC-ROC and AUC-PR results for in-hospital mortality with all sources (2002-2012). The baseline outperforms the DKL model on the test dataset for both metrics the AUC-ROC, and AUC-PR. Figure 3 shows the ROC and calibration curves for in-hospital mortality with all sources. Both of our models perform similarly on the ROC curve. The Brier score for the DKL is $0.082$ slightly better than the $0.084$ of the BiLSTM. In the Cox re-calibration the BiLSTM’s calibration intercept was -0.358 (-0.49, -0.229), and slope 0.802 (0.726, 0.88); compared to the DKL’s -0.066 (-0.185, 0.05), and 1.177 (1.062, 1.298). Unlike the BiLSTM the DKL showed no significant deviations from the ideal values of 0 and 1. The DKL was slightly more unsharp: 0.089 versus 0.081 for the BiLSTM. ## 4 Related Work Dürichen et al. [5] propose a multi-task GP that jointly models physiological variables for clinical time series. Cheng et al. [2] develop a real-time clinical prediction model based on a GP model. Aside from producing confidence regions in the predictions, the GP also scales to large patient databases, and produces interpretable relations across (clinical) variables. The interprtability is produced by inspecting the correlation across variables in the kernel function. Futoma et al. [6] propose a sepsis prediction model based on a pipeline with a GP that produces inputs for a NN classifier. The model takes into account uncertainty estimates and outperforms strong sepsis prediction baselines. On the other hand, our DKL model uses RNNs to model the time series physiological variables and feed the resulting features into the GP for prediction. Our work, however, is the first to investigate DKL in the context of temporal shift. ## 5 Conclusions and Future Work We investigated the DKL framework for the task of in-hospital mortality prediction under a temporal shift in population. The DKL shows competitive performance compared to a strong NN baseline, as well as a better calibration. However, when the test dataset is in the same distribution as the training both models show similar results. The GP component does not degrade the overall performance, and in addition, it provides extra guarantees such as uncertainty estimates. By contrasting the two experiments and inspecting the sharpness of the predictions we can ascribe the improved performance on the test set to the robustness of the GP when facing uncertainty. For future work, we will analyse different base kernels, evaluate the uncertainty estimate of the DKL, and use the framework described in [4] for better understanding of discrepancies in performance over time. ## References * Bengio [2009] Bengio, Y.: Learning deep architectures for AI. Foundations and Trends in Machine Learning 2(1), 1–127 (2009) * Cheng et al. [2020] Cheng, L.F., Darnell, G., Chivers, C., Draugelis, M.E., Li, K., Engelhardt, B.E.: Sparse multi-output Gaussian processes for medical time series prediction. BMC Medical Informatics and Decision Making 20(152) (2020) * Cox [1958] Cox, D.R.: Two further applications of a model for binary regression. Biometrika 45, 562–565 (1958) * Debray et al. [2015] Debray, T.P.A., Vergouwe, Y., Koffijberg, H., Nieboer, D., Steyerberg, E.W., Moons, K.G.M.: A new framework to enhance the interpretation of external validation studies of clinical prediction models. Journal of clinical epidemiology 68 3, 279–89 (2015) * Dürichen et al. [2014] Dürichen, R., Pimentel, M., Clifton, L., Schweikard, A., Clifton, D.: Multi-task gaussian processes for multivariate physiological time-series analysis. IEEE transactions on bio-medical engineering 62 (2014) * Futoma et al. [2017] Futoma, J., Hariharan, S., Heller, K.: Learning to detect sepsis with a multitask gaussian process rnn classifier. In: International Conference on Machine Learning. JMLR.org (2017) * Gardner et al. [2018] Gardner, J.R., Pleiss, G., Bindel, D., Weinberger, K.Q., Wilson, A.G.: Gpytorch: Blackbox matrix-matrix gaussian process inference with GPU acceleration. CoRR abs/1809.11165 (2018) * Ghassemi et al. [2015] Ghassemi, M., Pimentel, M.A.F., Naumann, T., Brennan, T., Clifton, D.A., Szolovits, P., Feng, M.: A multivariate timeseries modeling approach to severity of illness assessment and forecasting in icu with sparse, heterogeneous clinical data. In: Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence. p. 446–453. AAAI’15, AAAI Press (2015) * Harutyunyan et al. [2019] Harutyunyan, H., Khachatrian, H., Kale, D.C., Ver Steeg, G., Galstyan, A.: Multitask learning and benchmarking with clinical time series data. Scientific Data 6(1) (2019) * Hochreiter and Schmidhuber [1997] Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Computation 9(8), 1735–1780 (1997) * Johnson et al. [2016] Johnson, A.E., Pollard, T.J., Shen, L., Li-wei, H.L., Feng, M., Ghassemi, M., Moody, B., Szolovits, P., Celi, L.A., Mark, R.G.: Mimic-iii, a freely accessible critical care database. Scientific data 3, 160035 (2016) * Kingma and Ba [2014] Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization (2014), cite arxiv:1412.6980Comment: Published as a conference paper at the 3rd International Conference for Learning Representations, San Diego, 2015 * Mackay [1992] Mackay, D.J.C.: Bayesian Methods for Adaptive Models. Ph.D. thesis, USA (1992), uMI Order No. GAX92-32200 * Minne et al. [2012] Minne, L., Eslami, S., de Keizer, N., de Jonge, E., de Rooij, S.E., Abu-Hanna, A.: Effect of changes over time in the performance of a customized saps-ii model on the quality of care assessment. Intensive care medicine (2012) * Murphy and Winkler [1987] Murphy, A., Winkler, R.: A general framework for forecast verification. Monthly Weather Review 115, 1330–1338 (1987) * Nestor et al. [2019] Nestor, B., McDermott, M.B.A., Boag, W., Berner, G., Naumann, T., Hughes, M.C., Goldenberg, A., Ghassemi, M.: Feature robustness in non-stationary health records: Caveats to deployable model performance in common clinical machine learning tasks. In: Doshi-Velez, F., Fackler, J., Jung, K., Kale, D., Ranganath, R., Wallace, B., Wiens, J. (eds.) Proceedings of the 4th Machine Learning for Healthcare Conference. Proceedings of Machine Learning Research, vol. 106, pp. 381–405. PMLR, Ann Arbor, Michigan (2019) * Quiñonero Candela et al. [2007] Quiñonero Candela, J., Ramussen, C.E., Williams, C.K.I.: Approximation methods for gaussian process regression. Tech. Rep. MSR-TR-2007-124 (2007) * Rajkomar et al. [2018] Rajkomar, A., Oren, E., Chen, K., Dai, A.M., Hajaj, N., Liu, P.J., Liu, X., Sun, M., Sundberg, P., Yee, H., Zhang, K., Duggan, G.E., Flores, G., Hardt, M., Irvine, J., Le, Q.V., Litsch, K., Marcus, J., Mossin, A., Tansuwan, J., Wang, D., Wexler, J., Wilson, J., Ludwig, D., Volchenboum, S.L., Chou, K., Pearson, M., Madabushi, S., Shah, N.H., Butte, A.J., Howell, M., Cui, C., Corrado, G., Dean, J.: Scalable and accurate deep learning for electronic health records. CoRR abs/1801.07860 (2018) * Rasmussen and Williams [2005] Rasmussen, C.E., Williams, C.K.I.: Gaussian Processes for Machine Learning (Adaptive Computation and Machine Learning). The MIT Press (2005) * Roberts et al. [2012] Roberts, S., Osborne, M., Ebden, M., Reece, S., Gibson, N., Aigrain, S.: Gaussian processes for timeseries modelling. Philosophical Transactions of the Royal Society (2012) * Shickel et al. [2018] Shickel, B., Loftus, T.J., Ozrazgat-Baslanti, T., Ebadi, A., Bihorac, A., Rashidi, P.: Deepsofa: A real-time continuous acuity score framework using deep learning. CoRR abs/1802.10238 (2018) * Wilson et al. [2016a] Wilson, A.G., Hu, Z., Salakhutdinov, R.R., Xing, E.P.: Stochastic variational deep kernel learning. In: Lee, D.D., Sugiyama, M., Luxburg, U.V., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems 29, pp. 2586–2594. Curran Associates, Inc. (2016a) * Wilson et al. [2016b] Wilson, A.G., Hu, Z., Salakhutdinov, R., Xing, E.P.: Deep kernel learning. In: Gretton, A., Robert, C.C. (eds.) Proceedings of the 19th International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 51, pp. 370–378. PMLR, Cadiz, Spain (2016b) * Wilson and Nickisch [2015] Wilson, A.G., Nickisch, H.: Kernel interpolation for scalable structured gaussian processes (kiss-gp). In: International Conference on Machine Learning. JMLR.org (2015)
Characterizations Of Weakly Conformally Flat And Quasi Einstein Manifolds Ramesh Sharma University of New Haven, West Haven, CT 06516, USA, E-mail: <EMAIL_ADDRESS> ###### Abstract First, we show that a warped product of a line and a fiber manifold is weakly conformally flat and quasi Einstein if and only if the fiber is Einstein. Next, we characterize and classify contact (in particular, $K$-contact) Riemannian manifold satisfying weakly (and doubly weakly) conformally flat and quasi-Einstein ($\eta$-Einstein) conditions. Finally, we provide local classification and characterization of a semi-Riemannian (including the 4-dimensional spacetime) with harmonic Weyl tensor and a non-homothetic conformal (including closed) vector field, in terms of Petrov types and Bach tensor. ## 1 Introduction In [6] and [7], Chen and Yano introduced the idea of a Riemannian manifold with quasi constant curvature and defined it as a Riemannian manifold ($M,g$) which is conformally flat and whose Ricci tensor $Ric$ satisfies the condition $Ric=ag+bu\otimes u$ (1.1) where $a$ and $b$ are smooth functions and $u$ is a 1-form metrically equivalent to a unit vector field $U$ on $M$. Examples of quasi constant curvature manifolds are conformally flat hypersurfaces of a space of constant curvature and canal hypersurfaces (without singularities) of a Euclidean space. We let $g$ be semi-Riemannian and $g(U,U)=\epsilon$, such that $U$ is spacelike if $\epsilon=1$ and timelike if $\epsilon=-1$. Another example of a quasi constant curvature semi-Riemannian manifold is the warped product $I\times_{f}F$ of a real open interval $I$ and a Riemannian manifold (fiber) $F$ of constant curvature, with warping function $f>0$ on $I$. In case $I$ is timelike with line-element $-dt^{2}$, $I\times_{f}F$ is the well known Friedman - Robertson - Walker ($FRW$) cosmological model (see, for example, Hawking and Ellis [13]). Remark: The equation (1.1) defines a quasi Einstein manifold in the sense of Chaki and Maity [5], existence and examples of these manifolds were provided by De and Ghosh [9]. In the literature, there is another concept of a quasi Einstein manifold in terms of Bakry-Emery tensor, studied by Case, Shu and Wei [4]. In this paper, we follow the former definition [equation (1.1)]. Examples of a quasi Einstein manifold are $\eta$-Einstein Sasakian manifolds (in particular, Sasakian space forms and Heisenberg groups with left invariant metrics as expanding non-gradient Ricci solitons). For details, we refer to Blair [2], Chow et al. [8] and Ghosh and Sharma [12]. Another example with Lorentzian metric $g$ and a unit timelike vector field $U$, is the perfect fluid solution of Einstein’s field equations [13], described by $Ric=4\pi(\rho-p)g+8\pi(\rho-3p)$ on a 4-dimensional spacetime with energy density $\rho$ and pressure $p$. Henceforth we assume the dimension of a semi-Riemannian manifold ($M,g$) to be $>3$ and denote arbitrary vector fields on $M$ by $X,Y,Z$. Let us weaken the conformal flatness and say that a semi-Riemannian manifold ($M,g$) is weakly conformally flat if and only if $W(X,Y,V)=0,$ (1.2) for a non-zero vector field $V$ on $M$, and where $W$ denotes the Weyl tensor of type (1,3). In dimension 4, a weakly conformally flat ($M,g$) with a non-null $V$ (i.e. $g(V,V)\neq 0$) is conformally flat. This follows from the following result of Eardley et al. [10]. “If a 4-dimensional semi-Riemannian manifold has its Weyl tensor annihilating a non-null (spacelike or timelike) vector field, then the Weyl tensor vanishes identically”. Though this result has been proved for a spacetime, it holds in general, for any semi-Riemannian manifold. The proof is based on symmetry properties, tracelessness and combinatorial computations. In this paper we provide characterizations and classifications of manifolds that are either weakly conformally flat or quasi Einstein, or both within the frameworks of warped product, Lorentzian and contact geometries. First, we establish the following result. ###### Theorem 1.1 The semi-Riemannian warped product $(M,g)=I\times_{f}F$ of dimension $n\geq 4$, with metric $g=\epsilon dt^{2}+f^{2}(t)g_{F}$ ($g_{F}$ being the metric on a Riemannian manifold $F$) is weakly conformally flat and quasi-Einstein if any one of the following conditions hold: (i) $g_{F}$ is Einstein, (ii) $W(x,U,U,y)=0$ for arbitrary vector fields $x$ and $y$ tangent to $F$ and $U$ the lift of $\partial_{t}$ ($t$ being the standard coordinate on $I$ (spacelike or timelike)), (iii) Weyl tensor of $g$ is harmonic. Under any one of these conditions, $g$ is Bach flat. Remark: It follows from the above theorem that an example of a weakly conformally flat and quasi-Einstein is $I\times_{f}F$ with Einstein fiber $F$. In the 4-dimensional Lorentzian case ($\epsilon=-1$), it is the Friedman - Robertson - Walker ($FRW$) cosmological model, as mentioned earlier. Next, we recall from Blair [2] that a conformally flat contact Riemannian manifold whose Reeb vector field $\xi$ is everywhere an eigen vector of its Ricci tensor, is either of constant curvature 1, or flat (in which case the dimension is 3). Prompted by this result, we consider a $K$-contact manifold $M$ (for which we know that $\xi$ is Killing and is an eigen vector of the Ricci tensor) and check the impact of the doubly weakly conformally flat condition $W(X,\xi,\xi)=0$ and weakly conformally flat condition $W(X,Y)\xi=0$ . More precisely, we prove the following result. ###### Proposition 1.1 A $K$-contact manifold $M$ satisfies the doubly weakly conformally flat condition $W(X,\xi,\xi)=0$ if and only if $M$ is $\eta$-Einstein. If a $K$-contact manifold satisfies the weakly conformally flat condition $W(X,Y)\xi=0$, then it is Sasakian and $\eta$-Einstein (i.e. quasi Einstein with $U=\xi$). Now we provide a classification of a contact Riemannian manifold that is weakly conformally flat and quasi-Einstein with $U$ as the Reeb vector field, by establishing the following result. ###### Theorem 1.2 If a contact Riemannian manifold is weakly conformally flat and quasi-Einstein with $U$ as the Reeb vector field $\xi$, then it is Sasakian, flat (3-dimensional), or locally isometric to $SU(2)$ or $SL(2,\mathbb{R})$ with a left invariant metric. ## 2 Preliminaries Let us briefly review the basic formulas of the $n$-dimensional semi- Riemannian warped product $M=I\times_{f}F$ of an open real interval $I$ and an ($n-1$)-dimensional semi-Riemannian manifold $F$ (fiber), with warping function $f>0$ on $I$ such that the metric $g$ on $M$ is given by $g=\pi^{*}(g_{I})+(f\circ\pi)^{2}\sigma^{*}(g_{F})$ where $\pi$ and $\sigma$ are projections from $I\times_{f}F$ onto $I$ and $F$ respectively, $g_{I}$ and $g_{F}$ are the metrics on $I$ and $F$. Locally, we express $g$ as $\epsilon dt^{2}+f^{2}(t)(g_{F})_{ij}dx^{i}dx^{j}$ where $\epsilon=1$ or -1 according as $g(\partial_{t},\partial_{t})=1$ or -1, $t$ is the standard coordinate on $I$ and $x^{i}$ are local coordinates on $F$. If $U$ is the lift of $\partial_{t}$ and $x$, $y$, $z$, $w$ denote arbitrary vector fields tangent to $F$, then we have the following formulas $R(x,U)U=-\frac{\ddot{f}}{f}x,R(x,y)U=0$ (2.1) $R(x,U)y=\frac{\epsilon\ddot{f}}{f}g(x,y)U$ (2.2) $g(R(x,y,z),w)=g({{}^{F}}R(x,y,z),w)-\epsilon\bigg{(}\frac{\dot{f}}{f}\bigg{)}^{2}[g(y,z)g(x,w)-g(x,z)g(y,w)]$ (2.3) $Ric(U,U)=-\frac{n-1}{f}\ddot{f}$ (2.4) $Ric(U,x)=0$ (2.5) $Ric(x,y)={{}^{F}}Ric(x,y)-\bigg{[}\frac{\epsilon\ddot{f}}{f}+(n-2)\bigg{(}\frac{\dot{f}}{f}\bigg{)}^{2}\bigg{]}g(x,y)$ (2.6) $r=\frac{{{}^{F}}r}{f^{2}}-2\epsilon(n-1)\frac{\ddot{f}}{f}-(n-1)(n-2)\bigg{(}\frac{\dot{f}}{f}\bigg{)}^{2}$ (2.7) where an over-dot means derivative with respect to $t$, and $R$, $Ric$ and $r$ denote the curvature tensor, Ricci tensor and scalar curvature of $M$, while the symbols with a superscript $F$ denote the corresponding quantities of $F$. For details we refer to O’Neill [15]. Secondly, we briefly review contact geometry. A ($2m+1$)-dimensional smooth manifold $M$ is said to be a contact manifold if it carries a smooth 1-form $\eta$ such that $\eta\wedge(d\eta)^{m}\neq 0$ (where $\wedge$ is the exterior wedge product and $d$ is the exterior derivation). For a given contact form $\eta$, there exists a vector field (Reeb vector field) $\xi$ such that $\eta(\xi)=1$ and $(d\eta)(\xi,.)=0$. Polarizing $d\eta$ on the contact subbundle ($\eta=0$) we obtain a Riemannian metric $g$ and a (1,1)-tensor field $\varphi$ such that $(d\eta)(X,Y)=g(X,\varphi Y),\eta(X)=g(\xi,X),\varphi^{2}=-I+\eta\otimes\xi.$ (2.8) The Riemannian metric $g$ is an associated metric of $\eta$ and ($\varphi,\eta,\xi,g$) is a contact Riemannian structure on $M$. The (1,1)-tensor field $h=\frac{1}{2}\pounds_{\xi}\varphi$ ($\pounds$ denoting the Lie-derivative operator) on a contact Riemannian manifold is self-adjoint, trace-free and anti-commutes with $\varphi$. The following formulas hold on a contact Riemannian manifold: $\nabla_{X}\xi=-\varphi X-\varphi hX,$ (2.9) $Ric(\xi,\xi)=2m-|h|^{2}.$ (2.10) A contact Riemannian metric is said to be $K$-contact if $\xi$ is Killing, equivalently, $h=0$, or $R(X,\xi)\xi=X-\eta(X)\xi$ (2.11) It follows from (2.10) that $K$-contact metrics are maxima for the Ricci curvature along $\xi$. A contact Riemannian metric on $g$ on $M$ is called Sasakian if the almost Kaehler structure induced on the cone $C(M)$ over $M$ with metric $dr^{2}+r^{2}g$, is integrable (i.e. Kaehler). This Sasakian condition on a contact Riemannian manifold is equivalent to $R(X,Y)\xi=\eta(Y)X-\eta(X)Y.$ (2.12) A Sasakian metric is $K$-contact, however, the converse holds in general, only in dimension 3. A generalization of a Sasakian manifold is the ($k,\mu$)-contact manifold defined as a contact Riemannian manifold $M$ satisfying the nullity condition: $R(X,Y)\xi=k(\eta(Y)X-\eta(X)Y)+\mu(\eta(Y)hX-\eta(X)hY)$ (2.13) for real constants $k$, $\mu$. We know that $k\leq 1$, with equality when $M$ is Sasakian. For $k<1$ we have $\displaystyle Ric(X,Y)=$ $\displaystyle(2m-2-m\mu)g(X,Y)+(2m-2+\mu)g(hX,Y)$ (2.14) $\displaystyle+$ $\displaystyle(m(2k+\mu)-2m+2)\eta(X)\eta(Y),$ $r=[2m-2+k-m\mu].$ (2.15) Finally, a contact Riemannian manifold $M$ is said to be $\eta$-Einstein (Boyer, Galicki and Matzeu [3]) if there exist smooth functions $a$ and $b$ on $M$ such that $Ric=ag+b\eta\otimes\eta.$ (2.16) For a $K$-contact manifold (in particular, Sasakian manifold) of dimension $>$ 3, we know that $a$ and $b$ are constants. An example of an $\eta$-Einstein manifold is the Sasakian space-form which is a Sasakian manifold of constant $\varphi$-sectional curvature. Sasakian geometry has been extensively studied since its recently perceived relevance in string theory. Sasakian Einstein metrics have received a lot of attention in physics, for example, $p$-brane solutions in superstring theory, Maldacena conjecture (AdS/CFS duality) [14]. ## 3 Proofs of The Results Proof of Theorem 1.1 The Weyl tensor $W$ of type (1,3) is defined by $\displaystyle W(X,Y,Z)=$ $\displaystyle R(X,Y)Z+\frac{1}{n-2}[Ric(X,Z)Y-Ric(Y,Z)X+g(X,Z)QY$ (3.1) $\displaystyle-$ $\displaystyle g(Y,Z)QX]-\frac{r}{(n-1)(n-2)}[g(X,Z)Y-g(Y,Z)X]$ on an $n$-dimensional semi-Riemannian manifold ($M,g$), where $Q$ is the Ricci operator defined by $g(QX,Y)=Ric(X,Y)$. Using (3.1) and the formulas (2.1) through (2.7) and after a lengthy computation we obtain the following equations $g(W(x,U,U),y)=-\frac{\epsilon}{n-2}{{}^{F}}Ric^{0}(x,y)$ (3.2) $g(W(x,y,z),U)=0$ (3.3) $\displaystyle g(W(x,y,z),w)=$ $\displaystyle g({{}^{F}}W(x,y,z),w)-\frac{1}{(n-2)(n-3)}[g(y,w){{}^{F}}Ric^{0}(x,z)$ (3.4) $\displaystyle-$ $\displaystyle g(x,w){{}^{F}}Ric^{0}(y,z)+g(x,z){{}^{F}}Ric^{0}(y,w)$ $\displaystyle-$ $\displaystyle g(y,z){{}^{F}}Ric^{0}(x,w)]$ where $x$, $y$, $z$, $w$ are tangent to $F$, ${{}^{F}}W$ the Weyl tensor of $F$ and a superscript $0$ means the traceless part. Parts (i) and (ii) of the theorem follows from (2.6), (3.2) and (3.3). Part (iii) follows from the following result of Gebarowski [11], ”The Weyl tensor of the warped product $I\times_{f}F$ is harmonic (divergence - free) if and only if $F$ is Einstein”. Finally, under any one of the conditions (i), (ii) and (iii), the components of $W$ along $U$ are zero. The Bach tensor $B$ introduced by Bach [1] in the context of conformal relativity, is defined as a symmetric second order tensor with components $B_{ab}=\frac{1}{n-1}\nabla^{c}\nabla^{d}W_{cabd}+\frac{1}{n-2}R^{cd}W_{cabd}$ (3.5) where $R^{cd}$ are the components of the Ricci tensor. Under any one of the conditions $F$ is Einstein and hence by Gebarowski’s result, $W$ is harmonic, also $g$ is quasi-Einstein (implied by $g$ being Weakly conformally flat and quasi Einstein). Using all these facts in (3.5) shows that $B_{ab}=0$, completing the proof. Remark: Taking $x=\partial_{i}$, $y=\partial_{j}$ ($\partial_{i}$ is the coordinate basis $\mathfrak{X}(F)$, we can write $g(W(x,U,U),y)$ as $W_{i00j}$, because $U$ is the lift of $\partial_{t}$ and $t=x^{0}$ (the coordinate on $I$). The components $W_{i00j}$ are known as the components of the electric part of the Weyl tensor (see [17]) of the spacetime manifold (in which case $\epsilon=-1$). Proof of Proposition 1.1 Let ($M,\eta,\xi,g$) be a ($2m+1$)-dimensional $K$-contact manifold. Using the definition (3.1) of the Weyl tensor and the $K$-contact property (2.11) we find that $W(X,\xi)\xi=\frac{r-2m}{2m(2m-1)}[X-\eta(X)\xi]-\frac{1}{2m-1}[QX-2m\eta(X)\xi]$ (3.6) If $W(X,\xi)\xi=0$, then the above equation implies $QX=\bigg{(}\frac{r}{2m}-1\bigg{)}X+\bigg{(}2m+1-\frac{r}{2m}\bigg{)}\eta(X)\xi$ and hence $g$ is $\eta$-Einstein. Conversely, let $g$ be $\eta$-Einstein, i.e. (2.16) holds. Tracing it gives $r=(2m+1)a+b$ (3.7) Writing (2.16) as $QX=aX+b\eta(X)\xi$, substituting $\xi$ for $X$, and using the $K$-contact property: $Q\xi=2m\xi$ [2], we get $a+b=2m$ (3.8) Solving (3.7) and (3.8) for $a$ and $b$ and substituting them in (2.16) we get $QX=\bigg{(}\frac{r}{2m}-1\bigg{)}X+\bigg{(}2m+1-\frac{r}{2m}\bigg{)}\eta(X)\xi.$ (3.9) Plugging this value of $QX$ in (3.6) provides $W(X,\xi)\xi=0$. This proves the first part. For the second part, we have the stronger hypothesis $W(X,Y)\xi=0$ on a contact Riemannain manifold $M$. This implies $W(X,\xi,\xi)=0$, and hence applying the first part, we conclude that $g$ is $\eta$-Einstein, and hence weakly conformally flat and quasi Einstein. Now, a straight forward computation using (3.9) gives $W(X,Y)\xi=R(X,Y)\xi-\eta(Y)X+\eta(X)Y$ By our hypothesis, the left hand side is zero, and thus the above equation implies equation (2.12) and hence $g$ is Sasakian. This completes the proof. Proof of Theorem 1.2 By our hypothesis we have $W(X,Y,\xi)=0$ and equation (2.16), i.e. $Ric=ag+b\eta\otimes\eta$. Hence $r=(2m+1)a+b$, by contraction. Using these data we find, after a straight forward computation, that $R(X,Y)\xi=\frac{a+b}{2m}[\eta(Y)X-\eta(X)Y]$ (3.10) At this point, we recall the following Schur-type result (Sharma [16]), ”If $R(X,Y)\xi=k(\eta(Y)X-\eta(X)Y)$ for a smooth function $f$ independent of the choice of the vector fields $X$, $Y$ on a contact Riemannian manifold, then $k$ is constant on $M$”. Applying this result, in view of equation (3.10), we conclude that $\frac{a+b}{2m}$ is constant, say $k$. Therefore $M$ is a ($k,\mu$)-contact manifold with $\mu=0$. As $k\leq 1$, we have either $k=1$, in which case $M$ is Sasakian, or $k<1$. For $k<1$, comparing (2.16) with (2.14) and noting that $\mu=0$ and $a+b=2mk$, we have $(a-2m+2)g(X,Y)+(2m-2-a)\eta(X)\eta(Y)=2(m-1)g(hX,Y)$ (3.11) Contracting it at $X$ and $Y$, and noting that $tr.h=0$, we find $a=2(m-1)$. Consequently, equation (3.11) reduces to $(m-1)h=0$. So, $m=1$, because in this case $h\neq 0$ (non-Sasakian). Thus, (2.14) reduces to $Ric=2k\eta\otimes\eta$, i.e. Ricci tensor has rank 1. For $k=0$, $M$ is flat, and for $k\neq 0$, $M$ is locally isometric to $SU(2)$ ($k>0$) or $SL(2,\mathbb{R}$) ($k<0$) with a left invariant metric (see Blair [2]). This completes the proof. ## 4 Concluding Remark We may extend the definition (1.1) of a quasi-Einstein manifold for a semi- Riemannian metric by allowing the non-zero vector field $U$ to be null (light- like, i.e $g(U,U)=0$). An example for the null case is the null dust solution of Einstein’s field equations, in which the only mass-energy is due to some kind of massless radiation and whose stress-energy tensor is $T=\Phi k\otimes k$ where $k$ is a null vector field specifying the direction of motion of the radiation and $\Phi$ the intensity. This includes plane fronted waves with parallel rays ($pp$-waves) which, in turn include gravitational plane waves. For details, we refer to Stephani et al. [17]. ## References * [1] Bach, R., Zur Weylschen Relativitatstheorie und der Weylschen Erweiterung des Krummungstensorbegriffs, Math. Z. 9 (1921), 110-135. * [2] Blair, D. Riemannian geometry of contact and symplectic manifolds, Springer, 2010. * [3] Boyer, C., Galicki, K. and Matzeu, P. On eta-Einstein sasakian geometry. Commun. Math. Phys. 262 (2006), 177-208. * [4] Case, J., Shu, Y. and Wei, G. Rigidity of quasi-Einstein metrics. Diff. Geom. Appls. 29 (2011), 93-100. * [5] Chaki, M. and Maity, R.K. On quasi Einstein manifolds. Publ. Math. Debrecen. 57 (2000). 297-306. * [6] Chen, B. Hypersurfaces of a conformally flat space. Tensor, NS 26(1972), 315-321. * [7] Chen, B. and Yano, K. Special conformally flat spaces and canal hypersurfaces. Tohoku Math. J. 25 (1973), 177-184. * [8] Chow, B., Chu, S., Glickenstein, D., Guenther, C., Isenberg, J., Ivey, T., Knopf, D., Lu, P., Luo, F, and Ni, L., The Ricci flow: Techniques and Applications, Part I: Geometric Aspects, Mathematical Surveys and Monographs 135, American Math. Soc., 2004. * [9] De, U. C. and Ghosh, G. On quasi Einstein manifolds. Periodica Math. Hungarica. 48 (2004), 223-231. * [10] Eardley, D., Isenberg, J., Marsden, J. and Moncrief, V. Homothetic and conformal symmetries of solutions to Einstein’s equations. Commun. Math. Phys. 106 (1986), 137-158. * [11] Gebarowski, A. Nearly conformally symmetric warped product manifolds. Bull. Inst. Math. Acad. Sinica. 20 (1992), 359-377. * [12] Ghosh, A. and Sharma, R. Sasakian metric as a Ricci soliton and related results. J. Geom. Phys. 75 (2014), 1-6. * [13] Hawking, S. and Ellis, G. G.R., The large scale structure of space-time, Cambridge University Press, 1973. * [14] Maldacena, J., The large $N$ limit of superconformal field theories and supergravity, Adv. Theor. Math. Phys. 2 (1998), 231-252. * [15] O’Neill, B. Semi-Riemannian geometry with applications to relativity. Academic Press,1983. * [16] Sharma, R. On the curvature of contact metric manifolds. J. Geom.. 53 (1995), 179-190. * [17] Stephani, H., Kramer, D., MacCallum, M., Hoensselaers, C. and Herlt, E. Exact solutions of Einstein’s field equations, Cambridge Univ. Press, Cambridge, 2009. * [18] Yano, K., Integral formulas in Riemannian geometry, Marcel Dekker, New York, 1970.
Collider testable low-scale seesaw models predict pseudo-Dirac heavy neutrinos, that can produce an oscillating pattern of and events. We explore if such can be resolved at the $\HLLHC$. To that end, we employ the first ever full simulation of the oscillations, for several example benchmark points, and show under which conditions the experiment is able to discover them. The workflow builds on a FeynRules model file for the (pSPSS) and a patched version of MadGraph, able to simulate . We use the fast detector simulation Delphes and present a statistical analysis capable of inferring the significance of oscillations in the simulated data. Our results demonstrate that, for heavy neutrino mass splittings smaller than about [100]μeV, the discovery prospects for at the $\HLLHC$ are promising. § INTRODUCTION The observed light neutrino flavour oscillations <cit.> can be explained by introducing at least two mass splittings between the light neutrinos. It follows that at least two of the neutrinos have to be massive. However, the of particle physics lacks the corresponding mass terms. Therefore, it needs to be extended with a model able to generate these light neutrino masses, such as the type I seesaw mechanism <cit.>. In order to generate the two light neutrino masses in this model at least two sterile neutrinos have to be added to the particle content of the . In general, the type I seesaw mechanism is governed by two sets of parameters, that can be adjusted to obtain light neutrino masses at the right order of magnitude. Besides their Yukawa couplings, the second set is given by their Majorana masses $m_M^{}$, which are only possible since the additional neutrinos are singlets. After the Yukawa couplings result in Dirac mass terms $\vec m_D^{(i)}$ between each of the sterile neutrinos $N_i$ and the active neutrinos $\nu_\alpha$. [We indicate quantities with a suppressed vector index by bold font.] For the case of two heavy neutrinos, the generated light neutrino mass matrix then follows from the seesaw equation \begin{equation} \label{eq:seesaw} M_\nu = \frac{\vec m_D^{(1)} \otimes \vec m_D^{(1)}}{m_M^{(1)}} + \frac{\vec m_D^{(2)} \otimes \vec m_D^{(2)}}{m_M^{(2)}} \,. \end{equation} In order to obtain light neutrino masses, one could take the Yukawa couplings to be very small or choose the Majorana masses of the heavy neutrinos to be very large. Those two limits of the seesaw mechanism are referred to as the small coupling and the high scale limit, respectively. In both cases the parameters have to be taken to such extreme values that a direct detection of the sterile neutrinos at current collider experiments is not feasible. A third possibility, which leads to collider testable low-scale seesaw models, consists in heavy neutrino pairs forming a pseudo-Dirac structure. This leads to a cancellation between the two terms in (<ref>) and can be justified by an approximate symmetry. The symmetry can be realised as an extension of the lepton number and we refer to it as a <cit.>. It allows for the heavy neutrinos to be light enough, and at the same time have sufficiently large Yukawa couplings, to be directly observable at the <cit.>. When the is exact, the light neutrinos are massless and the two heavy neutrinos are precisely mass-degenerate, forming a single Dirac neutrino. However, if the is broken by small parameters, light neutrino masses are generated and typically a mass splitting between the heavy neutrinos is introduced. The corresponding seesaw scenario is referred to as <cit.>. The phenomenological effects of the slightly broken are captured by the introduced in <cit.>. Similar to the case of other neutral particles, this leads to particle-antiparticle oscillations <cit.>, which are in this context called <cit.>. The framework of external wave packets, which has been developed in <cit.> and applied to the case of in <cit.>, allows to describe this phenomenon including decoherence effects <cit.>. Although the small breaking of the generates processes, one can argue on principle grounds that for prompt heavy neutrino decays it is not possible to observe at the <cit.>. In contrast, generate observable amounts of when the heavy neutrinos are sufficiently long-lived compared to their oscillation period. When the heavy neutrinos have large enough lifetimes, such that their time of decay can be reconstructed, it can be possible to resolve the oscillation pattern of the . This allows to deduce the mass splitting of the heavy neutrinos, which might not be possible from measuring the amount of and processes alone. The possibility of resolving the at colliders has been estimated to be feasible during the in <cit.>. The goal of the present paper is to explore under which conditions it is possible to resolve at the at the reconstructed level, employing the process shown in <ref>. To this end, the FeynRules implementation of the <cit.> is used to simulate in MadGraph. Subsequently, Pythia is employed to simulate effects such as hadronisation, and the fast detector simulation Delphes is used to simulate the phase II detector. A cut based analysis of the generated events is performed using custom code. Finally, the prospects to resolve oscillation patterns for several scenarios of long-lived nearly mass degenerate heavy neutrinos at the are derived, using a detailed statistical analysis implemented in Mathematica. This paper is structured as follows: First, the and are briefly reviewed in <ref>. Subsequently, in <ref> the simulation of signal events containing these oscillations and relevant background processes are discussed. Afterwards, the statistical analysis is introduced in detail in <ref>. Finally, the results are presented in <ref> and the paper is concluded in <ref>. In <ref> we comment on the induction of residual oscillations through cuts on the transverse impact parameter. When considering two heavy neutrinos under the assumption of an intact , the only allowed additions to the Lagrangian are \begin{equation} \label{eq:symmetric Lagrangian} \mathcal L_{\SPSS}^L = \widebar{N_i^c} \ispd N_i^{} - y_{1\alpha} \widebar{N_1^c} \widetilde H^\dagger \ell_\alpha - \widebar{N_1^c} m_M^{} N_2^{} + \dots + \HC \,, \end{equation} where $N_1$ and $N_2$ are sterile neutrinos written here as left-chiral four-component spinor fields. The Higgs and lepton doublets of the are denoted by $H$ and $\ell$, respectively, and $\vec y_1$ is the Yukawa coupling vector with components $y_{1\alpha}$. The ellipses capture further contributions that can be generated by additional sterile neutrinos, but are assumed to be subdominant here. In this case the two terms in the seesaw formula (<ref>) cancel precisely. In addition, the following symmetry breaking terms can be introduced \begin{equation} \label{eq:broken Lagrangian} \mathcal L_{\SPSS}^{\cancel L} = - y_{2\alpha} \widebar{N_2^c} \widetilde H^\dagger \ell_\alpha - \mu_M^\prime \widebar{N_1^c} N_1^{} - \mu_M^{} \widebar{N_2^c} N_2^{} + \dots + \HC \,. \end{equation} When the Yukawa coupling vector $\vec y_2$ times the as well as the Majorana mass parameters $\mu_M^{}$ and $\mu_M^\prime$ are small compared to $m_M^{}$, the light neutrino masses are guaranteed to be small as well. Due to the approximate , it is possible for the heavy neutrinos to have a mass well below the $W$ boson mass, while at the same time their coupling can be large enough to obtain a sizable number of events at the , without violating constraints from light neutrino experiments such as searches for decay. In the with exact (<ref>), the neutrino mass matrix for the interaction eigenstates $n = \row{\nu_e, \nu_\mu, \nu_\tau, N_1, N_2}^\trans$ is given by \begin{equation} M_n^L = \begin{pmatrix} 0 & \vec m_D^{} & 0 \\ \vec m_D^\trans & 0 & m_M^{} \\ 0 & m_M^{} & 0 \end{pmatrix} \end{equation} where $\vec m_D = \vec y_1 v$ with the Higgs $v\approx\unit[174]{GeV}$. The mass term of the Lagrangian after is given by \begin{equation} \mathcal L_\text{mass} = - \frac12 \widebar{n^c} M_n n + \HC \,. \end{equation} This mass matrix does not generate neutrino masses. Furthermore, the two heavy neutrinos are mass degenerate with a relative phase of $-i$ in their Yukawa couplings and can therefore be described as a single Dirac fermion. However, in the presence of the small symmetry breaking terms (<ref>) the complete mass matrix reads \begin{equation} \label{eq:broken mass matrix} M_n^{\cancel L} = \begin{pmatrix} 0 & \vec m_D^{} & \vec \mu_D^{} \\ \vec m_D^\trans & \mu_M^\prime & m_M^{} \\ \vec \mu_D^\trans & m_M^{} & \mu_M^{} \end{pmatrix} \,, \end{equation} where $\vec \mu_D^{} = \vec y_2 v$. This mass matrix not only generates small masses for the light neutrinos, but additionally introduces a mass splitting between the two heavy neutrinos, that is also suppressed by the same small breaking parameters. Thus, the pair of two Majorana fermions can no longer be described as a pure Dirac particle, but as a pseudo-Dirac particle. The symmetry breaking terms can arise from specific low scale seesaw models, such as the linear or the inverse seesaw, which yield only the terms proportional to $\vec \mu_D^{}$ and $\mu_M^{}$, respectively. However, the complete mass matrix (<ref>) can be generated from more complicated seesaw models. Since the symmetry breaking terms are very small, their phenomenological effects can often be neglected in collider studies with the notable exception of , which can be phenomenologically significant as they are an interference effect. At , the are fully described by the mass splitting $\Delta m$ between the heavy neutrinos together with an additional damping parameter $\lambda$, capturing the potential decoherence effects discussed in <ref>. Therefore, instead of adding the terms (<ref>) to the Lagrangian (<ref>), in the , the mass splitting is directly introduced as a model parameter. Consequently, the masses of the heavy neutrinos are given by \begin{equation} \label{eq:Delta m mass splitting} m_{\nicefrac45}^{} = m_M^{} \left(1 + \frac12 \abs{\vec \theta}^2 \right) \mp \frac12 \Delta m \,, \end{equation} where $\vec \theta = \flatfrac{\vec m_D^{}}{m_M^{}}$ is the active-sterile mixing parameter. A detailed description of the can be found in <cit.>. $\Delta m / \unit{\mu eV}$ $c \tau_\text{osc} / \unit{mm}$ $R_{ll}$ 1 82.7 15 0.9729 2 207 6 0.9956 3 743 1.67 0.9997 [ points] All points have a mass of [14]GeV and an active-sterile mixing parameter satisfying $\abs{\theta_\mu}^2 = 10^{-7}$, which results for all points in a decay width of $\Gamma = \unit[13.8]{\mu eV}$. However, they vary in their mass splitting and consequently have different oscillation periods $\tau_\text{osc}$, which leads to different to ratios $R_{ll}$, defined in (<ref>). [Number of expected events at the $\HLLHC$] Contour lines for the number of expected displaced vertex events $N$ as well as bands for the over event ratio $R_{ll}$, definition (<ref>), for the three defined in <ref>. The contour lines for $N$ apply to the detector at the $\HLLHC$ with $\mathscr L = \unit[3]{\inv{ab}}$ and cuts as defined in <ref>. They have a nose-like shape and depend only on the heavy neutrino mass $m$ and the active-sterile mixing parameter $\theta_\mu$, with $\theta_e = \theta_\tau = 0$. $N$ is therefore identical for all three , which differ only in their mass splittings $\Delta m$. The position of the in the parameter plane is indicated by a purple cross. The different $\Delta m$ of the result in different $R_{ll}$ bands. The contour lines where $R_{ll} \in [0.1,0.9]$ are shown in yellow and orange for 1 and 3, respectively. The relative position of the cross to the bands shows that the have an $R_{ll}$ close to one. The grey area corresponds to the exclusion bounds from displaced vertex searches <cit.> and the shaded grey area to the bounds from searches for prompt processes <cit.>. Since the prompt searches rely on signals, they apply only to models with an $R_{ll}$ close to one. In order to obtain the significance with which could be observed at colliders, three that differ only in their mass splitting are used and given in <ref>. The Majorana mass parameter and the active-sterile mixing parameters for all points are chosen to be $m_M^{} = \unit[14]{GeV}$ and $\abs{\theta_\mu}^2 = \expno{-7}$, leading to a decay width of $\Gamma = \unit[13.8]{\mu eV}$. The active-sterile mixing parameter corresponds to a Yukawa coupling of $y_\mu = \expnumber{2.55}{-5}$, while the Yukawa couplings to the electron and $\tau$-lepton have been set to zero. The number of expected events with muons as final state leptons in conjunction with the cuts presented in <ref>, are shown in <ref>. It becomes clear that the chosen lie comfortably beyond the current bounds <cit.>. In the minimal linear seesaw model, only one pseudo-Dirac pair of heavy neutrinos is added to the . This results in the lightest neutrino being massless and the mass splitting of the heavy neutrinos $\Delta m$ being identical to the mass splitting of the light neutrinos, <cit.>. The mass splitting of 3 is chosen to represent this possibility, where the light neutrino mass splitting is taken from a recent global fit assuming inverse light neutrino mass hierarchy <cit.>. However, in top-down realisations of low scale seesaw models, the low scale linear seesaw models in <cit.>, it is common to have multiple pseudo-Dirac pairs of heavy neutrinos. Since the light neutrino masses get contributions from all these heavy neutrinos, it is possible for the pseudo-Dirac pairs to have smaller mass splittings than in 3 without the need of cancellations in the mass matrix. For simplicity it might be assumed that the collider phenomenology is dominated by only one of the pseudo-Dirac pairs. We introduce two additional points reflecting this possibility. §.§ Oscillations Oscillations of neutral particles, and thus , can be described in the framework of external wave packets. Compared to simple methods relying on plane waves, wave packets allow for a self-consistent description of oscillations. This is due to the finite uncertainty in energy-momentum and spacetime, that allows to produce a coherent superposition of non-degenerate mass eigenstates, while simultaneously makes it possible to introduce the notion of a travelled time and distance. The framework also allows to calculate the possible suppression of oscillations due to the loss of coherence between the propagating mass eigenstates. For phenomenological studies these effects can be captured by a damping parameter $\lambda$, which is thus included in the . At , the probability to obtain a or event is given by \begin{equation} P^{\nicefrac{\LNC}{\LNV}}_\text{osc}(\tau) = \frac{1 \pm \cos(\Delta m \tau) \exp(-\lambda)}2 \,. \end{equation} Therefore, the oscillation period, defined in the proper time frame, is [We use time for quantities in the proper time frame and length for quantities in the lab frame, independent from the units of those quantities.] \begin{equation} \label{eq:oscillation period} \tau_\text{osc} = \frac{2 \pi}{\Delta m} \,. \end{equation} Since the probability density of the heavy neutrino to decay is \begin{equation} P_\text{decay}(\tau) = - \dv \tau \exp\left(- \Gamma \tau\right) = \Gamma \exp\left(- \Gamma \tau\right) \,, \end{equation} the probability for a heavy neutrino to decay between the proper times $\tau_{\min}$ and $\tau_{\max}$ and forming an or event is given by \begin{equation} \label{eq:probability displacement} P_{ll}^{\nicefrac{\LNC}{\LNV}}(\tau_{\min}, \tau_{\max}) = \int_{\tau_{\min}}^{\tau_{\max}} P^{\nicefrac{\LNC}{\LNV}}_\text{osc}(\tau) P_\text{decay}(\tau) \d \tau \,. \end{equation} By integrating the oscillations from the origin to infinity, one derives the total ratio between and events <cit.> \begin{equation} \label{eq:Rll} R_{ll} = \frac{P_{ll}^{\LNV}}{P_{ll}^{\LNC}} = \frac{\Delta m^2}{\Delta m^2 + 2 \Gamma^2} \,, \end{equation} that can be directly deduced from cut and count based analyses. The expected number of events in a collider experiment is given by \begin{equation} \label{eq:event numbers} N^{\nicefrac{\LNC}{\LNV}} = \mathscr L \sigma \BR \int D(\vartheta, \gamma) P^{\nicefrac{\LNC}{\LNV}}_{ll}(\tau_{\min}(\vartheta, \gamma), \tau_{\max}(\vartheta, \gamma)) \d\vartheta \d\gamma\,, \end{equation} where the collider luminosity $\mathscr L$, the heavy neutrino production cross section $\sigma$, and the are used and $D(\vartheta, \gamma)$ is the probability density of the heavy neutrino to have a Lorentz factor $\gamma$ and an angle $\vartheta$ with respect to the beam axis. The detector geometry is incorporated in the parameters $\tau_{\min}$ and $\tau_{\max}$ when boosting to the laboratory frame via $\tau(\vartheta,\gamma) = (\gamma^2-1)^{-\nicefrac12} L(\vartheta)$. The in the have been discussed in detail in <cit.>. We have checked that for the parameter points under consideration decoherence effects can be neglected and we thus take $\lambda = 0$ in the following <cit.>. § SIMULATION [Diagram of heavy neutrino production, oscillation, and decay] Diagram depicting the production, oscillation, and decay of a heavy neutrino. The heavy neutrino interaction eigenstate $N$ is produced in association with a prompt antimuon. Subsequently, the mass eigenstates oscillate such that finally a neutrino or antineutrino interaction eigenstate decays into a displaced muon or antimuon, respectively. We indicate that the process is initiated by proton collisions and that, for our parameter points, the two final quarks, originating from the hadronic $W$ decay, immediately hadronise into a single jet. In order to simulate events exhibiting , we employ the FeynRules <cit.> implementation of the <cit.>. FeynRules is used to generate an output, passed to [2.9.10 (LTS)]MadGraph5_aMC@NLO <cit.> in order to generate events at parton level. We use the patch introduced in <cit.>, that modifies the function calculating the particle's , to implement the oscillations. When evaluating the matrix element, MadGraph treats the process as prompt, adding the information afterwards. Therefore, almost no is obtained when interference between diagrams with different mass eigenstates are taken into account. In order to circumvent this, the process is initially specified in such a way as to prevent interference between the heavy neutrino mass eigenstates. This is achieved by writing the heavy neutrinos explicitly as intermediate states. As a consequence the total cross sections for the and process are identical and the amount of generated and events is, except for statistical variations, the same. The correct interference effects, including , are then included via the patch described in <cit.>. The syntax to generate a process, as the one given in <ref>, is where, in addition to the jet j and proton p multi particles, containing quarks and gluons, the multi particles are used. Here, the initial $W$ boson is additionally taken to be on-shell, since we focus on heavy neutrinos with masses far below the $W$ boson mass. An additional hard initial jet is included by using where matching was enabled with the standard parameter choices of MadGraph to prevent double counting. MadGraph utilises [8.306]Pythia in order to hadronise and shower the parton level events <cit.>. Finally [3.5.0]Delphes is used with the standard phase II card CMS_PhaseII_0PU to simulate the detector effects <cit.>. Secondary vertex reconstruction and smearing Since Delphes neither simulates displaced tracks properly, nor reconstructs secondary vertices, we implemented a smearing function affecting the displaced vertices, with the idea to capture experimental uncertainties. To our knowledge, no results for the overall precision of the vertex reconstruction have been published by so far. Therefore, we vary it in our simulations between zero and [4]mm, which we assume to be a conservative parametrisation of our ignorance, relying on private communication with members on the experimental collaborations. We like to emphasise that it would be highly welcome if the experimental collaborations could provide such information and, ideally, if such uncertainties could be implemented in Delphes directly. In detail, it is assumed that each reconstructed vertex is distributed with a Gaussian, with a standard deviation of $\unit[n]{mm}$, around its actual value. The true value of the displaced vertex is obtained from the displaced muon. The results are presented for different values of $n$, demonstrating the effects of the uncertainty in the vertex reconstruction. §.§ Signal In order to observe via using the process given by the diagram presented in <ref>, the two leptons have to be measured. For the given in <ref> and indicated in <ref>, the final state quarks are soft and immediately hadronise into a multi pion state, which then forms a single jet that can be captured with a cone radius of $\Delta R = 0.4$. Furthermore, the points are chosen such that the heavy neutrinos are long-lived. Hence, the signal contains one prompt muon, one displaced muon, and one displaced jet. A cut is used, which requires the muon to have an impact parameter of $\abs{d_0} \leq \unit[100]{\mu m}$, to ensure cut:one prompt muonexactly one prompt $\mu$. Furthermore, this muon is required to have a transverse momentum $p_T^{}$ greater than [20]GeV, ensuring that the event can be triggered. Events that do not contain cut:one disp muonat least one displaced $\mu$ are excluded by a cut that requires at least one muon with an impact parameter of $\abs{d_0} \geq \unit[2]{mm}$. Furthermore, the displaced muon candidate is only valid if it emerges from the same vertex as the displaced jet. A jet is considered displaced if it contains at least two displaced tracks that originate from the same vertex within a radius of [100]μm. A track is considered displaced if it has an impact parameter of $\abs{d_0} \geq \unit[2]{mm}$. Signal events are required to have cut:one disp jetexactly one displaced $j$. In addition to those basic cuts that define the signal, we demand cut:muon isolation$\mu$ isolation by requiring that the displaced muon and displaced jet have a $\Delta R$ larger than $0.4$ in order to reject muons radiated from the jet. Furthermore, restricting the reconstructed heavy neutrino mass to the cut:N mass window$N$ mass window of $\pm \unit[2]{GeV}$ around its theoretical value results in a cleaner signal. Generally, displaced particles are only considered if their origin falls inside a cylinder with dimensions given by half the tracker size in each direction. For the phase II detector, this means that displaced tracks must have an origin with a transverse distance smaller than [60]cm, and a longitudinal distance smaller than [150]cm, from the primary vertex. This requirement ensures that the produced particles propagate through a volume of the tracker that should be large enough to facilitate detection. At the moment this is an optimistic assumption about the performance of the displaced track reconstruction in the tracker, however, we think that pushing this capability is a worthwhile goal for the collaboration (see also <cit.> and references therein). Although the reconstruction of discussed here mostly relies on displaced vertices appearing in the inner tracker, it is also possible to reconstruct muon tracks and vertices using the muon chambers <cit.>. §.§ Background Signal Background Heavy hadrons 2*Produced events Physical $2854$ $\expno[8.882]{7}$ Simulated $\expno[5]{4}$ $\expno[1.1]{7}$ 9*Cuts <ref> $-23196$ $-\expno[6.50]{6}$ <ref> $-21652$ $-\expno[4.48]{6}$ <ref> $-1396$ $-17411$ <ref> $-838$ $-45$ <ref> $0$ $0$ <ref> $0$ $0$ <ref> $-111$ $-3$ <ref> $-1211$ $0$ 2*Remaining events Simulated $1596$ $0$ Physical $91$ $0$ [Cut flow of signal and background events] Cut flow for the signal and the heavy hadron background events. The physical events are given at an integrated luminosity of $\mathscr L = \unit[3]{ab^{-1}}$ with a signal cross section of $\sigma = \unit[951]{ab}$. The cuts introduced above define the basic search strategy for the signal, which is based on reconstructing a displaced vertex. The main sources of background for such a process are given by heavy flavour processes generating long-lived heavy hadrons, interaction of particles with dense detector material, and cosmic ray muons <cit.>. The heavy hadrons can travel macroscopic distances before they decay, potentially forming displaced vertices. However, this is only possible if those heavy flavour particles, and consequently their decay products, are highly boosted. Since the selection rules that define the signal already require the displaced muon and the displaced jet to have a $\abs{d_0} > \unit[2]{mm}$, this background is strongly reduced <cit.>. A cut:W mass window$W$ mass window cut around the reconstructed $W$ boson mass of $ \pm \unit[20]{GeV}$ is employed to reduce the background even further. We simulated this background using the same programs as for the signal events. The MadGraph syntax we used is where the additional multi particle is defined. An additional hard initial jet was included by using where matching was enabled with the standard parameter choices of MadGraph. The whole process was simulated using the model sm-no_b_mass, which employs the five flavour scheme in the definitions of the protons and jets. After the parton level events are passed to Pythia and Delphes, the above mentioned cuts and event selection rules were used, and it has been found that the background is eliminated entirely. The cut flow for the simulated signal and background is given in <ref>. Interactions of particles with detector material can also result in a displaced vertex signature and thus have to be considered as part of the background. A map of regions containing dense detector material is required in order to accept only displaced vertices that are outside that region. Simulating this background requires detailed knowledge about the detector structure and is beyond the scope of this work. However, requiring a displaced vertex to be reconstructed from enough good tracks reduces this background. In an experimental analysis, a sophisticated track reconstruction algorithm judges which tracks are good, in the sense that the track is likely to be produced from a charged particle rather than detector noise. With no specific insight about this algorithm, we define the tracks as being good that pass our cut and selection rules, which ensure that the charged particles traverse at least half the tracker. With the cuts introduced to define the signal, it is ensured that each event contains at least three displaced tracks from which the displaced vertex can be reconstructed. In order to veto against cosmic ray muons, the cut:vertex directionvertex direction cut is implemented that requires the reconstructed heavy neutrino momentum to be in the same direction as the displaced vertex. For this the distance in $\row{\eta, \phi}$-space between the reconstructed momentum $\vec p_N^{}$ and the displaced vertex direction $\vec d$ is limited to $\Delta R \leq 1.5$. We assume that this background is eliminated when using this cut in combination with additional timing information, therefore we have not attempted to simulate it. With the cuts described above, the signal region can be assumed to be background free. Additionally we found that a cut:no prompt eprompt $e$ veto, where prompt is again defined as $\abs{d_0} < \unit[100]{\mu m}$, with a $p_T^{} \geq \unit[20]{GeV}$ does not affect the signal, such that it could be used to eliminate further background if that becomes necessary. § STATISTICAL ANALYSIS The number of events entering the cut based analysis is given by the luminosity times the cross section for the considered process. The number of expected events after the cut based analysis $N_\text{exp}$ is then given by \begin{align} \label{eq:number of events} N_\text{exp} &= \mathscr L \sigma f_\text{eff} \,, & f_\text{eff} &= \frac{N_\text{after cuts}}{N_\text{all events}}\,, \end{align} where the efficiency factor captures how the cuts summarised in <ref> impact the number of signal events. The set of events surviving all cuts is labeled $\mathcal D_\text{all}$, and can be divided into events $\mathcal D_{\LNC}$ and events $\mathcal D_{\LNV}$. These datasets are then binned in the proper time $\tau$ space and the number of events in the $i$-th bin are given by \begin{equation}\label{eq:bin counts} N_i = \abs*{\set*{E \in \mathcal D \suchthat \tau_E^{} \text{ in $i$-th bin}}}\,, \end{equation} where $\abs{}$ is the cardinality such that $N_\text{exp} = \abs{\mathcal D_\text{all}}$. The heavy neutrino $\tau_E^{}$ is defined as the proper time at which it decays in the event $E$. §.§ Hypotheses The shape of the histograms describing the heavy neutrino may be predicted by two hypotheses. In contrast to the null hypothesis, the alternative hypothesis features oscillations. Null hypothesis [Examples for the null and alternative hypotheses] Example histogram plot:M0 with [15]k events, demonstrating the proper time distribution after applying the cuts summarised in <ref>. The inverse Gaussian distribution $\mu_i^\text{all}(\mathcal M_0)$ described in (<ref>) is depicted plot:M0 fit together with the prediction of the null hypothesis $\mu_i^{\nicefrac{\LNC}{\LNV}}(\mathcal M_0)$ plot:M0 fit half. The inverse Gaussian is shown with the parameters given in <ref> on it:one, where the normalisation is approximated as $N_0 = 15500$. The distribution of the alternative hypothesis (<ref>) is shown for plot:M1 LNC and plot:M1 LNV events of 1, with example value $\alpha = 0.05$ for the washout parameter. The hypothesis without oscillations $\mathcal M_0$ is based on the assumption that, for each bin, the probability of the heavy neutrino superposition to decay in a process is equal to the probability to decay in a process. Therefore, the mean number of predicted events in the $i$-th bin is given by \begin{equation} \mu_i^{\nicefrac{\LNC}{\LNV}}(\mathcal M_0) = \frac12 \mu_i^\text{all}(\mathcal M_0) \,. \end{equation} The probability of a given bin count can then be computed assuming a Poisson distribution of bin counts around this mean. In principle, one would expect the predicted mean values of this hypothesis to follow an exponential due to the finite lifetime of the heavy neutrinos. However, the employed cuts change this distribution into a non trivial one, which can be approximated by the of a generalized inverse Gaussian, described by four free parameters, yielding \begin{equation} \label{eq:inverse Gaussian} \mu_i^\text{all}(\mathcal M_0) = \frac{N_0}{2} \frac{\tau_i^{\param\theta - 1}}{\param\mu^{\param\theta} K_{\param\theta}(\nicefrac{\param\lambda}{\param\mu})} \exp\left(-\frac{\param\lambda}{\param\mu^2} \frac{\tau_i^2 + \param\mu^2}{2 \tau_i}\right)\,, \end{equation} where $N_0$ denotes the overall normalisation, $\param\theta$ is the index parameter, $\param\mu$ is the mean of the distribution, and $\param\lambda$ is a scale parameter. Additionally, $\tau_i$ denotes the position of the middle point of the $i$-th bin, and $K_{\alpha}(z)$ is the modified Bessel function of the second kind. The distribution is shown in <ref>, where the parameters correspond to the best fit point estimated in <ref> on it:one. Alternative hypothesis The second hypothesis $\mathcal M_1$ includes oscillations, with an oscillation period given by (<ref>), on top of the distribution described by the first hypothesis. Due to the imperfect reconstruction of the Lorentz factor, an additional washout effect obscures the oscillation pattern for larger $\tau$. This effect is included by an exponential factor $\alpha$, suppressing the oscillation. The prediction for the mean number of expected events in a bin is then given by \begin{equation} \label{eq:model 1} \mu_i^{\nicefrac{\LNC}{\LNV}}(\mathcal M_1) = \left[1 \pm \cos(\Delta m \tau_i) e^{- \alpha \tau_i} \right] \mu_i^{\nicefrac{\LNC}{\LNV}}(\mathcal M_0) \,, \end{equation} where the two additional parameters $\Delta m$ and $\alpha$ are incorporated, resulting in a total of six free parameters. Again, a Poisson distribution of bin counts, around this mean, is assumed. One example of such oscillations is given in <ref>. §.§ Likelihood ratio test To test whether the hypothesis including oscillations is statistically preferred by the simulated data, we use a likelihood ratio test. The main idea is to decide if the null hypothesis, given by $\mathcal M_0$, can be discarded in favour of the alternative hypothesis, given by $\mathcal M_1$. The likelihood is denoted by $\lh(\set{N_i}, \mathcal M)$ and describes the probability of finding the measured bin counts $\set{N_i}$ given the hypothesis $\mathcal M$. It is given by the product of likelihoods for all bins in both, the and , cases. With the assumed Poisson distributed number of events around the mean value $\mu_i$ in the $i$-th bin, the likelihood of a single bin is given by \begin{equation} \lh(N_i, \mu_i) = \frac{{\mu_i}^{N_i} e^{-\mu_i}}{N_i !} \,, \end{equation} where, as introduced above, $N_i$ is the number of events in the $i$-th bin, which depends on the dataset and the binning. The best fit point for each hypothesis, given the bin counts $\set{N_i}$, is evaluated by maximising the corresponding likelihood. A hypothesis at its best fit point is denoted with a hat, $\widehat{\mathcal M}_0$. Therefore, the likelihood ratio can be computed, given the bin counts, and yields \begin{equation} \lhr(\set{N_i}) = \frac{\lh(\set{N_i}, \widehat{\mathcal M}_0)}{\lh(\set{N_i}, \widehat{\mathcal M}_1)} = \prod_i \frac{\lh(N_i^{\LNC}, \widehat{\mathcal M}_0) \lh(N_i^{\LNV}, \widehat{\mathcal M}_0)}{ \lh(N_i^{\LNC}, \widehat{\mathcal M}_1) \lh(N_i^{\LNV}, \widehat{\mathcal M}_1)} \,. \end{equation} Taking the washout parameter to infinity in the alternative hypothesis reproduces the null hypothesis. The two hypotheses are therefore nested and as a consequence the inequality This inequality can by violated if not a global but a local maximum of the likelihood is found, while searching for $\widehat{\mathcal M}_1$. Such a local maximum might result in a smaller likelihood for the alternative hypothesis than the likelihood obtained by a fit of the null hypothesis. \begin{equation} \lh(\set{N_i}, \widehat{\mathcal M}_0) \leq \lh(\set{N_i}, \widehat{\mathcal M}_1) \,, \end{equation} holds and hence the likelihood ratio is restricted by \begin{equation} 0 \leq \lhr(\set{N_i}) \leq 1 \,. \end{equation} A likelihood ratio close to zero means that the given binned data are much better fitted by the alternative hypothesis than by the null hypothesis. In contrast, a ratio close to one shows that there is no clear distinction between the two hypotheses for the given binned data. Since in practice the logarithm of the likelihood is better suited for numerical computations, we continue the discussion using the which is defined as \begin{equation} \llhr(\set{N_i}) = -2 \log(\lhr(\set{N_i})) \,. \end{equation} The ranges from zero to infinity, where now a value near zero states that both, the null hypothesis and the alternative hypothesis, produce an equally good fit of the binned data. Contrary, a high value states that the alternative hypothesis produces a better fit. $\llhr=5.70$, $p=\unit[21.5]{\%}$, $Z=\unit[0.79]{\sigma}$ Low significance $\llhr=13.59$, $p=\unit[0.82]{\%}$, $Z=\unit[2.4]{\sigma}$ Intermediate significance $\llhr=17.71$, $p=\unit[0.13]{\%}$, $Z=\unit[3.01]{\sigma}$ High significance [Statistical fluctuation of the null hypothesis] Three examples of statistical fluctuations in the null hypothesis, producing patterns that mimic . From panel fig:low to panel fig:high the oscillatory pattern becomes more distinct while the probability that the depicted histogram is generated decreases. Caused by statistical fluctuations of the bin counts around their predicted mean, the null hypothesis $\widehat{\mathcal M}_0$ can also feature an oscillation pattern, as demonstrated in <ref>. Therefore, the value of a alone does not contain enough information to decide whether oscillations in a given dataset are significant or not. To translate the into a significance, it is crucial to know the probability that the null hypothesis produces the same due to fluctuations. Per construction, the most likely produced by $\widehat{\mathcal M}_0$ are small, and the larger the the less likely it is to be produced. The distribution of , generated by statistical fluctuations, is the of the under the assumption that the null hypothesis is true. We label it $\pdf$. $\chi^2$ $\DOFs$ $k(\unit[5]{\sigma})$ 1 2.25 30.94 2 3.28 34.05 3 3.92 35.82 [Null hypothesis fluctuations] Statistical fluctuations in the bin counts, following the null hypothesis, lead to a $\chi^2$ distribution with given $\DOFs$. The distributions are computed following the <ref> on it:one. The corresponding to a $p$ value of [5]σ are found to be increasing from 1 to 3. The challenge is then to find a value $k_p$, for which the probability of obtaining a $\llhr(\set{N_i}) \geq k_p$ through statistical fluctuations, is smaller than $p$. This means that $p$ is the probability of rejecting the null hypothesis with respect to the alternative hypothesis, even though the null hypothesis is true. Given $\pdf$, the value of $k_p$ can be obtained via \begin{equation} \int_{k_p}^{\infty} \pdf(k) \d k = p \,. \end{equation} While in the limit of large sample sizes the is typically assumed to approach a $\chi^2$ distribution, the form of the for finite sample sizes is generally unknown. Therefore a simulation is performed, with the goal to sample the , as follows * All simulated events that survive the cut based analysis are used to estimate the true distribution of the null hypothesis $\widehat{\mathcal M}_0$, which is depicted in <ref>. For those events, the best fit parameters of $\widehat{\mathcal M}_0$ are given by $\param\mu = 10.81$, $\param\lambda = 17.28$, and $\param\theta = 0.71$. The overall normalisation parameter $N_0$ is not relevant for the following steps as they only depend on the , which is normalised to unity. To obtain a set of events that follows the null hypothesis, taking statistical fluctuations into account, are taken according to the from <ref>. This is done until the set contains $N_\text{exp}$ signal events. * Using these values, bin counts are computed using the binning parameters based on <ref>. * The is computed based on the bin counts from <ref>. By repeating <ref> the desired distribution can be obtained. The obtained distributions are well approximated by $\chi^2$ distributions, where the are treated as a free parameter. The values of the for the three are given in <ref>. The simulated events are based on 1 with an oscillation period of [15]mm. The best fit parameters are given by an oscillation period of $\unit[14.08^{+0.85}_{-0.71}]{mm}$ and a washout parameter of $\alpha = \expnumber{3.66^{+31.73}_{-3.66}}{-3}$. The $\LLR$ was found to be $51.0$, which in this case yields a significance of [6.66]σ. The simulated events are based on 3 with an oscillation period of [1.67]mm. The best fit parameters are given by an oscillation period of $\unit[1.63^{+0.03}_{-0.04}]{mm}$ and a washout parameter of $\alpha = \expnumber{7.44^{+17.76}_{-5.45}}{-2}$. The $\LLR$ was found to be $5.29$, which in this case yields a significance of [0.67]σ. [Examples for the statistical fit of the oscillations to the $\MC$ data] Examples for the best fit of the alternative hypothesis to the data. For each parameter point a luminosity of [3]ab is used, resulting in a total of about $90$ events after cuts. However, the number of events contributing to the fit is based on the range of $\tau$ values used as given in <ref> and therefore differs between the . The fit has been performed based on the binning options in <ref>. In these examples the secondary vertex smearing has been neglected. The bands around the oscillations depict the errors of one standard deviation, assuming a Poisson distribution for the event count in each bin. With the so obtained it is possible to compute the probability that a given produced from statistical fluctuations of the null hypothesis $\widehat{\mathcal M}_0$. This probability $p$ can be translated to a significance $Z$ by comparing it to a standard normal distribution. A smaller probability translates to a larger significance and a higher threshold $k_p$. A significance of [5]σ corresponds to a probability of $p \approx \expno[2.87]{-7}$, which for 1 corresponds to a threshold of $k_p \approx 30.94$. See <ref> for the values of the other . Therefore, in the case of 1, a greater than $30.94$ can be interpreted as a discovery, where oscillations have been found with a significance larger than [5]σ. For very small , it is possible that the obtained probability is greater than [50]%. Since for the translation into a significance a standard Gaussian is used, <cit.>, the corresponding significance would be smaller than zero . However, in cases where the are so small, the result should just be interpreted in the way that no oscillations could be proven in the given data. The simulated events and the resulting statistical fit for two example oscillations are shown in <ref>. §.§ Data pre-processing $c \tau_\text{max}/\unit{mm}$ Bins Used events 1 60 30 87 2 60 30 85 3 15 60 42 [Binning parameters] Binning parameters for the different points, as well as the number of events used in the computation of the $\LLR$. Note that for small samples of event surviving the cut based analysis, as it is the case here, the number of events participating in the computation of the $\LLR$ undergoes fluctuations, which contribute to the fluctuations in the obtained significances. Some pre-processing of the events is performed to stabilise the numerics. In principle, it is expected that for sufficiently large proper times alternative hypotheses with a washout parameter greater zero produce the same bin counts as the corresponding null hypotheses. Therefore, the are dominated by small proper times, for which the washout effect is also small. We have found that the fitting algorithm used by us gives more reliable results when we consider only these dominant oscillations. Therefore we do not use all events for the statistical analysis but consider only a window containing the first oscillations. The details of this restriction for each point are given in <ref>. To help reduce noise in the data a Gaussian filter of radius $r = 1$, which corresponds to a standard deviation of $\sigma = \flatfrac{r}{2}$ in bin space, is applied to the bin counts before the maximisation takes place. § RESULTS In this paper, the first ever analysis of at reconstructed level is performed. A detailed statistical analysis is employed to obtain a significance describing the feasibility to resolve the oscillations at the . To that end, three points with different oscillation periods, given in <ref>, have been simulated. All points feature a mean mass of the heavy neutrinos of [14]GeV as well as an active-sterile mixing parameter of $\abs{\theta_\mu}^2 = 10^{-7}$. This leads to a decay width of $\Gamma = \unit[13.8]{\mu eV}$. The parameters are chosen such that the bounds of current collider searches are well evaded. While 3 captures the mass splitting of the minimal linear seesaw model, producing the measured light neutrino data with an inverted hierarchy, 1 and 2 feature a smaller mass splitting as discussed in <ref>. Secondary vertex smearing Secondary vertex smearing for $90$ signal events. Lorentz factor reconstruction error in 3. [Significance as function of vertex smearing and Lorentz factor reconstruction] Panel fig:smearing: Significance of the three points at a luminosity of $\mathscr L = \unit[3]{ab^{-1}}$ as a function of the secondary vertex smearing. Panel fig:reconstruction error: Significance of 3, as function of the number of events surviving the analysis, for three different relative errors of the reconstruction of the Lorentz factor (<ref>). For this comparison no smearing has been taken into account. Since Delphes does not provide the experimental uncertainty to reconstruct secondary vertices, we have introduced a free parameter governing Gaussian smearing of the otherwise perfectly reconstructed secondary vertices. <Ref> shows the expected confidence with which oscillations could be observed in the data for a given smearing of the secondary vertex. With our points, the number of signal events that survive the cuts, shown in <ref>, is roughly $90$. This assumes a luminosity of $\unit[3]{ab^{-1}}$ which is the expected total integrated luminosity of the <cit.>. A factor of two in the number of events can easily be achieved by choosing points closer to the excluded region. Therefore we regard this analysis as conservative. For each data point in the plot, $100$ have been computed to obtain a mean value and a standard deviation for the significance. The figure shows that for an oscillation period of [15]mm in proper time space, corresponding to 1, a significance of [5.19]σ can be expected if no smearing is taken into account. Moreover, the significance is above [5]σ up to a smearing of [2]mm, dropping to [4.46]σ for a smearing of [4]mm. Parameter points with smaller oscillation periods are affected stronger by the smearing. This is expected since a smaller oscillation period in proper time space is related to a smaller oscillation length in lab space. A larger smearing therefore results in a stronger washout for smaller oscillation periods. Lorentz factor reconstruction Since it is crucial to reconstruct the Lorentz factor on an event by event basis, we show in <ref> how a better reconstruction of the Lorentz factor as well as a higher luminosity makes it possible to observe oscillations even for 3. The reconstruction error directly effects the quality of the reconstructed oscillations in the proper time frame. Therefore, if the error is too large for the mass splitting one wants to resolve oscillations for, a higher event numbers only yields a limited improvement. This is shown by the lowest line in <ref>. In contrast the higher lines, that represent a smaller reconstruction error of the Lorentz factor, benefit much more from a larger event number. The quality of the reconstruction of the Lorentz factor is measured using the relative error of the Lorentz factor, which is defined as \begin{equation} \label{eq:gamma error} \Delta \gamma = \frac{\abs{\gamma_\text{gen}^2 - \gamma_\text{reco}^2}}{\gamma_\text{gen}^2 - 1} \,, \end{equation} where $\gamma_\text{gen}$ is the true Lorentz factor of the heavy neutrino and $\gamma_\text{reco}$ is the reconstructed one. The set of events used for the analysis yields an exponential distribution of the relative errors of the Lorentz factor. Most events have a small $\Delta \gamma$, while only a small fraction of events have a large $\Delta \gamma$. The overall quality in the reconstruction can be measured by the standard deviation of that exponential distribution. While a large standard deviation corresponds to many events with a large relative error, and therefore to a poor reconstruction, the opposite is true for a small standard deviation. Without any improvement, with respect to the Lorentz factor achieved in this analysis, the events of 3 yield $\Delta \gamma$'s that result in an exponential distribution with a standard deviation of $0.16$. By improving the reconstruction such that the standard deviation is reduced to $0.0096$, the significance is improved from around zero to $\unit[(3.37 \pm 1.10)]{\sigma}$ for $90$ events. Additionally doubling the number of events yields a significance of $\unit[(5.13 \pm 2.28)]{\sigma}$. This scaling is justified since we assume that it is possible to improve the analysis presented here, such that more signal events survive while the background is still eliminated. Additionally, it is possible to choose a point closer to the experimentally excluded region as mentioned earlier. Furthermore, future collider experiments, such as the ones at the <cit.>, might yield much higher luminosities as well as better reconstruction possibilities of the Lorentz factor of the heavy neutrino. Mass splitting dependency Oscillations can be used to resolve very small mass splittings. However, from the discussion above it is clear that if the mass splitting becomes too large, the oscillation period becomes too small, the reconstruction of oscillation patterns will be challenging. This is shown in <ref> where the oscillation period has been varied, using a fast simulation described below. One can see that larger oscillation periods produce higher significances. It is expected that the significance drops again if the oscillation period reaches the mean lifetime of the heavy neutrinos, since then oscillations can not develop before the mass eigenstates decay. The fast simulation is based on the assumption that the kinematics of the events is independent of the oscillation period. With this assumption, the oscillation period has no impact on the cut based analysis. As a consequence, the sum of and events follows for all points the same distribution, given by the null hypothesis. Thus it is possible to obtain a large sample of valid signal events by combining the events passing all cuts of the three simulated points. It is then possible to give each event a new tag, describing if that events should be counted as or as . For this the of the heavy neutrino is computed on a per event basis, using the relation \begin{equation} \tau = \frac{\abs{\vec d}}{\sqrt{\gamma^2 - 1}}\,, \end{equation} where $\vec d$ is the position of the displaced vertex with respect to the primary vertex. The formula for the oscillation probability can then be used to tag the event based on the new oscillation period. At this point one has generated a sample of valid signal events with the new oscillation period. After that, one can pick a random subset of this sample containing the physical number of events, that can be computed using (<ref>). Subsequently, the analysis to obtain the significance can be applied to this subset of events. This strategy is orders of magnitude faster than performing the full simulation and cut based analysis for each oscillation period separately. § CONCLUSION [Significance as function of the oscillation period] Significance as function of the oscillation period for $90$ events. Three independently simulated points as well as eight points simulated using a fast simulation are depicted. Smearing of the secondary vertices is neglected. In this paper, we have performed a first full analysis of at the reconstructed level. The simulations are based on the FeynRules implementation of the introduced in <cit.>. After the generation of events at parton level with a patched version of MadGraph, hadronisation and showers are simulated using Pythia. Subsequently, a fast detector simulation of the detector has been preformed using Delphes. The uncertainty in the reconstruction of the displaced vertex has been implemented with a smearing function, that randomly selects a value of the displaced vertex around its true value, based on a Gaussian. The analysis of the events is performed using custom and Mathematica code. In our analysis we have focused on three within the , defined in <ref>, with heavy neutrino parameters conservatively chosen inside the region allowed by current experimental constraints. The differ by the heavy neutrino mass splitting, which is largest for 1 and smallest for 3. Simulating events containing heavy hadrons, we have shown that with the event selection rules and cuts as defined in <ref> the corresponding background is completely evaded. It has also been argued that with the given cuts other backgrounds that could not be simulated should also be evaded. Thus, the surviving signal events can be treated as background-free. The statistical method to obtain the significance with which oscillations can be found in the simulated data is described in <ref>. Our analysis shows that for small enough heavy neutrino mass splittings, corresponding to large enough oscillation periods, it is possible to discover with the detector at the assuming [3]ab integrated luminosity. The impact of the oscillation period, the displaced vertex smearing, the number of events, and the error in the reconstruction of the heavy neutrino Lorentz factor on the significance are shown in <ref>. For resolving the , it is important that the smearing is smaller than the oscillation length in lab space. Similarly, the variance of the , due to the error in reconstructing the Lorentz factor, should be smaller than the oscillation period. This is the case for 1, for which a significance of $\unit[(5.01 \pm 0.9)]{\sigma}$ is obtained, assuming a smearing of [2]mm and around $90$ total events relevant for the analysis, <ref>. For smaller oscillation periods, as in 2 and 3, the significance is below [3]σ even if smearing is not taken into account. However, we like to stress that smaller mass splittings might also be resolved with higher significance if the reconstruction of the Lorentz factor is improved. This would not only increase the significance itself but also improve the effect for larger event numbers as shown in <ref>. The event number could, , be increased by choosing a parameter point closer to the experimentally excluded region, <ref>. Additionally, it might be possible to increase the significance for smaller mass splittings by increasing the decay width of the heavy neutrinos, parameter points with increased Yukawa couplings or mass. Then the heavy neutrinos would decay faster and there would be more events in the first oscillation cycles such that resolving the pattern becomes more feasible. However, more events would be lost by the $d_0$ cut in this case. In order to study the interdependence between such considerations, a scan over a larger sample of benchmark parameters is necessary. In addition, the presented study might be improved by more sophisticated background reduction and by optimising the window of considered defined in <ref>. In summary, we have shown that the offers the exciting possibility to not only discover the induced by of long-lived heavy neutrinos, but also to resolve the pattern. Reconstructing the oscillation period would allow to measure this mass splitting and therefore discover the pseudo-Dirac nature of the heavy neutrino pair. This would provide deep insight into the mechanism of neutrino mass generation and could help to shed light on whether leptogenesis is able to generate the baryon asymmetry of the universe (as discussed in <cit.>). §.§ Acknowledgements The work of J.H. as partially supported by the Portuguese through the projects CFTP-Unit UIDB/00777/2020, UIDP/00777/2020, CERN/FIS-PAR/0002/2021, and CERN/FIS-PAR/0019/2021, which are partially funded through POCTI (FEDER), COMPETE, QREN and the EU. S.A. and J.R. acknowledge partial support from the Swiss National Science Foundation grant 200020/175502. § RESIDUAL OSCILLATIONS Impact of a $d_T$ cut oscillation pattern after a $d_0$ cut [Residual oscillations induced by a $d_0$ cut] Effects of the spin correlation sensitive cut $d_0$, in comparison to the non-sensitive cut $d_T$, on a set of $N_{\LNC} + N_{\LNV}$ events. For each of the three datasets, represented by histograms, $\expno[5]{5}$ generator level events, based on the values of 3, have been simulated. All histograms are normalised with the same factor, that ensures that the area under the uncut histogram in fig:dT sums to unity. Panel fig:dT: Comparison of the data before and after a $d_T$ cut of [4]mm. Panel fig:d0: Impact of a $d_0$ cut of [4]mm overlayed by the pattern of oscillations. Due to the angular dependence appearing in (<ref>) the $d_0$ cut does not only generate a delayed onset but additionally imprints an oscillatory pattern originating in the oscillations. The transverse impact parameter $d_0$, calculated for displaced tracks, is not only proportional to the transverse lifetime of the decaying particle, but additionally contains a component depending on the angle between the displaced vertex direction and the observed particle momentum. If the magnetic field can be neglected, which we explicitly checked to be the case for the points discussed in this paper, the transverse impact parameter is given by <cit.> \begin{equation} \label{eq:impact parameter} d_0 = d_T \sin(\varphi(\vec p_T^N, \vec p_T^\mu))\,, \end{equation} where $d_T$ is the transverse distance of the displaced vertex and the sine measures the angle in the transverse plane between the momenta of the heavy neutrino and the muon it decays into. This sine introduces an angular dependency, sensitive to spin correlations in the process under consideration. Since the and processes expose dissimilar spin correlations, the $d_0$ cut effects them differently. This leads to the observation of patterns in event samples that are a priory insensitive to the difference between and processes. As an example, <ref> shows the residual oscillations appearing in a large sample of $N_{\LNC} + N_{\LNV}$ events after introducing a $d_0$ cut. While the event sample with no cuts does not feature any oscillation pattern, it is shown that the $d_0$ cut results in residual oscillations, with peaks aligning with the ones of the oscillation pattern. It can be concluded that the $d_0$ cut affect the events more severely than the ones. By contrast, a cut on $d_T$ is independent of spin correlations and thus no residual oscillations appear. We found that this effect is subdominant for smaller event samples, such as in this analysis, and therefore neglected it in the main part of the paper. Missing 'biblatex' package The bibliography requires the 'biblatex' package. journaltitlePhys. Rev. Lett. titleDirect evidence for neutrino flavor transformation from neutral current interactions in the Sudbury Neutrino Observatory journaltitlePhys. Lett. B title$\mu \to e\gamma$ at a Rate of One Out of $10^9$ Muon Decays? reportnumberPrint-77-0182 (BERN) journaltitleConf. Proc. C titleComplex Spinors and Unified Theories journaltitlePhys. Rev. D titleNeutrino Masses in $\mathrm{SU}(2)\times\mathrm U(1)$ Theories journaltitlePhys. Rev. Lett. titleNeutrino Mass and Spontaneous Parity Nonconservation journaltitleProg. Theor. Phys. titleHorizontal Symmetry and Masses of Neutrinos journaltitlePhys. Rev. D titleNeutrino Decay and Spontaneous Violation of Lepton Number titleTesting sterile neutrino extensions of the Standard Model at future lepton colliders journaltitleInt. J. Mod. Phys. A titleSterile neutrino searches at future $e^-e^+$, $pp$, and $e^-p$ colliders titleSimulating lepton number violation induced by heavy neutrino-antineutrino oscillations at colliders titleSearch for heavy Majorana neutrinos with the ATLAS detector in $pp$ collisions at $ \sqrt{s} = 8$ TeV journaltitlePhys. Rev. Lett. titleSearch for heavy neutral leptons in events with three charged leptons in proton-proton collisions at $\sqrt{s} = 13$ TeV titleHeavy Neutrinos in displaced vertex searches at the LHC and HL-LHC journaltitlePhys. Rev. D titleSearch for long-lived, massive particles in events with a displaced vertex and a muon with large impact parameter in $pp$ collisions at $\sqrt{s} = 13$ TeV with the ATLAS detector titleSearch for long-lived heavy neutral leptons with displaced vertices in proton-proton collisions at $ \sqrt s = 13$ TeV journaltitlePhys. Rev. titleObservation of Long-Lived Neutral $V$ Particles journaltitlePhys. Lett. B titleObservation of $B^0-\bar B^0$ Mixing journaltitlePhys. Rev. Lett. titleObservation of $B^0_s - \bar{B}^0_s$ Oscillations journaltitlePhys. Rev. Lett. titleObservation of $D^0 - \bar D^0$ oscillations journaltitleMod. Phys. Lett. A titleResolvable heavy neutrino–antineutrino oscillations at colliders journaltitleAnnals of Physics titleInterference phenomena of neutral $K$ mesons journaltitlePhys. Rev. D titleOn the Quantum Mechanics of Neutrino Oscillation journaltitlePhys. Rev. D titleOn the treatment of neutrino oscillations without resort to weak eigenstates journaltitlePhys. Rept. titleOscillations of neutrinos and mesons in quantum field theory titleHeavy Neutrino-Antineutrino Oscillations in Quantum Field Theory titleDecoherence effects on lepton number violation from heavy neutrino-antineutrino oscillations journaltitlePhys. Rev. D titleRight-Handed Neutrinos at CERN LHC and the Mechanism of Neutrino Mass Generation subtitlePhenomenological symmetry protected seesaw scenario titleaddonFeynRules model file titleThe fate of hints: updated global analysis of three-flavor neutrino oscillations journaltitlePhys. Rev. Lett. titleNovel supersymmetric SO($10$) seesaw mechanism journaltitleNucl. Phys. B titleNon-universal $Z'$ from SO($10$) GUTs with vector-like family and the origin of neutrino masses titleQuasi-Dirac neutrinos at the LHC journaltitlePhys. Rev. D titleSame Sign versus Opposite Sign Dileptons as a Probe of Low Scale Seesaw Mechanisms journaltitleComput. Phys. Commun. subtitleA complete toolbox for tree-level phenomenology titleFeynRules 2.0 titleThe automated computation of tree-level and next-to-leading order differential cross sections, and their matching to parton shower simulations titleA comprehensive guide to the physics and usage of PYTHIA 8.3 reportnumberLU-TP 22-16,MCNET-22-04,FERMILAB-PUB-22-227-SCD subtitleA modular framework for fast simulation of a generic collider experiment titleDELPHES 3 titleQuasi-stable neutralinos at the LHC titleLong-lived higgsinos as probes of gravitino dark matter at the LHC titleSearch for long-lived particles that decay into final states containing two muons, reconstructed using only the CMS muon chambers journaltitlePhys. Rev. Lett. titleSearch for Displaced Supersymmetry in events with an electron and a muon with large impact parameters collaborationParticle Data Group titleReview of Particle Physics titleHigh-Luminosity Large Hadron Collider (HL-LHC): Technical design report journaltitleEur. Phys. J. ST titleFCC-ee: The Lepton Collider: Future Circular Collider Conceptual Design Report Volume 2 journaltitleEur. Phys. J. ST titleFCC-hh: The Hadron Collider: Future Circular Collider Conceptual Design Report Volume 3 titleProbing Leptogenesis at Future Colliders subtitleThe Lund Monte Carlo for high-$p_T^{}$ physic subtitleA Monte Carlo at next-to-leading order subtitleA Mathematica package to calculate Feynman rules subtitleA framework for fast simulation of a generic collider experiment
# The Characteristic Time Scale of Cultural Evolution Tobias Wand1,2∗ Daniel Hoyer3,4 (1Westfälische Wilhelms-Universität Münster, Insitut für Theoretische Physik 2Center for Nonlinear Science, Münster 3 George Brown College, Toronto 4 Evolution Institute, San Antonio $*$ Corresponding Author<EMAIL_ADDRESS> December 2022, Last Revision July 2023 ) ###### Abstract Numerous researchers from various disciplines have explored commonalities and divergences in the evolution of complex social formations. Here, we explore whether there is a ’characteristic’ time-course for the evolution of social complexity in a handful of different geographic areas. Data from the Seshat: Global History Databank is shifted so that the overlapping time series can be fitted to a single logistic regression model for all 23 geographic areas under consideration. The resulting regression shows convincing out-of-sample predictions and its period of extensive growth in social complexity can be identified via bootstrapping as a time interval of roughly 2500 years. To analyse the endogenous growth of social complexity, each time series is restricted to a central time interval without major disruptions in cultural or institutional continuity and both approaches result in a similar logistic regression curve. Our results suggest that these different areas have indeed experienced a similar course in the their evolution of social complexity, but that this is a lengthy process involving both internal developments and external influences. ##### Keywords: Cliodynamics, Cultural Evolution, Time Scale ## 1 Introduction ### 1.1 Motivation to Find Characteristic Time Scales Researchers from various disciplines have analysed commonalities and divergences in the evolution of complex social systems [1, 2, 3, 4, 5, 6, 7]. The recent emergence of Cliodynamics as a discipline has started the analysis of the dynamics of human societies and states with data-driven scrutiny and modelling approaches from natural sciences [8, 9]. Previous work established that a common set of factors associated with complex social formations typically moved in tandem across a wide variety of regions and time-periods; factors such as social scale, the use of informational media, administrative hierarchies, monetary instruments, and others [6]. These were interpreted as comprising the primary dimension of what could be called ’social complexity’ across cultures, though other dimensions can be adduced as well [3]. Various studies have already discussed or tried to identify the causal drivers of cultural evolution and evaluated the evidence for different theories of why cultures become more complex [7, 10, 11, 12, 13]. Beyond the causal similarities behind cultural evolution across cultures, researchers have also found evidence for temporal similarities and seemingly parallel time scales in the dynamics of various social structures. For example, models for societal collapse have been derived from demographic and fiscal data that show characteristic oscillation periods of a few centuries and a fine structure with a faster periodicity of approximately two human generations [4]. Other theories suggest that cultural evolution leads to the emergence of similar political institutions and schools of thought at roughly identical time intervals across different geographic regions [14, 15, 16]. Another recent study has evaluated the connection between the first emergence of complex societies in different world regions and the age of widespread reliance on agriculture in those areas [17], supporting the theory that agriculture is a necessary condition for the evolution of complex societies. While the time lag between the primary reliance on agriculture and the emergence of states was found to decrease over time, an average time lag of roughly 3,400 years for pristine states suggested the existence of a characteristic time scale, though this was not the explicit focus of that study. Similarly, the study on causal drivers in [7] also found that the time since the adoption of agriculture had a statistically significant effect as a linear predictor variable (called ’AgriLag’) for sociopolitical complexity, providing additional evidence for temporal regularities in the growth of complexity across different cultures and civilisations. Finally, using the same data as our article (cf. section 1.2), it was possible to identify periods of cultural macroevolution with either slow or rapid change in social complexity [5]. Nevertheless, as yet there is no consensus on whether there is a ’typical’ time scale for socio-political development cross-culturally, let alone what that time-course might be. Such characteristic time scales of dynamic systems are, however, well documented in different areas of the natural sciences such as physics and chemistry [18, 19]. Differentiating between fast and slow time scales in a dynamical system can lead to useful insights and can inform modelling assumptions for data analysis [20, 21]. In particular, Haken’s theory of the ’enslaving principle’ [22], according to which the dynamics of fast-relaxing modes are dominated (enslaved) by the behaviour of slowly relaxing modes in a dynamical system, pioneered the research on how dynamics on different time scales influence each other in the same observed system. The existence of temporal regularities among societal dynamics would suggest that cultural evolution not only occurs in similar developmental stages across geographic regions and time periods, but also in similar time intervals. This would add an important dimension to our understanding of how complex social formations evolve, and raise a number of critical questions about what drives these cross-cultural patterns. Here, we adapt some of the methods employed in the natural sciences in an attempt to identify characteristic time scales in the evolution of complex societies. We utilise data collected by the Seshat: Global History Databank [23, 24, 25], a large repository of information about the dynamics of social complexity across world regions from the Neolithic to the early modern period [26]. We find that, despite significant differences in the timing and intensity of major increases in social complexity reached by polities across the Seshat sample, there is a typical, quantitatively identifiable time course recognisable in the data. This result is robust to a variety of checks and covers polities from all major world regions and across thousands of years of history. Our findings offer a novel contribution to the study of cultural evolution, indicating the existence of a general, cross-cultural pattern in both the scale as well as the pace of social complexity development. ### 1.2 Seshat Databank The Seshat: Global History Databank includes systematically coded information on over 35 geographic areas and over 200 variables across up to 10,000 years in time steps of 100 years ([23, 26]; see also publicly available data at http://www.seshatdatabank.info/databrowser/). During the time interval captured by the Seshat databank, these NGAs are occupied by over 370 different identifiable polities, defined as an "independent political unit". This sample is constructed by identifying all known polities that occupied part or all of each NGA over time (see [23, 24, 25] for details). The recorded variables are aggregated into nine complexity characteristics (CCs) and a principal component analysis shows that 77% of the variation in the data can be explained by the first principal component (SPC1), which has almost equal contributions from all nine CCs [6]. In the case of missing data or expert disagreement in [6], multiple imputation [27] was used to create several data sets with the differently imputed values which were aggregated into the principle component analysis. The NGAs in the Seshat data cover a wide geographical range and different levels of social complexity, though it is important to note that the Seshat sample is focused laregly on relatively complex, sedentary societies (but not exclusively). Data on the CCs is sampled at century intervals, giving a time series of each polity’s estimated social complexity measure throughout its duration. Seshat data has allowed researchers to quantitatively test hypotheses on cultural evolution such as identifying drivers of social complexity and predictors of change in military technology, for example gauging the effect of moralising religions on cultural evolution or predicting historical grain yields [7, 28, 29, 30]. Further analysis of the Seshat data includes a discussion of ideas from biological evolutionary theory with respect to the tempo of cultural macroevolution, defined as "rates of change, including their acceleration and deceleration", concluding that "cultural macroevolution is characterized by periods of apparent stasis interspersed by rapid change" [5]. These results strongly relate to the question of the present article, whether there is some generality in the time scale of cultural evolution in the Seshat data. ### 1.3 Data on Culture/Polity Boundaries and Duration Each NGA’s time series can contain data about very different polities that succeeded each other. Sometimes, a gradual and continuous change between the polities justifies treating predecessor and successor polities as closely related; for instance, in the Latium NGA (modern-day central Italy), Seshat records three separate polities for the Roman Republic, indicating the Early, Middle, and Late phases. These phases are culturally and (to a signfiicant degree) institutionally continuous, so can be treated as a single polity- sequence. In other cases, there may have been an invasion or mass migration as a clear break-point between the two polity’s continuity; for instance, between the Ptolemaic Kingdom and Roman Principate polities in the Upper Egypt NGA. Data from [26] and other information recorded in the Seshat sample, notably information on the relationship between polities, is here used to establish a list of continuous polities. The continuity is evaluated either as cultural continuity or as political-institutional continuity and our cutout data for both approaches is published on [31]. ### 1.4 Organisation of this Article Section 2 explains how we transformed the time series data on each NGA in the Seshat sample to establish a common reference point to investigate the time course of changes in social complexity across NGAs. In short, we shift each NGA’s time series with respect to a single anchor time such that the transformed time variable RelTime shows major overlap between the RelTime-vs- SPC1-curves of all NGAs. Exploratory data analysis for the whole dataset reveals that there is a logistic relationship between RelTime and the SPC1 response variable. Section 3 identifies the time scale of growth from the lower to the upper plateau of the logistic curve via bootstrapping. The logistic curve is compared to a regression using only either the culturally or institutionally continuous time series and moreover, the duration of those continuous time series is compared to the estimated characteristic time scale. Finally, the results of the analyses are summarised and discussed in section 4. The mathematical methods and technical details are discussed in the appendices A and B and appendix C gives more details on the used data. ## 2 Approach: Data Transformation and Exploratory Analyses First, all raw SPC1 time series are rescaled via a min-max scaling, i.e. $SPC1=\frac{SPC1_{raw}-\min(SPC1_{raw})}{\max(SPC1_{raw})-\min(SPC1_{raw})}.$ (1) This has the advantage of making the interpretation of high and low SPC1 values much easier as high/low correspond to close to 1 or close to 0, respectively. It also makes the parametrisation of a logistic curve easier by restricting the observed data to a range between 0 and 1. ### 2.1 Anchor Time Considering that most NGAs have an SPC1 time series that starts at a low value barely above $0$ and ends at a high value close to $1$, a logistic regression model seems like a reasonable suggestion for the data. Although all NGAs experience a growth in SPC1 over time, they start at very different calendar years. Therefore, it is necessary to shift the time series via an anchor time so that in the new "relative" time, the growth phase in each NGA’s time series coincide. Then, one logistic regression can be used for all shifted time series (cf. figure 1; also A.1). Hence, each NGA $i$ needs an anchor time $T^{(i)}_{a}$ so that if all time series are shifted by $-T^{(i)}_{a}$, they roughly overlap. The shifted time series of the NGAs and the logistic fit are shown in the main part of figure 1. Figure 1: Main figure: Time series of RelTime vs. SPC1 for all 23 NGAs that cross $\textmd{SPC}1_{0}$ and the logistic regression. Marked in red is the area of growth between the two plateaus of the curve as identified in section 3. The various time series are shown in appendix C in multiple plots to make the identification easier. Insets: a) distribution of SPC1 for all 35 NGAs, the associated KDE (red) and the threshold $\textmd{SPC}1_{0}$ (vertical); b) residuals of the logistic regression. The anchor time can be chosen as the year during which the NGA $i\textmd{'s}$ SPC1 value crosses a threshold value. It has already been shown that there is a clear threshold $\textmd{SPC}1_{0}$ between high and low values of SPC1 in the data, which was used to define the RelTime variable in [29]. A similar methodology was also used in [32], but there, the authors used the emergence of a moralising religious belief as the "year zero" to shift each NGA’s time series. Copying the procedure from [29] to get the RelTime variable, $\textmd{SPC}1_{0}$ is chosen as the minimum between the two maxima in the kernel density estimation (KDE; explained in A.2) of the SPC1 values (figure 1, inset a). The anchor time $T^{(i)}_{a}$ is then selected as the first recorded data point when the NGA $i$ exceeds $\textmd{SPC}1_{0}$. An illustration of the anchor time shift is provided in the appendix B. Thus, the 12 NGAs that never exceed $\textmd{SPC}1_{0}$ are discarded from this analysis. On the one hand, this is not too problematic because their limited growth in SPC1 means that they would have only contributed little information to the estimation of SPC1’s characteristic growth time, but on the other hand, this discards all NGAs from the world region Oceania-Australia in the Seshat sample, meaning that it might introduce a bias. We discuss this and other possible limitations of the approach further in section 4 below. ### 2.2 Logistic Regression The RelTime-vs-SPC1 data is fitted to a logistic regression curve (cf. A.1) via the optimisation algorithm scipy.optimize.curve_fit from [33]. The quality of the regression curve is evaluated with the methods from A.3. With the exception of a few outlier observations occurring in several of the NGAs, all time series qualitatively agree with the regression curve fairly well. Also, the majority of the residuals shown in figure 1 (inset b) are distributed roughly symmetrically in a neighbourhood of zero. The distribution of the residuals and the rather low value of the root-mean-square-error $RMSE\approx 0.11$ both indicate that the logistic regression is a suitable model for the shifted SPC1 data. To further increase our trust in the quality of the regression, it is evaluated via the coefficient of determination $\rho^{2}$ in an out-of-sample prediction. The data is split randomly into equally sized training and testing data sets and a logistic regression curve $f_{i}$ is estimated by only using the training data. Then, $f_{i}$ is used to predict the values for the test data and the prediction is evaluated via the $\rho^{2}$ metric in A.4. The random training-test-split is repeated $i=1,\dots,100$ times, each time using the estimated parameters from the full time series as initial values, and the resulting $\rho^{2}$ values have an average of $\rho^{2}=0.81\pm 0.01$ far above 0 and therefore further strengthens our trust in the logistic model. ## 3 Analysis of Time Scales ### 3.1 Finding a Characteristic Time Scale Having established that the data can be accurately captured by a logistic curve, we can investigate our research question; namely, how many years did it typically take in these different regions to transition from a polity with low SPC1 to one with high SPC1? Or to reformulate the question: when does the curve leave the low plateau and when does it reach the high plateau? We attempt to answer these questions by estimating the heights of the plateaus and their respective uncertainties and by checking when the regression curve crosses these thresholds. We performed 1000 steps of bootstrapping by sampling from the list of NGAs and by estimating the regression parameters $(a_{i},b_{i},c_{i},d_{i})_{i=1,\dots,1000}$ for each sample (cf. A.5). According to the asymptotic behaviour in A.1, the plateaus are given by $b_{i}$ and $a_{i}+b_{i}$. In order to make conservative estimates instead of being influenced by noise, an upper boundary for the lower plateau’s value $Th_{1}$ and a lower boundary for the upper plateau’s value $Th_{2}$ are used as the thresholds. $Th_{1}$ is chosen as $Th_{1}=\mu(b)+3\sigma(b)$ of the bootstrapped distribution of $b$, $Th_{2}$ as $Th_{2}=\mu(a+b)-3\sigma(a+b)$. For each bootstrapped logistic curve $f_{i}(t)$, it is then determined at which RelTime values $t^{(i)}_{1}$ and $t^{(i)}_{2}$ it crosses the lower and upper thresholds $Th_{1}$ and $Th_{2}$. We can then understand the mean value $\mu\left(t^{(i)}_{2}-t^{(i)}_{1}\right)=\mu\left(t^{(i)}_{2}\right)-\mu\left(t^{(i)}_{1}\right)\approx$2500\text{\,}\mathrm{yr}$$ (2) as the characteristic time scale for the period of rapid cultural evolution between low and high plateaus of socio-political complexity, across geography and not in reference to any specific time period. Note that one can also choose less restrictive thresholds via $Th_{1}=\mu(b)+\sigma(b)$ and $Th_{2}=\mu(a+b)-\sigma(a+b)$. With these thresholds, the regression curve leaves the vicinity of the lower plateau rather quickly but needs much longer until it is close enough to the upper plateau to be considered as having reached the upper plateau. These $1\sigma$ thresholds would result in a longer time scale of roughly $\mu\left(t^{(i)}_{2}-t^{(i)}_{1}\middle|1\sigma\right)\approx$4000\text{\,}\mathrm{yr}$.$ (3) We can check the general validity of these results by explicitly identifying for each NGA $i$ the first time $\tau_{1}^{(i)}$ that their SPC1 value exceeds $Th_{1}$ and the first time $\tau_{2}^{(i)}$ they exceed $Th_{2}$. With the exception of the Ghanaian Coast, all NGAs cross $Th_{2}$ and therefore, this procedure yields 22 time durations $d^{(i)}=\tau_{2}^{(i)}-\tau_{1}^{(i)}$. Only for the two NGAs Kachi Plain and Middle Yellow River Valley (two ’pristine’ states, cf. the discussion in section 4.2) does the duration $d^{(i)}$ exceed the 4000 years estimated as an upper boundary in (3). Both the mean (approximately 2200 years) and median (2100 years) are in line with the main estimation in (2). ### 3.2 Continuous Polities There are two reasons why it makes sense to restrict the logistic regression only to a central part of each NGA’s time series, during which the polities in that NGA are not disrupted by external influence or major dislocations in socio-political structures. First, the logistic regression starts at a plateau of low values of SPC1 close to $0$ and ends at a plateau of high values close to $1$. Therefore, even a bad interpolation for the central part can achieve a good $RMSE$, if the plateaus of the high and low tails are sufficiently accurate. However, this would not be a reliable estimation to make an inference on the growth phase in the centre of the curve. Second, if the NGA’s polity is e.g. annexed by another, more developed polity, then it inherits the invading polity’s high SPC1 value and may make a sudden jump in the SPC1 curve. However, the logistic regression here is intended to model steady, uninterrupted growth like in [34] and not major transitions driven by developments experienced elsewhere, as through annexations by an external invader. Therefore, it makes sense to divide each NGA’s time series into intervals which are separated by sharp, discontinuous changes within each NGA and to restrict the analysis of the NGA to its central interval, i.e. to the time series from the polities that cross the $\textmd{SPC}1_{0}$ threshold. As mentioned earlier, there are two ways of identifying such discontinuous changes: either via cultural changes of via major institutional changes of the polity’s governance. Both approaches are analysed separately. The central time series for both methods and their resulting logistic regressions are shown in figure 2. Figure 2: Main figure: estimated logistic curves for the full data and the two cutout methods. Insets: a) each NGA’s central time series and resulting logistic curve for the culturally continuous time series; b) the same as subfigure a) for the institutionally continuous time series. #### 3.2.1 Cultural Continuity One set of sequences was determined by the absence of a major cultural dislocation; namely, the introduction of a new ideological and linguistic system, major population displacement, or major technological advance (the adoption of iron metallurgy, for instance). This is a very broad and lenient definition of continuity, as it allows for very different social formations to be part of a single sequence and can include significant developments. In Egypt, for instance, we treat nearly the entire Pharaonic period (from the Naqada period to the Achaemenid conquest) of over 3000 years, including the so-called Intermediate periods when central rule was fragmented (though many cultural and social features were retained), as a culturally continuous time period. For the 23 NGAs under consideration, the mean value of data points for the culturally continuous central interval is approximately $11.7$, i.e. there is on average a bit more than one millennium of data. While this is much shorter than the characteristic time scale of roughly $2500$ years, the longest continuous time series of the NGAs show a similar length to that of the characteristic time scale (cf. the left half of table 1). Hence, the logistic regression for these cutouts is rather close to the regression of the full data (cf. main part of figure 2) and in particular, the regression curves’ steepness (i.e. their time of growth) is quite similar. #### 3.2.2 Institutional Continuity For institutionally continuous time periods, we follow a similar procedure as above, though with different criteria for continuity leading to shorter sequences. Namely, we break each sequence at any significant political/institutional change, even if there was much continuity in socio- cultural forms. In Egypt, for instance, the institutional sequence starts at the 1st Dynasty period and ends a the end of the Old Kingdom period and the First Intermediate Period, which we call the ’Period of the Regions’. The mean value of data points for institutionally continuous central time series is only $5.9$ and represents approximately 500 years of data. Even the longest continuous sequences now do not last as long as the characteristic growth time of $2500$ years (cf. right half of table 1). Moreover, the logistic regression has only very little data for the parameter estimation (cf. inset b of figure 2) and hence, the logistic regression has a much lower SPC1 level for the upper plateau than the regression to the full data (main part of figure 2), because the cutout time series are too short to reach the high-SPC1 levels. NGA | Cultural Continuity Length | NGA | Institutional Continuity Length ---|---|---|--- Yellow River | 38 | Susiana | 17 Upper Egypt | 33 | Crete | 17 Kachi Plain | 22 | Konya Plain | 15 Susiana | 21 | Upper Egypt | 10 Table 1: For both methods of identifying continuous time sequences, the four longest continuous central time series are shown and the amount of data points they contain (given as their length). The data points are sampled at intervals of one century. The culturally continuous time series are much longer than the longest institutionally continuous sequences. ## 4 Summary and Discussion ### 4.1 Summary Exploratory data analysis shows in figure 1 that the logistic regression is a suitable model for the RelTime-vs-SPC1 time series. Bootstrapping allows us to narrow down the time interval of rapid SPC1 growth to approximately 2500 years, as highlighted in figure 1. Together, these results illustrate that there is a uniform behaviour in growth of social complexity represented by the time evolution of SPC1. If the data is restricted to the central part of each NGA’s time series without any discontinuous cultural or institutional transitions, the logistic regression is still a reasonable model and shows a similar shape to the full data as depicted in figure 2. In particular, the regression based on the culturally continuous time series show a very similar steepness (i.e. growth period) to the full regression curve. ### 4.2 Discussion of the Time Scale and Continuous Sequences Figure 2 shows that the culturally continuous and institutionally continuous time series result in a similar logistic regression to the full data. Notably, table 1 shows that the culturally continuous time series have a much longer duration than the institutionally continuous ones. In particular, in the Yellow River Valley, Upper Egypt, Kachi Plain and Susiana, the culturally continuous time series is approximately as long (or even longer) as the characteristic time scale of SPC1 growth. This is expected for regions that saw the emergence of large, complex states relatively early in history and without any precedent from neighbouring societies – the so-called ’pristine’ or ’primary’ states [35, 36] – which these regions all experienced. However, this is not the case for most other NGAs, indicating that in those NGAs, the growth from the lower to the higher SPC1 plateau did not take place over the course of just one culturally continuous era, but rather included developments across cultural spheres and, in most cases, including developments being ’brought in from the outside’ in the form of direct conquest or more indirect influence. The institutionally continuous time series are all significantly shorter than the characteristic growth time, as expected from the criteria used to generate those sequences. This is notable, as it suggests that in order to transition from low to high social complexity, major shifts in the NGA’s governing institutions are necessary to facilitate the increase in social complexity. In other words, our findings suggest that major transitions in social complexity are not feasible for a single polity to accomplish, but require multiple social formations or ’phases’ of rule building successively (but not monotonically, as the above figures illustrate) on prior developments. Nevertheless, the general similarity of the three regression curves in figure 2 shows that our analysis is stable with respect to the exact selection of time periods and different cutout criteria used. It is interesting to compare those NGAs that crossed the threshold $\textmd{SPC}1_{0}$ to those that failed to do so and stayed at lower complexity values. The latter group had a mean of only 6.4 recorded data points, i.e. there were only complex social formations coded as part of the Seshat sample for a period of roughly six centuries. On the other hand, the NGAs that did reach a high complexity and exceeded the threshold $\textmd{SPC}1_{0}$ had a mean of 57.3 recorded data points, corresponding to almost 6 millennia of observed data. Partly this is explained by different availability of historical and archaeological evidence in different regions, but it suggests also that cultural developments in the low complexity NGAs could have followed the same trajectory of logistic growth, if they had been given enough time. Unfortunately, the necessity to identify an anchor time for this analysis means that all NGAs from the Seshat world region Oceania- Australia had to be discarded for this research. The bias introduced by this has to be kept in mind while interpreting our results. ### 4.3 Interpretation and Comparison to Previous Work With the shifted time index RelTime, the logistic regression model of the SPC1 time series achieves a high accuracy in capturing the evolution of socio- political complexity measured by SPC1. Previous work has already demonstrated a significant amount of cross-cultural generality in the factors contributing to the evolution of socio-political complexity ([6], supplemented by findings in [3, 37]). Notably, a previous study has already identified a characteristic growth pattern of SPC1 and the second principal component SPC2 and found that a rapid period of scale is first followed by a growth of information processing and economic complexity and then by further growth in scale [3]. Here, we expand on this prior work by identifying that the time scales involved in these developments also exhibit a general, characteristic shape. Nevertheless, the evolution of social complexity is a lengthy and non- monotonous process; this emerges clearly from our analyses distinguishing the full regional time-series involved in the transition from low to high thresholds of SPC1 from sequences of cultural or institutional continuity. We see no examples of this evolution accomplished during a single institutionally-continuous sequence. Further, in all NGAs there are noisy periods during which SPC1 grows but also crises during which socio-political complexity sharply declines, only to recover later and continue increasing. These findings highlight both that different parts of the world experienced similar processes of social complexity growth, involving multiple phases of cultural and socio-political structures building off of (and occasionally recovering from) prior developments in each region. While the sample of past societies explored in this article is certainly not exhaustive, they comprise a fairly representative sample of regions from different parts of the world and include societies from different periods, cultures and different developmental experiences. Our results thus lends novel empirical support to the idea (from e.g. [14, 5]) that socio-cultural evolution does indeed occur in similar time scales across different cultures and geographies. Future research can expand these insights by including additional societies and exploring alternate thresholds of complexity to identify anchor times to include more NGAs from the original sample, because the current thresholding procedure in particular excluded some NGAs from modern-day Oceania-Australia from our analysis. In terms of the underlying approach, our study tries to single out the autocatalytic effect of social complexity growth. To this end, it not only focused on one NGA at a time, but also compares our regression results to the culturally and institutionally continuous periods for the respective NGA. Thus, we uncover an empirical pattern in the temporal evolution of SPC1 that has not yet been fully discussed by previous work, e.g in causal analyses of the drivers of social complexity like in [7]. Our methodology differs from e.g. the regression model in [7] by deliberately choosing a very simple model to single out the temporal evolution whilst disregarding possible drivers of the observed dynamics. We believe this approach can be utilised to answer other questions about long-run cultural evolution, for instance the processes by which key technologies (e.g. metallurgy, military technology, communications media etc.) are invented in certain locations and then adopted in others. While the autocatalytic growth model provides an elegant interpretation of our findings (the current level of complexity facilitates further growth until the presence of an upper boundary of complexity is approached), it has to be regarded with caution: We sought as far as possible to disentangle culturally and institutionally endogenous developments from those driven by interactions with other polities, though even the internal developments are not free from external influence. Previous work, for instance, shows the strong effect of military conflicts with other states on the growth of sociopolitical complexity [7, 28]. Hence, the autocatalytic model might be a useful low- dimensional description of the data, but not an exhaustive explanation. In short, our findings exposes a cross-cultural temporal pattern whose causes need to be fleshed out in future work. Finally, the findings of the present article can be used as a benchmark for future additions to the Seshat data: if a new NGA is added to the databank and shows a clear divergence from the logistic curve, it may be prudent to either check, if there are any mistakes in the data generation and interpolation, or if the divergences can be explained by historical developments. Such a benchmark may thus be useful for further expansion of the Seshat databank. #### Author Contributions TW performed all analyses and drafted the manuscript; DH assisted in conceptual development and drafting the manuscript. #### Acknowledgements Initial ideas behind this paper were developed at a workshop held by the Complexity & Collapse Research Group of the Complexity Science Hub, Vienna. The authors thank all the participants at that event, particularly Mateusz Iskrzyński for valuable contributions at early stages of this project. Financial support for this work was provided by the “Complexity Science” research initiative supported by the Austrian Research Promotion Agency FFG under grant #873927 and by the German Academic Scholarship Foundation (Studienstiftung des deutschen Volkes). ## References * [1] David Carballo, Paul Roscoe and Gary Feinman “Cooperation and Collective Action in the Cultural Evolution of Complex Societies” In _Journal of Archaeological Method and Theory_ 21.1, 2014, pp. 98–133 * [2] “Cultural Evolution” Boston: MIT Press, 2013 * [3] Jaeweon Shin et al. “Scale and information-processing thresholds in Holocene social evolution” In _Nature Communications_ 11.1, 2020, pp. 2394 * [4] Peter Turchin “Historical Dynamics: Why States Rise and Fall” Princeton University Press, 2003 * [5] Peter Turchin and Sergey Gavrilets “Tempo and Mode in Cultural Macroevolution” In _Evolutionary Psychology_ 19.4, 2021 * [6] Peter Turchin et al. “Quantitative historical analysis uncovers a single dimension of complexity that structures global variation in human social organization” In _PNAS_ 115.2, 2018, pp. E144–E151 * [7] Peter Turchin et al. “Disentangling the evolutionary drivers of social complexity: A comprehensive test of hypotheses” In _Science Advances_ 8.25, 2022 * [8] Peter Turchin “Arise ’cliodynamics”’ In _Nature_ 454, 2008, pp. 34–35 * [9] Patrick Manning, Pieter François, Daniel Hoyer and Vladimir Zadorozhny “Collaborative Historical Information Analysis”, Chapter 3.09 in Reference Module in Earth Systems and Environmental Sciences, Elsevier, 2018 * [10] V.. Childe “Man Makes Himself” Watts & Company, 1936 * [11] L.. White “The Evolution of Culture” McGraw-Hill, 1959 * [12] Elman R. Service “Origins of the state and civilization” New York, NY: WW Norton, 1975 * [13] Patrick Vinton Kirch “From chiefdom to archaic state: Hawai’i in comparative and historical context” In _How Chiefs Became Kings_ University of California Press, 2010, pp. 1–28 * [14] Oswald Spengler “Der Untergang des Abendlandes” Verlag Braumüller, 1918 * [15] David Engels “Kulturmorphologie und Willensfreiheit” In _Der lange Schatten Oswald Spenglers_ Lüdinghausen: Manuscriptum, 2018, pp. 79–102 * [16] David Engels “Oswald Spengler - Werk, Deutung, Rezeption” Stuttgart: Kohlhammer Verlag, 2021 * [17] Oana Borcan, Ola Olsson and Louis Putterman “Transition to agriculture and first state presence” In _Explorations in Economic History_ 82, 2021, pp. 101404 * [18] Elliot J. Carr “Characteristic time scales for diffusion processes through layers and across interfaces” In _Physical Review E_ 97.4 American Physical Society (APS), 2018 * [19] Eva-Maria Wartha, Markus Bösenhofer and Michael Harasek “Characteristic Chemical Time Scales for Reactive Flow Modeling” In _Combustion Science and Technology_ 193.16 Taylor & Francis, 2021, pp. 2807–2832 * [20] Todd L. Parsons and Tim Rogers arXiv, Dimension reduction for stochastic dynamical systems forced onto a manifold by large drift: a constructive approach with examples from theoretical biology, 2015 DOI: 10.48550/ARXIV.1510.07031 * [21] Clemens Willers and Oliver Kamps “Non-parametric estimation of a Langevin model driven by correlated noise” In _The European Physical Journal B_ 94.7 Springer ScienceBusiness Media LLC, 2021 * [22] Hermann Haken “Synergetik” New York, NY: Springer, 1990 * [23] Peter Turchin et al. “Seshat: The Global History Databank” In _Cliodynamics_ 6.1 California Digital Library (CDL), 2015 * [24] Pieter François et al. “A Macroscope for Global History. Seshat Global History Databank: a methodological overview” In _Digital Humanities Quarterly_ 10.4, 2016 * [25] Peter Turchin et al. “An Introduction to Seshat” Number: 1-2 In _Journal of Cognitive Historiography_ 5.1, 2018, pp. 115–123 * [26] Peter Turchin et al. “seshatdb (Equinox Packaged Data)” Zenodo, 2022 DOI: 10.5281/zenodo.6642230 * [27] Donald B Rubin “Multiple imputation for nonresponse in surveys” John Wiley & Sons, 2004 * [28] Peter Turchin et al. “Rise of the war machines: Charting the evolution of military technologies from the Neolithic to the Industrial Revolution” In _PLOS ONE_ 16.10 Public Library of Science, 2021, pp. 1–23 * [29] Peter Turchin et al. “Explaining the rise of moralizing religions: a test of competing hypotheses using the Seshat Databank” In _Religion, Brain & Behavior_ Routledge, 2022, pp. 1–28 * [30] Peter Turchin et al. “An integrative approach to estimating productivity in past societies using Seshat: Global History Databank” In _The Holocene_ 31.6 SAGE Publications, 2021, pp. 1055–1065 * [31] Tobias Wand and Daniel Hoyer “Seshat: Equinox Release with Continuous Polities” Zenodo, 2023 DOI: 10.5281/zenodo.8120128 * [32] Harvey Whitehouse et al. “Testing the Big Gods hypothesis with global historical data: a review and “retake”” In _Religion, Brain & Behavior_, 2022, pp. 1–43 * [33] Pauli Virtanen et al. “SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python” In _Nature Methods_ 17, 2020, pp. 261–272 * [34] P.-F. Verhulst “Notice sur la loi que la population poursuit dans son accroissement” In _Corresp. Math. Phys._ 10, 1838, pp. 113–121 * [35] Henri..M. Claessen “The emergence of pristine states” In _Social Evolution & History_ 15.1, 2016, pp. 3–57 * [36] Charles S. Spencer “Territorial expansion and primary state formation” In _PNAS_ 107.16, 2010, pp. 7119–7126 * [37] Timothy A Kohler, Darcy Bird and David H Wolpert “Social Scale and Collective Computation” In _Journal of Social Computing_ 3.1 TUP, 2022, pp. 1–17 * [38] David W. Hosmer and Stanley Lemeshow “Applied Logistic Regression” John Wiley & Sons, Ltd, 2000 * [39] Emanuel Parzen “On Estimation of a Probability Density Function and Mode” In _The Annals of Mathematical Statistics_ 33.3 Institute of Mathematical Statistics, 1962, pp. 1065–1076 * [40] Murray Rosenblatt “Remarks on Some Nonparametric Estimates of a Density Function” In _The Annals of Mathematical Statistics_ 27.3 Institute of Mathematical Statistics, 1956, pp. 832–837 * [41] Bradley Efron and Robert J. Tibshirani “An Introduction to the Bootstrap” Springer US, 1994 * [42] Peter Turchin et al. “The Equinox2020 Seshat Data Release” In _Cliodynamics_ 11.1, 2020 ## Appendix A Methods and Technical Details ### A.1 Logistic Regression Curve Logistic regression is used to model time series data which is mostly distributed at two plateaus with a transitory area between them [38]. It is based on the characteristic sigmoid curve of the logistic growth model described in [34], which models an exponential growth process constrained by a carrying capacity. The logistic curve has the functional form $f$ with an asymptotic behaviour $f(x)=\frac{a}{1+\exp(-c(x-d))}+b,\quad f(-\infty)=b\quad\textmd{ and }\quad f(\infty)=a+b.$ (4) Often, data is scaled such that $b=0$ and $a=1$, i.e. an asymptotic behaviour between two binary plateaus at height $0$ and $1$. #### A.1.1 Reversing the Direction Estimating the coefficients $(a,b,c,d)$ can lead to numerical instabilities because it is possible to transform a logistic curve with $c>0$ to an equivalent equation $\hat{f}$ with $c<0$. Consider e.g. $a=1,b=0,c=1$ and $d=0$, then $f(x)=\frac{1}{1+\exp(-x)}=\frac{\exp(x)}{\exp(x)+1}=\frac{\exp(x)+1-1}{\exp(x)+1}=1+\frac{-1}{1+\exp(x)}.$ (5) The last reformulation of $f$ can now be parametrised via $\hat{a}=-1,\hat{b}=1,\hat{c}=-1$ and $\hat{d}=0$. This ambiguity can lead to the regression algorithm yielding positive and negative results for $c$ during multiple runs. This can be prevented by setting an initial parameter guess with $c>0$, which locks the algorithm into positive values for $c$. ### A.2 Kernel Density Estimation (KDE) A KDE tries to reconstruct a probability density function based on a sample $x_{1},\dots,x_{n}$ of measurement data by smoothing the histogram of the data [39, 40]. The estimated density $\hat{\rho}(x)$ is modelled as a weighted sum of probability densities (kernels) centred around the measured $x_{i}$. In this article, the Gaussian density is used as the kernel via scipy.stats.gaussian_kde [33]. ### A.3 Residuals and Root Mean Squared Error For an algorithm $f$ which estimates values $\hat{y}$ from data $X$ with true values $y$, there are several methods to evaluate the accuracy of $f$. One of them is the root mean squared error $RMSE$. It is defined as $RMSE=\sqrt{\frac{1}{n}\sum_{i=1}^{n}r_{i}^{2}}$ (6) via the residuals $r_{i}=\hat{y}_{i}-y_{i}$. An $RMSE$ much smaller than the range of measured values $y_{i}$ means that the model shows only little deviation from the data. A roughly symmetric distribution of the residuals around $0$ indicates that the model does not have a bias towards particular values. ### A.4 Coefficient of Prediction $\rho^{2}$ Another method to evaluate the quality of an estimated function $f$ is the coefficient of prediction $\rho^{2}$ used in [6]. It takes the value of $\rho^{2}=1$, if the prediction is always exactly true, and $\rho^{2}=0$, if the prediction is only as accurate as always using the mean $\bar{y}$. It is defined by $\rho^{2}=1-\frac{\sum_{i=1}^{n}(\hat{y}_{i}-y_{i})^{2}}{\sum_{i=1}^{n}(\bar{y}-y_{i})^{2}}.$ (7) ### A.5 Bootstrapping Bootstrapping is used to estimate standard deviations and confidence intervals in a model-free approach. A sample $z_{1},\dots,z_{n}$ is re-sampled with replacement, i.e. a new sample $\tilde{Z}=z_{i_{1}},\dots,z_{i_{n}}$ is created that for some $j\neq k$ fulfils $i_{j}=i_{k}$. This procedure is repeated $N$ times so that there are $\tilde{Z}_{1},\dots,\tilde{Z}_{N}$ bootstrapped samples. If $N$ is large enough, then e.g. the mean $\tilde{\mu}(z)$ of the re-sampled data will converge to the true mean of the original sample, but the empirical distribution of the re-sampled means $\tilde{\mu}_{1}(z),\dots,\tilde{\mu}_{N}(z)$ enables the calculation of the confidence interval of the empirical mean [41]. This approach can be adapted to make inference on the standard deviation and CIs of any statistical property of the original sample. ## Appendix B Example of the Data Preprocessing We illustrate the preprocessing of the raw $\textmd{SPC}1$ time series using the NGA ’Latium’ (modern day central Italy) as an example. Figure 3 shows the shift between the original time series to the RelTime time frame relative to the anchor time $T^{(Latium)}_{a}=500\textmd{ BC}$. Additionally, this figure depicts how the $\textmd{SPC}1$ time series of Latium is dissected into culturally or institutionally continuous time series intervals of SPC values for the NGA. Note that the time index is given in $RelTime$ for the shifted data and real-world time for the original data. Figure 3: Illustration of how the $\textmd{SPC}1$ time series of the Latium NGA is shifted by its anchor time to the new relative time frame $RelTime$. Hence, at $RelTime=0$, the shifted time series is equal to the threshold value $\textmd{SPC}1_{0}$. The two recorded discontinuities for the Latium NGA (red) are used to divide its $\textmd{SPC}1$ time series into different intervals. The interval containing the threshold value $\textmd{SPC}1_{0}$ (the "Relevant Snippet" with the solid line) is used for further analysis whereas the rest (dotted) is discarded. Note that the times of the discontinuities do not line up with the 100-year time intervals of the $\textmd{SPC}1$ measurement which is why the transition between the black dashed/solid lines does not perfectly align with the red vertical lines. ## Appendix C Detailed Data The SPC1 data from figure 1 is shown in figure 4, but spread out onto several subplots to help the reader identify different NGAs. Figure 4: Spreading the various time series from figure 1 onto multiple plots to make it easier to distinguish the NGAs from each other. ### C.1 Data Collection and Availability As described in the main text, the data used in this paper is derived from [26], supplemented with information provided by the authors. The original data is described in [42]. It was gathered, cleaned, reviewed, and managed by members of the Seshat Databank project following standard project methods, as described in the references cited in the main text. For more details on the social complexity data utilised here and development of the SPC1 values, see especially [6] and [7]. The full dataset used in the analyses presented here is available on [31]. This shows: * • NGA: Name of the NGA * • PolID: Unique identifier for each polity in the sample * • AbsTime: The ’absolute’ or calendar time-point for each SPC1 value. Note that we sample at 100-year intervals, for as many intervals as there are polities in the sample occupying each NGA * • RelTime: The shifted time-series, such that RelTime = 0 in the century interval in each NGA during which SPC1 values crossed the calculated threshold, as described in the main text. Other rows in the NGA are express their relation before or after this threshold, in century intervals. Note that rows that fall outside of the central cultural or institutional sequence are not expressed * • SPC1: Raw SPC1 values for each polity at each century interval, calculated as described in the main text (but not yet min-max scaled) * • Culture.Sequence: Label indicating the central time-series identified as part of the culturally-continuous sequence that surrounds each NGA’s RelTime=0 threshold (labelled ’cultural.continuity’); all other century intervals in the NGA are labelled ’outside.central’ to indicate they fall outside of this central interval sequence * • Institutions.Sequence: Same as above, but indicating the institutionally continuous sequence.
# Reconstruction of a Hypersurface Singularity from its Moduli Algebra João Hélder Olmedo Rodrigues João Hélder Olmedo Rodrigues Instituto de Matemática e Estatística, Universidade Federal Fluminense, Rua Professor Marcos Waldemar de Freitas Reis, Campus Gragoatá, Bloco G - São Domingos 24210-201, Niterói, Rio de Janeiro Brazil<EMAIL_ADDRESS> (Date: August 28, 2024) ###### Abstract. In this paper we characterize ideals of the local ring $\mathscr{O}_{\mathbb{C}^{n},0}$ of germs of holomorphic functions at $0\in\mathbb{C}^{n}$ which arise as the moduli ideal $\langle f,\mathfrak{m}\,j(f)\rangle$, for some $f\in\mathfrak{m}\subset\mathscr{O}_{\mathbb{C}^{n},0}$. A consequence of our characterization is an effective solution to the long standing Reconstruction Problem of the hypersurface singularity from its moduli algebra. ###### Key words and phrases: Hypersurface singularity, Mather–Yau Theorem, Gaffney-Hauser Theorem, Tjurina algebras, Moduli algebras ###### 2010 Mathematics Subject Classification: Primary 32S05, 32S15, 14B05, 14B07, 14H20 Research partially supported by FAPERJ, ARC 211.361/2019 ###### Contents 1. 1 Introduction 2. 2 Preliminaries 3. 3 Computation of $\Delta_{1}$ 4. 4 Two properties of ideals in $\mathscr{O}_{\mathbb{C}^{n},0}$ 5. 5 Main Result ## 1\. Introduction This work comes from our interest in the classification of germs of hypersurface singularities in $(\mathbb{C}^{n},0)$. A germ of a complex hypersurface $(X,0)\subset(\mathbb{C}^{n},0)$ at the origin $0\in\mathbb{C}^{n}$ is defined as the zero set of some - non-trivial - principal ideal $I_{X}$ of the local ring $\mathscr{O}_{\mathbb{C}^{n},0}$, the ring of germs of holomorphic functions at $0\in\mathbb{C}^{n}$. A generator of $I_{X}$ \- which is well-defined, modulo multiplication by an invertible element in $\mathscr{O}_{\mathbb{C}^{n},0}$ \- is said to be an _equation_ for $(X,0)$; if $I_{X}=\langle f\rangle$ we often say that the germ _is defined by_ $f$ and we indicate this fact writing $(X_{f},0)\subset(\mathbb{C}^{n},0)$. We say that two germs of hypersurfaces $(X_{f},0)\subset(\mathbb{C}^{n},0)$ and $(X_{g},0)\subset(\mathbb{C}^{n},0)$ are _biholomorphically equivalent_ if there exist small open neighbourhoods $U$ and $V$ of the origin $0\in\mathbb{C}^{n}$, where $f,\,g$ converge and a (germ of) biholomorphism $\phi:U\rightarrow V$ \- which sends the origin to itself - such that $\phi(X_{f}\cap U)=X_{g}\cap V$. Put more algebraically, it is easy to verify that this holds if and only if there exists an invertible element $u\in\mathscr{O}_{\mathbb{C}^{n},0}$ such that $ug=\phi^{*}(f)$. It is often said in this case that the functions $f,g$ are _contact equivalent_ , because they lie in the same orbit of the action of the _contact group_ $\mathcal{K}$ on $\mathscr{O}_{\mathbb{C}^{n},0}$ (see the book [GLS] for definitions). The totality of germs of hypersurfaces all biholomorphically equivalent to one another is said to be a _biholomorphic class_. One of the most famous numerical invariants of a biholomorphic class is the _multiplicity_ of its elements. Let $\mathfrak{m}$ denote the unique maximal ideal of $\mathscr{O}_{\mathbb{C}^{n},0}$. We recall that if a system $\\{x_{1},\ldots,x_{n}\\}$ of generators of $\mathfrak{m}\subset\mathscr{O}_{\mathbb{C}^{n},0}$ is chosen and if $(X_{f},0)\subset(\mathbb{C}^{n},0)$ is a germ of hypersurface, its multiplicity $\operatorname{mult}(X_{f},0)=\operatorname{mult}(f)$ is the smallest degree $m$ of a non-zero homogeneous polynomial appearing in a series expansion $f(x_{1},\ldots,x_{n})=f_{m}(x_{1},\ldots,x_{n})+f_{m+1}(x_{1},\ldots,x_{n})+\ldots\,\,$ Clearly the multiplicity of a germ doesn’t depend on the choice of $\\{x_{1},\ldots,x_{n}\\}$. So, two biholomorphically equivalent germs of hypersurfaces in $(\mathbb{C}^{n},0)$ have the same multiplicity. Another important numerical invariant of a biholomorphic class is the _Tjurina number_ of its elements. We recall that the Tjurina number of $(X_{f},0)\subset(\mathbb{C}^{n},0)$ is defined to be the complex vector space dimension - whenever it is finite - $\tau(X_{f})$ of the _Tjurina algebra_ of $(X_{f},0)\subset(\mathbb{C}^{n},0)$, which is defined as the quotient algebra $A(X_{f})=\mathscr{O}_{\mathbb{C}^{n},0}/\langle f,j(f)\rangle,$ where $j(f)$ is the ideal generated by the partial derivatives of $f$. It turns out that $\tau(X_{f})$ is finite if and only if $(X_{f},0)\subset(\mathbb{C}^{n},0)$ has an _isolated singularity_ at $0\in\mathbb{C}^{n}$. Notice that having same multiplicity and same Tjurina number are necessary conditions for $(X_{f},0)\subset(\mathbb{C}^{n},0)$ and $(X_{g},0)\subset(\mathbb{C}^{n},0)$ to belong to the same biholomorphic class, but certainly are not sufficient. In fact, the search for a set of invariants separating biholomorphic classes of hypersurface singularities is a problem which seems far from being solved. Given the previous terminology, our research started with a famous result of J. Mather and S. Yau [MY] in the early eighties, which relates certain isomorphism classes of $\mathbb{C}$-algebras to biholomorphic classes of _isolated_ hypersurface singularities. Few years later the main Theorem of [MY] was generalized by T. Gaffney and H. Hauser [GH] to the case of non- isolated hypersurface singularities. These two results, therefore, throw important light into the problem of biholomorphic classification of hypersurface singularities. Let us introduce the notation necessary to explain Mather-Yau and Gaffney- Hauser Theorems: let $(X_{f},0)\subset(\mathbb{C}^{n},0)$ be a germ of hypersurface. We define the _moduli algebra_ of $(X_{f},0)$ \- or more accurately of $f$ \- as the quotient ring $B(X_{f})=B(f)=\mathscr{O}_{\mathbb{C}^{n},0}/\langle f,\mathfrak{m}\,j(f)\rangle.$ The ideal $\langle f,\mathfrak{m}\,j(f)\rangle$ appearing as the denominator will be called the _moduli ideal_ of $(X_{f},0)\subset(\mathbb{C}^{n},0)$ and we denote it as $T_{\mathcal{K}}(f)$. If $g$ is another generator of $I_{X_{f}}$ it is easy to check that $T_{\mathcal{K}}(f)=T_{\mathcal{K}}(g)$ and this shows that $B(X_{f})$ really doesn’t depend on the chosen generator for $I_{X_{f}}$. We observe, more generally, that when $(X_{f},0)$ and $(X_{g},0)$ are biholomorphically equivalent germs of hypersurfaces, then from a relation of type $ug=\phi^{*}(f)$ as above, it is straightforward to check that the moduli algebras $B(X_{f})$ and $B(X_{g})$ are _isomorphic_ as $\mathbb{C}$-algebras. The converse holds but it is much more subtle, being the essential part of the aforementioned results, which are stated here in a simplified manner: ###### Theorem (Mather, Yau; Gaffney, Hauser). Let $(X_{f},0)\subset(\mathbb{C}^{n},0)$ and $(X_{g},0)\subset(\mathbb{C}^{n},0)$ denote two germs of complex hypersurfaces. The statements are equivalent: 1. (1) $(X_{f},0)\subset(\mathbb{C}^{n},0)$ and $(X_{g},0)\subset(\mathbb{C}^{n},0)$ are biholomorphically equivalent; 2. (2) $B(X_{f})$ and $B(X_{g})$ are isomorphic as $\mathbb{C}$-algebras. ###### Remark 1.1. A few comments are in order: * • The original statement of Mather-Yau theorem (cf. [MY]) in the case of isolated hypersurface singularities says that (1) and (2) are also equivalent to “ _$A(X_{f})$ and $A(X_{g})$ are isomorphic as $\mathbb{C}$-algebras_”. In the general case (non-necessarily isolated hypersurface singularities) this is not true, as shown in [GH]. * • In [GH], the authors show, by means of the introduction of a quotient module which plays the role of the moduli algebra $B(X_{f})$, that a similar assertion holds beyond the case of hypersurface singularities. * • In [MY], the authors originally baptised $A(X_{f})$ and $B(X_{f})$ as the _moduli algebras_ of the germ of hypersurface $(X_{f},0)\subset(\mathbb{C}^{n},0)$, because the preceding result tells us that the problem of the classification of germs of (isolated) hypersurfaces singularities $(X_{f},0)\subset(\mathbb{C}^{n},0)$ up to biholomorphic equivalence is equivalent to that of the classification of their moduli algebras $A(X_{f})$ or $B(X_{f})$ up to $\mathbb{C}$-algebra isomorphism. * • The proofs presented in [MY] and [GH] are not constructive and until very recently there was the open problem, called the Reconstruction Problem, of reconstructing the (isolated) hypersurface singularity out of its Tjurina algebra. For solutions in special cases we refer to [Y], [IK], [E] and the very recent [ES]. In [OR] we solved the Reconstruction Problem, at least in the case where the hypersurface can be characterized by its Tjurina algebra. This is precisely the case where the hypersurface singularity is of _Isolated Singularity Type_ (cf. [GH] for details). This case includes - strictly - the case of isolated hypersurface singularities. The main purpose of this paper is to show that we can push our techniques a little further and reconstruct the hypersurface out of $B(X_{f})$, then closing the remaining cases. * • Connected to the Reconstruction problem is the _Recognition problem_ , which is to decide whether a quotient algebra $\mathscr{O}_{\mathbb{C}^{n},0}/I$ is isomorphic to the Tjurina algebra of some hypersurface singularity $(X_{f},0)\subset(\mathbb{C}^{n},0)$. This is of course equivalent to recognize whether the ideal $I$ is a Tjurina ideal $\langle f,j(f)\rangle$ for some $f\in\mathfrak{m}$. This was the approach taken in [OR]. To deal with the problem of reconstruction described above for algebras of type $B(X_{f})$ we introduce right away our guiding question through this paper, namely ###### Problem 1.2. Fixed $n\geqslant 1$, how to find necessary and sufficient conditions for a (proper) ideal $I\,\subset\mathscr{O}_{\mathbb{C}^{n},0}$ to be the moduli ideal of some $f\in\mathfrak{m}$? Clearly, for all $n\geqslant 1$, the zero ideal and the maximal ideal $\mathfrak{m}$ are the moduli ideals of $0$ and $x_{1}$, say, respectively. ###### Example 1.3. If $n=1$ then any ideal $I\subset\mathscr{O}_{\mathbb{C}^{1},0}=\mathbb{C}\\{x_{1}\\}$ is of the form $I=\mathfrak{m}^{k}=\langle x_{1}^{k}\rangle$, $k\geqslant 1$; this is the moduli ideal of $f(x_{1})=x_{1}^{k}$. So, if $n=1$, Problem 1.2 has a trivial solution. However, beginning at $n=2$, this is not true any more. Clearly a necessary condition for an ideal $I\,\subset\mathscr{O}_{\mathbb{C}^{n},0}$ to be the moduli ideal of some $f\in\mathscr{O}_{\mathbb{C}^{n},0}$ is that the minimal number of generators of $I$ should be at most $n^{2}+1$, but this is by no means sufficient as we see in the next ###### Example 1.4. For any $n\geqslant 2$ the ideal $I_{n}=\langle x_{1}^{2},x_{2},x_{3},\ldots,x_{n}\rangle\subset\mathscr{O}_{\mathbb{C}^{n},0}$ is not a moduli ideal. Indeed, since $\dim_{\mathbb{C}}\,\Big{(}\frac{\mathscr{O}_{\mathbb{C}^{n},0}}{I_{n}}\Big{)}=2$ the Tjurina number of a possible $f$ satisfying $I_{n}=T_{\mathcal{K}}(f)$ is at most $2$. Up to contact equivalence there are finitely many $f$ such that $\tau(X_{f})\leqslant 2$, namely $f_{0}=x_{1}$, $f_{1}=x_{1}^{2}+x_{2}^{2}+\ldots+x_{n}^{2}$ and $f_{2}=x_{1}^{3}+x_{2}^{2}+\ldots+x_{n}^{2}$. By direct inspection one checks that the corresponding quotients $\frac{\mathscr{O}_{\mathbb{C}^{n},0}}{T_{\mathcal{K}}(f_{i})}$ have $\mathbb{C}$-vector space dimensions $1$, $n+1$ and $n+2$ respectively. Since $n\geqslant 2$, none has $\mathbb{C}$-vector space dimension $2$. We now briefly describe the contents of the paper. In Section 2, along with some preliminaries on moduli ideals, we introduce and investigate some properties of the set $\Delta_{1}(I)$ of anti-derivatives of an arbitrary ideal $I$, which is a natural set of elements of $\mathscr{O}_{\mathbb{C}^{n},0}$ to look at in the search for a solution $f$ for the equation $I=T_{\mathcal{K}}(f)$. In Section 3 we suggest an easily applicable method for computation of $\Delta_{1}(I)$ in examples, with routines already implemented in SINGULAR [DGPS]. In Section 4 we introduce two easily checkable properties on arbitrary ideals of $\mathscr{O}_{\mathbb{C}^{n},0}$. Our main result, to be presented in Section 5, is a characterization of the ideals $I$ for which the equation $I=T_{\mathcal{K}}(f)$ admits a solution $f$. In other words, we characterize moduli ideals giving an explicit solution for Problem 1.2. In the same Section, we show by an example how to use SINGULAR to recognize a moduli ideal and to reconstruct the hypersurface singularity from it. ### Acknowledgments The author wishes to express his gratitude to T. Gaffney for patiently answering questions related to Mather-Yau type results; and to G.-M. Greuel for bringing to his attention an inaccuracy written in a previous preprint. ## 2\. Preliminaries Let $\mathscr{O}_{\mathbb{C}^{n},0}$ the local ring of germs of holomorphic functions at $0\in\mathbb{C}^{n}$ and let $\mathfrak{m}$ denote its maximal ideal. For some germ of holomorphic function $f\in\mathfrak{m}$ we let $j(f)$ denote the _Jacobian ideal_ of $f$, that is, the ideal generated by the (first order) partial derivatives of $f$ with respect to a chosen coordinate system $\\{x_{1},\ldots,x_{n}\\}$ \- minimal set of generators - for $\mathfrak{m}$. As indicated in the Introduction, we will be concerned with moduli ideals $T_{\mathcal{K}}(f)=\langle f,\mathfrak{m}j(f)\rangle$ of functions $f\in\mathfrak{m}$. It is easy to verify that for any biholomorphic change of coordinates $\phi:(\mathbb{C}^{n},0)\rightarrow(\mathbb{C}^{n},0)$ and any $f\in\mathscr{O}_{\mathbb{C}^{n},0}$, we have $T_{\mathcal{K}}(\phi^{*}(f))=\phi^{*}(T_{\mathcal{K}}(f))$. Hence we will fix a coordinate system and we will always compute $T_{\mathcal{K}}(f)$ with respect to this coordinate system. The properties below of $T_{\mathcal{K}}$ are immediate to verify: ###### Remark 2.1. For any $f,g\in\mathscr{O}_{\mathbb{C}^{n},0}$ we have: (i) $T_{\mathcal{K}}(f+g)\subseteq T_{\mathcal{K}}(f)+T_{\mathcal{K}}(g)$; (ii) $T_{\mathcal{K}}(fg)\subseteq T_{\mathcal{K}}(f)\,T_{\mathcal{K}}(g)$; (iii) $T_{\mathcal{K}}(f)=T_{\mathcal{K}}(g)$ if $\langle f\rangle=\langle g\rangle$. We will find it convenient to introduce the definition below: ###### Definition 2.2. Let $J\subset\mathscr{O}_{\mathbb{C}^{n},0}$ the ideal generated by $g_{1},\ldots,g_{q}\in\mathscr{O}_{\mathbb{C}^{n},0}$. We define the ideal $T_{\mathcal{K}}(J)=T_{\mathcal{K}}(g_{1})+\ldots+T_{\mathcal{K}}(g_{q})\subseteq\mathscr{O}_{\mathbb{C}^{n},0}.$ Notice that this is a well-posed definition since $T_{\mathcal{K}}(J)$ does not depend on the generators $g_{i}$ chosen but only on the ideal they generate, as the reader can check using $(i),\,(ii)$ in Remark 2.1. Although Problem 1.2 mentions moduli ideals of functions (or rather of principal ideals, cf.(iii) of Remark 2.1), it seems fair to refer to $T_{\mathcal{K}}(J)$ as the _moduli ideal_ of $J$. ###### Remark 2.3. The properties below are easy to verify. * (i) $J\subseteq T_{\mathcal{K}}(J)$; * (ii) If $J_{1}\subseteq J_{2}$ is an inclusion of ideals then $T_{\mathcal{K}}(J_{1})\subseteq T_{\mathcal{K}}(J_{2})$; * (iii) If $\\{J_{\lambda}\\}_{\lambda\in\Lambda}$ is any family of ideals of $\mathscr{O}_{\mathbb{C}^{n},0}$ then $T_{\mathcal{K}}(\sum_{\lambda}J_{\lambda})=\sum_{\lambda}T_{\mathcal{K}}(J_{\lambda})$; * (iv) $T_{\mathcal{K}}(J_{1}\cap J_{2})\subseteq T_{\mathcal{K}}(J_{1})\cap T_{\mathcal{K}}(J_{2})$. Now we introduce the main object - which is an adapted version of the object used in [OR] and - that will ultimately lead us to our solution to Problem 1.2. ###### Definition 2.4. Let $I\subset\mathscr{O}_{\mathbb{C}^{n},0}$ be an ideal. We define the _ideal of anti-derivatives_ of $I$ as $\Delta_{1}(I)=\\{f\in\mathscr{O}_{\mathbb{C}^{n},0}\,|\,T_{\mathcal{K}}(f)\subseteq I\\}.$ It is easy to show that $\Delta_{1}(I)$ does not depend on the choice of parameters $\\{x_{1},\ldots,x_{n}\\}$ used to compute Jacobian ideals. Moreover, the properties below are straightforward to check. ###### Remark 2.5. $\Delta_{1}(I)$ is an ideal of $\mathscr{O}_{\mathbb{C}^{n},0}$; $I^{2}\subseteq\Delta_{1}(I)\subseteq I$; If $\\{I_{\lambda}\\}_{\lambda\in\Lambda}$ is any family of ideals of $\mathscr{O}_{\mathbb{C}^{n},0}$ then $\Delta_{1}(\bigcap_{\lambda}I_{\lambda})=\bigcap_{\lambda}\Delta_{1}(I_{\lambda})$; If $I\subseteq J$ is an inclusion of ideals in $\mathscr{O}_{\mathbb{C}^{n},0}$ then $\Delta_{1}(I)\subseteq\Delta_{1}(J)$; ###### Example 2.6. Let $I=\langle f^{k}\rangle\subset\mathscr{O}_{\mathbb{C}^{n},0}$, $n\geqslant 2$, be a principal ideal with $k\geqslant 1$ and $f$ being an irreducible element. Then $\Delta_{1}(I)=\langle f^{k+1}\rangle$. Indeed, one inclusion is immediate. For the opposite inclusion, assume $g\in\Delta_{1}(I)$ and write $g=af^{k}$, for some $a\in\mathscr{O}_{\mathbb{C}^{n},0}$. For all $i,j=1,\ldots,n$, $x_{i}\frac{\partial g}{\partial x_{j}}=x_{i}f^{k}.\frac{\partial a}{\partial x_{j}}+akf^{k-1}x_{i}\frac{\partial f}{\partial x_{j}}$ is also a multiple of $f^{k}$. Hence $f$ must divide $ax_{i}\frac{\partial f}{\partial x_{j}}$, for all $i,j=1,\ldots,n$. Since $f$ is irreducible, for every $i,j=1,\ldots,n$, $f$ must divide $x_{i}$ or $\frac{\partial f}{\partial x_{j}}$ or $a$. If $f$ divides $\frac{\partial f}{\partial x_{j}}$ for all $j=1,\ldots,n$ then $\langle\frac{\partial f}{\partial x_{1}},\ldots,\frac{\partial f}{\partial x_{n}}\rangle\subseteq\langle f\rangle$ and we would deduce that $f=0\in\mathscr{O}_{\mathbb{C}^{n},0}$, which is not the case. Since $n\geqslant 2$, $f$ cannot divide all the $x_{i}$. So, for some $i,j$, $f$ does not divide $x_{i}\frac{\partial f}{\partial x_{j}}$. Being an irreducible element, $f$ must divide $a$. Hence $g\in\langle f^{k+1}\rangle$. ###### Example 2.7. Let $I=\langle f\rangle\subset\mathscr{O}_{\mathbb{C}^{n},0}$, $n\geqslant 2$, be a non-trivial principal ideal and $f=f_{1}^{k_{1}}\ldots f_{r}^{k_{r}}$ be a factorization of $f$ into irreducible, non-associated, elements with positive $k_{1},\ldots,k_{r}$. Then $I=\langle f_{1}^{k_{1}}\rangle\cap\ldots\cap\langle f_{r}^{k_{r}}\rangle$. It follows from Remark 2.5, (iii) and Example 2.6 that $\Delta_{1}(I)=\langle f_{1}^{k_{1}+1}\rangle\cap\ldots\cap\langle f_{r}^{k_{r}+1}\rangle=\langle f_{1}^{k_{1}+1}\ldots f_{r}^{k_{r}+1}\rangle$. ###### Example 2.8. Let $\mathfrak{m}^{k}$, with $k\geqslant 1$, denote the $k$-th power of the maximal ideal of $\mathscr{O}_{\mathbb{C}^{n},0}$. Let $g=x_{1}^{k_{1}}\ldots x_{n}^{k_{n}}$ with $k_{j}\geqslant 0$ and $\sum k_{j}=k$ be a monomial generator of $\mathfrak{m}^{k}$. For any pair $i,j=1,\ldots,n$ it is easy to check that $x_{i}\frac{\partial g}{\partial x_{j}}\in\mathfrak{m}^{k}$. Hence $T_{\mathcal{K}}(g)\subset\mathfrak{m}^{k}$, which shows $\Delta_{1}(\mathfrak{m}^{k})=\mathfrak{m}^{k}$. ###### Example 2.9. Let $I_{n}=\langle x_{1}^{2},x_{2},\ldots,x_{n}\rangle\subset\mathscr{O}_{\mathbb{C}^{n},0}$, $n\geqslant 2$, as in Example 1.4. Then $\Delta_{1}(I_{n})=\mathfrak{m}^{2}$. Indeed, since $\mathfrak{m}^{2}\subset I_{n}$ then according to Remark 2.5, item (iv) and the previous example we have $\mathfrak{m}^{2}=\Delta_{1}(\mathfrak{m}^{2})\subseteq\Delta_{1}(I_{n})$. For the other inclusion, if $g=a_{1}x_{1}^{2}+a_{2}x_{2}+\ldots+a_{n}x_{n}\in\Delta_{1}(I_{n})$, computing $\frac{\partial g}{\partial x_{j}}$, we see that $a_{j}\in(I_{n}:\mathfrak{m})=\mathfrak{m}$, for all $j\geqslant 2$. Hence $a_{2}x_{2}+\ldots+a_{n}x_{n}\in\mathfrak{m}^{2}$ and we conclude that $g\in\mathfrak{m}^{2}$. ## 3\. Computation of $\Delta_{1}$ Up to now we have obtained the ideal of anti-derivatives $\Delta_{1}(I)$ in a few specific cases (cf. Examples 2.7, 2.8, 2.9). In this section we suggest a method to compute $\Delta_{1}(I)$ for any ideal $I\subset\mathscr{O}_{\mathbb{C}^{n},0}$, then showing that it is accessible in concrete computations. We will illustrate our procedure with examples obtained using basic routines already implemented in the software SINGULAR, [DGPS]. Fix a basis $\\{x_{1},\ldots,x_{n}\\}$ of the maximal ideal $\mathfrak{m}\subset\mathscr{O}_{\mathbb{C}^{n},0}=\mathbb{C}\\{x_{1},\ldots,x_{n}\\}$. A free $\mathscr{O}_{\mathbb{C}^{n},0}$-module of certain rank $\ell$ will be denoted by $F_{\ell}$; for any $\underline{b}=(b_{1},\ldots,b_{\ell})^{t}$, $\underline{c}=(c_{1},\ldots,c_{\ell})^{t}\in F_{\ell}$ we put $\underline{b}\cdot\underline{c}$ to denote $\sum_{k=1}^{\ell}b_{k}c_{k}\in\mathscr{O}_{\mathbb{C}^{n},0}$. Assume the ideal $I$ given by generators $I=\langle f_{1},\ldots,f_{\ell}\rangle$. Then some element $g=\underline{a}\cdot\underline{f}=\sum_{k}a_{k}f_{k}\in I$ is an element of $\Delta_{1}(I)$ if and only if, for all $i,j=1,\ldots,n$, $x_{i}\frac{\partial g}{\partial x_{j}}=\sum_{k}a_{k}x_{i}\frac{\partial f_{k}}{\partial x_{j}}+\sum_{k}\frac{\partial a_{k}}{\partial x_{j}}x_{i}f_{k}=\underline{a}\cdot x_{i}\frac{\partial\underline{f}}{\partial x_{j}}+\frac{\partial\underline{a}}{\partial x_{j}}\cdot x_{i}\underline{f}$ belongs to $I$. In other words, $g=\underline{a}\cdot\underline{f}\in\Delta_{1}(I)$ if and only if $\underline{a}\cdot x_{i}\frac{\partial\underline{f}}{\partial x_{j}}\in I,$ for all $i,j=1,\ldots,n$. For each $i,j=1,\ldots,n$ we denote by $E^{i,j}$ the submodule of $F_{\ell}$ consisting of the elements $\underline{a}\in F_{\ell}$ such that $\underline{a}\cdot x_{i}\frac{\partial\underline{f}}{\partial x_{j}}\in I$. Then $E:=\bigcap_{i,j=1}^{n}\,E^{i,j}\subseteq F_{\ell}$ is a finitely generated $\mathscr{O}_{\mathbb{C}^{n},0}$-submodule of $F_{\ell}$ consisting of all elements $\underline{a}\in\,F_{\ell}$ such that $g=\underline{a}\cdot\underline{f}\in\Delta_{1}(I)$. Notation being as above we have ###### Proposition 3.1. Let $\underline{e}_{\,1},\ldots,\underline{e}_{\,q}\in F_{\ell}$ be generators of $E$. Then the ideal of anti-derivatives $\Delta_{1}(I)$ is generated by $\underline{e}_{\,t}\cdot\underline{f}$, for $t=1,\ldots,q$. ###### Proof. As already discussed, $g\in\Delta_{1}(I)$ if and only if $g=\underline{a}\cdot\underline{f}$ for some $\underline{a}\in E$. Since $E$ is generated by all the $\underline{e}_{\,t}$, the result follows from $\mathscr{O}_{\mathbb{C}^{n},0}$-linearity of the $(-)\cdot\underline{f}$ product. ∎ Here we show how we used the software SINGULAR, [DGPS] to compute the ideal of anti-derivatives of a given ideal. We claim no originality, since only routines already implemented by the software developers and collaborators were applied. ###### Example 3.2. Let $I=\langle yz,z^{3},xw,w^{2}\rangle\subset\mathscr{O}_{\mathbb{C}^{4},0}=\mathbb{C}\\{x,y,z,w\\}$. With the interface of SINGULAR open, type > ring r=0,(x,y,z,w),ds; This declares you are working over a field of characteristic zero, variables $x,y,z,w$ and set the corresponding ring of power series. We now declare the generators of the ideal $I$ by means of a matrix with one row and (in the present case) four columns: type > matrix B[1][4]=yz,z3,xw,w2; Now to compute the submodule $E^{1,1}$, (same notation as above), we declare a matrix with entries the partial derivatives of the given generators in terms of $x$ multiplied by $x$. We compute $E^{1,1}$ as follows: > matrix e11[1][4]=x*diff(B,x); def E11=modulo(e11,B); Likewise, compute $E^{1,2}$, $E^{1,3}$, $E^{1,4}$, $E^{2,1}$, $E^{2,2}$, $E^{2,3}$, $E^{2,4}$, $E^{3,1}$, $E^{3,2}$, $E^{3,3}$, $E^{3,4}$, $E^{4,1}$, $E^{4,2}$, $E^{4,3}$, $E^{4,4}$: > matrix e12[1][4]=x*diff(B,y); def E12=modulo(e12,B); > matrix e13[1][4]=x*diff(B,z); def E13=modulo(e13,B); ⋮ > matrix e44[1][4]=w*diff(B,w); def E44=modulo(e44,B); Now, we put > def > e=intersect(E11,E12,E13,E14,E21,E22,E23,E24,E31,E32,E33,E34,E41,E42,E43,E44); > def E=std(e); This defines $E$ and computes a standard basis, with respect to the given monomial order. Now, to obtain $E$, type > print(E); SINGULAR gives $\left(\begin{array}[]{ccccccccccccccc}0&0&0&0&0&0&0&0&yz&0&z^{2}&xw&zw&w^{2}&0\\\ 0&y&z&0&0&w&0&0&0&0&0&0&0&0&0\\\ 0&0&0&0&w&0&0&yz&0&0&0&0&0&0&z^{3}\\\ x&0&0&w&0&0&yz&0&0&z^{2}&0&0&0&0&0\\\ \end{array}\right)$ Hence, as in Proposition 3.1, we obtain generators of $\Delta_{1}(I)$ computing $(-)\cdot(yz,z^{3},xw,w^{2}),$ for all (transposed) columns of the previous matrix. We speed up calculations typing > ideal d=B*E; > ideal D=std(d); > D; The output lists the generators of $\Delta_{1}(I)$. In the case under consideration we obtain the ten generated monomial ideal below: $\Delta_{1}(I)=\langle xw^{2},w^{3},y^{2}z^{2},yz^{3},z^{4},xyzw,yz^{2}w,z^{3}w,yzw^{2},z^{2}w^{2}\rangle.$ ###### Example 3.3. Let $I=\langle 3xy^{2}+x^{6},y^{3},x^{5}y,x^{7}\rangle\subset\mathscr{O}_{\mathbb{C}^{2},0}=\mathbb{C}\\{x,y\\}$. Declare the ring and the generators of the ideal $I$ as > ring r=0,(x,y),ds; > matrix B[1][4]=3xy2+x6,y3,x5y,x7; Now we compute the submodules $E^{1,1}$, $E^{1,2}$, $E^{2,1}$, $E^{2,2}$: > matrix e11[1][4]=x*diff(B,x); def E11=modulo(e11,B); > matrix e12[1][4]=x*diff(B,y); def E12=modulo(e12,B); > matrix e21[1][4]=y*diff(B,x); def E21=modulo(e21,B); > matrix e22[1][4]=y*diff(B,y); def E22=modulo(e22,B); As before we define $E$ and compute a standard basis: > def e=intersect(E11,E12,E21,E22); > def E=std(e); > print(E); In this case, SINGULAR gives $\left(\begin{array}[]{cccccc}0&0&0&0&y&x^{3}\\\ 0&1&x&y&0&0\\\ 0&1&0&0&0&0\\\ 1&0&0&0&0&0\\\ \end{array}\right)$ Again, as in Proposition 3.1, generators of $\Delta_{1}(I)$ are obtained computing $(-)\cdot(3xy^{2}+x^{6},y^{3},x^{5}y,x^{7}),$ for all (transposed) columns of the previous matrix. We do this as > ideal d=B*E; > ideal D=std(d); > D; We obtain $\Delta_{1}(I)=\langle y^{3}+x^{5}y,x^{4}y^{2},x^{7},x^{6}y\rangle.$ ## 4\. Two properties of ideals in $\mathscr{O}_{\mathbb{C}^{n},0}$ In this section we introduce the two relevant conditions present in our characterization (see Theorem 5.1) of moduli ideals of hypersurface singularities in $\mathscr{O}_{\mathbb{C}^{n},0}$. ### $T_{\mathcal{K}}$-fullness Here we introduce $T_{\mathcal{K}}$-fullness, a quite natural notion related to our Problem 1.2. Recall the definition of the moduli ideal of an arbitrary ideal of $\mathscr{O}_{\mathbb{C}^{n},0}$ (cf. Definition 2.2) and observe that, in general, we have $T_{\mathcal{K}}(\Delta_{1}(I))\subseteq I$. ###### Definition 4.1. Let $I$ be an ideal of $\mathscr{O}_{\mathbb{C}^{n},0}$. We say that $I$ is _$T_{\mathcal{K}}$ -full_ if $T_{\mathcal{K}}(\Delta_{1}(I))=I$. ###### Example 4.2. Let $I_{n}=\langle x_{1}^{2},x_{2},\ldots,x_{n}\rangle\subset\mathscr{O}_{\mathbb{C}^{n},0}$, $n\geqslant 2$ as in Example 1.4. We have seen in Example 2.9 that $\Delta_{1}(I_{n})=\mathfrak{m}^{2}$. A routine calculation now shows that $T_{\mathcal{K}}(\Delta_{1}(I_{n}))=T_{\mathcal{K}}(\mathfrak{m}^{2})=\mathfrak{m}^{2}\subsetneq I_{n}$ so $I_{n}$ is not $T_{\mathcal{K}}$-full. The general significance of $T_{\mathcal{K}}$-fullness in our investigation is apparent in the next result, which is, in view of Example 4.2, an alternative proof that $I_{n}\subset\mathscr{O}_{\mathbb{C}^{n},0}$, as in Example 1.4, is not a moduli ideal. ###### Proposition 4.3. Let $I$ be an ideal of $\mathscr{O}_{\mathbb{C}^{n},0}$. If $I$ is a moduli ideal then $I$ is $T_{\mathcal{K}}$-full. ###### Proof. Let $\Delta_{1}(I)=\langle g_{1},\ldots,g_{q}\rangle$. If $I$ is a moduli ideal, then there exists $f\in\Delta_{1}(I)$ such that $I=T_{\mathcal{K}}(f)$. We may write $f=\sum_{k=1}^{q}r_{k}g_{k}$, for some $r_{k}\in\mathscr{O}_{\mathbb{C}^{n},0}$. Using Remark 2.1, we obtain $I=T_{\mathcal{K}}(f)\subseteq\sum_{k=1}^{q}T_{\mathcal{K}}(r_{k}g_{k})\subseteq\sum_{k=1}^{q}T_{\mathcal{K}}(g_{k})=T_{\mathcal{K}}(\Delta_{1}(I))\subseteq I$ and equality holds throughout. We conclude that $I$ is $T_{\mathcal{K}}$-full. ∎ ###### Example 4.4. (Example 3.2, continued) We have computed the anti-derivatives ideal of $I=\langle yz,z^{3},xw,w^{2}\rangle\subset\mathbb{C}\\{x,y,z,w\\}$ as $\Delta_{1}(I)=\langle xw^{2},w^{3},y^{2}z^{2},yz^{3},z^{4},xyzw,yz^{2}w,z^{3}w,yzw^{2},z^{2}w^{2}\rangle$. We check easily that $T_{\mathcal{K}}(\Delta_{1}(I))\subsetneq I$. Hence $I$ is not a moduli ideal because it is not $T_{\mathcal{K}}$-full. Being $T_{\mathcal{K}}$-full is, however, not sufficient for $I$ to be a moduli ideal. ###### Example 4.5. Let $k\geqslant 5$ and let $I=\mathfrak{m}^{k}\subset\mathscr{O}_{\mathbb{C}^{2},0}=\mathbb{C}\\{x,y\\}$. Certainly $I$ is not a moduli ideal because it is minimally generated by more than $5=2^{2}+1$ elements. However, we have seen (cf. Example 2.8) that $\Delta_{1}(I)=I$. It is equally easy to check that $T_{\mathcal{K}}(\Delta_{1}(I))=I$, which implies that $I$ is $T_{\mathcal{K}}$-full. ### $T_{\mathcal{K}}$-dependence We have shown above that $T_{\mathcal{K}}$-fullness is not sufficient for an ideal to be a moduli ideal. Here we explain the last ingredient which is needed, in addition to $T_{\mathcal{K}}$-fullness (cf. Definition 4.1), to characterize moduli ideals. We begin with some examples. ###### Example 4.6. Consider the ideals $\mathfrak{m}^{k}\subset\mathscr{O}_{\mathbb{C}^{2},0}=\mathbb{C}\\{x,y\\}$, $k\geqslant 1$. All of them are $T_{\mathcal{K}}$-full and Example 4.5 has shown that, if $k\geqslant 5$, they are not moduli ideals. For $k\leqslant 4$, we check easily that $\mathfrak{m}$, $\mathfrak{m}^{2}$, $\mathfrak{m}^{3}$ are moduli ideals of $x,\,xy,\,x^{3}+y^{3}$, respectively. We claim that $\mathfrak{m}^{4}$ is not a moduli ideal of any $f\in\mathbb{C}\\{x,y\\}$. Indeed, since $\dim_{\mathbb{C}}\,\Big{(}\frac{\mathbb{C}\\{x,y\\}}{\mathfrak{m}^{4}}\Big{)}=10$, the Tjurina number $\tau=\tau(X_{f})$ of a possible $f\in\mathbb{C}\\{x,y\\}$ such that $T_{\mathcal{K}}(f)=\mathfrak{m}^{4}$ is at most $10$. After a check to the Arnold’s lists in [Ar] we see that any $f$ with $\tau\leqslant 10$ is quasi-homogeneous; it follows that $\dim_{\mathbb{C}}\,\Big{(}\frac{\mathbb{C}\\{x,y\\}}{T_{\mathcal{K}}(f)}\Big{)}=\tau+2$. We deduce that the possible $f$ must define $A_{8}$, $D_{8}$ or $E_{8}$ singularities. Up to contact equivalence we can compute with the respective normal forms $x^{2}+y^{9}$, $x^{2}y+y^{7}$ and $x^{3}+y^{5}$ to obtain $T_{\mathcal{K}}(f)$ as $\langle x^{2},xy,y^{9}\rangle$, $\langle x^{3},x^{2}y,xy^{2},y^{7}\rangle$ and $\langle x^{3},x^{2}y,xy^{4},y^{5}\rangle$ respectively. All of them have elements of multiplicity smaller than $4$, opposite to $\mathfrak{m}^{4}$. We conclude that $\mathfrak{m}^{4}$ is not a moduli ideal but the reason is not anymore its minimal number of generators. We close this Example observing that the function germs $x,\,xy,\,x^{3}+y^{3}$ of which $\mathfrak{m}$, $\mathfrak{m}^{2}$, $\mathfrak{m}^{3}$ are moduli ideals are not the more general possible: indeed in these cases, _general_ $\mathbb{C}$-linear combinations of the generators of $\Delta_{1}(I)$ are function germs $f$ satisfying $I=T_{\mathcal{K}}(f)$ in each case, while in the case $I=\mathfrak{m}^{4}$ _every_ linear combination satisfies $T_{\mathcal{K}}(f)\subsetneq I$. We aim to give a precise meaning to the intuition, that of a general linear combination $f$ of the given generators of $\Delta_{1}(I)$ has the largest possible $T_{\mathcal{K}}(f)$ and could reveal whether a given ideal $I$ is a moduli ideal. Our results and definitions in this subsection are expressed geometrically, since it seemed to us more appropriate to explain these ideas. To this end, we use basic concepts on schemes, consistent with [H], Chapter II, to which we refer for terminology. We regard ideals in $\mathscr{O}_{\mathbb{C}^{n},0}$ as ideal sheaves on the affine scheme $\mbox{Spec}\,\mathscr{O}_{\mathbb{C}^{n},0}$. Let $J\subset\mathscr{O}_{\mathbb{C}^{n},0}$ be any ideal and assume that $J$ is given by generators: $J=\langle g_{1},\ldots,g_{q}\rangle\subset\mathscr{O}_{\mathbb{C}^{n},0}$; then we consider the projective $(q-1)$-space over $\mathscr{O}_{\mathbb{C}^{n},0}$, namely, $\mathbb{P}^{q-1}=\mbox{Proj}(S)$, where $S=\mathscr{O}_{\mathbb{C}^{n},0}[\alpha_{1},\ldots,\alpha_{q}]=\bigoplus_{d\geqslant 0}\,S_{d}$ is the standard polynomial ring over $\mathscr{O}_{\mathbb{C}^{n},0}$, with variables $\alpha_{i}$, graded so that $\deg\alpha_{i}=1$, for all $i$. Let $\pi:\mathbb{P}^{q-1}\rightarrow\mbox{Spec}\,\mathscr{O}_{\mathbb{C}^{n},0}$ be the natural morphism of $\mbox{Spec}\,\mathscr{O}_{\mathbb{C}^{n},0}$-schemes and let $\sigma$ denote the global section $\sum_{i=1}^{q}\,g_{i}\alpha_{i}$ of $\pi^{*}(J)\otimes\mathcal{O}_{\mathbb{P}^{q-1}}(1)$. Then there is a _moduli ideal sheaf_ of $\sigma$ on $\mathbb{P}^{q-1}$, namely $\mathscr{T}_{\mathcal{K}}(\sigma)$, the sheaf associated to the homogeneous ideal $\langle\sigma,\mathfrak{m}\frac{\partial\sigma}{\partial x_{1}},\ldots,\mathfrak{m}\frac{\partial\sigma}{\partial x_{n}}\rangle$ of $S$. Clearly $\mathscr{T}_{\mathcal{K}}(\sigma)$ is a subsheaf of $\pi^{*}(T_{\mathcal{K}}(J))$, so we can take the quotient sheaf $\mathscr{F}=\frac{\pi^{*}(T_{\mathcal{K}}(J))}{\mathscr{T}_{\mathcal{K}}(\sigma)}$ on $\mathbb{P}^{q-1}$. Since $\mathscr{F}$ is coherent and $\mathbb{P}^{q-1}$ is noetherian, the support $\mbox{Supp}\mathscr{F}$ of $\mathscr{F}$ is a closed subscheme of $\mathbb{P}^{q-1}$ given by the vanishing of $\Big{(}\mathscr{T}_{\mathcal{K}}(\sigma):\pi^{*}(T_{\mathcal{K}}(J))\Big{)}$. ###### Definition 4.7. We say that an ideal $J\subset\mathscr{O}_{\mathbb{C}^{n},0}$ is _$T_{\mathcal{K}}$ -dependent_ if $\pi^{-1}(\mathfrak{m})\not\subset\mbox{Supp}\mathscr{F}$. We check that $T_{\mathcal{K}}$-dependence for $J$ is a well-defined concept, being independent on the choice of generators for $J$. For, let the ideal $J$ be given by another system of generators, say $J=\langle h_{1},\ldots,h_{u}\rangle$. Let $\mathbb{P}^{u-1}=\mbox{Proj}(S^{\prime})$, being $S^{\prime}=\mathscr{O}_{\mathbb{C}^{n},0}[\,\beta_{1},\ldots,\beta_{u}]$. The above construction can be carried out with the obvious morphism $\pi^{\prime}:\mathbb{P}^{u-1}\rightarrow\mbox{Spec}\,\mathscr{O}_{\mathbb{C}^{n},0}$ and section $\sigma^{\prime}=\sum_{j=1}^{u}\,h_{j}\beta_{j}$ instead of $\pi$ and $\sigma$, obtaining a sheaf $\mathscr{F^{\prime}}$ on $\mathbb{P}^{u-1}$. ###### Lemma 4.8. Keeping notation as above, we have $\pi^{-1}(\mathfrak{m})\not\subset\mbox{Supp}\mathscr{F}$ if and only if $\pi^{\prime-1}(\mathfrak{m})\not\subset\mbox{Supp}\mathscr{F^{\prime}}$. ###### Proof. We will only show that $\pi^{-1}(\mathfrak{m})\not\subset\mbox{Supp}\mathscr{F}$ implies $\pi^{\prime-1}(\mathfrak{m})\not\subset\mbox{Supp}\mathscr{F^{\prime}}$. The proof of the other implication is analogous. Since the $g_{i}$’s and the $h_{j}$’s generate the same ideal $J$, we can write for all $i$, $g_{i}=\sum_{j}r_{ji}h_{j}$ for some $r_{ji}\in\mathscr{O}_{\mathbb{C}^{n},0}$ and at least one $r_{ji}$ is invertible in $\mathscr{O}_{\mathbb{C}^{n},0}$. We construct an homomorphism of graded $\mathscr{O}_{\mathbb{C}^{n},0}$-algebras $\Phi:S^{\prime}\rightarrow S$ (preserving degrees) given by $\beta_{j}\mapsto\sum_{i}r_{ji}\alpha_{i}$. This induces a natural morphism $\varphi:U\rightarrow\mathbb{P}^{u-1}$, of $\mathscr{O}_{\mathbb{C}^{n},0}$-schemes, where $U\subset\mathbb{P}^{q-1}$ is the complement of the indeterminacy locus of $\varphi$, given by the vanishing of $\langle\Phi(\beta_{1}),\ldots,\Phi(\beta_{u})\rangle$ in $\mathbb{P}^{q-1}$. Notice that $\pi^{-1}(\mathfrak{m})$ is not contained in the locus of indeterminacy of $\varphi$ because $\Phi(\beta_{j})\not\in\mathfrak{m}S$ for at least one $j$. Observe also that the construction of $\varphi$ implies both $\sigma|_{U}=\varphi^{*}(\sigma^{\prime})$ and $\pi|_{U}=\pi^{\prime}\circ\varphi$. In particular, $\mathscr{T}_{\mathcal{K}}(\sigma)|_{U}=\varphi^{*}(\mathscr{T}_{\mathcal{K}}(\sigma^{\prime}))$ and $\pi^{*}(T_{\mathcal{K}}(J))|_{U}=\varphi^{*}\pi^{\prime*}(T_{\mathcal{K}}(J))$. According to the definition of $\mathscr{F}$ and $\mathscr{F^{\prime}}$, we deduce $\mathscr{F}|_{U}=\varphi^{*}\mathscr{F}^{\prime}$. Assume $\pi^{-1}(\mathfrak{m})\not\subset\mbox{Supp}\mathscr{F}$. As observed above $\mbox{Supp}\mathscr{F}$ is closed in the irreducible $\mathbb{P}^{q-1}$. It follows that there exists some $P\in\pi^{-1}(\mathfrak{m})\cap U$, $P\not\in\mbox{Supp}\mathscr{F}$. Since $U\cap\mbox{Supp}\mathscr{F}=\varphi^{-1}(\mbox{Supp}\mathscr{F}^{\prime})$, we have $\varphi(P)\not\in\mbox{Supp}\mathscr{F^{\prime}}$ and $\mathfrak{m}S^{\prime}\subset\Phi^{-1}(\mathfrak{m}S)\subset\Phi^{-1}(P)=\varphi(P)$. Therefore $\varphi(P)\in\pi^{\prime-1}(\mathfrak{m})$, showing that $\pi^{\prime-1}(\mathfrak{m})\not\subset\mbox{Supp}\mathscr{F^{\prime}}$.∎ Now that we have defined $T_{\mathcal{K}}$-fullness and $T_{\mathcal{K}}$-dependence, we present some examples showing that the two notions are independent of each other. ###### Example 4.9. If $J\subseteq\mathscr{O}_{\mathbb{C}^{n},0}$ is a principal ideal then $J$ is $T_{\mathcal{K}}$-dependent. Indeed, in this case we see that $\mbox{Supp}\mathscr{F}\subset\mathbb{P}^{0}$ is empty, opposite to $\pi^{-1}(\mathfrak{m})$ which is not. It is easy to give examples of principal ideals which are not $T_{\mathcal{K}}$-full. ###### Example 4.10. Let $J=\langle x^{2},y\rangle\subset\mathbb{C}\\{x,y\\}$. We have seen before (cf. Example 4.2) that $J$ fails to be $T_{\mathcal{K}}$-full. Let us show that $J$ is $T_{\mathcal{K}}$-dependent. Here $\pi^{*}(T_{\mathcal{K}}(J))$ is the sheaf associated to $T_{\mathcal{K}}(J)S=\mathfrak{m}S.$ Let $\sigma=x^{2}\alpha_{1}+y\alpha_{2}$. Then the sections $\sigma,\,x\frac{\partial\sigma}{\partial x},x\frac{\partial\sigma}{\partial y},y\frac{\partial\sigma}{\partial x},y\frac{\partial\sigma}{\partial y}$ generate $\mathscr{T}_{\mathcal{K}}(\sigma)$. We investigate the support of $\mathscr{F}$ in $\mathbb{P}^{1}$. We can compute generators of $\Big{(}\mathscr{T}_{\mathcal{K}}(\sigma):\pi^{*}(T_{\mathcal{K}}(J))\Big{)}$ as $\langle\alpha_{1}x,\alpha_{2}\rangle\subset S$. Hence the fiber $\pi^{-1}(\mathfrak{m})$ is not contained in the support of $\mathscr{F}$, showing that $J$ is $T_{\mathcal{K}}$-dependent. ###### Example 4.11. Let $J=\mathfrak{m}^{4}\subset\mathbb{C}\\{x,y\\}$. Clearly $J$ is $T_{\mathcal{K}}$-full (cf. Examples 2.8, 4.5). Let us show that $J$ fails to be $T_{\mathcal{K}}$-dependent. Notations being as before, $\pi^{*}(T_{\mathcal{K}}(J))$ is the sheaf in $\mathbb{P}^{4}$ associated to $T_{\mathcal{K}}(J)S=JS.$ Let $\sigma=x^{4}\alpha_{1}+x^{3}y\alpha_{2}+x^{2}y^{2}\alpha_{3}+xy^{3}\alpha_{4}+y^{4}\alpha_{5}.$ Then $\sigma,\,x\frac{\partial\sigma}{\partial x},x\frac{\partial\sigma}{\partial y},y\frac{\partial\sigma}{\partial x},y\frac{\partial\sigma}{\partial y}$ are generators of $\mathscr{T}_{\mathcal{K}}(\sigma)$. We check that the support of $\mathscr{F}$ contains $\pi^{-1}(\mathfrak{m})$. To do this we show, with help of SINGULAR, that $\Big{(}\mathscr{T}(\sigma):\pi^{*}(T_{\mathcal{K}}(J))\Big{)}\subset\mathfrak{m}S$. We proceed as follows: > ring r=0,(x,y,a(1..5)),ds; > ideal m=x,y; > ideal M=std(m); > poly s=a(1)*x4+a(2)*x3y+a(3)*x2y2+a(4)*xy3+a(5)*y4; > ideal t=s,m*diff(s,x),m*diff(s,y); > ideal q=quotient(t,m4); > ideal Q=std(q); > reduce(Q,M); ## 5\. Main Result The purpose of this Section is to state and prove our result characterizing moduli ideals, then solving Problem 1.2. We also derive Corollary 5.2 which has as a consequence an explicit solution to both the Recognition and the Reconstruction problems mentioned in the Introduction, even for the case of non-isolated hypersurface singularities. ###### Theorem 5.1. Let $I\subset\mathscr{O}_{\mathbb{C}^{n},0}$ be an ideal. Then $I$ is a moduli ideal if and only if $I$ is $T_{\mathcal{K}}$-full and $\Delta_{1}(I)$ is $T_{\mathcal{K}}$-dependent. ###### Proof. Let $\Delta_{1}(I)=\langle g_{1},\ldots,g_{q}\rangle$ and let $\mathscr{F}$ denote the sheaf $\pi^{*}(T_{\mathcal{K}}(\Delta_{1}(I)))/\mathscr{T}_{\mathcal{K}}(\sigma)$ on $\mathbb{P}^{q-1}$, as described before. For any $\lambda=(\lambda_{1},\ldots,\lambda_{q})\in\mathbb{C}^{q}\setminus\\{0\\}$ we will denote by $p_{\lambda}\subseteq\mathbb{C}[\alpha_{1},\ldots,\alpha_{q}]$ the homogeneous prime ideal generated by $\\{\lambda_{k}\alpha_{\ell}-\lambda_{\ell}\alpha_{k}\\}_{k,{\ell}}$. If $P\in D_{+}(\alpha_{\ell})\subset\mathbb{P}^{q-1}$ one checks easily that $\mathscr{T}_{\mathcal{K}}(\sigma)_{P}=T_{\mathcal{K}}\Big{(}\sigma/\alpha_{\ell}\Big{)}S_{(P)}$, where parenthetical notation $-\,_{(P)}$ (here and in what follows) indicates the submodule of degree zero elements of the localization of a graded module at $P\in\mathbb{P}^{q-1}$. Moreover, if $p_{\lambda}S\subseteq P\in D_{+}(\alpha_{\ell})$ then $\lambda_{\ell}\in\mathbb{C}\setminus 0$ and we can use Remark 2.1 to obtain the following Estimate on ideals in the stalk of structural sheaf $\mathcal{O}_{\mathbb{P}^{q-1},P}=S_{(P)}$: (5.1) $T_{\mathcal{K}}\Big{(}\sum_{k}\Big{(}\lambda_{k}-\frac{\lambda_{\ell}\alpha_{k}}{\alpha_{\ell}}\Big{)}g_{k}\Big{)}S_{(P)}\subseteq\sum_{k}T_{\mathcal{K}}\Big{(}\Big{(}\lambda_{k}-\frac{\lambda_{\ell}\alpha_{k}}{\alpha_{\ell}}\Big{)}g_{k}\Big{)}S_{(P)}\subseteq\sum_{k}\Big{(}\lambda_{k}-\frac{\lambda_{\ell}\alpha_{k}}{\alpha_{\ell}}\Big{)}T_{\mathcal{K}}(g_{k})S_{(P)}\subseteq IP_{(P)}$ To prove the “$\Rightarrow$” assertion in our statement, assume $I$ to be a moduli ideal. We have seen (cf. Proposition 4.3) that $I$ is $T_{\mathcal{K}}$-full; hence $\mathscr{F}=\pi^{*}(I)/\mathscr{T}_{\mathcal{K}}(\sigma)$. Now we verify that $\Delta_{1}(I)$ is $T_{\mathcal{K}}$-dependent. Since this is clear if $I=\langle 0\rangle$, we assume $I=T_{\mathcal{K}}(f)\neq\langle 0\rangle$, with $f=\sum_{k}r_{k}g_{k}$ for some $r_{k}\in\mathscr{O}_{\mathbb{C}^{n},0}$. We write $r_{k}=\lambda_{k}+s_{k}$, for certain $\lambda_{k}\in\mathbb{C}$ and $s_{k}\in\mathfrak{m}$, for all $k=1,\ldots,q$. Since $I\neq\langle 0\rangle$ we can use Remark 2.5, (v) and Nakayama’s Lemma to check $f\not\in\mathfrak{m}\Delta_{1}(I)$. Hence, at least one of the $r_{k}$’s is an invertible element of $\mathscr{O}_{\mathbb{C}^{n},0}$. Using Remark 2.1 again we may assume $(r_{1},\ldots,r_{q})=(\lambda_{1},\ldots,\lambda_{q})\in\mathbb{C}^{q}\setminus\\{0\\}$. We consider $P=p_{\lambda}S+\mathfrak{m}S\in\mathbb{P}^{q-1}$. Then $P\in\pi^{-1}(\mathfrak{m})$, so that $\pi^{*}(I)_{P}=I_{\pi(P)}\otimes S_{(P)}=I_{\mathfrak{m}}\otimes S_{(P)}=IS_{(P)}$. On the other hand (say if $P\in D_{+}(\alpha_{\ell})$), our assumption $I=T_{\mathcal{K}}(f)$ together with Estimate (5.1) above produces $IS_{(P)}=T_{\mathcal{K}}\Big{(}\sum_{k}\lambda_{k}g_{k}\Big{)}S_{(P)}\subseteq T_{\mathcal{K}}\Big{(}\sum_{k}\Big{(}\lambda_{k}-\frac{\lambda_{\ell}\alpha_{k}}{\alpha_{\ell}}\Big{)}g_{k}\Big{)}S_{(P)}+T_{\mathcal{K}}\Big{(}\lambda_{\ell}\sigma/\alpha_{\ell}\Big{)}S_{(P)}\subseteq IP_{(P)}+\mathscr{T}_{\mathcal{K}}(\sigma)_{P}\subseteq IS_{(P)}.$ It follows that $IP_{(P)}+\mathscr{T}_{\mathcal{K}}(\sigma)_{P}=IS_{(P)}$. Now we use Nakayama’s Lemma to obtain $\pi^{*}(I)_{P}=IS_{(P)}=\mathscr{T}_{\mathcal{K}}(\sigma)_{P}$. Hence $P\not\in\mbox{Supp}\mathscr{F}$, which completes the proof that $\Delta_{1}(I)$ is $T_{\mathcal{K}}$-dependent. To prove the “$\Leftarrow$” assertion in the statement $I$ is assumed to be $T_{\mathcal{K}}$-full and $\Delta_{1}(I)$ is assumed to be $T_{\mathcal{K}}$-dependent. Then $I=T_{\mathcal{K}}(\Delta_{1}(I))$ and there exists some $P\in\pi^{-1}(\mathfrak{m})\setminus\mbox{Supp}\mathscr{F}$. It follows that $\pi^{*}(I)_{P}=\pi^{*}(T_{\mathcal{K}}(\Delta_{1}(I))_{P}=\mathscr{T}_{\mathcal{K}}(\sigma)_{P}.$ We make use of the decomposition $S=\mathbb{C}[\alpha_{1},\ldots,\alpha_{q}]+\mathfrak{m}S$ to simplify our argument. Indeed, we obtain $P=P\cap\mathbb{C}[\alpha_{1},\ldots,\alpha_{q}]+P\cap\mathfrak{m}S=P\cap\mathbb{C}[\alpha_{1},\ldots,\alpha_{q}]+\mathfrak{m}S$. Therefore, there exists some $(\lambda_{1},\ldots,\lambda_{q})\in\mathbb{C}^{q}\setminus\\{0\\}$, $p_{\lambda}\supseteq P\cap\mathbb{C}[\alpha_{1},\ldots,\alpha_{q}]$, such that the homogeneous prime ideal $p_{\lambda}S+\mathfrak{m}S$ also belongs to $\pi^{-1}(\mathfrak{m})\setminus\mbox{Supp}\mathscr{F}$. With this simplification - i. e., assuming $P=p_{\lambda}S+\mathfrak{m}S\in D_{+}(\alpha_{\ell})$, as before - the displayed equality of stalks above reads $IS_{(P)}=T_{\mathcal{K}}\Big{(}\lambda_{\ell}\sigma/\alpha_{\ell}\Big{)}S_{(P)}.$ We will prove $I=T_{\mathcal{K}}(f)$, for $f=\sum_{k}\lambda_{k}g_{k}\in\mathscr{O}_{\mathbb{C}^{n},0}$. Estimate (5.1) here gives $IS_{(P)}\subseteq T_{\mathcal{K}}\Big{(}\sum_{k}\Big{(}\lambda_{k}-\frac{\lambda_{\ell}\alpha_{k}}{\alpha_{\ell}}\Big{)}g_{k}\Big{)}S_{(P)}+T_{\mathcal{K}}\Big{(}\sum_{k}\lambda_{k}g_{k}\Big{)}S_{(P)}\subseteq IP_{(P)}+T_{\mathcal{K}}\Big{(}\sum_{k}\lambda_{k}g_{k}\Big{)}S_{(P)}\subseteq IS_{(P)}.$ We obtain $IP_{(P)}+T_{\mathcal{K}}(\sum_{k}\lambda_{k}g_{k})S_{(P)}=IS_{(P)}$ and by Nakayama’s Lemma again we deduce $T_{\mathcal{K}}(f)S_{(P)}=IS_{(P)}$. Since $T_{\mathcal{K}}(f)\subseteq I$, we should verify $I\subseteq T_{\mathcal{K}}(f)$. It suffices to check that if $h\in I$ and there are some $g\in T_{\mathcal{K}}(f)$ and $G_{1},G_{2}\in S$, homogeneous of same degree with $G_{2}\not\in P$ such that $G_{2}h=gG_{1}$, then $h\in T_{\mathcal{K}}(f)$. With this simplification, an argument with divisibility and degree in the unique factorization domain $S$ shows $\xi h=gH$, for some homogeneous element $H\in S$ and some degree zero factor $\xi$ of $G_{2}$. In particular, $\xi\in\mathscr{O}_{\mathbb{C}^{n},0}\setminus\mathfrak{m}$. Again by degree reasons, $H$ has degree zero, i.e., it belongs to $\mathscr{O}_{\mathbb{C}^{n},0}$. This shows that $h\in T_{\mathcal{K}}(f)$. Hence $I\subseteq T_{\mathcal{K}}(f)$ and we have shown that $I$ is a moduli ideal of $\mathscr{O}_{\mathbb{C}^{n},0}$.∎ We will keep the notation as in the proof of Theorem 5.1. Then we have the corollary below. ###### Corollary 5.2. Let $I\subset\mathscr{O}_{\mathbb{C}^{n},0}$ be an ideal and let $\Delta_{1}(I)=\langle g_{1},\ldots,g_{q}\rangle$. Then $I$ is a moduli ideal if and only if $I=T_{\mathcal{K}}(\sum_{k}\lambda_{k}g_{k})$ for $p_{\lambda}$ varying in a (non empty) Zariski open subset of $\mathbb{P}_{\mathbb{C}}^{q-1}$. ###### Proof. For the proof of the non-trivial part of the statement, assume that $I$ is a moduli ideal. The canonical inclusion $\mathbb{C}\rightarrow\mathscr{O}_{\mathbb{C}^{n},0}$ induces $\mathbb{C}[\alpha_{1},\ldots,\alpha_{q}]\rightarrow S$, which is a graded inclusion of $\mathbb{C}$-algebras; hence we obtain a proper and dominant morphism of schemes $\rho:\mathbb{P}^{q-1}\rightarrow\mathbb{P}_{\mathbb{C}}^{q-1}$. From the Theorem, we know that $\Delta_{1}(I)$ is $T_{\mathcal{K}}$-dependent, so that $\pi^{-1}(\mathfrak{m})\setminus\mbox{Supp}\mathscr{F}$ is a non empty Zariski open subset of $\pi^{-1}(\mathfrak{m})$. It follows that its image under $\rho$ is a non-empty open Zariski subset $U\subset\mathbb{P}_{\mathbb{C}}^{q-1}$. Homogeneous prime ideals $p_{\lambda}$ in $U$ correspond bijectively, via $\rho$, to homogeneous prime ideals of type $p_{\lambda}S+\mathfrak{m}S\in\pi^{-1}(\mathfrak{m})\setminus\mbox{Supp}\mathscr{F}$. Since the Theorem also guarantees that $I$ is $T_{\mathcal{K}}$-full, we can proceed as in its proof to show that $I=T_{\mathcal{K}}(\sum_{k}\lambda_{k}g_{k})$, for all $\lambda=(\lambda_{1},\ldots,\lambda_{q})$ such that $p_{\lambda}\in U$.∎ ###### Remark 5.3. The practical character and the usefulness of the above Corollary comes from its consonance with the fact outlined in Example 4.6. Precisely, it says that computing the moduli ideal of a sufficiently general $\mathbb{C}$-linear combination of generators of $\Delta_{1}(I)$ \- accessible in practical examples, as we have illustrated in Section 3 \- reveals whether $I\subset\mathscr{O}_{\mathbb{C}^{n},0}$ is a moduli ideal. We illustrate how one can use the software SINGULAR to check whether an ideal $I\subset\mathscr{O}_{\mathbb{C}^{n},0}$ is a moduli ideal, without knowing a priori the function germ which realizes it as such. We do this with the ideal presented in Example 3.3. ###### Example 5.4. (Example 3.3, continued): We have computed the anti-derivatives ideal of $I=\langle 3xy^{2}+x^{6},y^{3},x^{5}y,x^{7}\rangle\subset\mathbb{C}\\{x,y\\}$ as $\Delta_{1}(I)=\langle y^{3}+x^{5}y,x^{4}y^{2},x^{7},x^{6}y\rangle\subset I$. A SINGULAR check shows that $I$ is $T_{\mathcal{K}}$-full. We consider the corresponding sheaf $\mathscr{F}$ in $\mathbb{P}^{3}$ and investigate its support. In order to do so we use SINGULAR again: > ring r=0,(x,y,a(1..4)),ds; > ideal i=3xy2+x6,y3,x5y,x7; > poly s=a(1)*(y3+x5y)+a(2)*x4y2+a(3)*x7+a(4)*x6y; > ideal m=x,y; > ideal t=s,m*diff(s,x),m*diff(s,y); > ideal q=quotient(t,i); > ideal Q=std(q); > Q; The output is a list with twenty homogeneous generators of $\Big{(}\mathscr{T}_{\mathcal{K}}(\sigma):\pi^{*}(T_{\mathcal{K}}(J))\Big{)}$, the sixth element of which being $15\alpha_{1}^{3}\alpha_{3}+52x\alpha_{1}^{2}\alpha_{3}\alpha_{4}-121y\alpha_{1}\alpha_{2}\alpha_{3}\alpha_{4}+36y\alpha_{1}^{2}\alpha_{4}^{2}+15x^{2}\alpha_{1}\alpha_{3}\alpha_{4}^{2}+30xy\alpha_{1}^{2}\alpha_{2}^{2}+28x^{2}\alpha_{1}\alpha_{3}\alpha_{4}^{2}+\\\ 24xy\alpha_{1}\alpha_{4}^{3}+42x^{3}\alpha_{1}^{2}\alpha_{2}\alpha_{4}+24x^{2}y\alpha_{1}\alpha_{2}^{2}\alpha_{4}-54x^{4}\alpha_{2}^{2}\alpha_{3}\alpha_{4}+41x^{4}\alpha_{1}\alpha_{2}\alpha_{4}^{2}+4x^{3}y\alpha_{2}^{2}\alpha_{4}^{2}+14x^{5}\alpha_{2}\alpha_{4}^{3}.$ Clearly this element does not belong to $\mathfrak{m}[\alpha_{1},\alpha_{2},\alpha_{3},\alpha_{4}]$ due to the presence of the term $15\alpha_{1}^{3}\alpha_{3}$. It follows that $\Delta_{1}(I)$ is $T_{\mathcal{K}}$-dependent. According to the proof of Theorem, any $\mathbb{C}$-linear combination $\lambda_{1}(y^{3}+x^{5}y)+\lambda_{2}x^{4}y^{2}+\lambda_{3}x^{7}+\lambda_{4}x^{6}y$ with both $\lambda_{1}\neq 0$ and $\lambda_{3}\neq 0$ represents a function germ $f$ of which $I$ is a moduli ideal. ## References * [Ar] Arnold, V. I. - Local normal forms of functions. Inventiones mathematicae, 35(1), (1976): 87-109. * [DGPS] Decker, W.; Greuel, G.-M.; Pfister, G.; Schönemann, H. - Singular 4-3-1 — A computer algebra system for polynomial computations. http://www.singular.uni-kl.de (2022). * [E] Epure, R.-P. - Explicit and effective Mather-Yau correspondence in view of analytic gradings. 2020. * [ES] Epure, R. and Schulze, M. - Hypersurface singularities with monomial Jacobian ideal. Bulletin of the London Mathematical Society (2022). * [GH] Gaffney, T., and Hauser, H. - Characterizing singularities of varieties and of mappings. Inventiones mathematicae 81.3 (1985): 427-447. * [GLS] Greuel, G. M., Lossen, C., Shustin, E. I. - Introduction to singularities and deformations. Springer Science Business Media(2007). * [H] Hartshorne, R. - Algebraic geometry. Graduate texts in mathematics 52 (1977). * [IK] Isaev, A; Kruzhilin, N. - Explicit reconstruction of homogeneous isolated hypersurface singularities from their Milnor algebras. Proceedings of the American Mathematical Society, v. 142, n. 2, p. 581-590, 2014. * [MY] Mather, John N., and Stephen S-T. Yau. - Classification of isolated hypersurface singularities by their moduli algebras. Inventiones mathematicae 69.2: 243-251 (1982). * [OR] Olmedo Rodrigues, J. H. - On Tjurina Ideals of Hypersurface Singularities. arXiv:2205.03527. (To appear) * [Y] Stephen, S-T. Yau. - A necessary and sufficient condition for a local commutative algebra to be a moduli algebra: weighted homogeneous case. Complex analytic singularities 8: 687-697. (1987)
Measurement of the inclusive $t\bar{t}$ production cross section in the lepton+jets channel in $pp$ collisions at $\sqrt{s}=7$ TeV with the ATLAS detector using support vector machines A measurement of the top quark pair- production cross section in the lepton+jets decay channel is presented. It is based on 4.6 fb-1 of $\sqrt{s}=7$ TeV $pp$ collision data collected during 2011 by the ATLAS experiment at the CERN Large Hadron Collider. A three-class, multidimensional event classifier based on support vector machines is used to differentiate $t\bar{t}$ events from backgrounds. The $t\bar{t}$ production cross section is found to be $\sigma_{t\bar{t}}=168.5\pm 0.7$(stat.) ${}^{+6.2}_{-5.9}$(syst.) ${}^{+3.4}_{-3.2}$(lumi.) pb. The result is consistent with the Standard Model prediction based on QCD calculations at next-to-next-to-leading order. TOPQ-2017-08 CERN-EP-2022-191 Phys. Rev. D ###### Contents 1. 1 Introduction 2. 2 ATLAS detector 3. 3 Object definitions 4. 4 Event selection 5. 5 Data samples 6. 6 Signal and background modeling 1. 6.1 Multi-jet background / fake leptons 2. 6.2 Monte Carlo samples 3. 6.3 Signal and background classes 7. 7 Analysis method 1. 7.1 The SVM discriminant 2. 7.2 Physics observables 3. 7.3 SVM training 4. 7.4 Class templates 5. 7.5 Cross-section measurement 6. 7.6 Systematic uncertainties 8. 8 Results 1. 8.1 Top quark mass dependence 9. 9 Summary ### 1 Introduction In the Standard Model of particle physics [1, 2, 3], the top quark ($t$) and the bottom quark ($b$) belong to a doublet representation of the weak-isospin SU(2). The top quark is the most massive of the known elementary particles. Because its mass is close to the electroweak symmetry breaking scale, it may play a fundamental role in the mechanism of breaking of the SU(2) symmetry of the electroweak interaction. Top quark production is also the dominant background in many analyses looking for physics beyond the Standard Model at high mass scales at the Large Hadron Collider (LHC), and a good understanding of top quark production is a necessary step in many “new physics” searches. The cross section is one of the simplest observables that can be measured in the $t\bar{t}$ system. It allows one to make important comparisons with theoretical predictions available at next-to-next-to-leading order in perturbative QCD, including the soft-gluon resummation at next-to-next-to- leading-log order (NNLO + NNLL); see Ref. [4] and references therein. For $pp$ collisions at a center-of-mass energy of $\sqrt{s}=7$ TeV, the predicted $t\bar{t}$ production cross section is $\sigma_{t\bar{t}}^{\text{NNLO+NNLL}}=177^{+10}_{-11}$ pb. This theoretical value was calculated with the Top++ 2.0 program [5], including soft-gluon resummation, assuming a top quark mass value of 172.5 GeV, and using the PDF4LHC [6] procedures. According to the Standard Model, top quarks from $pp$ collisions at the LHC are produced mostly via the strong interaction as $t\bar{t}$ pairs, with each top quark decaying into a $W$ boson and a $b$-quark nearly $100\%$ of the time. The $t\bar{t}$ events in which one of the $W$ bosons decays into a quark pair and the other into a charged lepton and a neutrino, are classified as “lepton+jets”, as such events contain an electron or muon or $\tau$-lepton, a neutrino and typically four hadronic jets (two of which originate from the $b$-quark and $\bar{b}$-quark). In this paper, a measurement of the top quark pair-production cross section at $\sqrt{s}=7$ TeV using events with a single charged lepton (electron or muon) and jets in the final state is presented. The previously published result from the ATLAS Collaboration for the lepton+jets channel [7] uses 35 pb-1 of data and obtains a precision of 12%. The most precise CMS $t\bar{t}$ cross-section measurement in the same channel [8] has a precision of 7%. In the dilepton channel, the best ATLAS result [9] achieves 3.5% precision, while CMS [10] reaches 3.6%. These ATLAS and CMS dilepton results have been combined [11], resulting in an uncertainty of 2.6% in $\sigma_{t\bar{t}}$. The analysis presented in this paper is based on the full dataset collected with the ATLAS detector at the LHC in 2011, corresponding to an integrated luminosity of 4.6 fb-1, and attains statistical and systematic uncertainties that are significantly lower than in previous ATLAS measurements in this final state. In an extension of the usual application of binary multivariate classifiers, this analysis uses a large number of variables to train three different support vector machines (SVMs). The three SVMs are used to define a three-dimensional space in which a multi-class event discriminator is constructed to identify the $t\bar{t}$ events through a simultaneous profile likelihood fit in four independent regions of this space. ### 2 ATLAS detector The ATLAS detector is described in Ref. [12]. It is a multipurpose particle detector with forward-backward symmetry and a cylindrical geometry.111 ATLAS uses a right-handed coordinate system with its origin at the nominal interaction point (IP) in the center of the detector and the $z$-axis along the beam pipe. The $x$-axis points from the IP to the center of the LHC ring, and the $y$-axis points upwards. Cylindrical coordinates $(r,\phi)$ are used in the transverse plane, $\phi$ being the azimuthal angle around the $z$-axis. The pseudorapidity is defined in terms of the polar angle $\theta$ as $\eta=-\ln\tan(\theta/2)$, and the distance $\Delta R$ in the $\eta$–$\phi$ space is defined as $\Delta R\equiv\sqrt{(\Delta\eta)^{2}+(\Delta\phi)^{2}}$. The inner tracking detectors are surrounded by a thin superconducting solenoid, electromagnetic and hadronic calorimeters, and a muon spectrometer with a magnetic field generated by three superconducting toroidal magnets of eight coils each. The inner-detector system (ID), in combination with the 2 T magnetic field from the solenoid, provides precision momentum measurements for charged particles within the pseudorapidity range $|\eta|<2.5$. Moving radially outwards, it consists of a silicon pixel detector, a silicon microstrip detector, and a straw-tube tracker that also provides transition radiation measurements for electron identification. The calorimeter system covers the pseudorapidity range $|\eta|<4.9$. A high-granularity liquid-argon (LAr) sampling calorimeter with lead absorber measures electromagnetic showers within $|\eta|<3.2$. In the region matched to the ID, $|\eta|<2.5$, the innermost layer has fine segmentation in $\eta$ to improve the resolution of the shower position and direction measurements. Hadronic showers are measured by an iron/plastic-scintillator tile calorimeter in the central region, $|\eta|<1.7$, and by a LAr calorimeter in the endcap region, $1.5<|\eta|<3.2$. In the forward region, measurements of both electromagnetic and hadronic showers are provided by a LAr calorimeter covering the pseudorapidity range $3.1<|\eta|<4.9$. The muon spectrometer is instrumented with separate trigger and high-precision tracking chambers. It provides muon identification for charged-particle tracks within $|\eta|<2.7$. The combination of all ATLAS detector systems provides charged-particle measurement along with lepton and photon measurement and identification in the pseudorapidity range $|\eta|<2.5$. Jets are reconstructed over the full range covered by the calorimeters, $|\eta|<4.9$. A three-level trigger system [13] is used to select interesting events. The first-level (L1) trigger is implemented in hardware and uses a subset of detector information to reduce the event rate to a design value of at most $75\text{\,}\mathrm{kHz}$. This is followed by two software-based trigger levels which together reduce the event rate to about $200\text{\,}\mathrm{Hz}$. An extensive software suite [14] is used in data simulation, in the reconstruction and analysis of real and simulated data, in detector operations, and in the trigger and data acquisition systems of the experiment. ### 3 Object definitions Electrons Electron candidates are selected using the offline identification with tight requirements [15] within a fiducial region with transverse momentum $p_{\mathrm{T}}>25$ GeV and $|\eta|<2.47$, excluding the calorimeter transition region $1.37<|\eta|<1.52$. They are subjected to several other strict criteria including requirements on track quality, impact parameter, calorimeter shower shape, and track–cluster matching. The electron candidates are also required to be isolated. The transverse energy ($E_{\mathrm{T}}$) deposited in the calorimeter in a cone of size $\Delta R=0.2$ around the electron is calculated. Additionally, the scalar sum of the $p_{\mathrm{T}}$ of tracks in a cone of size $\Delta R=0.3$ is determined. Both of these quantities have selection requirements that depend on the $\eta$ and $E_{\mathrm{T}}$ of the electron candidate, and which ensure 90% efficiency for electrons from $W$ boson or $Z$ boson decays [16]. To avoid double counting of jets and electrons (which are reconstructed by independent algorithms), jets within $\Delta R=0.2$ of a reconstructed electron are removed. Finally, electrons lying within $\Delta R=0.4$ of a selected jet are discarded to reject leptons from heavy-flavor decays. Muons Muon candidates reconstructed from tracks in both the muon spectrometer and ID are selected with the MuID algorithm [17]. Only candidates satisfying $p_{\mathrm{T}}>20$ GeV and $|\eta|<2.5$ are selected. Muon candidates are required to have a sufficient number of hits in the ID. The impact parameter with respect to the primary vertex in the longitudinal direction along the beam axis is required to satisfy $|z_{0}|<2$ mm. The tight muon candidates used in this analysis are required to be isolated. The sum of the calorimeter transverse energy within $\Delta R=0.2$ of a muon is required to be below 4 GeV, and the sum of the $p_{\mathrm{T}}$ of all the tracks within $\Delta R=0.3$ (excluding the muon track) must be below 2.5 GeV. The efficiency of this combined isolation requirement varies between 95% and 97%, depending on the data-taking period. In order to reduce the background from muons produced by heavy-flavor decays inside jets, muons are required to be separated by $\Delta R>0.4$ from the nearest jet. Jets Jets are reconstructed from topological clusters [18] formed from energy deposits in the calorimeters using the anti-$k_{t}$ algorithm [19, 20] with a radius parameter of 0.4. Clusters are calibrated using the local cluster weighting (LCW), which differentiates between the energy deposits arising from electromagnetic and hadronic showers [21]. The jet reconstruction is done at the electromagnetic scale and then a scale factor is applied in order to obtain the jet energy at the hadronic scale. The jet energy scale (JES) corrections account for the calorimeter response to the true jet energy by using “truth jets” from simulation. The “truth jets” are formed through the application of the anti-$k_{t}$ algorithm to stable particles, with the exception of final-state muons and neutrinos. Jet calibration includes both the LCW and JES calibrations. In addition, the jets are corrected for distortions due to multiple $pp$ collisions per bunch crossing (pileup) using a method which estimates the pileup activity on an event-by-event basis, as well as the sensitivity of a given jet to pileup. With this method [21], a contribution to the jet transverse momentum equal to the product of the jet area in the $\eta$–$\phi$ plane and the average transverse energy density in a given event is subtracted. The effects of additional collisions in either the same bunch crossing or those adjacent in time are taken into account using corrections which depend on the average number of interactions per bunch crossing and the number of primary vertices. For this analysis, only jets in the central region of the detector, $|\eta|<2.5$, and with transverse momentum $p_{\mathrm{T}}>25$ GeV are considered. All jets, including those that do not pass the analysis selection requirements, are used in the procedure to remove overlaps with leptons. Identification of b-jets The identification of “$b$-jets” (jets arising from the decay of $B$-hadrons) is performed with the MV1 algorithm [22], which combines the outputs of three different tagging algorithms into a multivariate discriminant. Jets are defined to be “$b$-tagged” if the MV1 discriminant value is larger than a threshold (operating point) corresponding to 70% efficiency to identify $b$-quark jets in simulated $t\bar{t}$ events. Approximately 20% of jets originating from charm quarks are identified as $b$-jets, while light-flavor jets are mistagged as $b$-jets at the 1% level. Missing transverse momentum The missing transverse momentum is calculated [23] as the complement of the vector sum, in the transverse plane, of calorimeter cell energies within $|\eta|<4.9$, after all corrections are applied to the associated physics objects (including jets, electrons, and muons). A correction for significant energy deposits not associated with high-$p_{\mathrm{T}}$ physics objects is also included. The magnitude of the missing transverse momentum vector is denoted by $E_{\text{T}}^{\text{miss}}$, while its direction in the transverse plane is either denoted by an azimuthal angle $\phi$ or inferred through its vector components $E_{\text{T}}^{\text{miss}}$ x and $E_{\text{T}}^{\text{miss}}$ y. ### 4 Event selection This analysis considers the single-lepton decay channel for the $t\bar{t}$ pair. The selected events are required to have exactly one lepton (either an electron or a muon), a large amount of $E_{\text{T}}^{\text{miss}}$, and three or more hadronic jets. The number of $b$-tagged jets in an event must be two or less. Events must have at least one primary vertex, with five or more tracks with $p_{\textrm{T}}>150$ MeV. If there is more than one primary vertex, the one with the largest $\sum^{\text{tracks}}{p_{\textrm{T}}^{2}}$ is chosen. Events were collected using single-lepton triggers, and each lepton candidate must be matched to an appropriate lepton trigger. In the muon channel, events are selected with a single-muon trigger with a $p_{\mathrm{T}}$ threshold of 18 GeV. For the electron channel, a single- electron trigger is required with $p_{\mathrm{T}}>20$ GeV. This is increased to 22 GeV during high instantaneous luminosity periods. Three or more jets with $p_{\mathrm{T}}$ greater than 25 GeV are required in each event. A large amount of $E_{\text{T}}^{\text{miss}}$ is required to select events containing a neutrino. For electron events the $E_{\text{T}}^{\text{miss}}$ must be greater than 30 GeV, while for muon events the $E_{\text{T}}^{\text{miss}}$ is required to be greater than 20 GeV. In order to reduce the background due to multi-jet production containing misidentified or nonprompt leptons, an additional selection requirement is imposed. Typically, in an event arising from this background, the missing transverse momentum vector points in the same direction in the transverse plane as the charged-lepton candidate. Therefore, electron candidate events must pass a requirement that $m_{\mathrm{T}}(W)>30$ GeV, while muon candidate events must have $m_{\mathrm{T}}(W)+E_{\text{T}}^{\text{miss}}>60$ GeV. Here, $m_{\mathrm{T}}(W)$ is defined as $m_{\mathrm{T}}(W)=\sqrt{2\leavevmode\nobreak\ p_{\mathrm{T}}^{\ell}\leavevmode\nobreak\ E_{\text{T}}^{\text{miss}}\leavevmode\nobreak\ (1-\cos\Delta\phi)}\,,$ where $\Delta\phi$ is the difference in $\phi$ between the direction of the charged-lepton transverse momentum, $p_{\mathrm{T}}^{\ell}$, and the missing transverse momentum vector. ### 5 Data samples The data sample used in this analysis comes from the $\sqrt{s}=7$ TeV $pp$ collisions collected during LHC Run 1 in 2011, and was recorded during stable beam conditions with all relevant ATLAS subdetector systems operational. It corresponds to an integrated luminosity of 4.6 fb-1, with an uncertainty of 1.8% [24]. ### 6 Signal and background modeling Except for the background due to multi-jet production leading to misidentified leptons (which is estimated from the data), all signal and background samples are modeled using Monte Carlo (MC) simulated events in conjunction with factors to correct the simulations to data where required. #### 6.1 Multi-jet background / fake leptons The background from multi-jet events, in which a jet is misidentified as a muon or electron, or where a nonprompt lepton within a jet passes the tight lepton selection requirements, is sizable because of the large multi-jet production cross section. Events from these two sources of multi-jet background are referred to as fake-lepton events. This background is estimated using the so-called “matrix method”, which is based on the measurement of the selection efficiencies of leptons using data event samples satisfying relaxed identification criteria [25, 26]. Loose electrons are electrons satisfying the baseline selection criteria where the requirements on particle identification using transition radiation measurements and on the energy-to-momentum ratio ($E$/$p$) are eased, and no requirement on the isolation is imposed. For loose muons, isolation is not required, but all other selection criteria are applied. The “matrix method” is based on the measurement of the efficiencies for real and fake leptons in the loose lepton selection to pass the tight selection criteria. The real-lepton efficiencies are measured in $Z\rightarrow\ell\ell$ data samples, while the fake-lepton efficiencies are determined in data control regions with selection requirements designed to enhance the multi-jet content (1 lepton, $\geq 1$ jet, and only a small amount of $E_{\text{T}}^{\text{miss}}$). These efficiencies depend on both the lepton kinematics and event characteristics. To account for this, event weights are computed from the efficiencies parameterized as a function of a number of observables, which are then used to reweight the sample of data events with lepton candidates that satisfy the loose but not the tight selection criteria. The sums of these weights provide estimates of the multi-jet background. #### 6.2 Monte Carlo samples The samples used in this analysis were obtained from a simulation chain consisting of an event generator interfaced to a parton shower and hadronization model, the outputs of which were passed through a simulation of the ATLAS detector and trigger system [27], and then reconstructed with the same algorithms as the data. The ATLAS detector response was modeled with the ATLAS full simulation (FS) based on Geant4 [28]. For $t\bar{t}$ samples used to evaluate the signal modeling systematic uncertainties, the ATLAS fast simulation AtlasFast-II (AF) [27, 29] was used to model the response of the detector. The $t\bar{t}$ signal was simulated using the NLO Powheg-hvq (patch 4) matrix element generator [30] interfaced with Pythia 6.425 [31], using parameter values set according to the C variant of the Perugia 2011 tune [32] to model the underlying event and parton shower. The NLO CT10 [33] parton distribution function (PDF) set was used for the NLO matrix element part, and the LO PDF set CTEQ6L1 [34] was used with Pythia. The top quark mass was fixed at 172.5 GeV. This sample is referred to as Powheg+Pythia 6\. In order to evaluate the dependence on the choice of parton shower and fragmentation models, additional samples of $t\bar{t}$ events were created. The $\textsc{MC@NLO}{+}\textsc{Herwig}$ sample was created with the NLO MC@NLO [35] generator interfaced with Herwig [36] using the LO AUET2 tune [37]. The Powheg+Herwig sample was created with Powheg interfaced to Herwig using the LO AUET2 tune. The largest backgrounds to the $t\bar{t}$ events in the selected sample are from $W$+jets and $Z$+jets production. These were simulated with the LO event generator Alpgen 2.13 [38] with LO PDF set CTEQ6L1, and interfaced with Herwig 6.52. Alpgen calculates matrix elements (ME) for final states with up to six partons. The MLM [39] matching procedure was used to remove the overlaps between ME and parton shower products in samples with $N$ and $N+1$ final- state partons. In addition to the inclusive parton-flavor processes, separate ME samples of $W$+$b\bar{b}$+jets, $W$+$c\bar{c}$+jets, $W$+$c$+jets, and $Z$+$b\bar{b}$+jets were generated. The double counting of $b$\- and $c$-quarks in $W$/$Z$+jets that occurs between the shower of the inclusive samples and the ME of heavy-flavor production was eliminated using an overlap- removal algorithm based on parton-to-jet $\Delta R$ matching [40]. The $W$+jets and $Z$+jets event samples are referred to as $\textsc{Alpgen}{+}\textsc{Herwig}$. The single-top backgrounds were simulated at NLO using the MC@NLO generator with the NLO PDF set CTEQ6.6 [34], and interfaced with Herwig, except the $t$-channel samples, which were modeled with the LO AcerMC 3.8 generator [41] interfaced with Pythia. Dibosons ($WW$, $WZ$, and $ZZ$) were generated with Herwig using the LO PDF set CTEQ6L1. All samples generated with Herwig for the parton shower evolution and hadronization used Jimmy 4.31 [42] for the underlying-event model. The effects of pileup were modeled by overlaying simulated minimum-bias events on the hard-scattering events. The Monte Carlo events were then reweighted such that the distribution of the number of interactions per bunch crossing, $\langle\mu\rangle$, matched the shape and observed average of $\langle\mu\rangle=9.1$ in the 2011 data. #### 6.3 Signal and background classes The most challenging backgrounds (i.e., those which most resemble $t\bar{t}$ ) are single-top and $W$/$Z$+$b\bar{b}$+jets. Therefore, the $t\bar{t}$ cross- section measurement is expected to be affected most by the modeling of these backgrounds. To improve discrimination between the $t\bar{t}$ signal and the different types of background, this analysis separates the background events into two classes and treats them independently. The “Heavy” background class includes the Monte Carlo samples for single-top, $W$+$b\bar{b}$+jets, and $Z$+$b\bar{b}$+jets. All other types of background, including fake leptons, are assigned to the group designated as the “Light” class. Table 1 summarizes the composition of the classes and lists the datasets which are used to model them. Table 1: Class definitions and compositions are presented. In the Process column, “lf” is defined as any partons that are not $b$-quarks. The other columns show the source of the events and the fractional contribution to the given class. Class | Process | Source | Fraction ---|---|---|--- Signal | $t\bar{t}$ | Powheg+Pythia 6 | 100.0% Light | $W$+lf+jets | $\textsc{Alpgen}{+}\textsc{Herwig}$ | 74.9% Light | $Z$+lf+jets | $\textsc{Alpgen}{+}\textsc{Herwig}$ | 7.6% Light | Dibosons | Herwig | 1.4% Light | Fake $e$ | Data | 6.4% Light | Fake $\mu$ | Data | 9.7% Heavy | $W$+$b\bar{b}$+jets | $\textsc{Alpgen}{+}\textsc{Herwig}$ | 43.3% Heavy | $Z$+$b\bar{b}$+jets | $\textsc{Alpgen}{+}\textsc{Herwig}$ | 6.3% Heavy | Single-top | MC@NLO / AcerMC | 50.4% The expected numbers of signal and background events in the selected sample are presented in Table 2. The uncertainties shown include theoretical uncertainties in the production cross sections of the processes [4, 43, 44, 45, 46, 47]. The $W$/$Z$+jets and diboson uncertainties include a contribution derived from an event yield comparison with Sherpa [48] Monte Carlo samples. The uncertainty in the number of events with fake leptons is estimated to be 20% for muons and 50% for electrons [49, 50]. The observed number of events in data is in good agreement with the prediction. Table 2: The observed and expected numbers of events in the selected sample is shown. The first two columns list the contributions by physics process, and the two rightmost columns present events by class. The Heavy class includes the $W$/$Z$+$b\bar{b}$+jets and single-top processes, while the Light class includes all other backgrounds. Process | Events | Class | Events ---|---|---|--- $t\bar{t}$ | $86\,400$ p m 5700 | Signal | $86\,400$ p m 5700 $W$+jets | $184\,000$ p m 44000 | | $Z$+jets | $19\,000$ p m 4300 | | Dibosons | $3\,200$ p m 1600 | | Single-top | $11\,040$ p m 670 | | Fake leptons | $37\,500$ p m 8700 | | | | Light | $233\,000$ p m 44000 | | Heavy | $21\,900$ p m 1600 Total | $341\,000$ p m 45000 | | Observed | $344\,520$ | | ### 7 Analysis method The SVM is a binary learning algorithm [51]. For any two classes of events, the signed distance from a hyperplane that separates the events is the SVM discriminant. For the analysis presented in this paper, a system of three support vector machines is used to create a three-dimensional multi-class event classifier to distinguish signal events from two classes of background (i.e., Light and Heavy). For events from any dataset, the distances from the three hyperplanes, trained to distinguish between Signal vs. Light, Signal vs. Heavy, and Light vs. Heavy, are treated as the coordinates of points in a 3D decision space. The resulting templates of the prediction model are used in a binned likelihood fit to the analogous 3D distribution of the data events. The SVM was chosen as the binary classifier because it is linear, it has firm mathematical foundations, and it offers a simple geometrical interpretation. Because the SVM method provides the solution to a straightforward convex optimization problem, the minimum it finds is a global one. The stopping point at the training stage is well defined, which therefore makes the method robust against overtraining. The method also works well in problems involving a large number of observables. #### 7.1 The SVM discriminant Each event is described by $N$ observables (i.e., features), and can be represented as a point, $\vv{z}$, in an $N$-dimensional feature space. A linear binary classifier finds a hyperplane of dimension $N-1$ that separates the two classes. Once the separating hyperplane is found, its reconstruction only requires the vectors that lie closest to the plane. These are the support vectors from whence the method derives its name. If the two classes to be discriminated are not linearly separable in the original feature space, this $N$-dimensional space can be mapped, $\vv{z}\rightarrow\vv{\varphi}(\vv{z})$, into a higher-dimensional space in which the problem is linearly separable. Detailed knowledge of this mapping is not required when it is known how to calculate the inner product of the mapped vectors [52, 53]. The distribution of classes in the mapped space can be probed directly by analyzing the multidimensional space which takes as its mathematical basis the SVM solutions for different class pairings. The soft-margin SVMs [54] used in this analysis are constructed using a variant of the sequential minimal optimization algorithm [55] that includes the improvements suggested in Ref. [56]. In the case of a three-class problem like the one considered in this analysis, three different SVM classifiers are trained. Each SVM has the form $f(\vv{z})=\langle w|z\rangle-b=\sum_{i}^{\text{SVs}}\lambda_{i}y_{i}K(\vv{v_{i}},\vv{z})-b\,,$ which is the generalized equation of a plane. The $j^{\text{th}}$ SVM (with $j$ in {1, 2, 3}) has a normal vector given by $|w_{j}\rangle$, and a constant offset $b_{j}$ from the origin. The vectors $\vv{v}$ are the support vectors from training (all other training vectors find their $\lambda=0$), the $y$’s are their “truth” values ($\pm 1$), and the $\lambda$’s are parameters found in the training process along with $b_{j}$. Hence, $|w_{j}\rangle$ is a linear combination of training vectors mapped by $\vv{\varphi}$ to an alternative vector space. The bra-ket notation here serves as a reminder that these vectors belong to this mapped space. The inner product of two vectors in the mapped space given their non-mapped vectors $\vv{x_{1}}$ and $\vv{x_{2}}$ is determined via the kernel function $K(\vv{x_{1}},\vv{x_{2}})$. The SVMs in this analysis use the Gaussian kernel: $K(\vv{x_{1}},\vv{x_{2}})=\vv{\varphi}(\vv{x_{1}})\cdot\vv{\varphi}(\vv{x_{2}})=\exp(-|\vv{x_{1}}-\vv{x_{2}}|^{2}/2\sigma^{2})\,.$ The width $\sigma$ is an input parameter of the training process, along with an additional positive constant $C$ which limits the range of the $\lambda$’s and is necessary for soft-margin SVMs. In order to construct an orthonormal basis from the three trained SVMs, the Gram–Schmidt procedure [57] is used with their $|w_{j}\rangle$ vectors: $|{w^{\prime}_{1}}\rangle=|w_{1}\rangle\quad\text{,}\quad|{w^{\prime}_{2}}\rangle=|w_{2}\rangle-\frac{\langle w_{1}|w_{2}\rangle}{\langle w_{1}|w_{1}\rangle}|w_{1}\rangle\quad\text{,}\quad|{w^{\prime}_{3}}\rangle=|w_{3}\rangle-\frac{\langle w_{1}|w_{3}\rangle}{\langle w_{1}|w_{1}\rangle}|w_{1}\rangle-\frac{\langle w^{\prime}_{2}|w_{3}\rangle}{\langle w^{\prime}_{2}|w^{\prime}_{2}\rangle}|w^{\prime}_{2}\rangle\,.$ Using this basis, 3-tuples (X,Y, Z) for a decision space are created: $\textrm{X}(\vv{z})=\frac{\langle z|w^{\prime}_{1}\rangle}{\sqrt{\langle w^{\prime}_{1}|w^{\prime}_{1}\rangle}}\quad\text{,}\quad\textrm{Y}(\vv{z})=\frac{\langle z|w^{\prime}_{2}\rangle}{\sqrt{\langle w^{\prime}_{2}|w^{\prime}_{2}\rangle}}\quad\text{,}\quad\textrm{Z}(\vv{z})=\frac{\langle z|w^{\prime}_{3}\rangle}{\sqrt{\langle w^{\prime}_{3}|w^{\prime}_{3}\rangle}}\,.$ In this way, an input vector $\vv{z}$ describing an event has coordinates in the XYZ space given by calculating ( X$(\vv{z})$, Y$(\vv{z})$, Z$(\vv{z})$ ). It is these new coordinates in the decision space which are used to describe all events, and this is the space in which the 3D templates of the likelihood function are created. #### 7.2 Physics observables In this analysis, $21$ physics observables are used to distinguish $t\bar{t}$ events from background events (see Table 3). Twenty are kinematic variables, and one comprises the $b$-tagging information of the event. These include the electron or muon momentum, the number of jets in the event, the magnitude and direction of the missing transverse momentum vector, sums of the jet momenta components, the first five Fox–Wolfram moments (FWM), $H_{\mathrm{T}}$, the two largest eigenvalues of the normalized momentum tensor, and the mass of the lepton+jets system ($m_{\ell\mathrm{j}}$). The $H_{\mathrm{T}}$ is the scalar sum of $E_{\text{T}}^{\text{miss}}$, electron $p_{\mathrm{T}}$ or muon $p_{\mathrm{T}}$, and the $p_{\mathrm{T}}$ of all jets passing the selection requirements. Fox–Wolfram moments [58] were originally introduced for $e^{+}e^{-}$ colliders. The FWMs correspond to a decomposition of the event’s phase space into Fourier modes on the surface of a sphere. They were modified for use at the Tevatron and the LHC to characterize the complex shapes of final states at hadron colliders [59, 60]. They form a set of event shape variables, and the $l^{\text{th}}$ FWM ($H_{l}$) is defined in the following way: $H_{l}=\frac{4\pi}{2l+1}\sum_{m=-l}^{l}{\left|\sum_{i}^{\text{jets}}\frac{E_{\mathrm{T}}(i)}{E_{\mathrm{T}}(\text{total})}Y_{l}^{m}(\theta_{i},\phi_{i})\right|^{2}}\,.$ The $Y_{l}^{m}$’s are the spherical harmonics, $i$ runs over all selected jets in the event, and $E_{\mathrm{T}}{(\text{total})}$ represents the sum of the transverse energy from selected jets. The angles $\theta_{i}$ and $\phi_{i}$ indicate the direction of the $i^{\text{th}}$ jet. This analysis makes use of $H_{1}$ through $H_{5}$. The normalized momentum tensor uses the $E_{\text{T}}^{\text{miss}}$ and the momenta of the lepton and up to five jets, and has the following form: $P_{ij}=\frac{{E_{\text{T}}^{\text{miss}}}_{i}\leavevmode\nobreak\ {E_{\text{T}}^{\text{miss}}}_{j}}{|E_{\text{T}}^{\text{miss}}|^{2}}+\sum^{\text{lep}+5\text{jets}}\frac{p_{i}p_{j}}{|p|^{2}}\,.$ Here $i$ and $j$ run over the $x$, $y$, and $z$ components of momentum (for $E_{\text{T}}^{\text{miss}}$, only the $x$ and $y$ components are nonzero). The two largest eigenvalues of this “$p$-tensor” are used as SVM inputs. Because the lepton+jets decays are rotationally invariant in $\phi$, some variables are calculated with respect to the lepton direction in the plane transverse to the beam. Hence, for the momenta of jets, $p_{\parallel}$ and $p_{\bot}$ denote the components which are parallel and perpendicular to the direction of the lepton in the transverse plane. Similarly, the $\phi(E_{\text{T}}^{\text{miss}})$ variable is then the angle between the transverse momentum of the lepton and the missing transverse momentum vector, and $p_{\parallel}$ for the lepton corresponds to its entire transverse momentum. The SVMs treat each variable as one of the coordinates of a point in a $21$-dimensional space. The algorithm requires that each variable should fall roughly in the same numeric range so that all features contribute a similar weight when evaluating the distance from the separating hyperplane. The variables which have values outside the range of ($-1$, $+1$) are transformed such that they approximately meet this requirement. All input variables and the values that were used to scale them are listed in Table 3. Table 3: List of the 21 variables used as input to the SVMs. The variables were divided by the given values to make them all of similar magnitude. Number | Feature | Divided by ---|---|--- 11 | $E_{\text{T}}^{\text{miss}}$ [GeV] | 250 12 | $\phi(E_{\text{T}}^{\text{miss}})$ [radians] | $2\pi$ 13 | Lepton $E$ [GeV] | 400 14 | Lepton $p_{\parallel}$ [GeV] | 400 15 | Lepton $p_{\mathrm{z}}$ [GeV] | 400 16 | Mass(lepton+jets) [GeV] | 750 17 | Fox–Wolfram moment 1 | 1 18 | Fox–Wolfram moment 2 | 1 19 | Fox–Wolfram moment 3 | 1 10 | Fox–Wolfram moment 4 | 1 11 | Fox–Wolfram moment 5 | 1 12 | Sum all jets $E_{\mathrm{T}}$ [GeV] | 500 13 | Sum all jets $E$ [GeV] | 750 14 | Sum all jets $p_{\parallel}$ [GeV] | 750 15 | Sum all jets $p_{\bot}$ [GeV] | 750 16 | Sum all jets $p_{\mathrm{z}}$ [GeV] | 750 17 | $H_{\mathrm{T}}$ [GeV] | 500 18 | $p$-tensor eigenvalue 1 | 1 19 | $p$-tensor eigenvalue 2 | 1 20 | Number of jets | 10 21 | Number of $b$-tags | 10 #### 7.3 SVM training The SVMs are trained to separate three classes of events: the Light and Heavy backgrounds and the $t\bar{t}$ signal. In order to train the SVMs, the Monte Carlo simulation samples and a data sample representing the multi-jet background are split into two subsamples. For training purposes, events from each class are randomly selected from those passing the selection requirements. The remaining events are used to test how well the trained SVMs perform. Also, it is only these remaining events that are utilized in the subsequent analysis. The training process aims to find the set of support vectors that forms the optimal decision plane in the mapped space induced by the kernel function. As described in Section 7.1, there are two free parameters that need to be specified when training. These are the $\sigma$ parameter of the Gaussian kernel and $C$, the positive constant which constrains the $\lambda$’s in the solution. A search grid over the values of these parameters was implemented, and the performance of a given training was then evaluated based upon the area of the resulting receiver operating characteristic (ROC) curve created with the events not used in the training. As a result of this study, the values of 1.2 for $\sigma$ and 2.0 for $C$ were settled upon. The $t\bar{t}$ Signal class and the two background classes, Light and Heavy, each used 8,000 events for training, which is a small fraction of the total available events. Increasing the number of training events was not found to improve performance. It was also verified that the trained SVMs were not overtrained (i.e., that their discriminant distributions generalize well from the training set to the full class dataset). #### 7.4 Class templates Different physics processes can be distinguished by their distinctly different distributions in the XYZ decision space. These distributions are obtained by applying the Gram–Schmidt procedure to each event’s SVM output values. Histogramming the different physics processes in the resulting 3D decision space creates probability distribution functions (i.e., templates) that can then be used in the likelihood fit. Figure 1(a) shows a contour at a fixed value of the template function from each of the classes. This highlights the different regions in 3D decision space where the different event types congregate. As an alternative way of illustrating these distributions, Figure 1(b) shows a sampling of 10,000 events from each class. Also shown are the three trained decision planes, which serve to demonstrate the nonorthogonal nature of the basis defined by the SVMs. The SvL plane at $\text{X}=0$ is seen to separate Signal from Light as it extends downwards in Z. Similarly, the SvH and LvH planes separate their training classes. The multiple band structure in the templates arises because of training with the number of $b$-tags, which is a strong discriminant and is discrete. These template bands represent groupings of events with 0, 1, or 2 $b$-tags. (a) (b) Figure 1: Distribution of the three classes of events shown in the 3D orthonormalized decision space. 1(a) Contours from each class are shown together. 1(b) A view of the SVM decision space is shown with a random sampling of 10,000 events from each class. The three planes corresponding to each of the three trained SVMs are also depicted. SvL labels the plane which comes from the Signal vs. Light training. Similarly, SvH is the Signal vs. Heavy plane and LvH is the Light vs. Heavy solution. In order to minimize the potential effect of small fluctuations in the modeling, a small set of wider bins was constructed. The full 3D XYZ decision space was organized into four quadrants by dividing the space at $\text{Y}=0$ and $\text{Z}=-0.01$. Each quadrant was then further divided into bins along the X axis. The quadrants are designated YZ, Yz, yZ, and yz, where the capitalization of the letters indicates where the quadrant is located (e.g., Yz means $\text{Y}>0.00$ and $\text{Z}\leq-0.01$). The division points for X were chosen to keep a minimum of approximately 1,000 events in each bin while preserving the shapes of the distributions. It is these four binned distributions from the quadrants that are used for the final profile likelihood fit. #### 7.5 Cross-section measurement A binned profile likelihood function is used in a fit to determine the $t\bar{t}$ cross section from the data. In the likelihood fit, four templates are used: $t\bar{t}$, $W$/$Z$, single-top, and fake lepton. In evaluating the systematic uncertainties, particularly with respect to the modeling of $t\bar{t}$, it was observed that a large uncertainty occurs because of the similarity between the final states of $t\bar{t}$ events and single-top background events. To alleviate this effect the single-top backgrounds, which arise from electroweak processes rather than the strong interactions responsible for the production of $t\bar{t}$ pairs, are combined into a single template normalized to their predicted cross sections. The $W$/$Z$+$b\bar{b}$+jets, the light-flavor $W$/$Z$+jets, and the diboson backgrounds are combined into a $W$/$Z$ template. The normalizations of the $t\bar{t}$, $W$/$Z$, and fake-lepton templates are free parameters of the fit. The grouping of physics processes used when constructing the templates for the fit can differ from the class definitions used for training. At the training stage, the events are arranged in order to create SVMs that can distinguish between the $t\bar{t}$ signal, the Light backgrounds, and the Heavy backgrounds. After training, each physics process can be reassigned to templates. The chosen allocation of the physics processes to four templates ($t\bar{t}$, $W$/$Z$, single-top, and fake lepton) results in smaller expected uncertainties in the $t\bar{t}$ cross section. The likelihood function uses the templates (projections onto the X axis from each of the quadrants) that have been built in the XYZ decision space. Each template has an associated strength parameter $\theta$ in the likelihood. The maximum value of the likelihood is obtained in determining the central values of the $\theta$ parameters. The systematic uncertainties of the fit results are also included in the likelihood as nuisance parameters (NPs, $\alpha$’s, or collectively $\vv{\alpha}$) with Gaussian constraints. Each template is a function of the nuisance parameters in the likelihood, which is then able to capture the effects due to each source of systematic uncertainty. The likelihood of an unknown sample for an $n_{\text{T}}$ template problem is defined as: $L=\prod_{i}^{\text{bins}}P(\mathscr{E}_{i},o_{i})\times\prod_{j}^{\text{NPs}}G(\alpha_{j},\sigma_{j})\,.$ Here $G(\alpha_{j},\sigma_{j})$ is the Gaussian constraint for the $j^{\text{th}}$ NP, $\alpha_{j}$, with the corresponding uncertainty $\sigma_{j}$; and $P(\mathscr{E}_{i},o_{i})$ is the Poisson probability mass function for the $i^{\text{th}}$ bin given the observed number of events, $o_{i}$, and the expected number of events, $\mathscr{E}_{i}$ : $\mathscr{E}_{i}=\mathscr{E}(\vv{\theta},\vv{\alpha})_{i}=\sum_{j}^{n_{\text{T}}}{\theta_{j}N(\vv{\alpha})_{j}T_{j}(i,\vv{\alpha})}\,.$ The templates are constructed such that $T_{j}(i,\vv{\alpha})$ gives the fractional number of events in template $T_{j}$’s $i^{\text{th}}$ bin. Consequently, the sum over all bins of a given template is equal to $1$. The $N(\vv{\alpha})_{j}$ are defined to be the total number of events expected from the $j^{\text{th}}$ template assuming an integrated luminosity $\mathscr{L}$. To calculate this value, the following sum over all modeled processes belonging to the $j^{\text{th}}$ template is computed: $N(\vv{\alpha})_{j}=\sum_{k}^{\text{processes}}{\sigma_{k}\epsilon(\vv{\alpha})_{k}\mathscr{L}}\,.$ The $\sigma_{k}$ and $\epsilon_{k}$ are the cross section and acceptance for the $k^{\text{th}}$ physics process. They are derived from MC simulation. For the multi-jet background, $N_{j}$ is taken from the fake-lepton estimate. A maximum-likelihood fit is performed to extract the values of the $\theta$ and $\alpha$ parameters. The $1\sigma$ uncertainty for a given parameter is taken to be the change in the value of that parameter which causes $\ln L$ to decrease by $0.5$ away from $\ln L_{0}$, when $\ln L$ is maximized with respect to all other free parameters and where $\ln L_{0}$ is the global maximum. All $\theta$ and $\alpha$ parameters have both their $+1\sigma$ and $-1\sigma$ uncertainties determined in this way. The $\theta$ for the $t\bar{t}$ class, multiplied by the assumed cross section, gives the measured value of the $t\bar{t}$ cross section. Similarly, the uncertainty in $\theta_{t\bar{t}}$ from the fit, multiplied by the assumed cross section, gives the uncertainty in the $\sigma_{t\bar{t}}$ measurement. #### 7.6 Systematic uncertainties All systematic uncertainties were evaluated using the profile likelihood fit. The systematic effects are incorporated into the templates, and each template is associated with appropriate nuisance parameters ($\alpha$’s) in the likelihood. A nuisance parameter that takes a value of $0$ in the fit keeps the nominal template, while a value of $+1$ or $-1$ changes the template to look like the $+1\sigma$ / $-1\sigma$ effect. Templates at intermediate values of the $\alpha$’s are linearly interpolated. A Gaussian constraint is also applied to each $\alpha$ in order to propagate its controlled uncertainty when the data have no preference for that systematic effect. The profile likelihood fit then provides a simultaneous measurement of the $\theta$ and $\alpha$ parameters. In this way, the systematic effects are converted into a statistical framework that properly takes into account correlations and which can potentially lower the uncertainties in the measurement. The individual effects of various sources of systematic uncertainty are displayed in Table 4. They are obtained by leaving groupings of nuisance parameters out of the fit, and calculating the square of each effect as the difference of the squares of the total error and the residual error. Object modeling Systematic uncertainties in the lepton selection arise from uncertainties in lepton identification, reconstruction, and triggering. These are evaluated by applying tag-and-probe methods to $Z\rightarrow\ell\ell$ events [16]. Uncertainties due to the energy scale and resolution are also considered for electrons and muons. These effects are evaluated by assigning each of them a separate nuisance parameter in the likelihood so as to allow the error source to be shifted both upwards and downwards by its uncertainty. The resulting systematic effects are summarized in Table 4 as Leptons. For jets, the main source of uncertainty is the jet energy scale (JES). The JES and its uncertainty are evaluated using a combination of test-beam data, LHC collision data, and simulation [21]. As a result of the in situ analyses for the calibration of the full 2011 dataset, the correlations between various JES uncertainty components are encoded in 21 subcomponents. These include statistical and method uncertainties, detector uncertainties, modeling and theory uncertainties, mixed detector and modeling uncertainties, and pileup. The JES uncertainty is evaluated by assigning a separate NP to each of these 21 JES subcomponents. The jet energy resolution (JER) is separated by process ($t\bar{t}$, single-top, and $W$/$Z$+jets) and is assigned three corresponding NPs. These extra degrees of freedom allow differences in the kinematics and prevalence of $b$-quark, light-quark, and gluon jets in these processes to be better represented in the profile likelihood. The resulting uncertainties in $\sigma_{t\bar{t}}$ from these sources are indicated in Table 4 as Jets. The jet-flavor-dependent efficiencies of the $b$-tagging algorithm are calibrated using dijet events, and dilepton $t\bar{t}$ events from data. Differences in the $b$-tagging efficiency as well as $c$-jet and light-jet mistag rates between data and simulation are parameterized using correction factors, which are functions of $p_{\mathrm{T}}$ and $\eta$ [22]. The $b$-tag systematic uncertainties were evaluated by constructing nine NPs that correspond to unique bins in jet $p_{\mathrm{T}}$, as the uncertainties at low and high jet $p_{\mathrm{T}}$ should be largely uncorrelated. Single NPs were used for the $c$-tag and Mistag systematic uncertainties. These systematic effects appear in Table 4 as three uncertainties labeled $b$-tag, $c$-tag, and Mistag. During the variation of nuisance parameters related to jets and leptons, the $E_{\text{T}}^{\text{miss}}$ is recalculated in accordance with the changes caused by those systematic effects. In this way, the jet and lepton uncertainties are propagated to the $E_{\text{T}}^{\text{miss}}$. However, the $E_{\text{T}}^{\text{miss}}$ uncertainty due to calorimeter cells not assigned to any other physics object is evaluated individually. Also, an additional 6.6% uncertainty due to pileup is applied to $E_{\text{T}}^{\text{miss}}$. Both of these are given separate NPs in the profile likelihood, and they are listed in Table 4 under Missing transverse momentum. Modeling of $\boldsymbol{t\bar{t}}$ events Systematic uncertainties due to the choice of $t\bar{t}$ MC generator are evaluated by taking the full difference between Powheg+Herwig (AF) and $\textsc{MC@NLO}{+}\textsc{Herwig}$ (AF). The systematic uncertainty due to the choice of parton shower model is taken as the full difference between Powheg+Herwig (AF) and Powheg+Pythia 6 (AF). These are listed as Generator and Shower/hadronization in Table 4, respectively. The systematic error due to uncertainties in the modeling of initial- and final-state radiation (ISR/FSR) is evaluated using Alpgen interfaced to Pythia 6\. Monte Carlo samples were created in which the parameter that controls the amount of ISR/FSR in Alpgen was either halved or doubled. Half of the spread between the Alpgen samples with raised and lowered ISR/FSR parameter values is taken as the systematic error. The uncertainty due to renormalization and factorization scales is evaluated with two modified samples, generated with MC@NLO interfaced with Herwig, in which parameters controlling the renormalization and factorization scales, introduced to cure the ultraviolet and infrared divergences in ME calculations, are simultaneously either halved or doubled. The full difference between the two samples is taken as the Renormalization/factorization error. Each of the major $t\bar{t}$ modeling systematic uncertainties (Generator, Shower/hadronization, ISR/FSR, and Renormalization/factorization) is given a shape NP in each quadrant. The uncertainty in the normalization of events is assigned two NPs. One of these is used to track the migration of events between quadrants, and it mirrors the movement of events seen when comparing the nominal and systematically shifted samples. The second NP is taken as an overall normalization error which corresponds to the normalization difference seen for the full event selection (where all four quadrants are combined). Therefore, each of the $t\bar{t}$ modeling uncertainties mentioned above has six NPs (four for shape, and two for normalization). The underlying-event modeling error is evaluated by comparing two different $t\bar{t}$ MC event samples produced with varied parameters in Powheg+Pythia 6\. One was generated with the Perugia 2011 central tune, and the other with Perugia 2011 mpiHI [32]. Both of these samples use the P2011 CTEQ5L Pythia tune, and not P2011C CTEQ6L1, which applies to the nominal $t\bar{t}$ MC sample. Their full difference is used as the measurement uncertainty for the underlying event, using a single NP in the profile likelihood. All particles in the final state from the LHC $pp$ collisions must be color singlets. Different schemes for the color reconnection (CR) of the beam remnant and other outgoing hard collision objects are examined. The $t\bar{t}$ cross-section uncertainty due to this effect is estimated by comparing two different $t\bar{t}$ MC samples produced with Powheg+Pythia 6\. A reference sample was obtained using the Perugia 2011 central tune. The other sample was generated with Perugia 2011 noCR [32], and has modified color reconnection parameters. These samples use the P2011 CTEQ5L Pythia tune. The full difference between these two samples is taken as the CR uncertainty, using a single NP. To estimate the uncertainty due to the choice of parton distribution function, the CT10 PDF set parameterization is examined using its 26 upwards and downwards systematic variations. Each of the 26 CT10 eigenvector components is assigned a separate NP in the profile likelihood. Background modeling To estimate the error due to the shape of the $W$/$Z$+jets backgrounds and to assess the effect of any mismodeling, background samples are reweighted to match data for each of the following variables, taken one at a time, in a signal-depleted control region: lepton $E$, $\phi(E_{\text{T}}^{\text{miss}})$, $m_{\ell\mathrm{j}}$, $\sum_{\text{jets}}{p_{\parallel}}$, and $\sum_{\text{jets}}{p_{\bot}}$. This control region was defined as events matching the nominal selection, but containing exactly three jets, none of which are $b$-tagged. The reweighting functions were applied only to $W$/$Z$+jets samples (leaving $t\bar{t}$, single-top, and fake lepton untouched). Five NPs in the profile likelihood implement these functions such that the NPs turn the reweighting effects on and off, each according to the differences seen for these five variables in the data. In Table 4, these effects appear under the heading $W$/$Z$ reweighting. For the single-top shape, AcerMC samples with raised and lowered ISR/FSR parameter values are compared, and a single NP is assigned to this systematic uncertainty. The effects due to the uncertainty of the single-top, $W$+jets, and $Z$+jets cross sections are investigated by varying these cross sections within their theoretical errors. For the $W$+jets background, a 4% uncertainty is applied to the inclusive $W$ boson cross section, with an additional 24% uncertainty for each jet [39, 61]. This method was also applied to the $Z$+jets cross sections. To evaluate the systematic uncertainty in the $t\bar{t}$ cross section due to the theoretical uncertainties in the single-top cross section, the single-top cross section is varied in accordance with the theoretical results, taken from Refs. [45, 46, 47]. The relative normalization within the $W$/$Z$+jets MC sample is varied by raising and lowering the corresponding nominal relative yields of each jet multiplicity by their respective errors. Similarly, the relative normalization between fake-electron and fake-muon events is varied by raising and lowering their nominal predictions by their errors. The resulting effects are evaluated using appropriate NPs added in the profile likelihood fit. This uncertainty is quoted as $W$/$Z$ & fakes relative normalization in Table 4. Also, the uncertainty due to variations of the $W$+jets heavy-flavor fraction is included via three NPs in the profile likelihood. These NPs place an additional $25$% uncertainty on each of the assumed $W$+$b\bar{b}$+jets, $W$+$c\bar{c}$+jets, and $W$+$c$+jets cross sections. Table 4 summarizes these in the row labeled Heavy-flavor fraction. Template statistics / luminosity For the profile likelihood fit, an additional fit parameter is introduced for each bin. These parameters are used to represent the Poisson fluctuation of the predicted number of events in each bin as estimated from the size of the Monte Carlo samples. The error propagated to $\theta_{t\bar{t}}$ from these additional parameters is then an appropriate representation of the MC statistical error. The integrated luminosity measurement has an uncertainty of 1.8% [24], and therefore each physics process is assigned an uncertainty of this magnitude. This systematic error is controlled by a single nuisance parameter in the likelihood. The total measurement uncertainty, including individual groups of contributions, is listed in Table 4. The largest uncertainties are due to the lepton selection and luminosity, followed by the uncertainties due to JES, $b$-tagging, ISR/FSR, and other $t\bar{t}$ modeling. Beam energy The LHC beam energy during the 2012 $\sqrt{s}=8$ TeV $pp$ run was measured to be within $0.1$% of the nominal value of 4 TeV per beam, using the revolution frequency difference of protons and lead ions during $p+Pb$ runs in early 2013 combined with the magnetic model errors [62]. A similar uncertainty in the beam energy is applicable to the 2011 LHC run. The approach used in Ref. [63] was therefore applied to the measurement using the $\sqrt{s}=7$ TeV dataset. The uncertainty in the $t\bar{t}$ theoretical cross section due to this energy difference was calculated to be 0.27%, using the Top++ 2.0 program [5] and assuming that the relative change of the $t\bar{t}$ cross section for a $0.1$% change in $\sqrt{s}$ is as predicted by the NNLO + NNLL calculation. It is negligible compared to other sources of systematic uncertainty. Table 4: Summary table of the measurement uncertainties. Because the profile likelihood fit accounts for correlations, the total error is not simply the components added in quadrature. Individual effects were obtained by leaving groupings of NPs out of the fit, and calculating the square of each effect as the difference of the squares of the total error and the residual error. Source | $-1\sigma$ [pb] | $+1\sigma$ [pb] | $-1\sigma$ [%] | $+1\sigma$ [%] ---|---|---|---|--- Object Modeling | | | | Leptons | $-3.1$ | $+3.3$ | $-1.8$ | $+2.0$ Jets | $-2.9$ | $+3.0$ | $-1.7$ | $+1.8$ $b$-tag | $-1.9$ | $+2.0$ | $-1.1$ | $+1.2$ $c$-tag | $-0.4$ | $+0.4$ | $-0.3$ | $+0.3$ Mistag | $-0.3$ | $+0.3$ | $-0.2$ | $+0.2$ Missing transverse momentum | $-0.2$ | $+0.2$ | $-0.1$ | $+0.1$ $t\bar{t}$ Modeling | | | | Generator | $-1.6$ | $+1.8$ | $-1.0$ | $+1.1$ Shower/hadronization | $-2.4$ | $+2.6$ | $-1.4$ | $+1.5$ Renormalization/factorization | $-1.4$ | $+1.4$ | $-0.8$ | $+0.9$ ISR/FSR | $-2.4$ | $+2.5$ | $-1.4$ | $+1.5$ Underlying event | $-0.7$ | $+0.8$ | $-0.4$ | $+0.5$ Color reconnection | $-0.5$ | $+0.5$ | $-0.3$ | $+0.3$ PDF | $-1.8$ | $+1.9$ | $-1.1$ | $+1.1$ Background Modeling | | | | $W$/$Z$ reweighting | $-1.0$ | $+1.0$ | $-0.6$ | $+0.6$ $W$/$Z$ & fakes relative normalization | $-1.2$ | $+1.2$ | $-0.7$ | $+0.7$ Heavy-flavor fraction | $-1.1$ | $+1.2$ | $-0.7$ | $+0.7$ Single-top | $-1.0$ | $+1.0$ | $-0.6$ | $+0.6$ Other | | | | Data statistics | $-0.7$ | $+0.7$ | $-0.4$ | $+0.4$ Template statistics | $-1.0$ | $+1.0$ | $-0.6$ | $+0.6$ Luminosity | $-3.2$ | $+3.4$ | $-1.9$ | $+2.0$ Total | $-6.8$ | $+7.1$ | $-4.0$ | $+4.2$ ### 8 Results The top quark pair-production cross section for $pp$ collisions at a center- of-mass energy of $\sqrt{s}=7$ TeV in the lepton+jets channel is found to be $\sigma_{t\bar{t}}=168.5\pm 0.7$(stat.) ${}^{+6.2}_{-5.9}$(syst.) ${}^{+3.4}_{-3.2}$(lumi.) pb. This result includes all systematic uncertainties as evaluated with the profile likelihood fit, with the statistical and luminosity errors listed separately. Figure 2 shows a comparison between the observed and fitted numbers of events in each of the quadrants. A correlated $\chi^{2}$ test was used to check that there is good agreement between the data and the fit results within the combined statistical and systematic error bands. Figure 2: The fitted yields of signal and background processes compared with data, shown in four YZ quadrants divided along the X axis, as used in the fit. They are labeled quad1YZ, quad2Yz, quad3yZ, and quad4yz (the boundary letters are appended for easy reference). The lower panel shows the ratio of data to fit prediction. The shaded regions correspond to the statistical and systematic uncertainties. The first and last bins also contain any events found outside the range of the horizontal axis. Comparison plots between data and the fit prediction are shown for a few selected input variables in Figure 3 for all events. Analogous comparisons in signal-rich and background-rich regions of the XYZ space are shown in Figures 4 and 5, respectively. The signal-rich region is defined by $\text{X}>0$, $\text{Y}>0$, and $\text{Z}<0$, while the background-rich region lies in the opposite octant of XYZ space, and has $\text{X}<0$, $\text{Y}<0$, and $\text{Z}>0$. The X dimension corresponds to the Signal vs. Light decision hyperplane, while the Y and Z dimensions are linear combinations of the other SVM hyperplanes and are the directions orthogonal to X. Based on a correlated $\chi^{2}$ test, the data and the fit agree well within the combined statistical and systematic error bands for all 21 variables. The measured $t\bar{t}$ cross section is in good agreement with the theoretical predictions based on the NNLO + NNLL calculations of $\sigma_{t\bar{t}}^{\text{NNLO+NNLL}}=177^{+5}_{-6}\>\mathrm{(scale)}\,\pm 9\>\mathrm{(PDF+}\alpha_{\text{s}})$ pb $=177\leavevmode\nobreak\ ^{+10}_{-11}$ pb for $pp$ collisions at a center-of-mass energy of $\sqrt{s}=7$ TeV and a top quark mass of 172.5 GeV [4]. The ATLAS measurement of the $t\bar{t}$ cross section at 7 TeV in the dilepton channel [9] is $\sigma_{t\bar{t}}=182.9\pm 3.1$(stat.) $\pm$ 4.2(syst.) $\pm$ 3.6(lumi.) pb. Depending upon the assumptions made for the systematic uncertainty correlations between these two measurements, the significance of their discrepancy was found to be in the $1.9\sigma$ to $2.1\sigma$ range. (a) (b) (c) (d) (e) (f) Figure 3: The data distributions of six selected input variables are shown with their post-fit predictions in the selected sample. The predicted signal fraction is 24.8%. Shown are 3(a) the number of jets, 3(b) the number of $b$-tagged jets, 3(c) $H_{\mathrm{T}}$, 3(d) the $4^{\text{th}}$ Fox–Wolfram moment, 3(e) $E_{\text{T}}^{\text{miss}}$, and 3(f) the mass of the lepton and jets. Data are shown with the overlaid dots. The predicted events are shown for each of the templates used in modeling the data. The statistical and systematic error bands are given by the shaded regions. The first and last bins contain events found outside the range of the horizontal axis. (a) (b) (c) (d) (e) (f) Figure 4: The data distributions of six selected input variables are shown with their post-fit predictions in the selected sample for the signal-rich region with $\text{X}>0$, $\text{Y}>0$, and $\text{Z}<0$. The predicted signal fraction in this region is 79.3%. Shown are 4(a) the number of jets, 4(b) the number of $b$-tagged jets, 4(c) $H_{\mathrm{T}}$, 4(d) the $4^{\text{th}}$ Fox–Wolfram moment, 4(e) $E_{\text{T}}^{\text{miss}}$, and 4(f) the mass of the lepton and jets. Data are shown with the overlaid dots. The predicted events are shown for each of the templates used in modeling the data. The statistical and systematic error bands are given by the shaded regions. The first and last bins contain events found outside the range of the horizontal axis. (a) (b) (c) (d) (e) (f) Figure 5: The data distributions of six selected input variables are shown with their post-fit predictions in the selected sample for the background-rich region with $\text{X}<0$, $\text{Y}<0$, and $\text{Z}>0$. The predicted signal fraction in this region is 3.1%. Shown are 5(a) the number of jets, 5(b) the number of $b$-tagged jets, 5(c) $H_{\mathrm{T}}$, 5(d) the $4^{\text{th}}$ Fox–Wolfram moment, 5(e) $E_{\text{T}}^{\text{miss}}$, and 5(f) the mass of the lepton and jets. Data are shown with the overlaid dots. The predicted events are shown for each of the templates used in modeling the data. The statistical and systematic error bands are given by the shaded regions. The first and last bins contain events found outside the range of the horizontal axis. #### 8.1 Top quark mass dependence The result of the profile likelihood fit depends on the assumed mass of the top quark through differences in $t\bar{t}$ acceptance owing to lepton kinematics, and also from minor variations in the shape of the discriminant. The analysis in this paper assumes a top quark mass of $m_{\text{ref}}=172.5$ GeV. The 2014 average of Tevatron and LHC Run 1 measurements of the top quark mass [64] gives a value of $m_{t}=173.34\pm 0.27$(stat.) $\pm\leavevmode\nobreak\ 0.71$(syst.) GeV. The current ATLAS average from 2019 [65] yields $m_{t}=172.69\pm 0.25$(stat.) $\pm\leavevmode\nobreak\ 0.41$(syst.) GeV. The dependence of the $t\bar{t}$ cross section on the mass of the top quark was determined through alternative profile likelihood fits that assume different top quark masses. Monte Carlo samples for both $t\bar{t}$ and single-top with top quark mass values of 165.0, 170.0, 172.5, 175.0, 177.5, and 180.0 GeV were employed to measure the $t\bar{t}$ cross section assuming each of these masses. These measurements were then fitted to obtain the $t\bar{t}$ cross section’s top quark mass dependence. When constrained to go through this measurement’s central value at 172.5 GeV, the best-fit second- order polynomial for the $t\bar{t}$ cross section as a function of $\Delta m_{t}=m_{t}-m_{\text{ref}}$ is $\sigma_{t\bar{t}}(\Delta m_{t})=0.016\cdot\Delta m_{t}^{2}-0.75\cdot\Delta m_{t}+168.5$ pb, with $\Delta m_{t}$ in GeV. ### 9 Summary A measurement of the top quark pair-production cross section in the lepton+jets channel was performed with the ATLAS experiment at the LHC, using a multivariate technique based on support vector machines. The measurement was obtained with a three-class, multidimensional event classifier. It is based on $4.6\text{\leavevmode\nobreak\ fb}^{-1}$ of data collected during 2011 in $pp$ collisions at $\sqrt{s}=7$ TeV. The $t\bar{t}$ cross section is found to be $\sigma_{t\bar{t}}=168.5\pm 0.7$(stat.) ${}^{+6.2}_{-5.9}$(syst.) ${}^{+3.4}_{-3.2}$(lumi.) pb, which has a relative uncertainty of ${}^{+4.2}_{-4.0}$%. This measurement is consistent with the Standard Model prediction based on QCD calculations at next-to-next-to-leading order. ## Appendix: Fit visualization The fit to the data can be difficult to visualize because of the 3D nature of the decision space. A series of 2D and 1D projections of the 3D space have been created in order to better illustrate its characteristics. Projections are provided in the XY, XZ, and YZ planes to give a qualitative comparison of the fit results with the data in XYZ space, and also to help visualize the definitions of the signal-rich and background-rich regions employed in Figures 4 and 5. The 2D projections of the 3D decision space use the standard fit results in XYZ space, but project the constant binning obtained with cubes of edge length 0.008 in that space. On the right side of Figure 6 the 2D projection plots show the fractional relative difference between the number of events predicted by the fit and the observed data. The 2D plots on the left side of Figure 6 show the class composition of each bin when projected onto the XY, XZ, and YZ planes. The rectangle representing each 2D bin is colored from top to bottom such that the colors that fill it are in the same ratio as the predictions for each class in that bin. Contour lines, which each represent the concentration of events in the plane, are drawn on top of the colored bins. Contours exist for each of the following fractional values of the maximal 2D bin height: (0.05, 0.10, 0.20, 0.40, 0.60, 0.80, 0.90, 0.95). The 1D projections of the 3D decision space onto the X, Y, and Z axes are shown in Figure 7 for the results of the fit to the data. For visualization purposes, the full 3D decision space is projected onto each of these axes. These projections have the systematic error bands superimposed, and also include the ratio of data to fit prediction. --- Figure 6: (left) These 2D projections show the composition of each bin according to the template fit results. Bins are shaded in the same ratio as the fit prediction. The contours drawn on top of the bins represent the overall concentration of events. Contours are provided at 5%, 10%, 20%, 40%, 60%, 80%, 90%, and 95% of the maximal 2D bin height. (right) The 2D bins are shaded in accordance with the fractional difference between the observed data and the number of events predicted by the fit. The contours shown are the same as in the left column and provide a reference. (a) (b) (c) Figure 7: The 1D projections of the data and expected events after the likelihood fit. X, Y, and Z form an orthogonal basis in the decision space defined by the three trained SVMs. X corresponds to the Signal vs. Light plane. Y and Z are the directions orthogonal to X found with the remaining Signal vs. Heavy and Light vs. Heavy SVMs. The projections use the fit parameters obtained with the profile likelihood fit to four quadrants, but project a constant binning obtained with cubes of edge length 0.008 in XYZ space. In each of these plots, the prediction is broken down by the templates used in the fit. The first and last bins contain events found outside the range of the horizontal axis. ### References * [1] S.L. Glashow “Partial-symmetries of weak interactions” In _Nucl. Phys._ 22, 1961, pp. 579–588 DOI: 10.1016/0029-5582(61)90469-2 * [2] S. Weinberg “A model of leptons” In _Phys. Rev. Lett._ 19, 1967, pp. 1264–1266 DOI: 10.1103/PhysRevLett.19.1264 * [3] A. Salam “Elementary Particle Physics: Relativistic Groups and Analyticity” In _Eighth Nobel Symposium_ Stockholm: AlmquistWiksell, 1968, pp. 367 * [4] Michal Czakon, Paul Fiedler and Alexander Mitov “Total Top-Quark Pair-Production Cross Section at Hadron Colliders Through $O(\alpha_{S}^{4})$” In _Phys. Rev. Lett._ 110, 2013, pp. 252004 DOI: 10.1103/PhysRevLett.110.252004 * [5] Michal Czakon and Alexander Mitov “Top++: A program for the calculation of the top-pair cross-section at hadron colliders” In _Comput. Phys. Commun._ 185, 2014, pp. 2930 DOI: 10.1016/j.cpc.2014.06.021 * [6] Michiel Botje et al. “The PDF4LHC Working Group Interim Recommendations”, 2011 arXiv:1101.0538 [hep-ph] * [7] ATLAS Collaboration “Measurement of the top quark pair production cross-section with ATLAS in the single lepton channel” In _Phys. Lett. B_ 711, 2012, pp. 244 DOI: 10.1016/j.physletb.2012.03.083 * [8] CMS Collaboration “Measurement of the $t\bar{t}$ production cross section in $pp$ collisions at $\sqrt{s}=7\,\text{TeV}$ with lepton+jets final states” In _Phys. Lett. B_ 720, 2013, pp. 83 DOI: 10.1016/j.physletb.2013.02.021 * [9] ATLAS Collaboration “Measurement of the $t\bar{t}$ production cross-section using $e\mu$ events with $b$-tagged jets in $pp$ collisions at $\sqrt{s}=7$ and $8\,\text{TeV}$ with the ATLAS detector” In _Eur. Phys. J. C_ 74, 2014, pp. 3109 DOI: 10.1140/epjc/s10052-014-3109-7 * [10] CMS Collaboration “Measurement of the $t\bar{t}$ production cross section in the $e\mu$ channel in proton–proton collisions at $\sqrt{s}=7$ and $8\,\text{TeV}$” In _JHEP_ 08, 2016, pp. 029 DOI: 10.1007/JHEP08(2016)029 * [11] ATLAS and CMS Collaborations “Combination of inclusive top-quark pair production cross-section measurements using ATLAS and CMS data at $\sqrt{s}=7$ and $8\,\text{TeV}$”, 2022 arXiv:2205.13830 [hep-ex] * [12] ATLAS Collaboration “The ATLAS Experiment at the CERN Large Hadron Collider” In _JINST_ 3, 2008, pp. S08003 DOI: 10.1088/1748-0221/3/08/S08003 * [13] ATLAS Collaboration “Performance of the ATLAS Trigger System in 2010” In _Eur. Phys. J. C_ 72, 2012, pp. 1849 DOI: 10.1140/epjc/s10052-011-1849-1 * [14] ATLAS Collaboration “The ATLAS Collaboration Software and Firmware”, ATL-SOFT-PUB-2021-001, 2021 URL: https://cds.cern.ch/record/2767187 * [15] ATLAS Collaboration “Electron performance measurements with the ATLAS detector using the 2010 LHC proton–proton collision data” In _Eur. Phys. J. C_ 72, 2012, pp. 1909 DOI: 10.1140/epjc/s10052-012-1909-1 * [16] ATLAS Collaboration “Electron reconstruction and identification efficiency measurements with the ATLAS detector using the 2011 LHC proton–proton collision data” In _Eur. Phys. J. C_ 74, 2014, pp. 2941 DOI: 10.1140/epjc/s10052-014-2941-0 * [17] ATLAS Collaboration “Measurement of the muon reconstruction performance of the ATLAS detector using 2011 and 2012 LHC proton–proton collision data” In _Eur. Phys. J. C_ 74, 2014, pp. 3130 DOI: 10.1140/epjc/s10052-014-3130-x * [18] ATLAS Collaboration “Topological cell clustering in the ATLAS calorimeters and its performance in LHC Run 1” In _Eur. Phys. J. C_ 77, 2017, pp. 490 DOI: 10.1140/epjc/s10052-017-5004-5 * [19] Matteo Cacciari, Gavin P. Salam and Gregory Soyez “The anti-$k_{t}$ jet clustering algorithm” In _JHEP_ 04, 2008, pp. 063 DOI: 10.1088/1126-6708/2008/04/063 * [20] Matteo Cacciari, Gavin P. Salam and Gregory Soyez “FastJet user manual” In _Eur. Phys. J. C_ 72, 2012, pp. 1896 DOI: 10.1140/epjc/s10052-012-1896-2 * [21] ATLAS Collaboration “Jet energy measurement and its systematic uncertainty in proton–proton collisions at $\sqrt{s}=7\,\text{TeV}$ with the ATLAS detector” In _Eur. Phys. J. C_ 75, 2015, pp. 17 DOI: 10.1140/epjc/s10052-014-3190-y * [22] ATLAS Collaboration “Performance of $b$-jet identification in the ATLAS experiment” In _JINST_ 11, 2016, pp. P04008 DOI: 10.1088/1748-0221/11/04/P04008 * [23] ATLAS Collaboration “Performance of missing transverse momentum reconstruction in proton–proton collisions at $\sqrt{s}=7\,\text{TeV}$ with ATLAS” In _Eur. Phys. J. C_ 72, 2012, pp. 1844 DOI: 10.1140/epjc/s10052-011-1844-6 * [24] ATLAS Collaboration “Improved luminosity determination in $pp$ collisions at $\sqrt{s}=7\,\text{TeV}$ using the ATLAS detector at the LHC” In _Eur. Phys. J. C_ 73, 2013, pp. 2518 DOI: 10.1140/epjc/s10052-013-2518-3 * [25] ATLAS Collaboration “Measurement of the top quark-pair production cross section with ATLAS in $pp$ collisions at $\sqrt{s}=7\,\text{TeV}$” In _Eur. Phys. J. C_ 71, 2011, pp. 1577 DOI: 10.1140/epjc/s10052-011-1577-6 * [26] ATLAS Collaboration “Estimation of non-prompt and fake lepton backgrounds in final states with top quarks produced in proton–proton collisions at $\sqrt{s}=8\leavevmode\nobreak\ \text{TeV}$ with the ATLAS Detector”, ATLAS-CONF-2014-058, 2014 URL: https://cds.cern.ch/record/1951336 * [27] ATLAS Collaboration “The ATLAS Simulation Infrastructure” In _Eur. Phys. J. C_ 70, 2010, pp. 823 DOI: 10.1140/epjc/s10052-010-1429-9 * [28] GEANT4 Collaboration and S. Agostinelli “Geant4 – a simulation toolkit” In _Nucl. Instrum. Meth. A_ 506, 2003, pp. 250 DOI: 10.1016/S0168-9002(03)01368-8 * [29] Wolfgang Lukas “Fast Simulation for ATLAS: Atlfast-II and ISF” In _J. Phys.: Conf. Ser._ 396.2, 2012, pp. 022031 DOI: 10.1088/1742-6596/396/2/022031 * [30] Stefano Frixione, Giovanni Ridolfi and Paolo Nason “A positive-weight next-to-leading-order Monte Carlo for heavy flavour hadroproduction” In _JHEP_ 09, 2007, pp. 126 DOI: 10.1088/1126-6708/2007/09/126 * [31] Torbjorn Sjöstrand, Stephen Mrenna and Peter Z. Skands “PYTHIA 6.4 physics and manual” In _JHEP_ 05, 2006, pp. 026 DOI: 10.1088/1126-6708/2006/05/026 * [32] Peter Zeiler Skands “Tuning Monte Carlo generators: The Perugia tunes” In _Phys. Rev. D_ 82, 2010, pp. 074018 DOI: 10.1103/PhysRevD.82.074018 * [33] H.-L. Lai “New parton distributions for collider physics” In _Phys. Rev. D_ 82, 2010, pp. 074024 DOI: 10.1103/PhysRevD.82.074024 * [34] J. Pumplin “New Generation of Parton Distributions with Uncertainties from Global QCD Analysis” In _JHEP_ 07, 2002, pp. 012 DOI: 10.1088/1126-6708/2002/07/012 * [35] Stefano Frixione, Paolo Nason and Bryan R. Webber “Matching NLO QCD and parton showers in heavy flavour production” In _JHEP_ 08, 2003, pp. 007 DOI: 10.1088/1126-6708/2003/08/007 * [36] G. Corcella “HERWIG 6: an event generator for hadron emission reactions with interfering gluons (including supersymmetric processes)” In _JHEP_ 01, 2001, pp. 010 DOI: 10.1088/1126-6708/2001/01/010 * [37] ATLAS Collaboration “New ATLAS event generator tunes to 2010 data”, ATL-PHYS-PUB-2011-008, 2011 URL: https://cds.cern.ch/record/1345343 * [38] Michelangelo L. Mangano “ALPGEN, a generator for hard multiparton processes in hadronic collisions” In _JHEP_ 07, 2003, pp. 001 DOI: 10.1088/1126-6708/2003/07/001 * [39] J. Alwall “Comparative study of various algorithms for the merging of parton showers and matrix elements in hadronic collisions” In _Eur. Phys. J._ C53, 2008, pp. 473–500 DOI: 10.1140/epjc/s10052-007-0490-5 * [40] M.. Mangano, M. Moretti, F. Piccinini and M. Treccani “Matching matrix elements and shower evolution for top-quark production in hadronic collisions” In _JHEP_ 01, 2007, pp. 013 DOI: 10.1088/1126-6708/2007/01/013 * [41] B.P. Kersevan and E. Richter-Was “The Monte Carlo event generator AcerMC versions 2.0 to 3.8 with interfaces to PYTHIA 6.4, HERWIG 6.5 and ARIADNE 4.1” In _Comput. Phys. Commun._ 184.3, 2013, pp. 919–985 DOI: 10.1016/j.cpc.2012.10.032 * [42] J.. Butterworth, J.. Forshaw and M.. Seymour “Multiparton interactions in photoproduction at HERA” In _Z. Phys. C_ 72, 1996, pp. 637 DOI: 10.1007/s002880050286 * [43] Charalampos Anastasiou, Lance Dixon, Kirill Melnikov and Frank Petriello “High-precision QCD at hadron colliders: Electroweak gauge boson rapidity distributions at next-to-next-to leading order” In _Phys. Rev. D_ 69, 2004, pp. 094008 DOI: 10.1103/PhysRevD.69.094008 * [44] John M. Campbell and R. Ellis “Update on vector boson pair production at hadron colliders” In _Phys. Rev. D_ 60, 1999, pp. 113006 DOI: 10.1103/PhysRevD.60.113006 * [45] Nikolaos Kidonakis “Differential and total cross section for top pair and single top production”, DESY-PROC-2012-02/251 DOI: 10.3204/DESY-PROC-2012-02/251 * [46] M. Aliev et al. “HATHOR – HAdronic Top and Heavy quarks crOss section calculatoR” In _Comput. Phys. Commun._ 182, 2011, pp. 1034–1046 DOI: 10.1016/j.cpc.2010.12.040 * [47] P. Kant et al. “HatHor for single top-quark production: Updated predictions and uncertainty estimates for single top-quark production in hadronic collisions” In _Comput. Phys. Commun._ 191, 2015, pp. 74–89 DOI: 10.1016/j.cpc.2015.02.001 * [48] T. Gleisberg et al. “Event generation with SHERPA 1.1” In _JHEP_ 02, 2009, pp. 007 DOI: 10.1088/1126-6708/2009/02/007 * [49] ATLAS Collaboration “Simultaneous measurements of the $t\bar{t}$, $W^{+}W^{-}$, and $Z/\gamma^{*}\rightarrow\tau\tau$ production cross-sections in $pp$ collisions at $\sqrt{s}=7\,\text{TeV}$ with the ATLAS detector” In _Phys. Rev. D_ 91, 2015, pp. 052005 DOI: 10.1103/PhysRevD.91.052005 * [50] ATLAS Collaboration “Comprehensive measurements of $t$-channel single top-quark production cross sections at $\sqrt{s}=7\,\text{TeV}$ with the ATLAS detector” In _Phys. Rev. D_ 90, 2014, pp. 112006 DOI: 10.1103/PhysRevD.90.112006 * [51] B.. Boser, I.. Guyon and V.. Vapnik “A Training Algorithm for Optimal Margin Classifiers” In _Proceedings of the Fifth Annual Workshop on Computational Learning Theory (COLT’92)_ , 1992, pp. 144–152 DOI: 10.1145/130385.130401 * [52] J. Mercer “Functions of Positive and Negative Type, and their Connection with Theory of Integral Equations” In _Philos. Trans. Roy. Soc. London Ser. A_ 209, 1909, pp. 415–446 DOI: 10.1098/rsta.1909.0016 * [53] N. Aronszajn “Theory of Reproducing Kernels” In _Trans. Amer. Math. Soc._ 68, 1950, pp. 337–404 DOI: 10.1090/S0002-9947-1950-0051437-7 * [54] C. Cortes and V. Vapnik “Support-vector networks” In _Machine Learning_ 20.3, 1995, pp. 273 DOI: 10.1007/BF00994018 * [55] J.C. Platt “Sequential Minimal Optimization: A Fast Algorithm for Training Support Vector Machines”, MSR-TR-98-14, 1998 URL: https://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.43.4376 * [56] S.S. Keerthi “Improvements to Platt’s SMO Algorithm for SVM Classifier Design” In _Neural Comp._ 13.3, 2001, pp. 637 DOI: 10.1162/089976601300014493 * [57] Michael Reed and Barry Simon “Methods of Modern Mathematical Physics. I: Functional Analysis” New York: Academic Press, 1972 DOI: 10.1016/B978-0-12-585001-8.X5001-6 * [58] G.. Fox and S. Wolfram “Event Shapes in $e^{+}e^{-}$ Annihilation” In _Nucl. Phys. B_ 149, 1979, pp. 413 DOI: 10.1016/0550-3213(79)90003-8 * [59] C. Bernaciak, M… Buschmann, A. Butter and T. Plehn “Fox-Wolfram Moments in Higgs Physics” In _Phys. Rev. D_ 87, 2013, pp. 073014 DOI: 10.1103/PhysRevD.87.073014 * [60] L.. Spiller “Modification of Fox-Wolfram Moments for Hadron Colliders” In _JHEP_ 03, 2016, pp. 027 DOI: 10.1007/JHEP03(2016)027 * [61] ATLAS Collaboration “Measurement of the top quark pair production cross section in $pp$ collisions at $\sqrt{s}=7\,\text{TeV}$ in dilepton final states with ATLAS” In _Phys. Lett. B_ 707, 2012, pp. 459 DOI: 10.1016/j.physletb.2011.12.055 * [62] E. Todesco and J. Wenninger “Large Hadron Collider momentum calibration and accuracy” In _Phys. Rev. Accel. Beams_ 20, 2017, pp. 081003 DOI: 10.1103/PhysRevAccelBeams.20.081003 * [63] ATLAS Collaboration “Measurement of lepton differential distributions and the top quark mass in $t\bar{t}$ production in $pp$ collisions at $\sqrt{s}=8\,\text{TeV}$ with the ATLAS detector” In _Eur. Phys. J. C_ 77, 2017, pp. 804 DOI: 10.1140/epjc/s10052-017-5349-9 * [64] The ATLAS, CDF, CMS and D0 Collaborations “First combination of Tevatron and LHC measurements of the top-quark mass”, ATLAS-CONF-2014-008, 2014 arXiv:1403.4427 [hep-ph] * [65] ATLAS Collaboration “Measurement of the top quark mass in the $t\bar{t}\to$ lepton+jets channel from $\sqrt{s}=8\,\text{TeV}$ ATLAS data and combination with previous results” In _Eur. Phys. J. C_ 79, 2019, pp. 290 DOI: 10.1140/epjc/s10052-019-6757-9 * [66] ATLAS Collaboration In _Eur. Phys. J. C_ 76, 2016, pp. 642 DOI: 10.1140/epjc/s10052-016-4501-2